
null
null
Introduction In my previous post I tested eight academic RAG or RAG‑like tools—Elicit, scite assistant, SciSpace, Primo Research Assistant, Undermind, AI2 Scholar QA, and several "Deep Research" modes from OpenAI, Gemini, and Perplexity—to see how they handled a well‑publicised 2020 paper that has since been retracted : “The association between early‑career informal mentorship in academic collaborations and
Musing sabout librarianship - Substack A Home Full of Memories—And a New Door Opening When I first hit “Publish” on Musings About Librarianship on Blogger back in 2009, I was a wet behind the ears librarian nervously sharing half-baked thoughts. I never imagined that my tiny blogger corner would grow into a gathering place for thousands of curious colleagues around the world.
Academic Retrieval Augmented Systems (RAG) live or die on the sources they retrieve, so what happens if they retrieve retracted papers? In this post, I will discuss the ways different Academic RAG systems handle them, and I will end with some suggestions to vendors of such systems.Thanks for reading Aaron’s Musings about Librarianship!
May was a busy month for me in terms of output. [Article] Comparative review of Primo Research Assistant, Scopus AI and Web of Science Research Output First, I had two pieces of work published in the Katina Magazine that I am quite proud of. First, a comparative review of Primo Research Assistant, Scopus AI, and Web of Science Research Assistant—written by yours truly—was published.
As an academic librarian, I’m often asked: “Which AI search tool should I use?” With a flood of new tools powered by large language models (LLMs) entering the academic search space, answering that question is increasingly complex.
Following my recent talk for the Boston Library Consortium, many of you expressed a strong interest in learning how to test the new generation of AI-powered academic search tools.
I recently compared three academic AI search tools: Primo Research Assistant, Web of Science Research Assistant, and Scopus AI for a review article.
Introduction How do you test a tool that promises to revolutionize academic research? With AI-powered search engines popping up everywhere—from startups like Elicit.com to giants like Scopus AI—librarians, researchers, and developers need a reliable way to evaluate them.
Synopsis (Grok 3.0 aided) A new breed of "Deep Research" tools is reshaping how we tackle complex queries, moving beyond traditional search engines and quick AI summaries. From OpenAI’s Deep Research and Google’s Gemini counterpart to academic-focused players like Elicit, Undermind.ai and SciSpace, these agentic systems promise comprehensive reports and literature reviews—albeit with a much slower response.
Summary (summarised by Gemini 2.0-exp) We often think of AI-generated 'hallucinations' as blatant errors, but what happens when a RAG system produces information that's factually true but not explicitly stated in its sources?