<?xml version='1.0' encoding='UTF-8'?><?xml-stylesheet href="http://www.blogger.com/styles/atom.css" type="text/css"?><feed xmlns='http://www.w3.org/2005/Atom' xmlns:openSearch='http://a9.com/-/spec/opensearchrss/1.0/' xmlns:blogger='http://schemas.google.com/blogger/2008' xmlns:georss='http://www.georss.org/georss' xmlns:gd="http://schemas.google.com/g/2005" xmlns:thr='http://purl.org/syndication/thread/1.0'><id>tag:blogger.com,1999:blog-4727930222560708528</id><updated>2026-03-28T18:50:16.159+08:00</updated><category term="discovery"/><category term="tools"/><category term="personal"/><category term="open access"/><category term="large language model"/><category term="social media"/><category term="retrieval augmented generation"/><category term="web scale discovery"/><category term="analytics"/><category term="browser extension"/><category term="delivery"/><category term="future"/><category term="mobile"/><category term="open data"/><category term="OA discovery"/><category term="marketing"/><category term="twitter"/><category term="Google"/><category term="literature mapping"/><category term="machine learning"/><category term="open citations"/><category term="opinion"/><category term="google scholar"/><category term="survey"/><category term="AI"/><category term="citation-based literature mapping tools"/><category term="information literacy"/><category term="alert"/><category term="data discovery"/><category term="opensearch"/><category term="facebook"/><category term="humor"/><category term="rss"/><category term="iphone"/><category term="user experience"/><category term="catalogue"/><category term="institutional repository"/><category term="personal habits"/><category term="zotero"/><category term="bookmarklet"/><category term="conference"/><category term="presentation"/><category term="Summon"/><category term="friendfeed"/><category term="librarianship"/><category term="librarydayinthelife"/><category term="linked data"/><category term="search widget"/><category term="GPT3"/><category term="LLM"/><category term="award"/><category term="citation"/><category term="citation classification"/><category term="foursquare"/><category term="libx"/><category term="location-based"/><category term="mendeley"/><category term="meta"/><category term="netvibes"/><category term="ranking"/><category term="videos"/><category term="academic library"/><category term="app"/><category term="learning series"/><category term="lifestream"/><category term="portal"/><category term="reference"/><category term="reference manager"/><category term="search box"/><category term="video"/><category term="viral"/><category term="visualization"/><category term="widgets"/><category term="wikipedia"/><category term="&quot;google reader&quot;"/><category term="3d"/><category term="Elicit"/><category term="Internet tools"/><category term="ala11"/><category term="barcode"/><category term="bibliometrics"/><category term="blog"/><category term="boolean"/><category term="chat"/><category term="chatgpt"/><category term="customtoolbar"/><category term="data"/><category term="database"/><category term="delicious"/><category term="dipity"/><category term="endnote"/><category term="environment_scannings"/><category term="extension"/><category term="ezproxy"/><category term="flickr"/><category term="getFTR"/><category term="google analytics"/><category term="greasemonkey"/><category term="inspiration"/><category term="libguides"/><category term="mashup"/><category term="metadata"/><category term="mobile app"/><category term="musings"/><category term="news"/><category term="openaccess"/><category term="openrefine"/><category term="orcid"/><category term="orientation"/><category term="research"/><category term="smartkeyword"/><category term="sms"/><category term="startup pages"/><category term="statistics"/><category term="webmynd"/><category term="widget"/><category term="wizfolio"/><category term="yelp"/><category term="&quot;google custom search&quot;"/><category term="&quot;google place&quot;"/><category term="&quot;google wave&quot;"/><category term=". collection"/><category term="2d"/><category term="APC"/><category term="API"/><category term="Aggregator"/><category term="Alexa"/><category term="Aurasma"/><category term="Counter"/><category term="GPT4"/><category term="Grok"/><category term="IM"/><category term="Pageflakes"/><category term="Plan S"/><category term="Scopus"/><category term="Singapore"/><category term="Yuuguu"/><category term="addon"/><category term="android"/><category term="answers"/><category term="apple"/><category term="article processing charges"/><category term="aseo"/><category term="augmented reality"/><category term="backjtype"/><category term="bayesian"/><category term="bento"/><category term="bibliometrix"/><category term="bing"/><category term="blekko"/><category term="blogging"/><category term="book review"/><category term="boxcar"/><category term="branding"/><category term="buzzword"/><category term="calender"/><category term="camtasia"/><category term="cardstar"/><category term="channels"/><category term="china"/><category term="chrome"/><category term="citeulike"/><category term="conduit"/><category term="connected papers"/><category term="consalxv"/><category term="copyright"/><category term="crowd sourcing"/><category term="curation"/><category term="data analysis"/><category term="deep research"/><category term="delievery"/><category term="desktop"/><category term="dimensions"/><category term="disqus"/><category term="disruption"/><category term="divide"/><category term="doodle"/><category term="echo"/><category term="email"/><category term="encore"/><category term="etherpad"/><category term="exitreality"/><category term="ezpaarse"/><category term="fantasy"/><category term="feedback"/><category term="filter"/><category term="firefox"/><category term="free"/><category term="geolocation"/><category term="getsatisfaction"/><category term="google assistant"/><category term="google insights"/><category term="google local"/><category term="google profile"/><category term="google reader"/><category term="google+"/><category term="googlebooks"/><category term="googledoc"/><category term="googlenow"/><category term="guides"/><category term="hangout"/><category term="heroes"/><category term="history"/><category term="historypin"/><category term="holdings"/><category term="iatul2012"/><category term="ideas"/><category term="ifttt"/><category term="information dashboard"/><category term="insights"/><category term="ipad"/><category term="klout"/><category term="known item"/><category term="knowtro"/><category term="large language modted"/><category term="laytr"/><category term="lean library"/><category term="lemontree"/><category term="libanswers"/><category term="libday4"/><category term="libqual"/><category term="libraries"/><category term="library"/><category term="library2.0"/><category term="librarything"/><category term="librarything local"/><category term="literature review"/><category term="manyeyes"/><category term="map"/><category term="memes"/><category term="microsoft academic"/><category term="misc"/><category term="mobile app database"/><category term="mover"/><category term="moving"/><category term="mypoint"/><category term="netvibe"/><category term="nexus"/><category term="notification"/><category term="nypl"/><category term="oadoi"/><category term="open knowledge map"/><category term="openai"/><category term="opengoo"/><category term="operator"/><category term="pids"/><category term="plugin"/><category term="points of need"/><category term="policy"/><category term="pptplex"/><category term="prezi"/><category term="project management"/><category term="proquest"/><category term="proxy"/><category term="public library"/><category term="push"/><category term="qrcode"/><category term="qrcodes"/><category term="realtime"/><category term="recommender"/><category term="redlaser"/><category term="remote"/><category term="repositories"/><category term="resistance"/><category term="resolutions"/><category term="retraction"/><category term="roving"/><category term="schedule"/><category term="scispace"/><category term="scite"/><category term="screencast"/><category term="search"/><category term="search  box"/><category term="searching"/><category term="seo"/><category term="services"/><category term="sessions"/><category term="sgbeat"/><category term="siri"/><category term="skype"/><category term="slidecast"/><category term="slideshare"/><category term="smartwatch"/><category term="social networking"/><category term="source"/><category term="storify"/><category term="subject guides"/><category term="tagging"/><category term="tdm"/><category term="technical services"/><category term="text visualization"/><category term="time-line based"/><category term="timeline"/><category term="tips"/><category term="toolbar"/><category term="top 100"/><category term="training"/><category term="trends decline"/><category term="tungle"/><category term="tweetdeck"/><category term="tweetlist"/><category term="twiiter"/><category term="twitterleague"/><category term="uservoice"/><category term="visibility"/><category term="webex"/><category term="wiki"/><category term="wordtree"/><category term="xobni"/><category term="xtranormal"/><category term="yahoopipes"/><category term="yewno"/><category term="zoom"/><category term="zoomit"/><category term="zoterio"/><title type='text'>Aaron Tay&#39;s Musings about librarianship </title><subtitle type='html'>&lt;a href=&quot;http://musingsaboutlibrarianship.blogspot.com/p/subscribe-by-email.html&quot;&gt;(Subscribe by email)&lt;/a&gt;</subtitle><link rel='http://schemas.google.com/g/2005#feed' type='application/atom+xml' href='http://musingsaboutlibrarianship.blogspot.com/feeds/posts/default'/><link rel='self' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default?max-results=5&amp;redirect=false'/><link rel='alternate' type='text/html' href='http://musingsaboutlibrarianship.blogspot.com/'/><link rel='hub' href='http://pubsubhubbub.appspot.com/'/><link rel='next' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default?start-index=6&amp;max-results=5&amp;redirect=false'/><author><name>Aaron Tay</name><uri>http://www.blogger.com/profile/02750645621492448678</uri><email>noreply@blogger.com</email><gd:image rel='http://schemas.google.com/g/2005#thumbnail' width='32' height='32' src='//blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjNp5I56cmCnITp9u98mqHmOga9TIDbvdXeuRetlD5Lq7jfNQVbjCCMMyEkblX6PtSR34esdLJ6qarZjGFZC_pAuvyTr93fqVvnlAzrnm2DjjaaH1BTx3XN8lJH69-gnw/s220/profile.jpg'/></author><generator version='7.00' uri='http://www.blogger.com'>Blogger</generator><openSearch:totalResults>402</openSearch:totalResults><openSearch:startIndex>1</openSearch:startIndex><openSearch:itemsPerPage>5</openSearch:itemsPerPage><entry><id>tag:blogger.com,1999:blog-4727930222560708528.post-5019087510993509498</id><published>2025-06-10T08:25:00.006+08:00</published><updated>2025-06-12T01:06:10.358+08:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="personal"/><title type='text'>Musings About Librarianship Is Moving—Farewell Blogger, Hello Substack!</title><content type='html'>&lt;p&gt;&amp;nbsp;&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://aarontay.substack.com/feed&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;905&quot; data-original-width=&quot;993&quot; height=&quot;292&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjbnjOeQQEXc-bQckhdVM95RYgPqUipxucetPiMZ24eTNqEdiDHHnJNZRbvnln1N4jK22c2oJdgCs54VnLCbRtYiakBP32aulztgi-5E1_lxIJzeQ99Yvt7QKPkBeyLe2Cs0xpVvc3qC-nVtaBUqHai3-tvr7PpPV_nR5DradPuYFOQf2gYB2VFtbqCtzkC/s320/substack.png&quot; width=&quot;320&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://aarontay.substack.com/&quot;&gt;Aaron Tay&#39;s Musingsabout librarianship - Substack&lt;/a&gt; (https://aarontay.substack.com/feed) &lt;/div&gt;
&lt;p data-end=&quot;268&quot; data-start=&quot;255&quot;&gt;&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;h3 data-end=&quot;253&quot; data-start=&quot;203&quot;&gt;&lt;br /&gt;&lt;/h3&gt;&lt;h3 data-end=&quot;253&quot; data-start=&quot;203&quot;&gt;A Home Full of Memories—And a New Door Opening&lt;/h3&gt;&lt;div&gt;&lt;div&gt;When I first hit “Publish” on Musings About Librarianship on Blogger back in 2009, I was a wet behind the ears librarian nervously sharing half-baked thoughts. I never imagined that my tiny blogger corner would grow into a gathering place for thousands of curious colleagues around the world.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;Sixteen years later—after countless coffee-fueled midnight drafts and many generous words of encouragement—my little Blogger house is bursting at the seams. So, with a mix of nostalgia and fluttering excitement, I’m carrying every box, every post, every hard-won insight to substack-&amp;nbsp;&lt;a href=&quot;https://aarontay.substack.com/&quot;&gt;https://aarontay.substack.com/&lt;/a&gt;.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Why I Had to Move&lt;/b&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Many of you subscribed through &lt;a href=&quot;http://follow.it&quot;&gt;follow.it&lt;/a&gt;. Lately, that service wedged noisy ads between our conversations and hid full-text emails behind a paywall defeating the purpose of email subscriptions! The moment I saw this, I knew it was time to leave.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Substack, by contrast has been offering&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Clean, ad-free emails&lt;/li&gt;&lt;li&gt;A pleasant web reading experience—no pop-ups or paywalls&lt;/li&gt;&lt;li&gt;A modern interface for editing&lt;/li&gt;&lt;li&gt;An active community space for comments and threaded discussion (seems like many of you were already on substack)&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h3 data-end=&quot;1248&quot; data-start=&quot;1212&quot;&gt;What’s Changed (and What Hasn’t)&lt;/h3&gt;
&lt;div class=&quot;_tableContainer_16hzy_1&quot;&gt;&lt;div class=&quot;_tableWrapper_16hzy_14 group flex w-fit flex-col-reverse&quot; tabindex=&quot;-1&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;You still get every post whether by RSS, Email - 100% free&lt;/li&gt;&lt;li&gt;You still get the same old rambling pieces by me&lt;/li&gt;&lt;/ul&gt;&lt;div&gt;What Changed&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;The entire archive—every crazy or inspire idea, every typo—now &lt;a href=&quot;https://aarontay.substack.com/&quot;&gt;lives on Substack&lt;/a&gt;.&lt;/li&gt;&lt;li&gt;Future posts &lt;i&gt;will only be posted on Substack&lt;/i&gt;.&amp;nbsp;&lt;/li&gt;&lt;li&gt;The &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/&quot;&gt;old Blogger site (both posts and pages) &lt;/a&gt;will stand as a quiet museum, a place to wander down memory lane.&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;h2&gt;What you need to Do&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Email subscribers&lt;/b&gt;: Sit back—your address was imported automatically to stack. If it wasn&#39;t, &lt;a href=&quot;https://aarontay.substack.com/subscribe&quot;&gt;subscribe here&lt;/a&gt;.&lt;/li&gt;&lt;li&gt;&lt;b&gt;RSS followers: &lt;/b&gt;Update your feed to&amp;nbsp;&lt;a href=&quot;https://aarontay.substack.com/feed&quot;&gt;https://aarontay.substack.com/feed&lt;/a&gt;(the old feed will stop refreshing with new posts).&lt;/li&gt;&lt;li&gt;&lt;b&gt;Update your bookmarks&lt;/b&gt;:&amp;nbsp;&lt;a href=&quot;https://aarontay.substack.com/&quot;&gt;https://aarontay.substack.com/&amp;nbsp;&amp;nbsp;&lt;/a&gt;&lt;/li&gt;&lt;/ul&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;A Heartfelt Thank-You&lt;/h2&gt;&lt;div&gt;Thank You. You’ve read plenty of rambling post that struggled to make a point, peppered with typos and misconceptions (I am always learning!) and still stuck with me!&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Here’s to the next chapter—still free, still ad-free, still powered by curiosity and caffeinated late nights. I can’t wait to keep learning alongside you.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Hope to see you soon, and thank you for every click, comment, share and quiet read.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;With gratitude&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Aaron Tay&lt;/div&gt;&lt;div&gt;Musings About Librarianship&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;i&gt;Drafted and edited with help of ChatGPT 4o&lt;/i&gt;.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;</content><link rel='replies' type='application/atom+xml' href='http://musingsaboutlibrarianship.blogspot.com/feeds/5019087510993509498/comments/default' title='Post Comments'/><link rel='replies' type='text/html' href='http://www.blogger.com/comment/fullpage/post/4727930222560708528/5019087510993509498?isPopup=true' title='2 Comments'/><link rel='edit' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/5019087510993509498'/><link rel='self' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/5019087510993509498'/><link rel='alternate' type='text/html' href='http://musingsaboutlibrarianship.blogspot.com/2025/06/musings-about-librarianship-is.html' title='Musings About Librarianship Is Moving—Farewell Blogger, Hello Substack!'/><author><name>Aaron Tay</name><uri>http://www.blogger.com/profile/02750645621492448678</uri><email>noreply@blogger.com</email><gd:image rel='http://schemas.google.com/g/2005#thumbnail' width='32' height='32' src='//blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjNp5I56cmCnITp9u98mqHmOga9TIDbvdXeuRetlD5Lq7jfNQVbjCCMMyEkblX6PtSR34esdLJ6qarZjGFZC_pAuvyTr93fqVvnlAzrnm2DjjaaH1BTx3XN8lJH69-gnw/s220/profile.jpg'/></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjbnjOeQQEXc-bQckhdVM95RYgPqUipxucetPiMZ24eTNqEdiDHHnJNZRbvnln1N4jK22c2oJdgCs54VnLCbRtYiakBP32aulztgi-5E1_lxIJzeQ99Yvt7QKPkBeyLe2Cs0xpVvc3qC-nVtaBUqHai3-tvr7PpPV_nR5DradPuYFOQf2gYB2VFtbqCtzkC/s72-c/substack.png" height="72" width="72"/><thr:total>2</thr:total></entry><entry><id>tag:blogger.com,1999:blog-4727930222560708528.post-617629989690538701</id><published>2025-05-30T22:41:00.003+08:00</published><updated>2025-06-12T01:02:08.741+08:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="large language model"/><category scheme="http://www.blogger.com/atom/ns#" term="retrieval augmented generation"/><title type='text'>Comparative review on Primo Research Assistant, Scopus AI, Web of Science Research Assistant and a explainer for AI search for librarians</title><content type='html'>&lt;p&gt;May was a busy month for me in terms of output.&lt;/p&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;[Article] Comparative review of Primo Research Assistant, Scopus AI and Web of Science Research Output&lt;/h3&gt;&lt;p&gt;First, I had two pieces of work published in the &lt;a href=&quot;https://katinamagazine.org/&quot;&gt;Katina Magazine&lt;/a&gt; that I am quite proud of.&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://katinamagazine.org/content/article/reviews/2025/deep-dive-into-three-ai-academic-search-tools&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;625&quot; data-original-width=&quot;1839&quot; height=&quot;218&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh1rrlkhzPaq52T2FktZ5kDx5Bqcum0EpISk88ydTRZ3DgO7t2A65cDJ4KALg4avqLfehXGeJWDiM4ry2_OtXOYToKEkx-yrtN46M5KpEqrLWY0w4YQfkUKa7DcNXalOoXcyYWMDvdLHSVC_C1ACKV1OgjedH5KX_NOCiqucF4dt4LW0eNYlKUc0aw7rYXb/w640-h218/Deep%20Dive%20into%20Three%20AI%20Academic%20Search%20Tools%20_%20Katina%20Magazine.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;First, &lt;a href=&quot;https://katinamagazine.org/content/article/reviews/2025/deep-dive-into-three-ai-academic-search-tools&quot;&gt;a comparative review of Primo Research Assistant, Scopus AI, and Web of Science Research Assistant&lt;/a&gt;—written by yours truly—was published.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;These are three of the most commonly used academic search tools in the library world, so I was honoured to be invited to write the piece.&lt;/div&gt;&lt;blockquote&gt;&lt;a href=&quot;https://knowledge.exlibrisgroup.com/Summon/Product_Documentation/Searching_in_The_Summon_Service/Search_Features/Getting_Started_with_Summon_Research_Assistant&quot;&gt;Summon Research Assistant &lt;/a&gt;was released just as I was finishing the article. From a quick glance, it appears functionally equivalent to Primo Research Assistant.&lt;/blockquote&gt;&lt;p&gt;&lt;/p&gt;&lt;p&gt;This was one of the most technically complex pieces I’ve written. The original draft was more than twice as long. Eventually, with the help of editors and copyeditors, we decided to split the article into two parts. The first part focuses solely on how these three systems generate direct answers using Retrieval-Augmented Generation (RAG).&lt;/p&gt;&lt;p&gt;A second part will be forthcoming, comparing other features such as finding seminal papers, topic maps, and top experts.&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;h3&gt;[Article] A &quot;explainer&quot; of AI search for librarians&lt;/h3&gt;&lt;p&gt;Even after splitting the review into two, the writing was still too long and unwieldy because I kept digressing into technical explanations.&lt;/p&gt;&lt;p&gt;Eventually, the editors had the brilliant idea of pulling out all the technical concept detours and publishing them separately as an “explainer” piece titled &lt;a href=&quot;https://katinamagazine.org/content/article/reviews/2025/a-librarians-guide-to-ai-in-academic-search-tools&quot;&gt;A Librarian’s Guide to AI in Academic Search Tools.&lt;/a&gt;&amp;nbsp;&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://katinamagazine.org/content/article/reviews/2025/a-librarians-guide-to-ai-in-academic-search-tools&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;598&quot; data-original-width=&quot;1601&quot; height=&quot;240&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjBuDo9EuCLLztZhmQlr8nA8baMMaVG56PluXDwl5XYF2pWiWLV2H8Q9YfBqBCIy8RW1BLrC2Hdog-QCMOFg_W33WqHwmC2-UA6JDaQAFND8vaRnd_paF-4JzdIUEtoZvyWgTn47eiaKDtEMuvEKfsIXsSJZO97TBcaM3qU8jWA0PE0aysgIY2m_0T7F1si/w640-h240/A%20Librarian%E2%80%99s%20Guide%20to%20AI%20in%20Academic%20Search%20Tools%20_%20Katina%20Magazine.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;p&gt;To be honest, when I first saw the title, I thought it was a bit overreaching. Understanding “AI”—even within just the field of information retrieval—is no easy task. I’m not even sure I fully understand it myself.&lt;/p&gt;&lt;p&gt;I expanded the piece a little to smooth out the rough edges (remember, it started as essentially footnotes to the comparative review). It covers the following topics:&lt;/p&gt;&lt;ol style=&quot;background-color: white; box-sizing: border-box; color: #484848; font-family: Inter; font-size: 16px; list-style-image: initial; list-style-position: outside; margin: 0px 0px 40px; padding: 0px 0px 0px 20px; position: relative;&quot;&gt;&lt;li lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box; line-height: 1.5; margin: 0px 0px 10px; padding-left: 5px;&quot;&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-SG&quot; style=&quot;box-sizing: border-box;&quot;&gt;Constructing an Answer with Retrieval Augmented Generation (RAG)&lt;/span&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;span data-ccp-props=&quot;{&amp;quot;134233117&amp;quot;:false,&amp;quot;134233118&amp;quot;:false,&amp;quot;201341983&amp;quot;:0,&amp;quot;335551550&amp;quot;:0,&amp;quot;335551620&amp;quot;:0,&amp;quot;335559738&amp;quot;:0,&amp;quot;335559739&amp;quot;:0,&amp;quot;335559740&amp;quot;:257}&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;/li&gt;&lt;li lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box; line-height: 1.5; margin: 0px 0px 10px; padding-left: 5px;&quot;&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-SG&quot; style=&quot;box-sizing: border-box;&quot;&gt;Understanding (Vector) Embedding Search&lt;/span&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;span data-ccp-props=&quot;{&amp;quot;134233117&amp;quot;:false,&amp;quot;134233118&amp;quot;:false,&amp;quot;201341983&amp;quot;:0,&amp;quot;335551550&amp;quot;:0,&amp;quot;335551620&amp;quot;:0,&amp;quot;335559738&amp;quot;:0,&amp;quot;335559739&amp;quot;:0,&amp;quot;335559740&amp;quot;:257}&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;/li&gt;&lt;li lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box; line-height: 1.5; margin: 0px 0px 10px; padding-left: 5px;&quot;&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-SG&quot; style=&quot;box-sizing: border-box;&quot;&gt;Why the Use of Embeddings in Retrieval Reduces Interpretability&lt;/span&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;span data-ccp-props=&quot;{&amp;quot;134233117&amp;quot;:false,&amp;quot;134233118&amp;quot;:false,&amp;quot;201341983&amp;quot;:0,&amp;quot;335551550&amp;quot;:0,&amp;quot;335551620&amp;quot;:0,&amp;quot;335559738&amp;quot;:0,&amp;quot;335559739&amp;quot;:0,&amp;quot;335559740&amp;quot;:257}&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;/li&gt;&lt;li lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box; line-height: 1.5; margin: 0px 0px 10px; padding-left: 5px;&quot;&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-SG&quot; style=&quot;box-sizing: border-box;&quot;&gt;Embedding Search in Practice&lt;/span&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;span data-ccp-props=&quot;{&amp;quot;134233117&amp;quot;:false,&amp;quot;134233118&amp;quot;:false,&amp;quot;201341983&amp;quot;:0,&amp;quot;335551550&amp;quot;:0,&amp;quot;335551620&amp;quot;:0,&amp;quot;335559738&amp;quot;:0,&amp;quot;335559739&amp;quot;:0,&amp;quot;335559740&amp;quot;:257}&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;/li&gt;&lt;li lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box; line-height: 1.5; margin: 0px 0px 10px; padding-left: 5px;&quot;&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-SG&quot; style=&quot;box-sizing: border-box;&quot;&gt;Why Embedding Search Leads to Less Reproducible Results&lt;/span&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;span data-ccp-props=&quot;{&amp;quot;134233117&amp;quot;:false,&amp;quot;134233118&amp;quot;:false,&amp;quot;201341983&amp;quot;:0,&amp;quot;335551550&amp;quot;:0,&amp;quot;335551620&amp;quot;:0,&amp;quot;335559738&amp;quot;:0,&amp;quot;335559739&amp;quot;:0,&amp;quot;335559740&amp;quot;:257}&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;/li&gt;&lt;li lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box; line-height: 1.5; margin: 0px 0px 10px; padding-left: 5px;&quot;&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-SG&quot; style=&quot;box-sizing: border-box;&quot;&gt;Reranking with Embedding Search&lt;/span&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;span data-ccp-props=&quot;{&amp;quot;134233117&amp;quot;:false,&amp;quot;134233118&amp;quot;:false,&amp;quot;201341983&amp;quot;:0,&amp;quot;335551550&amp;quot;:0,&amp;quot;335551620&amp;quot;:0,&amp;quot;335559738&amp;quot;:0,&amp;quot;335559739&amp;quot;:0,&amp;quot;335559740&amp;quot;:257}&quot; style=&quot;box-sizing: border-box;&quot;&gt;&lt;/span&gt;&lt;/li&gt;&lt;li lang=&quot;EN-US&quot; style=&quot;box-sizing: border-box; line-height: 1.5; margin: 0px; padding-left: 5px;&quot;&gt;&lt;span data-contrast=&quot;auto&quot; lang=&quot;EN-SG&quot; style=&quot;box-sizing: border-box;&quot;&gt;Hybrid Search and Rerankers&lt;/span&gt;&lt;/li&gt;&lt;/ol&gt;&lt;p&gt;In some ways, writing this piece stressed me out even more. Leaving aside imposter syndrome (I’m self-taught!), I struggled to strike the right balance between conciseness, technical accuracy, and accessibility for librarians with no background in information retrieval.&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;blockquote&gt;&lt;p&gt;A “fun” game I played was pasting chunks of text into frontier LLMs like o3 and Gemini 2.5 Pro and asking them to critique my writing. They were merciless—and made it clear how much expertise is required to be both concise and technically accurate! In case you’re worried about hallucinations, I’ve found that top LLMs are generally solid on the fundamentals of information retrieval—likely because much of the content is open access and there&#39;s a wealth of teaching material available.&lt;/p&gt;&lt;/blockquote&gt;&lt;p&gt;In the end, I came to a conclusion: when writing technical explanations, you can only really choose two out of three of the following:&lt;/p&gt;&lt;p&gt;a) Concise&lt;/p&gt;&lt;p&gt;b) Comprehensive and Accurate&lt;/p&gt;&lt;p&gt;c) Comprehensibility to laypersons&lt;/p&gt;&lt;p&gt;I chose to sacrifice (b) in favour of being concise and comprehensible—even to librarians without a background in AI or information retrieval.&lt;/p&gt;&lt;p&gt;I’m not sure I succeeded. Even now, I’m unhappy with the opening sentence and I think should have changed some of the examples and analogies I used in the piece. I also really itch to revise and add more sections.&lt;/p&gt;&lt;p&gt;Still, even if the piece isn’t 100% technically precise, I believe it points in approximately the right direction for any librarian who wants to understand these emerging tools.&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;h3 style=&quot;background-color: #fafafa; color: #333333; font-family: &amp;quot;Helvetica Neue Light&amp;quot;, HelveticaNeue-Light, &amp;quot;Helvetica Neue&amp;quot;, Helvetica, Arial, sans-serif; font-weight: bolder; margin: 0.75em 0px; padding: 0px; position: relative; text-align: left;&quot;&gt;[ADV] &amp;nbsp;Want to Go Deeper into AI and Information Retrieval?&lt;/h3&gt;&lt;div&gt;As I write this, I’ve just wrapped up my 1.5-hour “Master Class” – Understanding the Fundamentals of AI in Academic Search. In total, 127 librarians and researchers from around the world registered—thank you for your support!&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;But 1.5 hours barely scratches the surface.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;That’s why I’ve teamed up with my colleague, Senior Librarian Bella Ratmelia, to offer a more comprehensive course titled AI-Powered Search in Libraries: A Crash Course on Understanding the Fundamentals for Library Professionals.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;It will be conducted online as part of&lt;a href=&quot;https://force11.org/fsci/post/fsci-2025-courses-abstracts/#L15&quot;&gt; FSCI 2025 (FORCE11 Scholarly Communication Institute in partnership with UCLA Library)&lt;/a&gt; from &lt;i&gt;July 22–24, 6-9 pm (Pacific Time)&lt;/i&gt;, in three sessions of three hours each.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Bella and I co-designed the course using sound pedagogical practices and plenty of interactive activities to help you build your intuition for technical concepts like embeddings, LLMs, and RAG. The extra time will also allow us to experiment with AI academic search tools. Unlike my previous talks, we’ll also dedicate an entire session to testing methodology. No coding knowledge is required.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://whova.com/portal/registration/HDzKwbPHTjc9qUc8a0D3/&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;1080&quot; data-original-width=&quot;1320&quot; height=&quot;524&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgiuZy8Hc7LWV69TMVmYR8Xd1vXeabH_O3P_bUlCGyR3W_vHM0DJ65gstyVS3QYDVa0_JYeZqWnaVrhVoPAPT7kmVfRFrQlUljYF9anHIyo5Df7a8aPerpnAla3n3n1S3_MmcrCkpr0pdGvXd1qf2Kp7381pJ2tixvjRWvqHc7cKBljGQbjWl2ilZkv73hf/w640-h524/2025-05-30%2021_58_11-Presentation%2060.pptx.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;p&gt;I’m excited to have more time to share and learn together. If you’re interested, &lt;a href=&quot;https://whova.com/portal/registration/HDzKwbPHTjc9qUc8a0D3/&quot;&gt;please register here.&lt;/a&gt;&lt;/p&gt;&lt;p&gt;&lt;a href=&quot;https://force11.org/fsci-2025&quot;&gt;Scholarships are available for participants from the Global South&lt;/a&gt;, but applications must be submitted by &lt;i&gt;June 20, 2025&lt;/i&gt;.&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;&lt;b&gt;[Recording]&amp;nbsp;&lt;/b&gt;&lt;span style=&quot;font-size: 18.72px;&quot;&gt;Playing Devil’s Advocate on AI Search Engines&lt;/span&gt;&lt;/h3&gt;&lt;p&gt;This month, I was invited to give several talks. Some weren’t recorded, and others were private sessions.&lt;/p&gt;&lt;p&gt;However, &lt;a href=&quot;https://youtu.be/2cpLsjdaGW8?si=AQDqxXvIaursfW3S&quot;&gt;my keynote at the Librarian Futures Virtual Summit (hosted by Technology from Sage) was recorded and is available for public viewing&lt;/a&gt;.&lt;/p&gt;&lt;div&gt;&lt;div&gt;In this talk, I took a different approach—playing devil’s advocate on the topic of AI search. This is a technique I’ve used before in my career,&amp;nbsp;blogging about then new developmenets like&amp;nbsp;&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2021/06/playing-devils-adovcate-why-newly.html&quot;&gt;citation based mapping services (e.g. Connected papers)&lt;/a&gt;,&amp;nbsp;&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2016/08/are-institutional-repositories-failing.html&quot;&gt;Institutional Repositories&lt;/a&gt;,&amp;nbsp;&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2012/12/playing-devils-advocate-why-you.html&quot;&gt;Web Scale Discovery Services (e.g. Summon)&lt;/a&gt;,&amp;nbsp;&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2010/09/a-few-heretical-thoughts-about-library.html&quot;&gt;mobile related library services&lt;/a&gt;&amp;nbsp;and more in the past.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;I actually have content for a longer version of this talk, which I’ll likely revisit in a future blog post.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Conclusion&amp;nbsp;&lt;/h2&gt;&lt;div&gt;I’ve been exploring information retrieval and AI for almost three years now. This year, I decided to try sharing what I’ve learned in a more structured and deliberate way. I hope my work helps librarians better navigate and understand this fascinating—if sometimes bewildering—topic.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;i&gt;Text has been copyedited with the help of gpt4o.&lt;/i&gt;&lt;/div&gt;</content><link rel='replies' type='application/atom+xml' href='http://musingsaboutlibrarianship.blogspot.com/feeds/617629989690538701/comments/default' title='Post Comments'/><link rel='replies' type='text/html' href='http://www.blogger.com/comment/fullpage/post/4727930222560708528/617629989690538701?isPopup=true' title='0 Comments'/><link rel='edit' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/617629989690538701'/><link rel='self' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/617629989690538701'/><link rel='alternate' type='text/html' href='http://musingsaboutlibrarianship.blogspot.com/2025/05/comparative-review-on-primo-research.html' title='Comparative review on Primo Research Assistant, Scopus AI, Web of Science Research Assistant and a explainer for AI search for librarians'/><author><name>Aaron Tay</name><uri>http://www.blogger.com/profile/02750645621492448678</uri><email>noreply@blogger.com</email><gd:image rel='http://schemas.google.com/g/2005#thumbnail' width='32' height='32' src='//blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjNp5I56cmCnITp9u98mqHmOga9TIDbvdXeuRetlD5Lq7jfNQVbjCCMMyEkblX6PtSR34esdLJ6qarZjGFZC_pAuvyTr93fqVvnlAzrnm2DjjaaH1BTx3XN8lJH69-gnw/s220/profile.jpg'/></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh1rrlkhzPaq52T2FktZ5kDx5Bqcum0EpISk88ydTRZ3DgO7t2A65cDJ4KALg4avqLfehXGeJWDiM4ry2_OtXOYToKEkx-yrtN46M5KpEqrLWY0w4YQfkUKa7DcNXalOoXcyYWMDvdLHSVC_C1ACKV1OgjedH5KX_NOCiqucF4dt4LW0eNYlKUc0aw7rYXb/s72-w640-h218-c/Deep%20Dive%20into%20Three%20AI%20Academic%20Search%20Tools%20_%20Katina%20Magazine.png" height="72" width="72"/><thr:total>0</thr:total></entry><entry><id>tag:blogger.com,1999:blog-4727930222560708528.post-8245258870244415347</id><published>2025-05-12T01:48:00.002+08:00</published><updated>2025-05-12T01:49:31.537+08:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="large language model"/><category scheme="http://www.blogger.com/atom/ns#" term="retrieval augmented generation"/><title type='text'>Ai2 Paper Finder and Futurehouse PaperQA2: More transparent Deep Search for Scholars?</title><content type='html'>&lt;p&gt;As an academic librarian, I’m often asked: “Which AI search tool should I use?” With a flood of new tools powered by large language models (LLMs) entering the academic search space, answering that question is increasingly complex. Both startups and established vendors are rushing to offer &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2025/02/the-rise-of-agent-based-deep-research.html&quot;&gt;“Deep Search” or “Deep Research” solutions&lt;/a&gt;, leading to a surge in requests to test these products. As a working academic librarian even one with a focus on AI search, I have limited time to invest in these new tools, and only ones that stand out in performance, novelty or price are likely to catch my attention.&lt;/p&gt;&lt;p&gt;Honestly, many are unstable or perform poorly, making them unready for widespread use. My current benchmark for performance is Undermind.ai. If a &quot;Deep Search&quot; product performs significantly worse than Undermind in retrieving relevant papers for my sample queries, I quickly lose interest. I know my sample tests aren’t foolproof, but I have to draw the line somewhere!&lt;/p&gt;&lt;p&gt;In today’s blog post, I’ll highlight two academic deep search/research tools that caught my attention:&lt;a href=&quot;https://allenai.org/blog/paper-finder&quot;&gt; Ai2 Paper Finder &lt;/a&gt;and &lt;a href=&quot;https://platform.futurehouse.org/&quot;&gt;Futurehouse’s PaperQA2-based search&lt;/a&gt;. These two tools are currently free, and perform at roughly the same level as my current favourite Deep Search Academic tool - &lt;a href=&quot;http://Undermind.ai&quot;&gt;Undermind.ai&lt;/a&gt;.&amp;nbsp;&lt;/p&gt;&lt;p&gt;More importantly, while Undermind offers a polished, user-friendly interface, its process feels like a black box, revealing little about how it retrieves results.&lt;/p&gt;&lt;p&gt;In contrast, Ai2 Paper Finder displays every step—query analysis, search workflows, and relevance judgments—making it easier for librarians to trust and explain results to researchers.&amp;nbsp;&lt;/p&gt;&lt;p&gt;Similarly, Futurehouse’s Reasoning tab provides detailed insights into the papers found and the evidence considered, empowering users to understand the search process.&amp;nbsp;&lt;/p&gt;&lt;p&gt;For academic librarians, transparent tools foster trust and enhance our ability to guide researchers in navigating and evaluating AI-driven search results effectively.&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;[ADV] Want to learn about the fundamentals of AI in academic search from me?&lt;/h2&gt;&lt;div&gt;&lt;div&gt;I rarely do this, but this is likely of interest to my readers.&amp;nbsp;&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;I am offering a &lt;a href=&quot;https://eventregistration.smu.edu.sg/AISearch&quot;&gt;&quot;Master Class&quot; on May 29, 2025, from 4 PM to 5:30 PM (SGT)&lt;/a&gt;, tailored for librarians navigating the new era of academic AI search. Prioritizing general principles and understanding over specific tools, this session will equip you with the foundational knowledge to excel.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgIrlwZNFjnCJziNS4h3WH8M7FUYxTJvjaeGGiVyjLEsklvaMG9NnyiVxLIsu4C_BJCfoUdDlXVos3JF6NiOzljYbdfqGThKT3bFR97WkZyIDeWjkG00OchJATIy4U4IKqS6XikTLNOTI5tlgJv9oSiK_MYNgQdub06s5rHAZC2UhAwzt1yrxQlfLuI77uy/s1920/MasterClass_Poster_L_AI_Search_Linkedin%20(1).png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;1080&quot; data-original-width=&quot;1920&quot; height=&quot;360&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgIrlwZNFjnCJziNS4h3WH8M7FUYxTJvjaeGGiVyjLEsklvaMG9NnyiVxLIsu4C_BJCfoUdDlXVos3JF6NiOzljYbdfqGThKT3bFR97WkZyIDeWjkG00OchJATIy4U4IKqS6XikTLNOTI5tlgJv9oSiK_MYNgQdub06s5rHAZC2UhAwzt1yrxQlfLuI77uy/w640-h360/MasterClass_Poster_L_AI_Search_Linkedin%20(1).png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;p&gt;As I write this post, there are now less than 10 seats left out of the 50 for my &quot;Master Class.&quot; Can’t attend live? You get recordings! Register &lt;a href=&quot;https://eventregistration.smu.edu.sg/event/AISearch/summary&quot;&gt;here&lt;/a&gt;.&lt;/p&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;h1 style=&quot;text-align: left;&quot;&gt;1. Ai2 Paper Finder&lt;/h1&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjklTHLvPSD8FV6jaIOB-mtM73akeZW8VDm-MUmKoi4lDDEFURmfmTTntOAykCoIIYdhm1z4ofd7d8jmGcUnLTaepQG5sIy_i-O_fUd8JwSsoK9aMagC1nc6bMtSqG0GLmICRR_bVf-NAoTqIlCGS3iDlqsFx6ZPO4oeyVDPi3gv6RjE65Kb9BCIiut7HUy/s793/paperfinder.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;682&quot; data-original-width=&quot;793&quot; height=&quot;550&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjklTHLvPSD8FV6jaIOB-mtM73akeZW8VDm-MUmKoi4lDDEFURmfmTTntOAykCoIIYdhm1z4ofd7d8jmGcUnLTaepQG5sIy_i-O_fUd8JwSsoK9aMagC1nc6bMtSqG0GLmICRR_bVf-NAoTqIlCGS3iDlqsFx6ZPO4oeyVDPi3gv6RjE65Kb9BCIiut7HUy/w640-h550/paperfinder.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Regular readers of this blog are familiar with Ai2, or the &lt;a href=&quot;https://allenai.org/&quot;&gt;Allen Institute for AI&lt;/a&gt;, and in particular the work they do on their Semantic Scholar search engine. Arguably more significant is that they provide much of the dataset behind Semantic Scholar openly for use, and many—if not most—of the startups in the academic search space, such as Elicit.com and Undermind.ai, use this dataset as a base for their search engines.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;More recently Ai2 have also been developing their own fully &lt;a href=&quot;https://allenai.org/language-models&quot;&gt;open LLMs&lt;/a&gt;&amp;nbsp;such as &lt;a href=&quot;https://allenai.org/olmo&quot;&gt;OLMo 2&amp;nbsp;&lt;/a&gt;&amp;nbsp;and now&lt;/div&gt;&lt;div&gt;Ai2 themselves have gotten into the act applying LLMs to search, and in March 2025, they launched &lt;a href=&quot;https://ai2-scholar-qa.allen.ai/&quot;&gt;Ai2 Scholar QA&lt;/a&gt;, their version of Deep Research that generates long-form reports. See &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2025/02/the-rise-of-agent-based-deep-research.html&quot;&gt;my coverage of academic Deep Research tools here&lt;/a&gt;.&lt;/div&gt;&lt;div&gt;&amp;nbsp;&lt;/div&gt;&lt;div&gt;But Deep Research tools are extremely hard to evaluate, requiring not only careful reading of all citation statements to ensure they are faithful to the source but also subjective assessments of the coherence of the writing. As such, my current interest lies in evaluating results from Deep Search alone.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgocae5VB75qNeTg-mygqTlTPc1ysFpzeEkeD-lSjcT9W4xjUtm-00rKc41z0kqXkHRDoeZljaQXxMuLoQu7jzGaJGs9ZIHOBn4Ff5_-njBS14DG9Z4d5xtYaKsVLRsibyp3DG8_CNZ_f0Z8fN7tapMaHZylyM_KO-buSkors7lpWb92QP7OUXQS33PBvY5/s933/deepsearchvsdeepresearch.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;327&quot; data-original-width=&quot;933&quot; height=&quot;224&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgocae5VB75qNeTg-mygqTlTPc1ysFpzeEkeD-lSjcT9W4xjUtm-00rKc41z0kqXkHRDoeZljaQXxMuLoQu7jzGaJGs9ZIHOBn4Ff5_-njBS14DG9Z4d5xtYaKsVLRsibyp3DG8_CNZ_f0Z8fN7tapMaHZylyM_KO-buSkors7lpWb92QP7OUXQS33PBvY5/w640-h224/deepsearchvsdeepresearch.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;This generally involves examining the list of results generated by the Deep Search engine and checking whether most of the known gold-standard results are included.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;I will discuss this further in a future post, but using articles from published systematic reviews as a gold standard often makes the &quot;test&quot; too easy for such Deep Search systems. They will inevitably find the systematic review and, through iterative citation searching, identify most of your gold-standard papers.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Indeed, &lt;a href=&quot;https://allenai.org/blog/paper-finder&quot;&gt;Ai2 has launched a seperate deep search called Ai2 Paper Finder&lt;/a&gt; which recognises that&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;&lt;i&gt;literature search is a multi-step process that involves learning and iterating as you go.&lt;/i&gt;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;It creates an agent that simulates this process by&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;&lt;i&gt;break[ing] down your query into relevant components, searches for papers, follows citations, evaluates for relevance, runs follow-up queries based on the results, and then presents not only the papers, but also short summaries of why the paper is relevant to your specific query.&lt;/i&gt;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;If this sounds familiar, it’s exactly what Undermind does, and Ai2 even name-checks Undermind as&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;blockquote&gt;&lt;i&gt;working in the same domain with a similar goal of finding good and exhaustive sets of papers on a topic.&amp;nbsp;&lt;/i&gt;&lt;/blockquote&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Ai2 Paper Finder: Overview&lt;/h2&gt;&lt;div&gt;So why do I spotlight Ai2 Paper Finder?&lt;/div&gt;&lt;p&gt;First, I believe it is intended to be free (as in free beer) for the foreseeable future. Second, they aim to be as &quot;open as possible.&quot; Although they do not currently open-source their code, they&amp;nbsp;&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;blockquote&gt;&lt;p&gt;&lt;i&gt;&amp;nbsp;plan to release more of our&lt;/i&gt;(their)&lt;i&gt; source code in the future.&lt;/i&gt;&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;/blockquote&gt;&lt;br /&gt;Even without releasing their code, Ai2 Paper Finder is far more transparent about its processes.&lt;br /&gt;&lt;br /&gt;Unlike Undermind.ai, which hides its process behind a generic placeholder screen, Ai2 Paper Finder acts like an open book. It displays every step—think of it as a librarian showing you exactly how they tracked down your sources—making it easier to trust and understand the results.&lt;br /&gt;&lt;br /&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both;&quot;&gt;&amp;nbsp;&amp;nbsp;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEitiymwGAOxaQ38UTa8wEma98c4sp8Ie0sRiauiaPt1J7PgamvPIZCuZsA1h8vBM1qZVCMJ9_zJQatg-IBilqO_0bvL3EKp46NFyQP-JdMGpo6HV7_V10iXau_PwSN0DP5jseYUJ4a0mUKVNxpCDtFuAPbEtwknqk99QPLoYEeop1BuHH4VLDWli7QfMxPx/s715/paperfinder2.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;715&quot; data-original-width=&quot;315&quot; height=&quot;640&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEitiymwGAOxaQ38UTa8wEma98c4sp8Ie0sRiauiaPt1J7PgamvPIZCuZsA1h8vBM1qZVCMJ9_zJQatg-IBilqO_0bvL3EKp46NFyQP-JdMGpo6HV7_V10iXau_PwSN0DP5jseYUJ4a0mUKVNxpCDtFuAPbEtwknqk99QPLoYEeop1BuHH4VLDWli7QfMxPx/w282-h640/paperfinder2.png&quot; width=&quot;282&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Query Analysis and Intent&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;For more context, the&amp;nbsp;&lt;a href=&quot;https://allenai.org/blog/paper-finder&quot;&gt;introductory blog post to the tool&amp;nbsp;&lt;/a&gt;&amp;nbsp;provides a fairly detailed explanation of the inner workings of Ai2 Paper Finder, and I’ll quote liberally from it in italics below.&lt;/div&gt;&lt;p&gt;First, the query analyzer goes to work and does two main things. It determines the query intent, which essentially has two modes:&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Searching for a specific known paper&lt;/li&gt;&lt;li&gt;Searching for a set of papers on a topic&lt;/li&gt;&lt;/ul&gt;&lt;p&gt;&lt;/p&gt;&lt;br /&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjjuljAVvD-DAHSyNvp6gKwGBQ87JEYCIzvuCp5bO_RP13ofG9S6P0tJMCDhAp647wrPDRxreNffZs6PUvQWwkrobspIZ1vLUVakwCbNGdlbD0lSqQjF0kMUNssvsRDoppAAFgUSx3IUHFOqYwHyVWMFwXCw8PT5UD6biGNHj6hazkRaPswqtSmz8jzxhHU/s571/paperfinder-specific.png&quot; style=&quot;margin-left: 1em; margin-right: 1em; text-align: center;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;571&quot; data-original-width=&quot;480&quot; height=&quot;320&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjjuljAVvD-DAHSyNvp6gKwGBQ87JEYCIzvuCp5bO_RP13ofG9S6P0tJMCDhAp647wrPDRxreNffZs6PUvQWwkrobspIZ1vLUVakwCbNGdlbD0lSqQjF0kMUNssvsRDoppAAFgUSx3IUHFOqYwHyVWMFwXCw8PT5UD6biGNHj6hazkRaPswqtSmz8jzxhHU/s320/paperfinder-specific.png&quot; width=&quot;269&quot; /&gt;&lt;/a&gt;&lt;br /&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both;&quot;&gt;&lt;img border=&quot;0&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjX6UyjUvVTGzanTxBQCBRoA1pPcTVm44W90sXLuxy-DZbfee9f-5YB8WQhHm9idN-DBwixvGYaf4vLnV67Zn-XkoqbdLHWko0hHmknS1TiI8o2P6qLOLtfzyNtj1xJe7XYIU0qmfzI20VUkLgpobZ7iginBSjlSPn_hUZjm0-9wzA_Epj_Iw5Of0rH3-4Z/s320/paperfinder-findasetofpapers.png&quot; /&gt;&lt;/div&gt;&lt;br /&gt;&lt;p&gt;&lt;b&gt;My observation :&lt;/b&gt; In my early testing of Ai2 Paper Finder, it erroneously assumed I was searching for a specific known paper too often, which caused failures and occasional crashes. After providing feedback to the Ai2 Paper Finder team, this issue seems mostly resolved.&lt;/p&gt;&lt;p&gt;I find it interesting that unlike most Deep Search or Deep Research tools like Undermind.ai, Gemini Deep Research, OpenAI Deep Research, it doesn&#39;t ask for clarification and goes straight off to search. This could help the system decide which mode to go into.&lt;/p&gt;&lt;p&gt;Second, the query analyzer checks if the query string includes metadata criteria like author, year, or journal. It can also recognize terms like &quot;central,&quot; &quot;popular,&quot; or &quot;recent.&quot;&lt;/p&gt;&lt;p&gt;For example as seen below, it recognizes &quot;Accounting Review&quot; as a journal and the years of publication specified.&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgYeGtaeiOFBfa289aSCfJQCaMPSaj3k5O3vkXTAK7rS6Y1o1pMUfZXMegTNfV45yaYJJKDLdP8pMxPqXLDddZezO5PTY_pLd4ak67fWJ1IUuy1g99V7bnjkhBIcZ1MnHMch-XOm4DFH6x3vc-3cD6jQLQgvjoOSDgfBiYeaYylLk7ADG_9Zl0uX-4Gc8KL/s665/paperfinder3.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;665&quot; data-original-width=&quot;323&quot; height=&quot;640&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgYeGtaeiOFBfa289aSCfJQCaMPSaj3k5O3vkXTAK7rS6Y1o1pMUfZXMegTNfV45yaYJJKDLdP8pMxPqXLDddZezO5PTY_pLd4ak67fWJ1IUuy1g99V7bnjkhBIcZ1MnHMch-XOm4DFH6x3vc-3cD6jQLQgvjoOSDgfBiYeaYylLk7ADG_9Zl0uX-4Gc8KL/w310-h640/paperfinder3.png&quot; width=&quot;310&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;My observation :&amp;nbsp;&lt;/b&gt;It doesn’t yet work for metadata like institutional affiliation.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Search Workflows and Sub-Flows&lt;/h2&gt;&lt;p&gt;&lt;span style=&quot;white-space-collapse: preserve;&quot;&gt;The analyzed query is then passed to the &quot;query planner,&quot; which launches several predefined &quot;workflows,&quot; including:&lt;/span&gt;&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Specific paper search&lt;/li&gt;&lt;li&gt;Semantic search with potential metadata constraints&lt;/li&gt;&lt;li&gt;Pure-metadata queries&lt;/li&gt;&lt;li&gt;Queries involving an author name&lt;/li&gt;&lt;/ul&gt;&lt;p&gt;&lt;/p&gt;&lt;blockquote&gt;&lt;p&gt;&lt;i&gt;Each of these sub-flows (that can be thought of as &quot;sub-agents&quot;) return a set of paper-ids, and in sub-flows that include a semantic criteria, each paper-id is also associated with a list of evidence, and a ranking score reflecting its judged adherence to the semantic criteria.&lt;/i&gt;&amp;nbsp;&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;/blockquote&gt;&lt;p&gt;
&lt;/p&gt;&lt;blockquote dir=&quot;auto&quot;&gt;
&lt;p dir=&quot;auto&quot; style=&quot;white-space: pre-wrap;&quot;&gt;&lt;/p&gt;&lt;/blockquote&gt;&lt;p&gt;&lt;/p&gt;&lt;p dir=&quot;auto&quot; style=&quot;white-space: pre-wrap;&quot;&gt;Like any good system today, reranking is performed, and result sets are:&lt;/p&gt;&lt;blockquote&gt;&lt;p&gt;&lt;i&gt;re-ranked according to a formula that combines the semantic relevance with metadata criteria, such as prioritizing more recent and more highly cited papers. This is influenced by query modifiers such as &quot;recent&quot;, &quot;early&quot;, &quot;central&quot; or &quot;classic&quot; that increase the emphasis on the metadata criteria over the semantic one.&lt;/i&gt;&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;/blockquote&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;The specific paper subflow&lt;/h2&gt;&lt;p&gt;The specific paper sub-flow is designed to help you find the one specific paper you’re thinking of but can’t recall the title. It employs three strategies in parallel:&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Searches the&amp;nbsp;&lt;a href=&quot;https://www.semanticscholar.org/product/api/tutorial&quot;&gt;Semantic Scholar title-search API&lt;/a&gt;&lt;/li&gt;&lt;li&gt;Search for sentences containing the core search terms, focusing on those that also contain a citation and looking for what the majority of them cite (for example, in the query “the alphageometry paper”, it will search for “alphageometry” and see what is cited in its vicinity)&lt;/li&gt;&lt;li&gt;Asks the LLM directly for a paper and verifies its existence via the&amp;nbsp;&lt;a href=&quot;https://www.semanticscholar.org/product/api/tutorial&quot;&gt;Semantic Scholar title-search API&lt;/a&gt;&lt;/li&gt;&lt;/ul&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Observation :&lt;/b&gt;&amp;nbsp;As researchers, you’ve likely experienced wanting to cite a paper you vaguely remember but can’t find. This sub-flow seems designed to address this use case.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Initially, I even mistakenly thought the name Ai2 &quot;Paper Finder&quot; implied it was only for this purpose!&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;While the idea is good, I find it often fails for me. I suspect this is because what a paper actually says may differ from what it’s commonly cited for, which may differ from what you want to cite it for. For example, a paper might be cited for a certain conclusion, but you found a nuance in its research method interesting and want to cite it for that. Given the algorithm, you can see why Ai2 Paper Finder struggles here.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&amp;nbsp;&amp;nbsp;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;&lt;b&gt;Semantic search sub-flow&lt;/b&gt;&lt;/h2&gt;&lt;p&gt;The explanation in the blog post isn’t entirely clear when I read it, but here’s OpenAI&#39;s o3&#39;s interpretation based on the text.&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&amp;nbsp;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjFLnDikHgP0L6rznv4AWNKOUQ7Sk4ecVg7yjpr3SJVbJxxdcZlYHhF1Cfr63YgL8-UnpQLTjyL8wj-sK7DilaTVH1hOo0jj2cCZR-0YfNSX9Z5KYz8-Og2UgryklqOkDz9WbvZEk2cB9s5duBTs__-NSM9iQNIyKRL-_6_1Gnnv9AnB2xufDIyvHYGl1ch/s798/ai2paperfinder-semanticsearchalgo.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;720&quot; data-original-width=&quot;798&quot; height=&quot;578&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjFLnDikHgP0L6rznv4AWNKOUQ7Sk4ecVg7yjpr3SJVbJxxdcZlYHhF1Cfr63YgL8-UnpQLTjyL8wj-sK7DilaTVH1hOo0jj2cCZR-0YfNSX9Z5KYz8-Og2UgryklqOkDz9WbvZEk2cB9s5duBTs__-NSM9iQNIyKRL-_6_1Gnnv9AnB2xufDIyvHYGl1ch/w640-h578/ai2paperfinder-semanticsearchalgo.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;It sound plausible to me and should be close enough I guess...&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Iterative searching&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;&lt;blockquote&gt;&lt;div&gt;&lt;i&gt;The search process then enters another round, based on the most relevant papers so far. In this round, an LLM reformulates more queries based on the original query and the text of the found relevant papers, and sends them to the above-mentioned indices. Additionally, it does both forward and backward citation tracking based on the most relevant papers, which is again followed by LLM-based relevance judgment. The process continues for several rounds and stops when it either finds enough papers or scans too many candidates.&lt;/i&gt;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Observation:&amp;nbsp;&lt;/b&gt;Nothing surprising here, but the fact that it performs both forward and backward citation tracking on the most relevant papers explains why using published systematic review papers as a gold standard is not a good test. The system will quickly find the systematic review, identify it as highly relevant, and citation searching will uncover the rest. (I’ve seen this in practice.)&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;h2&gt;How relevance is judged&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;&lt;span&gt;&lt;/span&gt;&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;&lt;i&gt;&lt;span style=&quot;background-color: #faf2e9; color: #0a3235; font-family: __telegraf_df7567, __telegraf_Fallback_df7567; font-size: 16px;&quot;&gt;Developing a reliable method for relevance judgment was challenging. While the system is still evolving, we found a &quot;mini breakthrough&quot; that significantly improved results and usability: ask an LLM to break the user&#39;s semantic criteria into multiple semantic sub-criteria, to be verified separately. For example, in the above query about dialog datasets, these sub-criteria would be “introducing an unscripted dialog dataset”, “English language”, “annotated speaker properties” and “relation between dialog and annotation”. Each of these titles also includes a brief elaboration. Then, the relevance-judging LLM is asked to first provide adherence to each of the individual sub-criteria, which are only then combined into a final score and final judgement.&lt;/span&gt;&lt;/i&gt;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Observation :&lt;/b&gt; This seems to involve asking an LLM to assess each candidate paper based on sub-criteria, somewhat similar to how LLMs are used for screening in systematic reviews, where they score title/abstracts based on PICO (Population, Intervention, Comparison, Outcome).&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEg9y6XfVULu2CDL_ItRQMRHVUYskfEmjBu6fdjZ0fgp616dznXqEx8MIZNTlnDuwp67SfyYEHvABJxAoz5-MfTaGEpmyGZEz1L1CxjqI0iGToIIrdGSOO15c56RpNqot8Fqum_ty5JxQhhYXCpHc7DmwOnIU4Or-knx9w9SDN6MY2w1ywTYW1v0qSVTlgv1/s618/aipaper2-relevance-sub-1.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;618&quot; data-original-width=&quot;365&quot; height=&quot;640&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEg9y6XfVULu2CDL_ItRQMRHVUYskfEmjBu6fdjZ0fgp616dznXqEx8MIZNTlnDuwp67SfyYEHvABJxAoz5-MfTaGEpmyGZEz1L1CxjqI0iGToIIrdGSOO15c56RpNqot8Fqum_ty5JxQhhYXCpHc7DmwOnIU4Or-knx9w9SDN6MY2w1ywTYW1v0qSVTlgv1/w378-h640/aipaper2-relevance-sub-1.png&quot; width=&quot;378&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh_bt9O8xvVpVSyjLRbRS6-0AoRolgtiW5E4nJL0ENwyShfTjR2zbbjozDyuo-Hx0EkGss6j6l4xvgsHGaiKd1q5swyBloOzmjciG8QW9-MEOLkg9igi9XzT3l6ydSaPFwhvMhxVLKX2005POZ68kfgmCH9HRfmTfKuzFXxpZT-6n9lrPdqEJTiUzHxrXxq/s1392/aipaper2-relevance-sub.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;411&quot; data-original-width=&quot;1392&quot; height=&quot;188&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh_bt9O8xvVpVSyjLRbRS6-0AoRolgtiW5E4nJL0ENwyShfTjR2zbbjozDyuo-Hx0EkGss6j6l4xvgsHGaiKd1q5swyBloOzmjciG8QW9-MEOLkg9igi9XzT3l6ydSaPFwhvMhxVLKX2005POZ68kfgmCH9HRfmTfKuzFXxpZT-6n9lrPdqEJTiUzHxrXxq/w640-h188/aipaper2-relevance-sub.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;In the above example, the query is &quot;Paper showing Google Scholar can be used for systematic review instead of using multiple databases&quot; and&amp;nbsp; three sub criteria used by the LLM to judge relevancy are&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Google Scholar as a primary database - judged &quot;perfectly relevant&quot;&lt;/li&gt;&lt;li&gt;Systematic Review Methodology -judged &quot;somewhat relevant&quot;&lt;/li&gt;&lt;li&gt;Comparison to multiple databases - judged &quot;perfectly relevant&quot;&lt;/li&gt;&lt;/ul&gt;&lt;div&gt;with overall judgement of &quot;relevant&quot;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;This is somewhat similar to how LLMs are used for screening in systematic reviews and they are instructed to score title abstracts based on PICO.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Impact of citing papers&lt;/h2&gt;&lt;div&gt;&lt;blockquote&gt;&lt;i&gt;In the relevance judgment phase, an LLM goes over all the candidate papers and judges how well they match the semantic criteria in the user&#39;s query. For each candidate paper, we format it as markdown text that includes the paper&#39;s title, abstract, and the returned snippets from the search, where the snippets are ordered by the paper order, and include the section titles of the sections in which they appeared. If the returned snippets include snippets that cite this paper, these snippets are added as well.&lt;/i&gt;&lt;/blockquote&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;p&gt;&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEi9h-Dc7BXUJiyyZUI4gjoxmMVrcx4fQu8DiGLiARHZkxAbawTRXitDaGg98tck1hyphenhyphenOjhmtvMW1d098gDpMVRJo15PDK_Lu83uCtptnqarDMBHuLxpT2DQ8q6-KnfW-GGY4x3gsKe3TrLzrVbnS3P938cQRyn2VK39qbTBpML5xyesVPa3GGaB4nc9037gh/s905/paperfinder-relevantpassages.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;600&quot; data-original-width=&quot;905&quot; height=&quot;424&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEi9h-Dc7BXUJiyyZUI4gjoxmMVrcx4fQu8DiGLiARHZkxAbawTRXitDaGg98tck1hyphenhyphenOjhmtvMW1d098gDpMVRJo15PDK_Lu83uCtptnqarDMBHuLxpT2DQ8q6-KnfW-GGY4x3gsKe3TrLzrVbnS3P938cQRyn2VK39qbTBpML5xyesVPa3GGaB4nc9037gh/w640-h424/paperfinder-relevantpassages.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&amp;nbsp;&lt;p&gt;&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEhI8Dnv2CFBbcPVbMa-1cTU1qo1MirjGA5WEjb4cP6WUcc1dQL9kSkUjAEH4yEUD2fHjlWENceFdykfGtkFAu7yfc7K_FElhrWHSTX9OY_ZpaYM4oGaRkjnLHgBXOHwOu1bAXyHuaj3tsRVuqV3MzZTY2d6ZF8tXJlw0_4vJFHs02Id2pO22OBrCnXeHfhb/s1126/aipaper2-relevance-citing.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;487&quot; data-original-width=&quot;1126&quot; height=&quot;276&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEhI8Dnv2CFBbcPVbMa-1cTU1qo1MirjGA5WEjb4cP6WUcc1dQL9kSkUjAEH4yEUD2fHjlWENceFdykfGtkFAu7yfc7K_FElhrWHSTX9OY_ZpaYM4oGaRkjnLHgBXOHwOu1bAXyHuaj3tsRVuqV3MzZTY2d6ZF8tXJlw0_4vJFHs02Id2pO22OBrCnXeHfhb/w640-h276/aipaper2-relevance-citing.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;b&gt;Observation :&amp;nbsp;&lt;/b&gt;There’s a “Relevant evidence from 4 citing papers” section, presumably affecting relevance?&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Fast mode vs work harder mode&lt;/h2&gt;&lt;p&gt;&lt;i&gt;As you may imagine, the full process is effective but rather lengthy... For this reason, we also introduce a fast mode that does less work: it retrieves fewer papers in the initial stage based on the user&#39;s semantic criteria without additional reformulations and without the follow-up iterative procedure.&lt;/i&gt;&lt;/p&gt;&lt;p&gt;&lt;i&gt;This fast mode is the mode that runs by default, so you don&#39;t wait two or three minutes for each response. Based on the results, you can then ask Paper Finder to &quot;work harder&quot; in which case it will invoke the more exhaustive mode described above. You can also invoke the exhaustive mode directly by asking Paper Finder for &quot;an extensive set of paper about X&quot; or something similar in the original query. This way you can get good and (relatively) fast answers to 80% of your queries, while getting higher quality and exhaustiveness for the 20% of queries that require the exhaustive mode.&lt;/i&gt;&lt;/p&gt;&lt;div&gt;&lt;b&gt;Observation :&lt;/b&gt; Fast mode works well, but I&#39;ve found asking it to “work harder” often doesn’t significantly improve results which is by design I suspect.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Limitations&lt;/h2&gt;&lt;/div&gt;&lt;/div&gt;&lt;blockquote style=&quot;border: none; margin: 0px 0px 0px 40px; padding: 0px;&quot;&gt;&lt;div&gt;&lt;div&gt;&lt;div style=&quot;text-align: left;&quot;&gt;&lt;i&gt;As for the semantic queries, while we get top results on academic benchmarks such as LitSearch and Pasa, there is still a lot to do. In particular, we’ve already identified several areas which are particularly challenging: queries when the user does not know the right vocabulary, overly long and rambling queries where the user enters a very long, paragraph-length description of their intents, some queries that involve a combination of multiple semantic criteria where each of them appears in different part of the paper, and queries that search for things that are inherently hard to search for using an index (e.g. numeric ranges such as in &quot;training techniques for models with more than 7b parameters&quot;, or negated semantic criteria as in &quot;fairness papers that do not discuss race or gender&quot;)...&amp;nbsp;Finally, the system is now strong but quite rigid, and while it is influenced by LLM decisions, the flows are predominantly shaped by the researchers and engineers in our team. This is powerful and effective but also limiting (as an almost trivial example, a query like &quot;the bert paper and the roberta paper&quot; is currently not handled well, and could be trivially supported by a more dynamic, LLM-controlled flow). Going forward, we&#39;d like to see more and more decisions delegated to the LLM, supporting more dynamic and ad-hoc flows.&lt;/i&gt;&lt;/div&gt;&lt;/div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;p&gt;&lt;b&gt;Observation :&lt;/b&gt;&lt;a href=&quot;https://leehanchung.github.io/blogs/2025/02/26/deep-research/&quot;&gt; In The Differences between Deep Research, Deep Research, and Deep Research&lt;/a&gt;&amp;nbsp;, the author classifies deep research tools in two dimensions.&lt;/p&gt;&lt;p&gt;a) Depth of search - Shallow vs Deep&lt;/p&gt;&lt;p&gt;b) Hand crafted vs Trained&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjJAC0a1fQPTqQseAroMA9A2rqyq8HWqG9fOlQCWLhIVw4LjlVfGpB9xXk8cWVcKgaDZl5sALb9hstw00fQrvOfwBYw9VCRvjr8_oHLUD84WLgn1NGrXMPUfHSbBWrI_Gw9F-eu7V0scovsmOSTz-MoxjdcvnGYPpM9ubL__gIYUL3ZP5YZO73upL2h1TX4/s819/handcratfedvstrained.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;720&quot; data-original-width=&quot;819&quot; height=&quot;281&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjJAC0a1fQPTqQseAroMA9A2rqyq8HWqG9fOlQCWLhIVw4LjlVfGpB9xXk8cWVcKgaDZl5sALb9hstw00fQrvOfwBYw9VCRvjr8_oHLUD84WLgn1NGrXMPUfHSbBWrI_Gw9F-eu7V0scovsmOSTz-MoxjdcvnGYPpM9ubL__gIYUL3ZP5YZO73upL2h1TX4/s320/handcratfedvstrained.png&quot; width=&quot;320&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;p&gt;Ai2 Paper Finder is clearly on the handcrafted side, with specifically defined sub-flows (rather than the LLM being trained via reinforcement learning), making it vulnerable to performing poorly on unexpected scenarios, as noted above (e.g., long, rambling queries).&lt;/p&gt;&lt;blockquote style=&quot;border: none; margin: 0px 0px 0px 40px; padding: 0px;&quot;&gt;&lt;p style=&quot;text-align: left;&quot;&gt;&lt;i&gt;Another area we recently started to explore is interactivity and multi-turn interactions. Real world search is not a one-shot process: once there are results, the searcher may like to refine the query. This refinement may refer to the returned results (&quot;these are great but now focus on memory efficiency&quot; or &quot;the third and fourth are great, can you find more like these&quot;), and we&#39;d like the follow-up queries to take this into account.&amp;nbsp;&lt;/i&gt;&lt;/p&gt;&lt;/blockquote&gt;&lt;p&gt;&lt;b&gt;&amp;nbsp;Observation : &lt;/b&gt;Agreed&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;p&gt;&lt;b&gt;Overall assessment :&amp;nbsp;&lt;/b&gt;I spent considerable time testing Ai2 Paper Finder against Undermind, and its performance is very close, though it can be less robust, with occasional unexpected failures (hopefully now corrected). I suspect Ai2 Paper Finder is still more sensitive to unexpected query inputs than Undermind.&amp;nbsp;&lt;/p&gt;&lt;p&gt;UI-wise, Undermind.ai offers a polished experience, while Ai2 Paper Finder has room for improvement, though I’m pleased they adopted my suggestion to add more filters for criteria and relevance.&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiZNDyY-myH6OHTlw_Gz_qC6Spa-oij0dPV0AaoROcEQyeXCKgKEaAmCdWhAirlZ0EhxxcYAHGEjdqG5bR1ERHma-gv_KB2877vAMXEqI_9Ki0dLqqzet1Gz4QzVc_hJeFZdeXnR4S0ELDEdOX1F6Y_RP5NXkYe5Pw98dkM3DVh96B6KDJxSEjtKLCtjMBy/s555/paperfinder-filter.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;555&quot; data-original-width=&quot;298&quot; height=&quot;320&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiZNDyY-myH6OHTlw_Gz_qC6Spa-oij0dPV0AaoROcEQyeXCKgKEaAmCdWhAirlZ0EhxxcYAHGEjdqG5bR1ERHma-gv_KB2877vAMXEqI_9Ki0dLqqzet1Gz4QzVc_hJeFZdeXnR4S0ELDEdOX1F6Y_RP5NXkYe5Pw98dkM3DVh96B6KDJxSEjtKLCtjMBy/s320/paperfinder-filter.png&quot; width=&quot;172&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div style=&quot;font-weight: bold;&quot;&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;/div&gt;&lt;div&gt;Criteria filters are particularly helpful for focusing on the most important aspects of your query that must match.&lt;/div&gt;&lt;div style=&quot;font-weight: bold;&quot;&gt;&lt;br /&gt;&lt;/div&gt;&lt;p&gt;&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;h1 style=&quot;text-align: left;&quot;&gt;&lt;b&gt;2. Futurehouse Platform search&lt;/b&gt;&lt;/h1&gt;&lt;div&gt;Besides the commercial Undermind.ai, which dates back to early 2024, Futurehouse’s PaperQA2 was another early “agentic search” in the academic search space. An open-source project, you can read details in the &lt;a href=&quot;https://arxiv.org/abs/2409.13740&quot;&gt;preprint&lt;/a&gt; or &lt;a href=&quot;https://www.futurehouse.org/research-announcements/paperqa2-achieves-sota-performance-on-rag-qa-arena-science-benchmark&quot;&gt;this blog post&lt;/a&gt;.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both;&quot;&gt;More recently, they launched &lt;a href=&quot;https://platform.futurehouse.org/&quot;&gt;a platform offering “AI agents for scientific discovery.”&lt;/a&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both;&quot;&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both;&quot;&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEignGa9FAttjjGUwQEyQGscfQgfZ6ms6OA0i-yAZ_fUx8HQZHfbY-Mko-yt1GGxby6szKnqgh7jmhVIfUxiHrdS7xSJeyHMwGL4hV8OIXu8mFPg1cG2_K4hKVIYmOjBcKXDKUpaMhv0m0unrJOxpCZ6QVqoUADmy2gA7JJxm8S4G9Wcwv_tcMMpuvLJBSQX/s958/futurehouse.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;666&quot; data-original-width=&quot;958&quot; height=&quot;444&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEignGa9FAttjjGUwQEyQGscfQgfZ6ms6OA0i-yAZ_fUx8HQZHfbY-Mko-yt1GGxby6szKnqgh7jmhVIfUxiHrdS7xSJeyHMwGL4hV8OIXu8mFPg1cG2_K4hKVIYmOjBcKXDKUpaMhv0m0unrJOxpCZ6QVqoUADmy2gA7JJxm8S4G9Wcwv_tcMMpuvLJBSQX/w640-h444/futurehouse.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;p&gt;&lt;/p&gt;&lt;p&gt;They offer three types of searches (leaving aside “Phoenix,” which is designed for chemistry-only tasks):&lt;/p&gt;&lt;p&gt;1.&amp;nbsp;&lt;b&gt;Crow&lt;/b&gt;: Based on the original PaperQA2, good for specific questions.&lt;/p&gt;&lt;p&gt;2.&amp;nbsp;&lt;b&gt;Owl&lt;/b&gt;: Used for precedent search, ideal for checking if something has been done before. This reminds me of Ai2 Paper Finder’s “specific paper” mode.&lt;/p&gt;&lt;p&gt;3.&amp;nbsp;&lt;b&gt;Falcon&lt;/b&gt;: Used for Deep Search, producing long reports with many sources, likely comparable to Ai2 Paper Finder and Undermind in standard topic searches.&lt;/p&gt;&lt;p&gt;The data source they are using is&amp;nbsp; &quot;38 million papers on PubMed, 500,000+ clinical trials&quot; and Open Access papers.&lt;/p&gt;&lt;p&gt;Note that while, PaperQA2 is open source, these new agents are not. You can access them via an API or the free web interface.&amp;nbsp;&lt;/p&gt;&lt;p&gt;&lt;br /&gt;&lt;/p&gt;&lt;h2&gt;So why do I spotlight Futurehouse?&lt;/h2&gt;&lt;div&gt;It’s &lt;i&gt;currently&lt;/i&gt; free, and the results are good, though I haven’t tested it as extensively as Ai2 Paper Finder.&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;How does it work?&lt;/h2&gt;&lt;div&gt;My experience with “Owl” for finding specific papers I vaguely recall has been underwhelming, for the same reasons as Ai2 Paper Finder. So, I’ll focus on Falcon (Deep Search) and Crow (Concise Search).&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Falcon Deep Search vs Crow Concise Search&lt;/h2&gt;&lt;div&gt;Interestingly, when I tested the same query, Crow (Concise Search) took longer than Falcon (Deep Search), contrary to my expectations!&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjMytZ2qgaTxARRhzQ6LYFQMEgwOjZwiq9vRrcSeCJUWXqR0P9rxsdeOVSYoAmBrtiWJaXkGLIthxPbN4fC_kHyorm9LbJSvlQDQkg-Algs2N7lRxmvKnMP45-m7n09tAfQaj5kHtSzeBZajzLChLh1cKFAaDlAFPevttGJVqPqU0_FFwDFWbRNDJxmBYsh/s1260/falconvscrow1.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;294&quot; data-original-width=&quot;1260&quot; height=&quot;150&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjMytZ2qgaTxARRhzQ6LYFQMEgwOjZwiq9vRrcSeCJUWXqR0P9rxsdeOVSYoAmBrtiWJaXkGLIthxPbN4fC_kHyorm9LbJSvlQDQkg-Algs2N7lRxmvKnMP45-m7n09tAfQaj5kHtSzeBZajzLChLh1cKFAaDlAFPevttGJVqPqU0_FFwDFWbRNDJxmBYsh/w640-h150/falconvscrow1.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;For the same query, Falcon (Deep Search) took 5 minutes, while Crow (Concise Search) took over double the time—12 minutes!&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjXMmEIQaUeN5AbnZB6Fft_wHQHAMlOpupa_CAtJfoW0spJx3EF30PSIia69CITtH7qzbscrA_7Bahv4nvhXCksCjt89esgKnyaG-gHgh5uDoOk6ZIIfJJrGa1EZIwSLU1YWdcsitX031F0FsEF9bgE3BFUqTfxAfrPt3KflrJdmhe_wE5GVN71UMQdeumz/s1227/falconvscrow2.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;276&quot; data-original-width=&quot;1227&quot; height=&quot;144&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjXMmEIQaUeN5AbnZB6Fft_wHQHAMlOpupa_CAtJfoW0spJx3EF30PSIia69CITtH7qzbscrA_7Bahv4nvhXCksCjt89esgKnyaG-gHgh5uDoOk6ZIIfJJrGa1EZIwSLU1YWdcsitX031F0FsEF9bgE3BFUqTfxAfrPt3KflrJdmhe_wE5GVN71UMQdeumz/w640-h144/falconvscrow2.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Despite spending less time, Falcon ran more queries (71 vs. 48). In terms of sources considered, Falcon evaluated 45, found 20 relevant, and referenced 14, while Crow considered 110, found 4 relevant, and referenced only 2!&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;I think Falcon isn’t necessarily “deeper” but produces a broader literature review than Crow. For example,&lt;a href=&quot;https://platform.futurehouse.org/trajectories/4de1444a-a205-4486-8d72-12c39f15ef4c&quot;&gt; Falcon’s output&lt;/a&gt; spanned 22,144 characters, while &lt;a href=&quot;https://platform.futurehouse.org/trajectories/8e83c5e6-5cee-4275-bc3e-a6aa25d69ed7&quot;&gt;Crow’s output&lt;/a&gt; was only 3,247 characters, reflected in the “result tokens” used. Crow’s answer dives straight into studies estimating Google Scholar’s index size, while Falcon provides a longer, scene-setting report.&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;In this case, Falcon also performed better, finding 20 relevant papers vs. Crow’s 4. To be fair, Falcon’s broader literature review makes it easier to find relevant papers. Still, looking at the actual output Falcon clearly outperformed Crow in surfacing studies estimating Google Scholar’s index size, so I’ll focus on Falcon Deep Search for the rest of this post.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Falcon Deep Search interface&lt;/h2&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEi4sTlefYHHpaAzFvBhdwsmi-vblqIByDP0pI2KZB3KxW5oqUhNR612pgTEKxL2pTRjTTZQw5vO2FkxFyXeb0fWh55bxQJ8CBJ9ftANHtI2kA6w1RmrtXcpFNOr_p_yNcRXsBU9spvJjjQrhU8tmgKsulvIiLKvaySD37cl__Rto5bMq9YI2JclQFwaDn0i/s1228/falconinterface.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;778&quot; data-original-width=&quot;1228&quot; height=&quot;406&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEi4sTlefYHHpaAzFvBhdwsmi-vblqIByDP0pI2KZB3KxW5oqUhNR612pgTEKxL2pTRjTTZQw5vO2FkxFyXeb0fWh55bxQJ8CBJ9ftANHtI2kA6w1RmrtXcpFNOr_p_yNcRXsBU9spvJjjQrhU8tmgKsulvIiLKvaySD37cl__Rto5bMq9YI2JclQFwaDn0i/w640-h406/falconinterface.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;div&gt;Falcon’s interface is the same as Crow’s and Owl’s.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Besides task details (which reveal the LLM models used), there are three tabs: “Results,” “Reasoning,” and “References.”&lt;/div&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;References Tab&lt;/h2&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiv7SdQhS-3B1xqymw5lP7M5kP5rzoZ4q7UJBlVGvFutunejF-fpb58yQ9Fajbr4R6Gf6F1Z7jTXtLw2sZJgTcFTweHxclOQWmziUqN9lr3b_A6qQSemM1TCvl7HEbGpxK5o9EIRdp9FEML84Q78YcEPSqJm0NjVkDVfIYenTt4B1JiwtvQ8p1An0w2rkFH/s1251/falconinterface-reference.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;692&quot; data-original-width=&quot;1251&quot; height=&quot;354&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiv7SdQhS-3B1xqymw5lP7M5kP5rzoZ4q7UJBlVGvFutunejF-fpb58yQ9Fajbr4R6Gf6F1Z7jTXtLw2sZJgTcFTweHxclOQWmziUqN9lr3b_A6qQSemM1TCvl7HEbGpxK5o9EIRdp9FEML84Q78YcEPSqJm0NjVkDVfIYenTt4B1JiwtvQ8p1An0w2rkFH/w640-h354/falconinterface-reference.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;div&gt;The References tab lists the references used in the generated answers. Two things stood out: some references are tagged with labels like “Domain leading,” “Highest quality,” or “Peer reviewed.”&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Under each paper, contexts (both used and unused) are listed (e.g., 1.1, 1.2, 1.3).&amp;nbsp;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjWhl42WMKlnsUd6JJPX8jkDzPNzzyB_eSna_c7u08QwhAva_C8qTVHrJGBkog_B4KMNNkWDAdMTdSWxWrAqTMWl8jwHPEGMnDMZeVeLk8h8TP7RzztEh_JoPBh-2h9I5-rMOypsqpohFXwTZEPDD-isHh-3AEuERcrtaAgrDvptP2WtzND4TEgPR8V3f9e/s972/falconinterface-reference2.png&quot; style=&quot;margin-left: 1em; margin-right: 1em; text-align: center;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;670&quot; data-original-width=&quot;972&quot; height=&quot;442&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjWhl42WMKlnsUd6JJPX8jkDzPNzzyB_eSna_c7u08QwhAva_C8qTVHrJGBkog_B4KMNNkWDAdMTdSWxWrAqTMWl8jwHPEGMnDMZeVeLk8h8TP7RzztEh_JoPBh-2h9I5-rMOypsqpohFXwTZEPDD-isHh-3AEuERcrtaAgrDvptP2WtzND4TEgPR8V3f9e/w640-h442/falconinterface-reference2.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;br /&gt;&lt;br /&gt;&lt;div&gt;Mousing over them shows what appears to be a “reasoning trace”—typically a summary like “The excerpt from Author (year) discusses…” Each context (1.1, 1.2, etc.) may start similarly but generally differs.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;h2&gt;Result Tab&lt;/h2&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjsd4hDwU3mhvVXNYZ4y33_Pu8MkI2mvy4v2dWxnIG82-EV4sKvYM9YSp4Oo_b480_vybkOpl0u1cIqSBiaPA0zeki7S9_yhRN12MutKhR09VBhziYQm8tbqmUrpnTcjGVHNm7k9uDpL7xY40MufnS5hoSX76jBTNbsD4bc7HRfogwCjrx3uNm5v9RYcu4V/s1239/falconinterface-reference3.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;836&quot; data-original-width=&quot;1239&quot; height=&quot;432&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjsd4hDwU3mhvVXNYZ4y33_Pu8MkI2mvy4v2dWxnIG82-EV4sKvYM9YSp4Oo_b480_vybkOpl0u1cIqSBiaPA0zeki7S9_yhRN12MutKhR09VBhziYQm8tbqmUrpnTcjGVHNm7k9uDpL7xY40MufnS5hoSX76jBTNbsD4bc7HRfogwCjrx3uNm5v9RYcu4V/w640-h432/falconinterface-reference3.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;&amp;nbsp;&lt;br /&gt;Reasoning tab&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;The Reasoning tab is the most interesting, offering insight into the system’s steps, including papers found and “evidence found.”&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgx0JDqfQE4OSKayDAv7WJ1h-sGZ-Kuntjr8HDASyYUlJTAsRFsCNouBdiVE5Al2jST9il6VFGKbdZznIClR8o8DaUhuNWl3YUQ1vBRgqiXBpGeCXvrgU2BQDgN12GiIOrvf_YG2m3PYtLkFtzZdPJtGhH5ABc_ElO2AbfBKHmQPlTzPGjXpRBY8nNnrryC/s1249/falconinterface-reasoning1.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;816&quot; data-original-width=&quot;1249&quot; height=&quot;418&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgx0JDqfQE4OSKayDAv7WJ1h-sGZ-Kuntjr8HDASyYUlJTAsRFsCNouBdiVE5Al2jST9il6VFGKbdZznIClR8o8DaUhuNWl3YUQ1vBRgqiXBpGeCXvrgU2BQDgN12GiIOrvf_YG2m3PYtLkFtzZdPJtGhH5ABc_ElO2AbfBKHmQPlTzPGjXpRBY8nNnrryC/w640-h418/falconinterface-reasoning1.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEhCmGpqGn8FWhrBZJHHfXX-8thTJNsLAYDYipEaWgBPW33AadXnazrWpQx_AXwKjm4eazkMOlnXpklI7QsMOlx_ibZfvhRXyHAe5hoRQUCRIQgJIFyaNsxUzAdad5ch75OV0rxBnBLAK9dnm1uobsp6MGFUZKGwZ5kEUM0QxUSMQJevHGiNGiOGvFzpfodL/s1272/falconinterface-reasoning2.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;687&quot; data-original-width=&quot;1272&quot; height=&quot;346&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEhCmGpqGn8FWhrBZJHHfXX-8thTJNsLAYDYipEaWgBPW33AadXnazrWpQx_AXwKjm4eazkMOlnXpklI7QsMOlx_ibZfvhRXyHAe5hoRQUCRIQgJIFyaNsxUzAdad5ch75OV0rxBnBLAK9dnm1uobsp6MGFUZKGwZ5kEUM0QxUSMQJevHGiNGiOGvFzpfodL/w640-h346/falconinterface-reasoning2.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h1 style=&quot;text-align: left;&quot;&gt;Conclusion&lt;/h1&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;As the landscape of AI-powered academic search tools continues to evolve, the proliferation of &quot;Deep Search&quot; and &quot;Deep Research&quot; products presents both opportunities and challenges for academic librarians and researchers.&amp;nbsp; To me, Ai2 Paper Finder and Futurehouse’s PaperQA2-based search stand out in a crowded market by offering robust performance, transparency, and free access, with performance that rivals my favourite Undermind.ai.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;In my view, while Undermind retains an edge in user experience and polish, the openness of Ai2 Paper Finder’s processes and Futurehouse’s detailed reasoning traces provide valuable insights into how these tools operate, fostering trust and enabling users to better understand their search workflows.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;These tools demonstrate that effective &quot;deep search&quot; capabilities are becoming more accessible, and the push for transparency could become a key differentiator in this rapidly evolving market. Although both have areas for refinement, from Ai2 Paper Finder’s occasional sensitivity to query inputs and UI improvements, to the need for more extensive testing of Futurehouse&#39;s offerings, their current performance and open approach are commendable. Their emergence signals a healthy dynamism in the field, offering powerful, free alternatives that challenge established players and empower users with greater insight into the AI-driven discovery process. As these and other tools continue to mature, the quest for the ideal AI academic search companion – one that is effective, transparent, and user-friendly – remains an exciting one to watch.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;br /&gt;&lt;/div&gt;</content><link rel='replies' type='application/atom+xml' href='http://musingsaboutlibrarianship.blogspot.com/feeds/8245258870244415347/comments/default' title='Post Comments'/><link rel='replies' type='text/html' href='http://www.blogger.com/comment/fullpage/post/4727930222560708528/8245258870244415347?isPopup=true' title='0 Comments'/><link rel='edit' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/8245258870244415347'/><link rel='self' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/8245258870244415347'/><link rel='alternate' type='text/html' href='http://musingsaboutlibrarianship.blogspot.com/2025/05/ai2-paper-finder-and-futurehouse.html' title='Ai2 Paper Finder and Futurehouse PaperQA2: More transparent Deep Search for Scholars?'/><author><name>Aaron Tay</name><uri>http://www.blogger.com/profile/02750645621492448678</uri><email>noreply@blogger.com</email><gd:image rel='http://schemas.google.com/g/2005#thumbnail' width='32' height='32' src='//blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjNp5I56cmCnITp9u98mqHmOga9TIDbvdXeuRetlD5Lq7jfNQVbjCCMMyEkblX6PtSR34esdLJ6qarZjGFZC_pAuvyTr93fqVvnlAzrnm2DjjaaH1BTx3XN8lJH69-gnw/s220/profile.jpg'/></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgIrlwZNFjnCJziNS4h3WH8M7FUYxTJvjaeGGiVyjLEsklvaMG9NnyiVxLIsu4C_BJCfoUdDlXVos3JF6NiOzljYbdfqGThKT3bFR97WkZyIDeWjkG00OchJATIy4U4IKqS6XikTLNOTI5tlgJv9oSiK_MYNgQdub06s5rHAZC2UhAwzt1yrxQlfLuI77uy/s72-w640-h360-c/MasterClass_Poster_L_AI_Search_Linkedin%20(1).png" height="72" width="72"/><thr:total>0</thr:total></entry><entry><id>tag:blogger.com,1999:blog-4727930222560708528.post-5632526957483615302</id><published>2025-05-02T03:56:00.000+08:00</published><updated>2025-05-02T03:56:11.727+08:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="large language model"/><category scheme="http://www.blogger.com/atom/ns#" term="retrieval augmented generation"/><title type='text'>Testing AI Academic Search Engines - What to find out and how to test (2)</title><content type='html'>&lt;p&gt;Following my recent talk for the Boston Library Consortium, many of you expressed a strong interest in learning how to test the new generation of AI-powered academic search tools. Specifically, evaluating systems using Retrieval-Augmented Generation (RAG) was the top request, surpassing interest in learning more about semantic search or LLMs alone.&lt;/p&gt;&lt;p&gt;This is a crucial topic, as these tools are rapidly entering our landscape. This post outlines my current thinking on practical ways librarians can evaluate and understand them through a series of questions with answers you might be able to obtain via reading the documentation, asking the vendor or light-weight testing of the system directly.&lt;/p&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;What Are We Evaluating? AI Academic Search with RAG&lt;/h2&gt;&lt;p&gt;First, let&#39;s clarify the type of tool we&#39;re discussing (different types of tools may requre different approaches). I&#39;m focusing on systems that:&lt;/p&gt;&lt;p&gt;1. &lt;b&gt;Function as Search Engines: &lt;/b&gt;They use your query specifically to search an academic corpus (unlike general chatbots that might optionally search - as they have &quot;search as a tool&quot; functionality).&lt;/p&gt;&lt;p&gt;2. &lt;b&gt;Generate Summaries with Citations:&lt;/b&gt; They use RAG to produce a synthesized text answer (a paragraph or more) based on retrieved documents, including citations linking back to those sources.&lt;/p&gt;&lt;p&gt;Examples include Elicit, Scite Assistant, Consensus, Scopus AI, and the research assistants for Primo and Web of Science.&amp;nbsp;&lt;/p&gt;&lt;p&gt;These tools often present results with a generated summary alongside the list of source documents, though the exact layout varies&lt;/p&gt;&lt;p&gt;&lt;img border=&quot;0&quot; height=&quot;640&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh4aXcD6jIVxBWYV7Av26ZNerbwoaIA9vldd-lWDYQ9dpvTLdCIM0UjtBsJ-k26rGoM-kyyS9A2aK-MSWecMbqDFjbRf19h442RY4JLV6Q3rIlHcoqbmtVm_A0X2O1EztuaGXidBBZmR9av21sSS1NVXPYy6aj9bHJCh79905vGbZjucQ0e-2uNvdvx5CeU/w426-h640/RAGwireframe-2.png&quot; style=&quot;text-align: justify;&quot; width=&quot;426&quot; /&gt;&lt;/p&gt;&lt;div style=&quot;text-align: justify;&quot;&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;This is of course a simplified layout, there are still many UI decisions to be made. &lt;a href=&quot;https://www.nature.com/articles/s41467-025-58551-6&quot;&gt;Given how latest studies still show RAG systems generate sentences that are not fully supported by valid citations&lt;/a&gt;, it is important for the system to carefully consider how the citations should be displayed in the generated answers and to make it easy to verify generated statements with citations. e.g. Are they inline numbers [1], hover-over highlights?&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Below shows another possible layout from Scite Assistant with the references on the right panel.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgZamnINeHUIZyR7Ej7_1OqiWo-MpdjFJ1Jd2b8KEZXfSBE3xuoR-UvrKpa_zzSMZuE5AoUiRzXKJ4yMdeGqWqiEd1iPitKtrb1-gBqVOa4S1IZpnro8rb5yEVRFsrOjj_eRuNvUzKUksymZKAihpjN98nKvb5oprXSBckqYD_duPfObDVHTc4KQt-K5STt/s1901/sciteassistant-layout.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;690&quot; data-original-width=&quot;1901&quot; height=&quot;232&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgZamnINeHUIZyR7Ej7_1OqiWo-MpdjFJ1Jd2b8KEZXfSBE3xuoR-UvrKpa_zzSMZuE5AoUiRzXKJ4yMdeGqWqiEd1iPitKtrb1-gBqVOa4S1IZpnro8rb5yEVRFsrOjj_eRuNvUzKUksymZKAihpjN98nKvb5oprXSBckqYD_duPfObDVHTc4KQt-K5STt/w640-h232/sciteassistant-layout.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;Examples of what I consider AI academic search includes&amp;nbsp;&lt;a href=&quot;http://Elicit.com&quot;&gt;Elicit.com&lt;/a&gt;, &lt;a href=&quot;http://scite.ai&quot;&gt;Scite assistant*&lt;/a&gt;, &lt;a href=&quot;https://consensus.app/&quot;&gt;Consensus&lt;/a&gt;, &lt;a href=&quot;https://www.elsevier.com/products/scopus/scopus-ai&quot;&gt;Scopus AI&lt;/a&gt;, &lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Product_Documentation/020Primo_VE/Primo_VE_(English)/015_Getting_Started_with_Primo_Research_Assistant&quot;&gt;Primo Research Assistant&lt;/a&gt;, &lt;a href=&quot;https://clarivate.com/academia-government/scientific-and-academic-research/research-discovery-and-referencing/web-of-science/web-of-science-research-assistant/&quot;&gt;Web of Science Research Assistant&amp;nbsp;&lt;/a&gt;&amp;nbsp;and many more (&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/p/list-of-academic-search-engines-that.html&quot;&gt;see list here&lt;/a&gt;)&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;But what should we test and how?&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h2&gt;The RAG Challenge&lt;/h2&gt;&lt;br /&gt;RAG systems work in two main stages:&lt;br /&gt;&lt;br /&gt;&lt;div&gt;1. &lt;b&gt;Retrieval:&lt;/b&gt; Finding relevant documents based on your query.&lt;br /&gt;2. &lt;b&gt;Generation: &lt;/b&gt;Using a Large Language Model (LLM) to synthesize an answer based only on the retrieved documents.&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjyEXpMwWlyUpG1Y7NFbDtdLRM6QF50r-gLaPTaForZDzzwT0W3dXFDkWiQsfQyMkluIa8dpPVQFEREhIOo3A5lrh-pW5tlZEXrV-fFMWSwYBz8p51mw6a0FChemtLXy74iXHBjAy0s1zpvr9Ml-a51kvRIPLwq7CZ4RJDUPThvcBMMUwwMQmDtew2IswMg/s821/ragmodelexample.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;416&quot; data-original-width=&quot;821&quot; height=&quot;324&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjyEXpMwWlyUpG1Y7NFbDtdLRM6QF50r-gLaPTaForZDzzwT0W3dXFDkWiQsfQyMkluIa8dpPVQFEREhIOo3A5lrh-pW5tlZEXrV-fFMWSwYBz8p51mw6a0FChemtLXy74iXHBjAy0s1zpvr9Ml-a51kvRIPLwq7CZ4RJDUPThvcBMMUwwMQmDtew2IswMg/w640-h324/ragmodelexample.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;The special prompt that is built-in to the RAG system could uses a prompt like this&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiJdhyphenhyphenzr7kbdDDnCC-LpdyG7pKhxaMHuYeiHSmrDxPzhMCmOSHmxhFLbJhh72ZbXRVb6rvF-SFKKL9iHvU9NCw3UJcXLP2PZHe24v-ZVDQiHiEtWbCtOmTYA43A7JJ5i55tPnuGPpd1Vzqc27s-xip4UmmAh5hChnSJ-KPCpGQjzZ0Q9akMCyOgiajxW9jS/s624/ragprompt-1.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;237&quot; data-original-width=&quot;624&quot; height=&quot;244&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiJdhyphenhyphenzr7kbdDDnCC-LpdyG7pKhxaMHuYeiHSmrDxPzhMCmOSHmxhFLbJhh72ZbXRVb6rvF-SFKKL9iHvU9NCw3UJcXLP2PZHe24v-ZVDQiHiEtWbCtOmTYA43A7JJ5i55tPnuGPpd1Vzqc27s-xip4UmmAh5hChnSJ-KPCpGQjzZ0Q9akMCyOgiajxW9jS/w640-h244/ragprompt-1.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;The quality of the final answer heavily depends on Stage 1: Retrieval. If the system fails to find the right information, the LLM has nothing good to work with, leading to weak, incomplete, or even misleading answers. Therefore, evaluating the retrieval component is critical.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;A Practical Approach for Librarians&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;While academic researchers have rigorous (and time-consuming) methods for evaluating information retrieval (like TREC), librarians need practical approaches. We need to understand these tools well enough to guide users and make informed decisions, without needing weeks of formal testing.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;The &lt;i&gt;preliminary &quot;framework&quot;&lt;/i&gt; of questions presented here focuses specifically on questions designed to help librarians understand the functional performance and underlying mechanics of these RAG systems.&lt;/div&gt;&lt;div&gt;You can investigate the answers to these questions through checking the documentation and targeted testing. I&#39;ve marked suggestions for hands-on testing with&lt;span style=&quot;color: red;&quot;&gt; &lt;b&gt;[TEST]&lt;/b&gt;&lt;/span&gt;.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;I tested this framework during a recent comparitive review of Primo Research Assistant, Web of Science Research Assistant and Scopus AI (to be published in &lt;a href=&quot;https://katinamagazine.org/content/resource-reviews&quot;&gt;Katina resource reviews&lt;/a&gt;).&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;Key Questions for Evaluating RAG Academic Search Engines&lt;/h3&gt;&lt;div&gt;To structure our evaluation, we can draw inspiration from comprehensive frameworks designed for assessing academic search systems. For instance, &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/31614060/&quot;&gt;Gusenbauer &amp;amp; Haddaway (2020)&lt;/a&gt; provide extensive criteria for evaluating the suitability of search tools, particularly for demanding tasks like systematic reviews.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;The questions below adapt and simplify elements to ones relevant mostly to RAG systems only, focusing on practical insights librarians can gain through documentation review and targeted testing.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Also it&#39;s important to note that this practical testing framework deliberately excludes broader, though equally critical, evaluation areas that can be asked such as:&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Ethical Considerations:&lt;/b&gt;&amp;nbsp;Questions around copyright and ip implications of the training and use of LLMs&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Environmental Impact:&amp;nbsp;&lt;/b&gt;Assessing the computational resources and energy consumption associated with running these often complex AI models.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Let&#39;s break down the evaluation into understanding the retrieval process, the generation process, and the user interface for verification. I&#39;ve marked suggestions for hands-on testing with &lt;span style=&quot;color: red;&quot;&gt;&lt;b&gt;[TEST].&lt;/b&gt;&lt;/span&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Part 1: Understanding the Retrieval Component&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;What content is being searched (The Index)?&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters&lt;/b&gt;: The RAG answer depends entirely on what&#39;s retrieved. Knowing the source index reveals potential coverage gaps.&amp;nbsp;&lt;/li&gt;&lt;ul&gt;&lt;li&gt;Does it include only openly scholarly metadata from Semantic Scholar or OpenAlex (like Undermind.ai, Elicit.com) or does it use a proprietary index (like Scopus or Web of Science core collections)? Does it retrieve over just metadata only or over full-text (open access only, or includes some paywalled?)&lt;/li&gt;&lt;ul&gt;&lt;li&gt;Does it use the whole source or filters away any subsets? This could either be by choice (e.g. Semantic Scholar fuels both Undermind.ai and Elicit.com, yet they show different number of indexed items due to different ingestion criteria and updating strategies, while Scopus AI searches only over Scopus data from 2003 only) or optout by content owner ((like&lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Product_Documentation/020Primo_VE/Primo_VE_(English)/015_Getting_Started_with_Primo_Research_Assistant#:~:text=Any%20collections%20from%20the%20following%20content%20providers%3A%20APA%2C%20DataCite%2C%20Elsevier%2C%20JSTOR%2C%20and%20Conde%20Nast.&quot;&gt;&amp;nbsp;Elsevier in Primo Research Assistant&lt;/a&gt;)&amp;nbsp;&lt;/li&gt;&lt;li&gt;Does it search only holding&#39;s licensed by the institution (e.g. Web of Science Core Collection owned) or the entire index regardless of holding (like Primo Research Assistant)?&amp;nbsp;&amp;nbsp;&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;li&gt;&lt;b&gt;How to check&lt;/b&gt;: Review documentation.&lt;/li&gt;&lt;li&gt;&lt;span style=&quot;color: red;&quot;&gt;&lt;b&gt;[TEST]&lt;/b&gt;&lt;/span&gt;: If unclear whether full text is used, design queries where the answer likely resides only in the full text and see if the system can answer accurately. Check and verify for known content opt-outs &amp;nbsp;by purposely searching for such content.&lt;br /&gt;&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;How does the search actually work (Retrieval Mechanism)?&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters:&lt;/b&gt; Most RAG systems encourage typing your input in natural query. &lt;a href=&quot;https://medium.com/@aarontay/boolean-vs-keyword-lexical-search-vs-semantic-keeping-things-straight-95eb503b48f5&quot;&gt;While keyword search will probably be a part of the retrieval mechanism, it is likely to use other methods&lt;/a&gt;.&lt;b&gt;&amp;nbsp;&lt;/b&gt;This affects relevance performance, interpretability and reproducibility of search.&amp;nbsp;&lt;/li&gt;&lt;ul&gt;&lt;li&gt;Does it use an LLM to translate your natural language query into a Boolean string (like Scopus AI, WoS RA, Primo RA)?&amp;nbsp;&lt;/li&gt;&lt;li&gt;Does it use &quot;semantic&quot; search (like dense vector embeddings or learned sparse retrieval like Elicit)?&lt;/li&gt;&lt;li&gt;Does it do any form of two stage re-ranking? (e.g. Primo Research Assistant reranks the top 30 results with embedding search)&lt;/li&gt;&lt;li&gt;Or Is it a hybrid approach combining multiple methods?&amp;nbsp;&lt;/li&gt;&lt;li&gt;See l&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2025/04/the-reproducibility-and.html&quot;&gt;ast blog post for more details&lt;/a&gt;&lt;/li&gt;&lt;/ul&gt;&lt;li&gt;&lt;b&gt;How to check&lt;/b&gt;: Check documentation though not all details are always mentioned. Still Vendors, particularly traditional providers often highlight if they use natural language processing or semantic search.&lt;/li&gt;&lt;li&gt;&lt;span style=&quot;color: red;&quot;&gt;&lt;b&gt;[TEST]&lt;/b&gt;&lt;/span&gt;: You might check the interface for clues (e.g., some tools show you the generated Boolean query if it uses LLM to generate boolean search strategies and you can check for example to see &lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Product_Documentation/020Primo_VE/Primo_VE_(English)/015_Getting_Started_with_Primo_Research_Assistant#:~:text=Query%20Conversion%20%E2%80%93%20The%20user%27s%20question%20is%20sent%20to%20the%20LLM%2C%20where%20it%20is%20converted%20to%20a%20Boolean%20query%20that%20contains%20a%20number%20of%20variations%20of%20the%20query%2C%20connected%20with%20an%20OR.%20If%20the%20query%20is%20non%2DEnglish%2C%20some%20of%20the%20variations%20will%20be%20in%20the%20query%20language%2C%20and%20the%20other%20variations%20will%20be%20in%20English.&quot;&gt;if Primo Research Assistant really creates boolean search strategies the way the documents describes it.&lt;/a&gt;). You can also check if there is any additional reranking by checking the default relevancy sort of the keyword search against the actual top results. If they are different, there is likely some type of additional reranking.&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Are the search results consistent and explainable (Reproducibility &amp;amp; Interpretability)?&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters&lt;/b&gt;:Interpretability of search results is important for certain use cases (e.g. Systematic reviews). This is mostly a matter of the retrieval mechanism. With regards to reproducbility, if the same query returns different source documents each time, the generated RAG answer will of course also vary!&amp;nbsp;&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2025/04/the-reproducibility-and.html&quot;&gt;LLM generated search strategies are interpretable but semantic search based methods tend to be black boxes&lt;/a&gt;&lt;/li&gt;&lt;li&gt;&lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2025/04/the-reproducibility-and.html&quot;&gt;LLM-generated search strategies and some semantic search methods can be less deterministic than traditional keyword search.&amp;nbsp;&lt;/a&gt;&amp;nbsp;&lt;/li&gt;&lt;li&gt;LLMs might also be used in other parts of the retrieval mechanism such as for query expansion which can also lead to more randomness.&lt;/li&gt;&lt;/ul&gt;&lt;li&gt;&lt;b&gt;How to check:&lt;/b&gt; Understand the retrieval mechanism (see above).&lt;/li&gt;&lt;li&gt;&lt;span style=&quot;color: red;&quot;&gt;&lt;b&gt;[TEST]&lt;/b&gt;&lt;/span&gt;: Assuming the RAG system uses LLM to generate boolean. Run the exact same query multiple times (e.g., 5 times in quick succession). Look at the generated Boolean query in the interface (like Scopus AI, WoS Research Assistant), how much does the generated Boolean change? (I found WoS Research Assistant fairly consistent - perhaps having a different query 1 in 5 times, while Scopus AI varied more).&lt;/li&gt;&lt;li&gt;&lt;span style=&quot;color: red;&quot;&gt;&lt;b&gt;[TEST]&lt;/b&gt;&lt;/span&gt;:Compare the list of top N source documents retrieved each time (where N is the number used for generation, see below). Does the order or composition change significantly? (when tested 5 times in quick succession, you may want to repeat each batch of 5 a few times)&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Does the natural language search understand search constraints (Metadata Parsing)?&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters:&lt;/b&gt; Can you refine your search using natural language for common fields? For example, does &quot;peer-reviewed articles on climate change from 2020-2024&quot; correctly apply filters for publication type and publication date?&lt;/li&gt;&lt;li&gt;&lt;b&gt;How to check: &lt;/b&gt;Documentation might list supported natural language commands. While both RAG search tools have limited support e.g. Primo Research Assistant supports year of publication and limited article types, &lt;a href=&quot;https://webofscience.zendesk.com/hc/en-us/articles/31437630410129-Web-of-Science-Research-Assistant&quot;&gt;Web of Science Research Assistant supports a dazzling array of metadata queries using natural language&lt;/a&gt;.&lt;/li&gt;&lt;li&gt;&lt;span style=&quot;color: red; font-weight: bold;&quot;&gt;[TEST]:&lt;/span&gt; Verify claimed support of metadata parasing by try queries incorporating dates, author names, affiliations, publication types, citation counts, etc. (e.g. for Web of Science Research Assistant try &quot;Papers by Patel from MIT on genomics since 2022,&quot; &quot;Review articles on solar panels&quot;). See if the results reflect these constraints accurately.&amp;nbsp;&amp;nbsp;&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Does it handle non-English queries?&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters:&lt;/b&gt; Can users query in languages other than English, even if the underlying corpus is primarily English? This is often possible with LLM-based query interpretation or multilingual embedding models.&lt;/li&gt;&lt;li&gt;&lt;b&gt;How to check:&lt;/b&gt; Documentation might state language support.&lt;/li&gt;&lt;li&gt;&lt;span style=&quot;color: red;&quot;&gt;&lt;b&gt;[TEST]&lt;/b&gt;&lt;/span&gt;: Input a few queries in another language you understand (e.g., French, Spanish, Chinese) and see if the system retrieves relevant English-language documents.&amp;nbsp;&amp;nbsp;&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Part 2: Understanding the Generation Component&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;What is the exact RAG method used&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters: &lt;/b&gt;RAG is a general term these days and there are many different variants and techniques such as GraphRAG, RAGFusion etc that can lead to quite different results&lt;/li&gt;&lt;li&gt;&lt;b&gt;How to check&lt;/b&gt; : In general, the main way is to check the documentation or ask the vendor.&lt;/li&gt;&lt;/ul&gt;&lt;li&gt;How many retrieved results feed the summary (Top N)?&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters: &lt;/b&gt;This indicates the breadth of information the LLM uses to generate the answer. Is it fixed (e.g., top 5 for Primo RA, top 8 for Scopus AI) or dynamic? &lt;i&gt;Note that the final answer might not cite all N documents if some weren&#39;t deemed relevant by the LLM.&lt;/i&gt;&lt;br /&gt;&lt;b&gt;How to check: &lt;/b&gt;Documentation usually states this.&lt;/li&gt;&lt;li&gt;&lt;span style=&quot;color: red;&quot;&gt;&lt;b&gt;[TEST]:&lt;/b&gt;&lt;/span&gt; Run a search and look at the interface which will typically list all the retrieved references even those that are not used in the answer.&amp;nbsp;&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;What LLM is used for generation?&amp;nbsp;&lt;/li&gt;&lt;ul&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters&lt;/b&gt;: Different LLMs have varying capabilities (though this is often a black box not stated explictly). Some tools might offer choice of models to use (like Scite Assistant).&lt;/li&gt;&lt;li&gt;&lt;b&gt;How to check:&lt;/b&gt; Usually only knowable if stated in the documentation. Often, vendors don&#39;t disclose the specific model.&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Can the generated answer be in non-English?&lt;/li&gt;&lt;ul&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters: &lt;/b&gt;If a user queries in French, will the generated summary also be in French, or will it default to English? Systems vary (e.g., Scopus AI answers in English; Primo RA/WoS RA answer in the query language).&lt;/li&gt;&lt;li&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;How to check: &lt;/b&gt;Documentation might specify.&lt;/li&gt;&lt;li&gt;&lt;b&gt;&lt;span style=&quot;color: red;&quot;&gt;[TEST]:&lt;/span&gt; &lt;/b&gt;Use a non-English query (as tested in Part 1) and observe the language of the generated summary.&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;b&gt;Part 3: Evaluating the User Interface and Verification&lt;/b&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;How are citations displayed and linked?&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;b&gt;Why it matters:&lt;/b&gt; Verification is key. How easily can a user connect and check a specific statement in the generated text back to the supporting evidence in the source document? Are citations inline numbers [1], hover-overs, linked phrases, or listed separately? Is it clear which part of the source supports the claim?&lt;/li&gt;&lt;li&gt;&lt;b&gt;How to check: &lt;/b&gt;Examine the user interface.&lt;/li&gt;&lt;li&gt;&lt;b&gt;&lt;span style=&quot;color: red;&quot;&gt;[TEST]:&lt;/span&gt; &lt;/b&gt;Try to trace a few claims in the generated text back to the cited sources. How easy and accurate is this process? Can you quickly access the source abstract or full text?&lt;/li&gt;&lt;/ul&gt;&lt;/ul&gt;&lt;/div&gt;&lt;/div&gt;&lt;div style=&quot;text-align: left;&quot;&gt;&lt;br /&gt;&lt;/div&gt;&lt;div style=&quot;text-align: left;&quot;&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Important Note: What We Haven&#39;t Covered Yet – Evaluating Output Quality&lt;/h2&gt;It&#39;s crucial to recognize that the questions and tests outlined in this post primarily help us understand how these AI academic search engines are built and function. We&#39;ve focused on dissecting their components: the underlying index, the retrieval methods used, the generation process parameters, and the UI mechanisms for verification. This provides a foundational understanding of the system&#39;s mechanics and potential capabilities or limitations.&lt;/div&gt;&lt;div style=&quot;text-align: left;&quot;&gt;&lt;br /&gt;However, we have deliberately not yet addressed the direct evaluation of the quality of the final output – the generated RAG answer itself. Assessing critical aspects such as:&lt;/div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Accuracy and Faithfulness: Does the generated text correctly represent the information found in the cited sources? Are there factual errors or &quot;hallucinations&quot;?&lt;/li&gt;&lt;li&gt;Relevance: How well does the generated answer actually address the user&#39;s specific query?&lt;/li&gt;&lt;li&gt;Completeness and Conciseness: Does the summary capture the key information from the sources without being overly verbose or missing crucial points?&lt;/li&gt;&lt;li&gt;Usefulness: Is the generated summary genuinely helpful for the user&#39;s information need?&lt;/li&gt;&lt;/ul&gt;&lt;h2&gt;&lt;span style=&quot;font-size: small;&quot;&gt;&lt;span style=&quot;font-weight: 400;&quot;&gt;Evaluating these performance aspects requires different approaches, drawing from both long-established Information Retrieval (IR) evaluation techniques (like assessing the relevance of retrieved documents which heavily affects the RAG result) and newer methods specifically developed for evaluating RAG systems (such as measuring faithfulness, answer relevance, and guarding against hallucination).&amp;nbsp;&lt;/span&gt;&lt;/span&gt;&lt;/h2&gt;&lt;h2&gt;&lt;span style=&quot;font-size: small;&quot;&gt;&lt;span style=&quot;font-weight: 400;&quot;&gt;Measuring this crucial aspect of RAG system performance will be the focus of the next part of this blog series.&lt;/span&gt;&lt;/span&gt;&lt;/h2&gt;&lt;h2&gt;Conclusion&lt;/h2&gt;&lt;h2&gt;&lt;/h2&gt;&lt;h2&gt;&lt;span style=&quot;font-size: small;&quot;&gt;&lt;span style=&quot;font-weight: 400;&quot;&gt;Evaluating these AI academic search tools requires a blend of understanding their mechanics and practical testing. By asking the targeted questions detailed above and performing simple tests, librarians can gain a much deeper feel for a system&#39;s potential strengths, weaknesses, coverage gaps, and operational characteristics. This foundational knowledge is essential for guiding our users effectively and making informed decisions as we integrate these powerful, but still evolving, tools into our library services. Stay tuned for the next post where we will tackle the challenge of evaluating the quality of the answers these systems produce.&lt;/span&gt;&lt;/span&gt;&lt;/h2&gt;&lt;div&gt;&lt;i&gt;This blog post has been edited with the help of Gemini 2.5 Pro.&lt;/i&gt;&lt;/div&gt;&lt;h2&gt;&amp;nbsp;&lt;/h2&gt;&lt;div&gt;&lt;b&gt;Technical note 1 :&amp;nbsp; Interpretability and reproducibility of search&lt;/b&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Based on your understanding of how retrieval works you should already have a sense how non-deterministic or misinterpretable the search is. Roughly you would expect keyword search to be most interpretable and reproducible, with Dense Embedding search being the least interpretable and reproducbilty.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgTwPx-WWJZVuEfLUcCrbcVFgLHKpc417J_Zq-DMZK9OI32dbXhTLqrNu59RnpxkdJ6MedARH7innJMtlxzyjxR5lQEtsTWk6Tp6fj9SI2VHrJ9q9nEuPEZYI5lq1zDma5awirAIdOe5TIbJnW2iRB1KO3WpNDWchv9GxE6v0IgRtHPT3O2V_J1wufJSW7p/s1091/reprodinterpret.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;484&quot; data-original-width=&quot;1091&quot; height=&quot;284&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgTwPx-WWJZVuEfLUcCrbcVFgLHKpc417J_Zq-DMZK9OI32dbXhTLqrNu59RnpxkdJ6MedARH7innJMtlxzyjxR5lQEtsTWk6Tp6fj9SI2VHrJ9q9nEuPEZYI5lq1zDma5awirAIdOe5TIbJnW2iRB1KO3WpNDWchv9GxE6v0IgRtHPT3O2V_J1wufJSW7p/w640-h284/reprodinterpret.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;See &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2025/04/the-reproducibility-and.html&quot;&gt;this blog post for more information&lt;/a&gt;, since retrieval systems may use a blend of such systems, things can be even less clear.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;div&gt;&lt;div&gt;&lt;b&gt;Technical note 2 :&amp;nbsp; multi-lingual support&lt;/b&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;Even though your sources might be mostly in English, many RAG systems are capable of &quot;understanding&quot; your query even if it is inputted in any of the dozen non english languages.&amp;nbsp;&lt;br /&gt;&lt;br /&gt;Why do these new AI search systems generally work with non-english language? This is actually one of the benefits from moving away from just a keyword based search. &lt;br /&gt;&lt;br /&gt;Firstly if it uses LLMs to generate Boolean search strategies, this clearly works even if you input your query into say Chinese as most modern LLMs are multi-lingual and are capable of &quot;understanding&quot; your query and creating an approproiate Boolean in English.&lt;br /&gt;&lt;br /&gt;How about the more common dense embedding models? While not guaranteed (e.g. there are monolingual embeddings like MiniLM) a lot of the dense embeddings used are are built upon transformer models that are pretrained multilingually (e.g. BERT). They learn to map text from multiple languages into a shared vector space where similar concepts are represented close together, regardless of the language used.&lt;br /&gt;&lt;br /&gt;You essentially just need one index to handle multiple languages. Compare this to lexical keyword based methods like BM25 where you need a seperate inverted index for each language! &lt;br /&gt;&lt;br /&gt;&lt;div&gt;&lt;/div&gt;&lt;/div&gt;</content><link rel='replies' type='application/atom+xml' href='http://musingsaboutlibrarianship.blogspot.com/feeds/5632526957483615302/comments/default' title='Post Comments'/><link rel='replies' type='text/html' href='http://www.blogger.com/comment/fullpage/post/4727930222560708528/5632526957483615302?isPopup=true' title='0 Comments'/><link rel='edit' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/5632526957483615302'/><link rel='self' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/5632526957483615302'/><link rel='alternate' type='text/html' href='http://musingsaboutlibrarianship.blogspot.com/2025/05/testing-ai-academic-search-engines-what.html' title='Testing AI Academic Search Engines - What to find out and how to test (2)'/><author><name>Aaron Tay</name><uri>http://www.blogger.com/profile/02750645621492448678</uri><email>noreply@blogger.com</email><gd:image rel='http://schemas.google.com/g/2005#thumbnail' width='32' height='32' src='//blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjNp5I56cmCnITp9u98mqHmOga9TIDbvdXeuRetlD5Lq7jfNQVbjCCMMyEkblX6PtSR34esdLJ6qarZjGFZC_pAuvyTr93fqVvnlAzrnm2DjjaaH1BTx3XN8lJH69-gnw/s220/profile.jpg'/></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh4aXcD6jIVxBWYV7Av26ZNerbwoaIA9vldd-lWDYQ9dpvTLdCIM0UjtBsJ-k26rGoM-kyyS9A2aK-MSWecMbqDFjbRf19h442RY4JLV6Q3rIlHcoqbmtVm_A0X2O1EztuaGXidBBZmR9av21sSS1NVXPYy6aj9bHJCh79905vGbZjucQ0e-2uNvdvx5CeU/s72-w426-h640-c/RAGwireframe-2.png" height="72" width="72"/><thr:total>0</thr:total></entry><entry><id>tag:blogger.com,1999:blog-4727930222560708528.post-5056661138239098138</id><published>2025-04-14T22:04:00.004+08:00</published><updated>2025-04-17T19:41:16.742+08:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="discovery"/><category scheme="http://www.blogger.com/atom/ns#" term="large language modted"/><title type='text'>The reproducibility and interpretability of academic Ai search engines like Primo Research Assistant, Web of Science Research Assistant, Scopus Ai and more</title><content type='html'>&lt;p&gt;I recently compared three academic AI search tools:&lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Product_Documentation/020Primo_VE/Primo_VE_(English)/015_Getting_Started_with_Primo_Research_Assistant&quot;&gt; Primo Research Assistant&lt;/a&gt;,&lt;a href=&quot;https://clarivate.com/academia-government/scientific-and-academic-research/research-discovery-and-referencing/web-of-science/web-of-science-research-assistant/&quot;&gt; Web of Science Research Assistant&lt;/a&gt;, and &lt;a href=&quot;https://www.elsevier.com/products/scopus/scopus-ai&quot;&gt;Scopus AI&lt;/a&gt; for a review article.&lt;/p&gt;&lt;p&gt;&lt;b&gt;Why these three?&lt;/b&gt; Mainly because they are add-ons to extremely well-established academic search engines or databases:&lt;/p&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Primo&lt;/b&gt;: Owned by Exlibris (a Clarivate company), Primo is one of the four major discovery systems used by universities, often serving as the default library search box. It is also, the only one of the three bundled in &quot;free&quot; with Primo.&lt;/li&gt;&lt;li&gt;&lt;b&gt;Web of Science &lt;/b&gt;(WoS): Provided by Clarivate, WoS is the pioneer and oldest of the &quot;Big Three&quot; citation indexes.&lt;/li&gt;&lt;li&gt;&lt;b&gt;Scopus:&lt;/b&gt; Developed by Elsevier, Scopus is a major competitor to Web of Science and is currently used in important global university rankings, such as the Times Higher Education (THE) World University Rankings and the QS World University Rankings.&lt;/li&gt;&lt;/ul&gt;&lt;div&gt;&lt;/div&gt;&lt;p&gt;(Note: Summon, the sister discovery service to Primo, also from Exlibris/Clarivate, launched its &lt;a href=&quot;https://knowledge.exlibrisgroup.com/Summon/Product_Documentation/Searching_in_The_Summon_Service/Search_Features/Getting_Started_with_Summon_Research_Assistant&quot;&gt;Summon Research Assistant&lt;/a&gt; last month. From all appearances, it seems identical to Primo Research Assistant, but I will not discuss it further here.)&lt;/p&gt;&lt;div&gt;&lt;div&gt;I won&#39;t delve into a full comparison of how these three tools are similar and different from each other or from other academic AI search tools in this post. Instead, &lt;i&gt;I want to focus on the reproducibility and interpretability of the search results they provide.&lt;/i&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;Reproducibility of the Search results vs Retrival Augmented Generation Answer&lt;/h3&gt;&lt;div&gt;In general, &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2024/05/retrieval-augmented-generation-and.html&quot;&gt;Retrieval-Augmented Generation (RAG) answers use a Large Language Model (LLM) to synthesize information from top retrieved documents&lt;/a&gt;. Because LLMs are involved, the generated answer can vary slightly even if the underlying search consistently retrieves the exact same top items.&amp;nbsp;&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;The non-deterministic nature of Transformer-based LLMs is well-known to anyone who uses ChatGPT, where the exact same input prompt can yield slightly different responses. &lt;a href=&quot;https://rumn.medium.com/setting-top-k-top-p-and-temperature-in-llms-3da3a8f74832&quot;&gt;Advanced users employing LLM APIs know you can reduce this randomness by adjusting settings like temperature, Top P, and Top K&lt;/a&gt;. However, even with settings aimed at maximum consistency (e.g., temperature=0, Top P=0, Top K=1), some small degree of variability often remains.&amp;nbsp;&amp;nbsp;I understand &lt;a href=&quot;https://www.pamelatoman.net/blog/2023/08/nondeterminism-in-llms/?utm_source=chatgpt.com&quot;&gt;this is caused by factors like parallelization in computation, rounding errors in floating points and more&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Of course, if the retrieved results themselves differ, the RAG-generated answer will almost certainly differ as well. As you will see unlike keyword search based retrieval techniques, non-keyword search based techniques often bring interpretability and reproducibility issues.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;High level view of how retrieval works in Academic AI Search Engines&amp;nbsp;&lt;/h2&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiCgJBjjrUcJiAfa3T9Fb-YM_hCLF2QL4DJ89ktCxPgA2Jga3NlJwGE8_F_X4eFsnsFDGL30_Lq398WuXDkABSynTmIstOo9q2YNHbJLfDcZP2rTTzt1M-HnIn27O01B3ahaz45DwphJNHBVKPG2_LfU6GarkoSzmt6ONLcsrsqBKPlJVi6fyHoYyTdfBAy/s1205/bafkreiglsgw5x7yyloq56ooiqrkhmxc3anynykrg6rn6irilpmzojl7ujm.jpg&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;642&quot; data-original-width=&quot;1205&quot; height=&quot;340&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiCgJBjjrUcJiAfa3T9Fb-YM_hCLF2QL4DJ89ktCxPgA2Jga3NlJwGE8_F_X4eFsnsFDGL30_Lq398WuXDkABSynTmIstOo9q2YNHbJLfDcZP2rTTzt1M-HnIn27O01B3ahaz45DwphJNHBVKPG2_LfU6GarkoSzmt6ONLcsrsqBKPlJVi6fyHoYyTdfBAy/w640-h340/bafkreiglsgw5x7yyloq56ooiqrkhmxc3anynykrg6rn6irilpmzojl7ujm.jpg&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;div&gt;Let&#39;s consider how various academic AI search engines might use your natural language input to find relevant literature. The above image shows my best understanding of how Primo Research Assistant, Web of Science Research Assistant and Scopus Ai work based on the documentation and some tests.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;The approach taken significantly impacts the reproducibility and interpretability of the results. This generally depends on the search mechanism(s) employed.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;This can look very technical, if you having been looking into details of information retrieval but let me try to explain.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;br /&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh6VXau5GtZNYrO1c2uV1aWtqjqQ5L78EClligsrGtfJA0e1N5wPx4WK6rdXPPbyeS1cewb2a-WvFFSzE4-6b6fRmC8meQmCTnoac4WV5Jhgfq60ptA_sPb69TDLcl1m0CRBPN4HTMRO-WHC0e15nb-3Oic_6DXmJx8Wwt39J9Q8FhyMUDmVGBj9oydi7Ok/s1153/reproducbilitysearch.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;553&quot; data-original-width=&quot;1153&quot; height=&quot;306&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh6VXau5GtZNYrO1c2uV1aWtqjqQ5L78EClligsrGtfJA0e1N5wPx4WK6rdXPPbyeS1cewb2a-WvFFSzE4-6b6fRmC8meQmCTnoac4WV5Jhgfq60ptA_sPb69TDLcl1m0CRBPN4HTMRO-WHC0e15nb-3Oic_6DXmJx8Wwt39J9Q8FhyMUDmVGBj9oydi7Ok/w640-h306/reproducbilitysearch.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;&lt;div&gt;Technical Note: &lt;a href=&quot;https://pubmed.ncbi.nlm.nih.gov/31614060/&quot;&gt;Gusenbauer &amp;amp; Haddaway (2020)&lt;/a&gt; distinguish and test both &quot;Reproducibility of search results at different times&quot; and &quot;Reproducibility of search results at different locations.&quot; In this post, reproducibility refers to the first meaning – getting the same results when running the same search again, whether seconds apart or in different sessions.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;1. Keyword-Based Methods (e.g., Boolean + TF-IDF/BM25 Ranking)&lt;/h3&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Description: &lt;/b&gt;This is the traditional method used by most conventional academic databases. It relies on matching keywords from your query within the documents. Most common is Boolean for matching items and relevancy ranking using&lt;a href=&quot;https://zilliz.com/learn/tf-idf-understanding-term-frequency-inverse-document-frequency-in-nlp&quot;&gt; TF-IDF&lt;/a&gt;/&lt;a href=&quot;https://www.youtube.com/watch?v=ruBm9WywevM&amp;amp;t=55s&quot;&gt;BM25&lt;/a&gt;&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Interpretability&lt;/b&gt;: High. With strict Boolean logic (AND, OR, NOT), you can understand precisely why a document is included or excluded. Standard relevance ranking algorithms like &lt;a href=&quot;https://emschwartz.me/understanding-the-bm25-full-text-search-algorithm/&quot;&gt;TF-IDF or BM25&lt;/a&gt;, while technical, are somewhat intuitive – you can often grasp why some results rank higher by looking at the frequency and placement of matched terms (e.g. title matches worth more).&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Reproducibility:&lt;/b&gt; Generally high. Running the exact same Boolean query usually yields the exact same set of results (assuming the underlying database index hasn&#39;t changed).&lt;/li&gt;&lt;/ul&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;&amp;nbsp;2. LLM-Generated Keyword/Boolean Search&lt;/h3&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjk4eYgtkomG0Y_aHVp1abBLd16rlFn3baxJfvDiqUbNaaO4NPxrEW2fh-rKhpGP-iS2hxm-02yqCNHRC_SJQYl_yZfca3Zr2yncUdORMdcmU8ODhwcMbHl49WcJGNMMNXMnv9VIRtEhYg9cgK8zQphhg7cTcryF8zzdnmzTWFrylC3eXJgRaZDuDJ8xni9/s1390/wos-ra-retrieval.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;523&quot; data-original-width=&quot;1390&quot; height=&quot;240&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjk4eYgtkomG0Y_aHVp1abBLd16rlFn3baxJfvDiqUbNaaO4NPxrEW2fh-rKhpGP-iS2hxm-02yqCNHRC_SJQYl_yZfca3Zr2yncUdORMdcmU8ODhwcMbHl49WcJGNMMNXMnv9VIRtEhYg9cgK8zQphhg7cTcryF8zzdnmzTWFrylC3eXJgRaZDuDJ8xni9/w640-h240/wos-ra-retrieval.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div style=&quot;text-align: center;&quot;&gt;&lt;i&gt;Example of Web of Science Research Assistant using LLM to generate Boolean Search&lt;/i&gt;&lt;/div&gt;&lt;div style=&quot;text-align: center;&quot;&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Description:&lt;/b&gt; An LLM takes your natural language query and translates it into a structured Boolean or keyword search strategy, which is then run against a traditional database inverted index.&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Interpretability:&lt;/b&gt; Remains high, assuming the system displays the generated Boolean query. You can see exactly which terms and logic were used to retrieve the results.&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Reproducibility: &lt;/b&gt;Lower than pure keyword search. Because an LLM is involved in generating the query, the non-deterministic nature of LLMs means the generated search strategy itself might differ slightly even when given the exact same input query. This variation in the search strategy directly affects the reproducibility of the final result set.&lt;/li&gt;&lt;/ul&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;b&gt;3. Non-Keyword / Embedding-Based Methods (Knowned as Vector Search, Semantic Search, etc.)&lt;/b&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Description:&lt;/b&gt; &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2024/04/a-conceptual-view-of-information.html&quot;&gt;These methods generally convert both your query and the documents into numerical representations (vectors or embeddings) in a high-dimensional space. The system then finds documents whose vectors are &quot;closest&quot; or most similar to the query vector&lt;/a&gt;. Terminology varies: you might see &quot;vector search&quot;, &quot;(dense) embeddings&quot;, &quot;neural search&quot;, &quot;dense retrieval&quot;, or &quot;semantic search.&quot; There are many technical variants (e.g., bi-encoders, cross-encoders, late-interaction/multi-vector approaches like ColBERT. Also learned sparse methods like SPLADE.)&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Interpretability: &lt;/b&gt;Generally low. These methods often function as &quot;black boxes.&quot; It&#39;s hard, if not impossible, to explain precisely why a specific document is retrieved or ranked higher than another, beyond pointing to a calculated similarity score between the query and the documents. Some advanced methods (like &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2024/06/can-semantic-search-be-more.html&quot;&gt;ColBERT or SPLADE&lt;/a&gt;) have the potential for more interpretability by showing which components contribute most to the relevance score, but they are not yet widely implemented or exposed in interfaces. (To my knowledge, &lt;a href=&quot;https://blog.elicit.com/semantic-search/&quot;&gt;Elicit is one academic AI tool using SPLADE&lt;/a&gt;, but its interface doesn&#39;t currently break down the score components for users).&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Reproducibility:&lt;/b&gt; Often lower than keyword methods. Several factors contribute, but a major one is the common use of &lt;a href=&quot;https://rockset.com/articles/nearest-neighbor-search/&quot;&gt;Approximate Nearest Neighbor (ANN) algorithms&lt;/a&gt;. ANN speeds up the computationally intensive process of finding similar vectors in massive datasets but often introduces some randomness, meaning results can vary slightly even for identical queries.&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;Let&#39;s now examine how our three example tools blend these methods, leading to differences in their interpretability and reproducibility.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&amp;nbsp;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjKWHPo3ppY9kpu9rhOqTk2oM2r6lgW0dXs1J2c_3lkaJ2ztAJ066MsXNpx9h-FR1ywg6izENw_JEu-4rBwgpm9hXulRd1xKhufNQahif05tzntF8KLmO9NUO-mL2BIN0PAuH1PenQ0sCOdoPugDzDXyylTdAGo_hqKXJxoNWpILaJ6qqlT685Wd6Jv6LpY/s850/reproducbilitysearch-example.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;702&quot; data-original-width=&quot;850&quot; height=&quot;528&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjKWHPo3ppY9kpu9rhOqTk2oM2r6lgW0dXs1J2c_3lkaJ2ztAJ066MsXNpx9h-FR1ywg6izENw_JEu-4rBwgpm9hXulRd1xKhufNQahif05tzntF8KLmO9NUO-mL2BIN0PAuH1PenQ0sCOdoPugDzDXyylTdAGo_hqKXJxoNWpILaJ6qqlT685Wd6Jv6LpY/w640-h528/reproducbilitysearch-example.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;Web of Science Research Assistant&lt;/h3&gt;&lt;div&gt;&lt;div&gt;This is perhaps the most straightforward of the three. Based on my testing and the available &lt;a href=&quot;https://clarivate.libguides.com/ld.php?content_id=79642031&quot;&gt;documentation&lt;/a&gt;:&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;1. It feeds your natural language query into an LLM (at the time of writing, likely a model like GPT-4o mini).&lt;/div&gt;&lt;div&gt;2. The LLM is prompted with instructions to generate a complex Boolean search query suitable for Web of Science.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;For example, if you enter:&lt;/div&gt;&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;&quot;impact of climate change on biodiversity&quot;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;The Assistant might generate a query like this:&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;(climate change OR global warming OR climate variation OR climatic changes OR climate variability OR climate crisis OR climate emergency OR greenhouse gases OR anthropogenic climate change OR carbon emissions) AND (biodiversity OR species diversity OR ecosystem diversity OR biological diversity OR ecological diversity) AND (impact OR effect OR influence OR consequence)&lt;/div&gt;&lt;/blockquote&gt;&lt;p&gt;While we don&#39;t know the exact prompt used, observing examples suggests instructions along these lines:&lt;/p&gt;&lt;p&gt;1. Extract the main concepts from the user query.&lt;/p&gt;&lt;p&gt;2. For each concept, generate relevant synonyms and related terms, combining them with OR.&lt;/p&gt;&lt;p&gt;3. Combine the resulting concept blocks using AND.&lt;/p&gt;&lt;div&gt;&lt;div&gt;Crucially, the Web of Science Research Assistant displays the generated Boolean query and provides a link to run it directly in the standard Web of Science interface.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Testing confirms that the top 8 results shown by the Research Assistant are exactly the same results, in the same order, as running the generated Boolean query directly in Web of Science using the default relevance sort (which likely uses BM25/TF-IDF).&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Interpretability:&lt;/b&gt; Very high. You know exactly which Boolean search query was used. You can also understand why the top 8 results were retrieved and ranked as they are – they directly correspond to a standard WoS search using that generated query and default relevance ranking.&lt;/div&gt;&lt;div&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;b&gt;Reproducibility:&lt;/b&gt; Moderately high for an AI system. Since an LLM generates the Boolean query, some non-determinism is expected. In my tests, running the same natural language query multiple times resulted in a different generated Boolean search strategy roughly 1 out of 5 times. While this indicates variability, it&#39;s notably more consistent than some other AI search tools.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;h3 style=&quot;text-align: left;&quot;&gt;&lt;b&gt;Primo Research Assistant&lt;/b&gt;&lt;/h3&gt;&lt;div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjXUQCrN5PVzR6KpOot52a95_gwdl7ja1ncajbHe9O4V88xesmu9M0iKd-vaiunY3PQjzYWoLTC5Ya-xADE67Gvnog7xcqjeyNBf0w1wrdsb1M3w_HSXUw1fxiqLM4iugcPWNZRuRtcFLU0nUuFVccPn1QzFg-xfEk_VHlltZVtDVTBJGRL5srlEdDY26Df/s992/primora-searchcreated.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;177&quot; data-original-width=&quot;992&quot; height=&quot;114&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjXUQCrN5PVzR6KpOot52a95_gwdl7ja1ncajbHe9O4V88xesmu9M0iKd-vaiunY3PQjzYWoLTC5Ya-xADE67Gvnog7xcqjeyNBf0w1wrdsb1M3w_HSXUw1fxiqLM4iugcPWNZRuRtcFLU0nUuFVccPn1QzFg-xfEk_VHlltZVtDVTBJGRL5srlEdDY26Df/w640-h114/primora-searchcreated.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div style=&quot;text-align: center;&quot;&gt;Boolean search generated by LLM for Primo Research Assistant&lt;/div&gt;&lt;/div&gt;&lt;div style=&quot;text-align: center;&quot;&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;/div&gt;&lt;div&gt;As I &lt;a href=&quot;https://musingsaboutlibrarianship.blogspot.com/2024/09/primo-research-assistant-launches-first.html&quot;&gt;explained in a previous post&lt;/a&gt;, Primo Research Assistant functions similarly to the WoS tool in that it also uses an LLM to construct a Boolean search strategy. However, the style of the generated query and the subsequent processing differ.&amp;nbsp;&lt;/div&gt;&lt;p&gt;Using the same example query:&lt;/p&gt;&lt;blockquote&gt;&quot;impact of climate change on biodiversity&quot;&lt;/blockquote&gt;&lt;p&gt;Primo RA might generate a query structured like this:&lt;/p&gt;&lt;blockquote&gt;climate change biodiversity impact) OR (effects of climate change on ecosystems) OR (biodiversity loss due to climate change) OR (climate change species extinction) OR (impact of global warming on wildlife) OR (effects of climate change on ecosystems and species diversity) OR (how climate change impacts wildlife and biodiversity) OR (climate change consequences for biological diversity) OR (relationship between climate change and loss of biodiversity) OR (climate change threats to flora and fauna diversity) OR &lt;b&gt;(impact of climate change on biodiversity)&lt;/b&gt;&lt;/blockquote&gt;&lt;blockquote&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;&lt;div&gt;Based on documentation and observation, the LLM appears to be instructed to:&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;1.&lt;span&gt;&amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/span&gt;Generate multiple query variant strings based on the original input.&lt;/div&gt;&lt;div&gt;2.&amp;nbsp; &amp;nbsp; Do it 10 times&lt;/div&gt;&lt;div&gt;&lt;span&gt;3.&amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/span&gt;Combine these 10 generated strings and the original query input using the OR operator.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;So far, it seems analogous to WoS RA, just with a different query generation strategy. However, Primo RA adds another layer: &lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Product_Documentation/020Primo_VE/Primo_VE_(English)/015_Getting_Started_with_Primo_Research_Assistant#:~:text=Re%2Dranking%20%E2%80%93%20The%20top%20results%20(up%20to%2030)%20are%20re%2Dranked%20using%20embeddings%20to%20identify%20five%20sources%20that%20best%20address%20the%20user%27s%20query.&quot;&gt;&lt;b&gt;it reranks the top 30 results retrieved by the generated Boolean query. &lt;/b&gt;This reranking step uses embedding models&lt;/a&gt; – a form of the less interpretable dense/vector retrieval discussed earlier.&lt;/div&gt;&lt;div&gt;&lt;blockquote&gt;&lt;div&gt;PubMed also works similarly, though it runs &lt;a href=&quot;https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6112631/&quot;&gt;a standard keyword search (BM25) first followed by reranking the top 500 using a machine learning-based reranking method - LambdaMART&lt;/a&gt;, making results somewhat less interpretable if you want to explain the order of results.&lt;/div&gt;&lt;/blockquote&gt;&lt;/div&gt;&lt;div&gt;Because of this reranking step, the top results displayed by Primo Research Assistant will not necessarily be in the same order as those you&#39;d get by simply running the generated Boolean query in the standard Primo interface.&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;blockquote&gt;&lt;div&gt;Furthermore, the results might differ because Primo Research Assistant searches Primo&#39;s Central Discovery Index (CDI), &lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Product_Documentation/020Primo_VE/Primo_VE_(English)/015_Getting_Started_with_Primo_Research_Assistant#:~:text=We%20use%20the,their%20stakeholder%20groups.&quot;&gt;with specific exclusions (e.g., content owners opting out) that might not apply identically to a user&#39;s standard Primo interface search configured by their institution.&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;/div&gt;&lt;/blockquote&gt;&lt;div&gt;&lt;b&gt;Interpretability: &lt;/b&gt;Quite high, but slightly less than WoS Research Assistant. Like WoS RA, it displays the generated Boolean query, so you understand the initial retrieval logic. However, the final order of the top results is influenced by the embedding-based reranking step, which is harder to interpret. You can still understand why an item was retrieved at all (it matched the Boolean query), but explaining its exact rank among the top results is more difficult than with WoS RA.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Reproducibility:&lt;/b&gt; Similar to WoS Research Assistant in terms of the generated Boolean query. In my tests, the generated query also changed roughly 1 out of 5 times for the same input. The reproducibility of the final ranked list might be slightly lower due to potential non-determinism in the reranking step, although this is harder to isolate.&lt;/div&gt;&lt;p&gt;&lt;/p&gt;&lt;blockquote&gt;&lt;p&gt;&amp;nbsp;&lt;/p&gt;&lt;/blockquote&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Scopus AI&lt;/h2&gt;&lt;p&gt;Scopus AI employs the most complex retrieval mechanism of the three.&lt;/p&gt;&lt;p&gt;Like the others, it uses an LLM to generate a Boolean search strategy. For our example query:&lt;/p&gt;&lt;p&gt;&quot;impact of climate change on biodiversity&quot;&lt;br /&gt;&lt;/p&gt;&lt;p&gt;It might generate something like:&lt;/p&gt;&lt;p&gt;(&quot;climate change&quot; OR &quot;global warming&quot; OR &quot;climate crisis&quot; OR &quot;climatic change&quot;) AND (&quot;biodiversity&quot; OR &quot;species diversity&quot; OR &quot;ecosystem diversity&quot; OR &quot;biological diversity&quot;) AND (&quot;impact&quot; OR &quot;effect&quot; OR &quot;influence&quot; OR &quot;consequence&quot;) AND (&quot;conservation&quot; OR &quot;preservation&quot; OR &quot;protection&quot; OR &quot;sustainability&quot;) AND (&quot;habitat&quot; OR &quot;ecosystem&quot; OR &quot;environment&quot; OR &quot;biome&quot;) AND (&quot;adaptation&quot; OR &quot;mitigation&quot; OR &quot;resilience&quot; OR &quot;response&quot;)&lt;br /&gt;&lt;/p&gt;&lt;p&gt;Interestingly, this query seems to incorporate more concepts (perhaps 4-6 distinct blocks) compared to the WoS RA example (3 blocks). The Scopus AI documentation mentions a &quot;Copilot&quot; that ensures &quot;complex queries are broken down into their component parts,&quot; hinting at some initial query analysis or expansion (which add new aspects or concepts) before the LLM generates the final Boolean query.&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh3AQ9UfacWUroFTyhbNmgQvyHj9RXVGkTzy_d10EtlEEMktEoypIFo3Nsq2zo9ANr_Ofs3B8C6gJPBTvTyIxAE_GGXTqgsTTXpFWG8J0eR_uXCxgQqAju5io57RD3CwbO3WGIG7snLGIcu8Ivtu8UXTSel-Ubs4tEVdrz9OU9vUfKQXSLLeUe9Fo59X2Tv/s1156/scopus-ai-hybrid1.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;441&quot; data-original-width=&quot;1156&quot; height=&quot;244&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEh3AQ9UfacWUroFTyhbNmgQvyHj9RXVGkTzy_d10EtlEEMktEoypIFo3Nsq2zo9ANr_Ofs3B8C6gJPBTvTyIxAE_GGXTqgsTTXpFWG8J0eR_uXCxgQqAju5io57RD3CwbO3WGIG7snLGIcu8Ivtu8UXTSel-Ubs4tEVdrz9OU9vUfKQXSLLeUe9Fo59X2Tv/w640-h244/scopus-ai-hybrid1.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;p style=&quot;text-align: center;&quot;&gt;&lt;i&gt;Scopus AI does both &quot;natural language search&quot; and keyword Boolean search&lt;/i&gt;&lt;/p&gt;&lt;p style=&quot;text-align: center;&quot;&gt;&lt;br /&gt;&lt;/p&gt;&lt;p&gt;The major difference, however, is that Scopus AI also performs a &quot;natural language search&quot; simultaneously with the Boolean search. &lt;a href=&quot;https://elsevier.libguides.com/Scopus/ScopusAI&quot;&gt;The documentation describes this as a Vector Search&lt;/a&gt;. Unlike Primo RA&#39;s reranking step, Scopus AI uses vector search as part of the initial retrieval.&lt;/p&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjrfQJPJR1ztrvyn9PYECDT4ylEM_Kf73evdrocQ8AcSq5p9Q2qfYZrNyK75fDJ5SLfH8N-pPW2S2sCTG93EM21hnXB_5pY1SC0BAJNYVMlKBYe3CAszZJd6E7lVXPsXMVoW-hWgk31OBJsZzTIzwhJrdMaITAvMCle1Y1nagCQ_bGqwEV8mminV12TQYEc/s707/how%20copilot%20works.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;334&quot; data-original-width=&quot;707&quot; height=&quot;302&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjrfQJPJR1ztrvyn9PYECDT4ylEM_Kf73evdrocQ8AcSq5p9Q2qfYZrNyK75fDJ5SLfH8N-pPW2S2sCTG93EM21hnXB_5pY1SC0BAJNYVMlKBYe3CAszZJd6E7lVXPsXMVoW-hWgk31OBJsZzTIzwhJrdMaITAvMCle1Y1nagCQ_bGqwEV8mminV12TQYEc/w640-h302/how%20copilot%20works.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;p&gt;In the documentation the natural language search is described as a Vector Search. This is akin to the reranking step down by Primo Research Assistant, except this is done at the initial retrieval step.&lt;/p&gt;&lt;p&gt;In essence, Scopus AI uses a &quot;hybrid search&quot; system:&lt;/p&gt;&lt;div&gt;1.&amp;nbsp;It runs the LLM-generated Boolean query against the Scopus index.&lt;/div&gt;&lt;div&gt;2.&amp;nbsp;It runs the original natural language query (or a processed version) through an embedding-based vector search against the Scopus index.&lt;/div&gt;&lt;div&gt;3.&amp;nbsp;It takes the top results from both searches, deduplicates them, and combines/reranks them (likely using a method like Reciprocal Rank Fusion) to produce the final list.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;b&gt;Interpretability: &lt;/b&gt;Clearly much lower than the other two. While the generated Boolean query component is displayed and interpretable, the simultaneous use of vector embedding search during initial retrieval makes the overall results significantly less interpretable. The vector search part acts as a black box; there&#39;s no straightforward way to explain why certain results were retrieved via that path or how they contributed to the final ranking, beyond abstract similarity scores&lt;/div&gt;&lt;div&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;b&gt;Reproducibility: &lt;/b&gt;Very low, based on my testing.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;The generated Boolean search query itself is highly variable. It changed almost every other time (roughly 1 in 2 times) I ran the same natural language query. Potential reasons include randomness introduced by the initial &quot;Copilot&quot; query analysis step, or the LLM generating the Boolean query might be configured with parameters allowing for more variation (e.g., higher temperature).&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;The reproducibility of the overall search results (the final list) is also very low. This is likely due to the combination of the variable Boolean query and the inherent non-determinism of the vector search component (potentially using ANN algorithms). Even if the Boolean query happened to be the same between two runs, the vector search part could still return slightly different results, leading to a different final combined list.&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;b style=&quot;color: #e2e2e5; font-family: &amp;quot;Google Sans Text&amp;quot;, &amp;quot;Helvetica Neue&amp;quot;, sans-serif; font-size: 14px;&quot;&gt;&amp;nbsp;&amp;nbsp;&lt;/b&gt;&lt;/div&gt;&lt;h2&gt;Other Academic AI Search Systems &amp;amp; General Trends&lt;/h2&gt;&lt;div style=&quot;text-align: left;&quot;&gt;My tests here are informal. Eventually, we will hopefully see formal evaluations of AI search tools, similar to Gusenbauer &amp;amp; Haddaway&#39;s (2020) study of 26 academic search systems (&lt;a href=&quot;https://onlinelibrary.wiley.com/doi/full/10.1002/jrsm.1378&quot;&gt;&quot;Which academic search systems are suitable for systematic reviews or meta-analyses?&lt;/a&gt;&quot;), but adapted for these new complexities. For now, let me speculate based on available information and trends:&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Earlier tools like &lt;a href=&quot;https://www.youtube.com/watch?v=KfSYyNW7b1s&amp;amp;t=2285s&quot;&gt;Scite.ai Assistant &lt;/a&gt;and the experimental &lt;a href=&quot;https://arxiv.org/abs/2307.04683&quot;&gt;CORE-GPT&lt;/a&gt; also use LLMs to generate search strategies. &lt;a href=&quot;https://www.youtube.com/watch?v=KfSYyNW7b1s&amp;amp;t=2285s&quot;&gt;Scite.ai&#39;s generated queries sometimes resemble Primo RA&#39;s style (multiple OR&#39;d phrases)&lt;/a&gt;, while CORE-GPT might lean towards the WoS RA style (concept blocks with synonyms). Their specific reproducibility and interpretability would depend on their exact implementation (e.g., do they rerank? do they also use vector search?).&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEij3HTQ1TQygt9dlFmhPJEQuZhE1ALJw3YYRjk18rcMiwLWKJIOF0FJEFD2RHxMT2xxJDEC-w1tfOYbrlMMwSpATzuaxGvNKvwQN3yqr3XigghFKCVj54u_-OVvIoDdKwrS-_ModEhC0e3azOEWZD6UTnt4Gvflg1QceCAY1GPobAeYFciSyUkT6itUncF8/s678/scitesearchstrategy-2.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;211&quot; data-original-width=&quot;678&quot; height=&quot;200&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEij3HTQ1TQygt9dlFmhPJEQuZhE1ALJw3YYRjk18rcMiwLWKJIOF0FJEFD2RHxMT2xxJDEC-w1tfOYbrlMMwSpATzuaxGvNKvwQN3yqr3XigghFKCVj54u_-OVvIoDdKwrS-_ModEhC0e3azOEWZD6UTnt4Gvflg1QceCAY1GPobAeYFciSyUkT6itUncF8/w640-h200/scitesearchstrategy-2.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;&lt;div&gt;&amp;nbsp;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEheOaKXyOsbmcM1X8kWIL3DZ204jWr77R2LvrvxKtxg_BS5w031IJSz5pLBddDlaYPk9puVvQzQtH_wZjpOgrgPoY7-UxlrAfw9iPp2Y9qBAa5IUp5-bhjRtFDfz7wTEfG5AmI1IYsVVGOhzQjZsq6_UnTkpE9LsId278nxxzjhG2pdLid824Ghp8D60Hnr/s806/coregpt-searchstrategy.png&quot; style=&quot;margin-left: 1em; margin-right: 1em;&quot;&gt;&lt;img border=&quot;0&quot; data-original-height=&quot;671&quot; data-original-width=&quot;806&quot; height=&quot;532&quot; src=&quot;https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEheOaKXyOsbmcM1X8kWIL3DZ204jWr77R2LvrvxKtxg_BS5w031IJSz5pLBddDlaYPk9puVvQzQtH_wZjpOgrgPoY7-UxlrAfw9iPp2Y9qBAa5IUp5-bhjRtFDfz7wTEfG5AmI1IYsVVGOhzQjZsq6_UnTkpE9LsId278nxxzjhG2pdLid824Ghp8D60Hnr/w640-h532/coregpt-searchstrategy.png&quot; width=&quot;640&quot; /&gt;&lt;/a&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;a href=&quot;https://arxiv.org/abs/2307.04683 &quot;&gt;&lt;i&gt;CORE-GPT: Combining Open Access research and large language models for credible, trustworthy question answering&lt;/i&gt;&lt;/a&gt;&lt;/div&gt;&lt;div class=&quot;separator&quot; style=&quot;clear: both; text-align: center;&quot;&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Typical &quot;AI Academic Search&quot;: It&#39;s often hard for us as librarians to know precisely what&#39;s happening under the hood, as vendor documentation varies in technical detail. However, based on informal discussions and industry trends, my guess is that many newer academic AI search engines, especially those from startups, are implementing hybrid search similar in principle to Scopus AI (running keyword/Boolean and vector searches in parallel, then combining results).&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Beyond the use of LLMs to generate Boolean search strategies, LLMs can and have been used in other aspects of retrieval including being used outright to evaluate relevancy of the papers (e.g. Undermind.ai, Elicit Research Reports)&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;Incumbent Vendor Choices: Established vendors with massive, existing, highly optimized keyword-based indexes (like Exlibris/Clarivate with Primo/Summon and WoS) might initially prefer the &quot;LLM-generates-Boolean-query&quot; approach seen in WoS RA and Primo RA. The main advantage is likely cost and infrastructure. Building, maintaining, and constantly updating a parallel vector database for billions of items to support embedding search is significantly more resource-intensive than leveraging their existing infrastructure with an LLM front-end. This might explain why Exlibris/Clarivate have adopted this method for tools &lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Content_Corner/Central_Discovery_Index/Documentation_and_Training/Documentation_and_Training_(English)/CDI_-_The_Central_Discovery_Index/010An_Overview_of_the_Ex_Libris_Central_Discovery_Index_(CDI)#:~:text=It%20contains%20over%205%20billion%20records%C2%A0and%20many%20different%20resource%20types%20from%20thousands%20of%20publishers%2C%20aggregators%2C%20and%20repositories.&quot;&gt;searching huge indexes like Primo&#39;s Central Discovery Index (containing 5 billion records&lt;/a&gt; - but of course &lt;a href=&quot;https://knowledge.exlibrisgroup.com/Primo/Product_Documentation/020Primo_VE/Primo_VE_(English)/015_Getting_Started_with_Primo_Research_Assistant#:~:text=GPT%2D%2D4o%20mini.-,Content%20Scope,the%20entirety%20of%20CDI%20metadata%20and%20abstracts%20with%20the%20following%20exceptions%3A,-News%20content%20(Newspaper&quot;&gt;not all records can be used in Primo Research Assistant&lt;/a&gt;).&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;div&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Interpretability&lt;/b&gt;: Many emerging &quot;AI academic search&quot; tools that heavily rely on embedding-based vector search (especially in hybrid models) suffer from lower interpretability compared to traditional keyword search or even the LLM-generated Boolean approach (when the query is shown). Complex multi-stage retrieval pipelines used by some systems can make it even harder to understand precisely why you&#39;re seeing the results you get.&lt;/li&gt;&lt;/ul&gt;&lt;ul style=&quot;text-align: left;&quot;&gt;&lt;li&gt;&lt;b&gt;Reproducibility&lt;/b&gt;: While variable, the reproducibility of typical academic AI search engines is generally expected to be lower than that of traditional, deterministic keyword-based academic search. The degree of reproducibility varies significantly depending on the specific methods used (LLM settings, use of ANN in vector search, hybrid ranking strategies, etc.).&lt;/li&gt;&lt;/ul&gt;&lt;/div&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;&lt;br /&gt;&lt;/h2&gt;&lt;h2 style=&quot;text-align: left;&quot;&gt;Conclusion: Beyond Reproducibility and Interpretability – The Challenge of Bias&lt;/h2&gt;&lt;div&gt;&lt;div&gt;The advent of modern AI in search&amp;nbsp; presents multifaceted challenges extending beyond the interpretability and reproducibility issues discussed here. Perhaps the most profound, and certainly harder to investigate thoroughly, lies in detecting and measuring algorithmic bias.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;As platforms increasingly replace explicit keyword control with AI-driven relevance judgments—often using neural networks for ranking or retrieval—they risk reflecting and potentially amplifying biases inherent in their underlying data and design.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;a href=&quot;https://app.undermind.ai/report/96d1ce264f5b976eac434514d16e2529a99968d6928b225d57859617b14beca1&quot;&gt; Indeed, early studies suggest that neural rankers, particularly transformer-based models like BERT, can intensify the retrieval of certain content—for instance, potentially favouring male-associated sources—compared to traditional keyword systems such as BM25.&lt;/a&gt;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;While known biases can sometimes be mitigated with effort, it is the unknown biases that pose a more significant concern. Detecting these hidden skews remains a formidable challenge, particularly in the current dynamic landscape. Is there a way to detect them? I certainly have no clue.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;Currently, no single AI academic search tool or algorithm dominates (with the possible exception of Google Scholar). We are in a &#39;Wild West&#39; phase where the information retrieval field lacks consensus on optimal ranking methods, leading to a variety of approaches. This situation is both a boon and a curse.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;It&#39;s a curse because, without consensus, biases identified in one tool might be specific to its unique implementation, making systemic solutions difficult. Conversely, it&#39;s a boon because the diversity of tools means users aren&#39;t locked into a single potentially flawed system, and employing multiple tools might help mitigate the impact of any one tool&#39;s specific bias.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;This inherent opacity presents a significant challenge. If the inner workings of these AI systems remain largely obscure, and unknown biases are difficult, if not impossible, to detect definitively, what practical approach can librarians and users take? Perhaps the most crucial skill we can currently cultivate and teach is fundamental skepticism.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;On the other hand, navigating information systems with opaque algorithms and potential biases isn&#39;t entirely unprecedented. Librarians and researchers have long relied on tools like Google (&lt;a href=&quot;https://blog.google/products/search/search-language-understanding-bert/&quot;&gt;which as far back as 2019 already incorporated BERT neural rankings for 10% of queries!&lt;/a&gt;) and Google Scholar, despite limited insight into their ranking mechanisms and awareness of potential commercial or systemic biases influencing results.&amp;nbsp;&amp;nbsp;&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;However, the context and claims of dedicated academic AI search tools arguably raise the stakes. These tools are explicitly marketed for scholarly discovery, potentially leading users to place greater trust in their outputs. Therefore, while the challenge of dealing with black-box systems isn&#39;t new, the need for active skepticism might be even more critical now. This translates into practical strategies: consistently questioning surprising results, habitually comparing outputs across different AI search tools (leveraging algorithmic diversity as a necessity), and remaining acutely aware that these systems are prone to biases, even if invisible.&lt;/div&gt;&lt;div&gt;&lt;br /&gt;&lt;/div&gt;&lt;div&gt;&amp;nbsp;Edit: Article was written by a human but edited with help of Grok3 and Gemini 2.5 pro&lt;/div&gt;&lt;/div&gt;&lt;/div&gt;&lt;p&gt;&lt;/p&gt;</content><link rel='replies' type='application/atom+xml' href='http://musingsaboutlibrarianship.blogspot.com/feeds/5056661138239098138/comments/default' title='Post Comments'/><link rel='replies' type='text/html' href='http://www.blogger.com/comment/fullpage/post/4727930222560708528/5056661138239098138?isPopup=true' title='0 Comments'/><link rel='edit' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/5056661138239098138'/><link rel='self' type='application/atom+xml' href='http://www.blogger.com/feeds/4727930222560708528/posts/default/5056661138239098138'/><link rel='alternate' type='text/html' href='http://musingsaboutlibrarianship.blogspot.com/2025/04/the-reproducibility-and.html' title='The reproducibility and interpretability of academic Ai search engines like Primo Research Assistant, Web of Science Research Assistant, Scopus Ai and more'/><author><name>Aaron Tay</name><uri>http://www.blogger.com/profile/02750645621492448678</uri><email>noreply@blogger.com</email><gd:image rel='http://schemas.google.com/g/2005#thumbnail' width='32' height='32' src='//blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjNp5I56cmCnITp9u98mqHmOga9TIDbvdXeuRetlD5Lq7jfNQVbjCCMMyEkblX6PtSR34esdLJ6qarZjGFZC_pAuvyTr93fqVvnlAzrnm2DjjaaH1BTx3XN8lJH69-gnw/s220/profile.jpg'/></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiCgJBjjrUcJiAfa3T9Fb-YM_hCLF2QL4DJ89ktCxPgA2Jga3NlJwGE8_F_X4eFsnsFDGL30_Lq398WuXDkABSynTmIstOo9q2YNHbJLfDcZP2rTTzt1M-HnIn27O01B3ahaz45DwphJNHBVKPG2_LfU6GarkoSzmt6ONLcsrsqBKPlJVi6fyHoYyTdfBAy/s72-w640-h340-c/bafkreiglsgw5x7yyloq56ooiqrkhmxc3anynykrg6rn6irilpmzojl7ujm.jpg" height="72" width="72"/><thr:total>0</thr:total></entry></feed>