Rogue Scholar Posts

language
Published in Aaron Tay's Musings about librarianship
Author Aaron Tay

In the last blog post , I argued that despite the advancements in AI thanks to transformer based large language models, most academic search still are focused mostly in supporting exploratory searches and do not focus on optimizing recall and in fact trade off low latency for accuracy.

Published in Aaron Tay's Musings about librarianship
Author Aaron Tay

One of the tricks about using the newer "AI powered" search systems like Elicit, SciSpace and even JSTOR experiment search is that they recommend that you type in your query or what you want in full natural language and not keyword search style (where you drop the stop words) for better results. So for example do

Published in Aaron Tay's Musings about librarianship
Author Aaron Tay

Earlier related pieces - How Q&A systems based on large language models (eg GPT4) will change things if they become the dominant search paradigm - 9 implications for libraries In the ever-evolving landscape of information retrieval and library science, the emergence of large language models, particularly those based on the transformer architecture like GPT-4, has opened up a Pandora's box of possibilities and challenges.

Published in Aaron Tay's Musings about librarianship
Author Aaron Tay

List of academic search engines that use Large Language models for generative answers (for the latest version - see this page) This is a non-comprehensive list of academic search engines that use generative AI (almost always Large language models) to generate direct answers on top of list of relevant results, typically using Retrieval Augmented Generation (RAG) Techniques. We expect a lot more!

Published in Aaron Tay's Musings about librarianship
Author Aaron Tay

I was asked to give a talk at CILIP Conference 2023 for the Data & AI panel. I was given 10 minutes to send a recording which I eventually delivered.  But here's an extended 40-minute recording before I cut it down. As I was just practicing, the delivery might not be the best, and in fact the final 10-minute version might be better!