Using Google NotebookLM with “Understanding…”
Post de Olivier Ezratty du 20 octobre 2024 - Tags : Actualités,Quantique | No Comments
I published the open-source book “Understanding Quantum Technologies 2024” late September 2024. Since then, I have updated it many times to follow the news, add a couple companies and also correct various spelling, presentation and other errors. I try to keep up with the scientific and other news until I publish this edition on arXiv and then on AWS in a paperback edition in 5 volumes (or parts), around November 2024.
The book length may be intimidating (1,552 pages) and you may have no time to read it entirely. I provide a 26-page zipped version which contains the key highlights from each part. But you may have many questions who need answers. Sometimes simple ones, but which require digging into many parts of the book to get a sound answer.
Some of you may have created a GPT with uploading the book or its five volumes (or “parts”). It can help you navigate in the book and ask questions about it. Recently, Google released an experimental tool named NotebookLM. With it, you can also create workspaces where you upload up to 20 documents and then can ask any question related to the uploaded documents. In the past, I had tried similar and more structured tools like Perplexity but I was less impressed.
Its performance seems very good and better than with a GPT. It uses two main techniques: a RAG (Retrieval-Augmented Generation) to access the documents and an LLM to format your questions and the RAG answers. RAG vectorizes your document content in a database. NotebookLM provides responses to your questions with series of numbers pointing to the text extracts from your documents that were used to craft it. As a result, answers are traceable and explainable. If it makes mistakes, you’ll see why.
The results are quite spectacular. Its capability to consolidate information scattered in different parts of the book is really cool. Sometimes, LLM-based chatbots give the impression they can reason and to understand what they are talking about. It is not (yet) the case, particularly with a RAG coupled with a LLM. Reasoning still exists in ChatGPT and the like due to their use of agents and chains of thoughts. The recent o1-preview from ChatGPT is able to decompose a question into sub-questions, showing basic reasoning skills. One must avoid projecting some anthropomorphic skills on these tools. These are machines using brute force to consolidate information from many sources.
Book ingestion
NotebookLM is a prototype solution that is free to use. You just need to login with your Google/Gmail account. Of course, it is not advised to upload confidential information in such a tool. My book being public, there is no such problem here.
I uploaded the 5 volume PDFs from “Understanding Quantum Technologies” to see what it could achieve. The files ingestion lasted less than a minute.
Questioning “Understanding Quantum Technologies”
So, here are the first questions I tested with a rating from 1 (bad) to 5 (very good). You can try this yourself using this link which I hope will last long.
The detailed responses are in this PDF. At the end, I added a couple other questions that you can try on your own. It is now up to use your imagination… and information needs.
At this point, NotebookLM is “emulating” me to some extent but not entirely, hopefully. It is a useful tool to consolidate information coming from long documents or many documents. It helps uncover some myths and misconceptions, beyond the famous “quantum hype” that creates a reality distorsion field around the real and potential capabilities of quantum computers. But NotebookLM won’t really reason and it can make mistakes, lacking “judgment” in some situations. You need to know the rules of the game here and always fact-check what these tools are producing.
Extended version
The next step to this trial would be to create an extended version of this setup. How about uploading the 5600+ scientific references from the book’s bibliography? It is not possible with the end-user tools ala NotebookLM and ChatGPT GPTs. You need to create your custom solution and host it somewhere on the cloud and to use some open source LLM like Llama or Mistral. This is another, more complicated project.
Podcast generation
On top of that, NotebookLM can create a podcast with two (American) speakers summarizing key points from your documents in a very realistic way. I tried it with the book and the result is so-so regarding the content but spectacular with regards to its text-to-speech capability.
You can even add a prompt when creating this audio. I used the following one: focus on the part on quantum computing hardware and software, be as technical as possible, describe the various challenges to build a scalable quantum computer, identify the challenges to do this for the main qubit modalities (superconducting, silicon, ions, cold atoms, photons), be specific on the key figures of merit (fidelities, number of qubits, etc.) and discuss about quantum error correction. It didn’t really follow all these instructions as you can hear.
Figures captioning
I tried NotebookLM in other situations like with digging into the history of quantum foundations and comparing quantum physics interpretations (Qbism, RQM, MWI, CSM, etc) and it fared very well. To do that, I extracted from my book about 85 PDFs from the related chapter bibliography. I consolidated it into 4 PDF files using a PDF file merger tool, due to the NotebookLM file upload limit of 50 items.
I also discovered recently that it can create clear captions for cryptic charts and figures in scientific papers. I tried it on “Magic state cultivation: growing T states as cheap as CNOT gates” by Craig Gidney, Noah Shutty, and Cody Jones, Google, arXiv, September 2024 (33 pages). The result is very convincing. It seems that NBLM consolidates information from the paper text, its caption and even its own content. This is a game changer.
To be sure I was not just lucky once, I tried it several times, like with the following table that is in my book as an example of bad captioning. The source is A Short Review on Novel Approaches for Maximum Clique Problem: from Classical algorithms to Graph Neural Networks and Quantum algorithms by Raffaele Marino et al, arXiv, March 2024 (24 pages). And here is NotebookLM’s response. It is really good, again!
So not only these tools can summarize long arXiv and other scientific papers, but it can help you decipher their cryptic content. It is a boon for all scientists!
More on this at Quantum+AI
That was kind of a teaser for my presentation on October 29th, 2024, in New York City at the conference Quantum+AI. I will present some insights on how LLM-based tools are already changing the life of scientists and others in the quantum ecosystem and where we may be going. My presentation will be a follow-up to the paper I published on this topic in January 2024: How can AI, LLMs and quantum science empower each other?.
See you there, maybe!
Lien du blog Opinions Libres : https://www.oezratty.net/wordpress
Lien de l'article : https://www.oezratty.net/wordpress/2024/using-google-notebooklm-with-understanding/
Cliquez ici pour imprimer
(cc) Olivier Ezratty - http://www.oezratty.net