Educational Applications of AI in Web Browsers #108
Replies: 4 comments 1 reply
-
Hi Adam, Great questions. Some answers to get you started:
Yes, here's a prompt asked of -003:
I think so, i.e., you can ask the AI to prioritize different content in its responses. I believe sections and other docs would be fine. This would be worth experimenting for speicfics though:
You could include the docs & articles in your prompts. Another thought is using a mix of native and semantic code to chunk your docs up, summarize pieces of them, and ask the AI to adjust priority based on those summaries. Skills I think you might be interested in: HTTPSkill, Summarizer skill. I suspect you will want to customize what is there, using your own custom prompts as you build our your app above. I suspect though the inbox skills will get you through a POC and give you a feel for what you'd want to change / add. HTH- TIm |
Beta Was this translation helpful? Give feedback.
-
Adding @Stevenic here to see if he can help @AdamSobieski with any ideas on prompts for this great use case. |
Beta Was this translation helpful? Give feedback.
-
@AdamSobieski to pile on to what @timlaverty suggested you're probably going to want to ground the model by showing it all or at least part of the document you want to discuss. The key issue here is that you have only so many tokens to work with and passing in a large document could easily chew up most of your budget. As Tim suggested you could break your document up into chunks and then use Semantic Search to identify the most relevant portions of the document in the prompt. To do that the document chunks need to be converted into embeddings and the users question needs to be converted to embeddings. You can then use cosine similarity to compare each chunk to the users query and stack the rank the results. It's probably a research topic in itself to identify how to best break up the source document... |
Beta Was this translation helpful? Give feedback.
-
One tricky bit will be how do you identify new document chunks to pull into the prompt... You want your students to feel like they can discuss any part of the document or text regardless of its size. This could mean that as the conversation progresses new parts of the document might become more relevant... To help with that you could take not just the users current query but maybe the last 5 user message and generate embeddings for that and compare that to your document chunks. That way you're building a sliding window of sort over the document that can be moved around as the conversation progresses... I'd probably also feed the document fragments into the prompt in source order and give the model a since of where the gaps are. For example your prompt might look like this:
This will give the model a sense of what information it's not seeing. Just keep in mind it might try to fill in those gaps on its own. If the document has a summary it might be useful to always include that so that if the model does try and stray from the facts, it at least a hope of synthesizing something that could be true. |
Beta Was this translation helpful? Give feedback.
-
Hello. I am interested in educational applications of AI in Web browsers.
I am looking forward to students being able to ask questions about and to engage in dialogues about Web documents, e.g., Wikipedia articles, and to do so with respect to PDF documents displayed in their Web browsers, e.g., arXiv, Google Scholar, Semantic Scholar, etc. Interestingly, these technologies could also benefit students by interoperating with their digital textbooks.
With respect to interoperations between AI and Web browsers, one can envision the page and scroll positions of documents as well as any content selections as being relevant to end-users’ questions and dialogues. One can also envision AI components being able to turn pages, scroll documents, and highlight selections of content while responding to end-users. Soon, AI systems may be able to open documents for end-users and subsequently turn pages, scroll to, and highlight relevant content while communicating.
I would like to request more information and documentation about how the document-related skills work. I am eager to learn more about these topics to be able to better brainstorm about educational applications of AI in Web browsers and about how new Web standards could be of use.
For starters: Can developers provide one or more documents' URLs and, potentially, content selectors to LLMs? Can developers ask questions while indicating for LLMs to consider one or more documents, or chapters or sections thereof, as being prioritized above others while answering those questions? Would this, instead, require uploading portions of documents before or during end-users' questions or dialogue moves? Thank you for any more information about LLMs and document-related skills!
Beta Was this translation helpful? Give feedback.
All reactions