## Summary Notes ### Wednesday Jan 10th, 2024 9:00 AM Nirav's introduction. Mithun discussion: What are LLM? - what is intelligence? - ChatGPT is not intelligent - What are LLM? - What is a Turing test? - Can chatGPT pass a test? - Why are we here today? - What are the issues we will come up against? - Privacy concerns: all information passed to ChatGPT is public, not good for HIPAA, FERPA and HSR proteted data - "No guardrails" in terms of biases 10:00 AM - [Ollama](https://github.com/jmorganca/ollama) Nirav demo running in CyVerse (running on [GPU 06:](http://gpu06.cyverse.org:8080)) - Goal for the codeathon: create an open source chat bot that resembles ChatGPT. - Allows: - Professors to upload data - Students to ask questions - Timeline of deliverables: - Wed: model definition, GUI, Cloud Infrastructure, Auth, Course info - Thu: minimum viable product - Fri: testing - Testing material: - Class RNR355 material is going to be used for testing - vs ChatGPT4: - Known limitation using openAI: max 10 files, 25MB per file, USD20 subscription - Thoughts: - Cost - Data ingenstion (mathematics, pdfs, images) - Efficiency- does every information retreival need LLMs? - The realities of time and effort constraints- important to focus on interpretability existing tools - Focus on quality of answers. 10:45 Tyson - Winter project: set up information retreival chatbot for course materials - React app that interfaces with OpenAI backend - ChatGPT 4.0 chat is accurate, accuracy falls and speed increases if ChatGPT 3 is used - Tutorbot, vs Teacherbot- future stretch goal of creating something that can work with private data between teachers and students for personalized teaching materials - [Flowise AI](https://flowiseai.com/) - Could be used to create what we aim to do - Can work with both open access and paywalled models 11:00 Nirav - Storyboarding for the chatbot and use case, for a walk through - Two use cases- course materials, website - Grouper- how to restrict chatbot access to each course, so authorization is restricted - What other universities are doing? - Reading Math equations - Is LLM training data private or public? - How deep to follow URL links in a document? - Use Local API for privacy - Low code-no code for Professors - Internal - Need to integrate other commercial solutions if plausible - D2L is moving to Canvas (LMS chatbot) - Office 365 (high prices for including chatbots) Mithun - Laying down basis for day achievable goals - Described some initial tasks for the Team leads - Prompt engineering - Each team decides more realistic tasks 1:30PM ### Propeller Team (LLM) - how a pipeline works. 1. Create Vector DB; Chunks of documents, pass them through LLM in order to create Vector DB. 2. Retrieval: Get the closest document to the query 3. Generate: Depending on the task: Summary/ QA/ Translating Document etc. Terminology: Retrieval Augmented Generation (RAG) - prime the LLM with context, which can help combat hallucinations - entire text book is too long to provide in the prompt, rather we use the vector db to pick relevant chunks for dynamic prompt *Technologies:* - [Chroma](https://www.trychroma.com/) - [LangChain](https://www.langchain.com/) - [Nvidia Nemo](https://www.nvidia.com/en-us/ai-data-science/products/nemo/get-started/) ### Rudder (Back-End) - items ### Crow's nest (GUI) - items - ### Anchor (Testing/QA) - items -