Skip to content

CLASS/rag-llm-workshop

main
Switch branches/tags

Name already in use

A tag already exists with the provided branch name. Many Git commands accept both tag and branch names, so creating this branch may cause unexpected behavior. Are you sure you want to create this branch?
Code

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 

Developing a Web-based Q&A system using a RAG-based approach to LLMs - June 25, 2024

Course Information

  • Title: Developing a Web-based Q&A system using a RAG-based approach to LLMs
  • Instructor: Sujee Maniyam, Node 51
  • Time: Tuesday, June 25, 8.00 am PDT/11am pm EDT till 2.00pm PDT / 5pm EDT
  • Cost: $300 for NET+ AWS and GCP Subscribers, $400 for Internet2 members(see the list of member institutions here), $500 for non-members
  • Registration: https://tinyurl.com/3tnfey9n

Introduction

We have all seen how powerful and impactful Large Language Models (LLMs) like ChatGPT are. These very powerful AI models are enabling a new generation of conversational applications. This 6-hour hands-on workshop introduces the audience to LLMs, data augmentation using the Retrieval Generation Augmentation (RAG) approach, and how to how to build web using open-source tools.

Skills Level and Target Audience

  • Introductory to Intermediate
  • Research IT, Enterprise IT

What You Will Learn

  • Embeddings and embedding models; searching text for meanings / context using embeddings
  • Vector databases and performing text search with vector databases
  • LLMs and using them (via API and locally)
  • Developing applications using LLM
  • Developing RAG applications
  • Deploying containerized applications

Prerequisites

  • Comfortable with python programming and notebooks
  • Access to Google Colab
  • A free subscription to a vector database like MongoDB Atlas
  • Sign up for Mistral AI
  • GPU access to TBD

Details

1 - Embeddings

  • Understanding embeddings
  • Various embedding models
  • Semantic text search using embeddings
  • evaluating a various embedding models

2 - Vector Databases

  • Introduction to vector databases
  • getting started with MongoDB Atlas
  • Loading data into database and populate embeddings
  • Vector search with database and embeddings

3 - LLMs

  • Introduction to LLMs, the eco system
  • Access LLMs via API
  • Run LLMs locally using llama-cpp, oolama, lm-studio, jen
  • Experiment with different LLMs

4 - Running LLMs

  • We will use a framework like llama-cpp, llama-index and Langchain to run local LLMs
  • Run a local LLM. Experiment with various open LLMs like Mistral, Llama

5 - Develop a custom application using LLM

  • Use frameworks like streamlit, flask to build sample applications

6 - Building RAG Applications

  • Here we will query PDF documents using LLMs
  • Index documents with embeddings. Use various embedding models (OpenAI, Mistral, open source models)
  • query the documents using various LLMs (OpenAI, Mistral, LLama)

7 - Deploying RAG Applications

  • Containerizing applications
  • Model serving
  • Scalability

8 - Workshop / Project

  • Attendees will a sample application using data of interest, and LLM of their choice

About

Repository for Internet2's CLASS RAG, LLM and Cloud Workshop

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published