Day-1: Foundational ML for Generative AI
|
Session
|
Description
|
Remarks
|
1
|
Fundamentals of Machine Learning for LLM Dr. Hamza Zidoum, Dr. Noushath Shaffi
- Introduction to Machine Learning Concepts: Supervised and unsupervised learning, data splits (training/testing), and model evaluation (accuracy, precision, recall).
- Neural Network Basics: Overview of neural networks, including Perceptrons, activation functions, and layers.
Hands-On Exercise: Model building for classification/regression
|
Outcome: Participants will understand the basics of Machine Learning
Pre-requisite:
Python Programming
|
2
|
Fundamentals of Deep Learning for LLM
Dr. Abdelhamid, Dr. Abdur Rahman.
Hands-On Exercise: Image or Text Classification using Keras or Pytorch.
|
Outcome: Participants will understand basics of LLM essentials such as tokenizing and training them for representing natural language
|
3
|
Looking Inside Large Language Models
Dr. Fatma Al Raisi, Dr. Abdur Rahman
- An overview of Transformer Models
- Concepts of self-attention and the multi-head attention
- Brief intro to tools like Hugging Face for accessing and deploying models.
Hands-On Exercise: Small exercise generating text with a pre-trained transformer model
|
Outcome: Participants will understand the concept of Attention and what is fueling the
|
Day 2: Retrieval Augment Generative based Large Language Models for Document Question and answering
|
Session
|
Description
|
Remarks
|
1
|
Understand the basics of LLMs and how they can be used for document-based question answering
- Key concepts: tokenization, embeddings, attention mechanism.
- Overview of how LLMs understand and process documents.
- Introduction to document question answering (QA) using LLMs.
- Demo: Basic question-answering with a pre-trained LLM.
Hands-on Activity
- Setting up a basic LLM using Hugging Face or OpenAI API to answer simple questions from a document.
- Explore LLM outputs for different document inputs.
|
Outcome: Participants will understand the basics of LLM and how it can be used for document question and answering.
Pre-requisite
Python programming
Basics of machine learning
Duration: 2 hours
|
2
|
Retrieval-Augmented Generation (RAG) for Document QA
- Why RAG? Bridging retrieval with generation.
- RAG architecture: How retrieval and LLM generation work together.
- Chunking documents into retrievable pieces.
- Storing and retrieving document chunks using vector embeddings.
- Demo: Setting up a basic RAG pipeline using FAISS or PostgreSQL for embeddings.
Hands-on Activity
- Building a simple RAG system for document QA.
- Experiment with document chunking and embeddings for better results.
|
Outcome: Participants will understand the basics of RAG based LLM and how it can be used for document question and answering. How it is different from LLM and what are it advantages over LLM based Q&A.
Pre-requisite:
Python programming
Basics of machine learning
Duration: 2 hours
|
3
|
Develop a web-based interface for users to upload documents and ask questions.
- Introduction to Streamlit for rapid prototyping of web apps.
- Designing an interface for document upload and question input.
- Integrating the RAG model with Streamlit for real-time QA.
- Displaying results and real-time feedback as questions are answered.
- Demo: Uploading documents and querying them via a web interface.
Hands-on Activity
- Build a simple Streamlit application for document QA.
- Test the app by uploading Word/PDF documents and asking questions.
|
Outcome: How to build a quick prototype using Streamlit for DAQ.
Pre-requisite:
Python programming
Duration: 2 hours
|