Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Milestone 4 - Subtask 3: Finetune model with Q/A format dataset #108

Open
TobeTek opened this issue Oct 10, 2024 · 3 comments
Open

Milestone 4 - Subtask 3: Finetune model with Q/A format dataset #108

TobeTek opened this issue Oct 10, 2024 · 3 comments
Assignees

Comments

@TobeTek
Copy link
Collaborator

TobeTek commented Oct 10, 2024

Potential Data Format that gives courses to take and reasoning. Will be replace with LPI Dataset courses/infoy

input: I want to learn about llm and how to finetune them. I'm intermediate and i want to build a rag pipeline
output: Given your intermediate level and interest in building a RAG pipeline using LLMs, I recommend the following courses and modules:

Course/Module Source Level Duration (Estimate) Keywords Reason
Introduction to Large Language Models (LLMs) Various online courses (e.g., Coursera, edX) Beginner/Intermediate Varies (4-8 weeks) LLMs, transformers, attention mechanisms, language modeling, tokenization Provides foundational knowledge of LLMs, their architecture, and capabilities. Essential before tackling fine-tuning or RAG.
Natural Language Processing (NLP) Fundamentals Various online courses (e.g., Stanford NLP) Intermediate Varies (6-10 weeks) NLP, text preprocessing, word embeddings, sentiment analysis, named entity recognition Necessary for understanding how LLMs process and understand text. Many RAG techniques rely on NLP for data preprocessing and query understanding.
Fine-tuning LLMs Hugging Face Course, Papers with Code Intermediate/Advanced Varies (2-4 weeks) Fine-tuning, transfer learning, hyperparameter tuning, model evaluation Teaches you how to adapt pre-trained LLMs to specific tasks, crucial for building a high-performing RAG system. Learn techniques like prompt engineering.
Retrieval Augmented Generation (RAG) Techniques Research Papers, Blogs, Tutorials Advanced Varies (Ongoing Study) RAG, vector databases, embedding generation, knowledge retrieval, question answering Focuses on the architecture and implementation of RAG pipelines. You'll learn to select and integrate components like vector databases (e.g., Pinecone, Weaviate) and retrieval methods.
Python for Data Science (if needed) DataCamp, Codecademy, Fast.ai Intermediate Varies (2-4 weeks) Python, pandas, numpy, scikit-learn Reinforce your Python skills for data manipulation and model building within your RAG pipeline.
Vector Databases (if needed) Pinecone, Weaviate documentation Intermediate Varies (1-2 weeks) Vector databases, similarity search, indexing, scalability Understanding vector databases is crucial for efficient knowledge retrieval in a RAG pipeline. Learn about different databases and their strengths.
@Tacoman99
Copy link
Collaborator

Tacoman99 commented Oct 21, 2024

Description moved to #108 (comment)

@Tacoman99 Tacoman99 self-assigned this Oct 21, 2024
@neomatrix369
Copy link
Owner

neomatrix369 commented Oct 23, 2024

Description moved to #108 (comment)

As discussed on Discord, you create a new dataset based on the LPI dataset and use that for your training and other purposes.

LPI dataset is still needed for our other purposes (and DOES NOT need to be replaced), plus it holds metadata about the courses which is added information (if can or need to be used).

Can you pls amend your wordings to remove any confusions to others who are also referring or reading this issue.

@neomatrix369
Copy link
Owner

@Tacoman99 just moved the ticket to In Progress and also added it to the LPI project

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants