LLM Q&A #1: Prompting vs Fine-Tuning, More vs Fewer Sources for RAG, Prompting vs LLMs as a System

Subscribers:
5,330
Published on ● Video Link: https://www.youtube.com/watch?v=oapjIiQAsUk



Duration: 7:11
176 views
9


I have been receiving some questions lately on how to use LLMs. Here are some of my brief thoughts on the matter:

Q1: Should I prompt or fine-tune an LLM?
Prompt before fine-tuning. It is possible to use both actually if needed. Prompting is about 1000 examples of fine-tuning.

Q2: Should I use more or fewer sources for Retrieval Augmented Generation?
Use as few as possible. A few small models with short context is better than a longer context one.

Q3: Should I focus my efforts on prompting techniques?
Not really. Focus on LLMs as a system so you get to augment the capabilities with tools.

~~~~~~~~~~~~~~~~~~~~~~

References:
How many data points is a prompt worth?
https://arxiv.org/pdf/2103.08493.pdf

Retrieval meets long context large language models
https://browse.arxiv.org/pdf/2310.03025.pdf

RECOMP: Improving retrieval-augmented LMs with compression and selective augmentation
https://arxiv.org/pdf/2310.04408.pdf

LLMs as Optimisers (DeepMind)
https://arxiv.org/pdf/2309.03409.pdf

PromptBreeder (DeepMind)
https://arxiv.org/pdf/2309.16797.pdf

LLMs as a System to solve the ARC Challenge
https://github.com/tanchongmin/TensorFlow-Implementations/blob/main/Paper_Reviews/LLMs%20as%20a%20System%20for%20the%20ARC%20Challenge.pdf

~~~~~~~~~~~~~~~~~~~~~~~

0:00 Q1: Prompting vs Fine-tune an LLM
2:05 Q2: More or Fewer Sources for RAG
4:27 Q3: Prompting vs LLMs as a System

~~~~~~~~~~~~~~~~~~~~~~~

AI and ML enthusiast. Likes to think about the essences behind breakthroughs of AI and explain it in a simple and relatable way. Also, I am an avid game creator.

Discord: https://discord.gg/bzp87AHJy5
LinkedIn: https://www.linkedin.com/in/chong-min-tan-94652288/
Online AI blog: https://delvingintotech.wordpress.com/
Twitter: https://twitter.com/johntanchongmin
Try out my games here: https://simmer.io/@chongmin




Other Videos By John Tan Chong Min


2023-11-20Tutorial #11: Virtual Persona from Documents, Multi-Agent Chat, Text-to-Speech to hear your Personas
2023-11-14A Roadmap for AI: Past, Present and Future (Part 3) - Multi-Agent, Multiple Sampling and Filtering
2023-11-07Learning, Fast and Slow: My Landmark Idea for fast, adaptable agents (ICDL 2023 Best Paper Finalist)
2023-11-06A roadmap for AI: Past, Present and Future (Part 2): Fixed vs Flexible, Memory Soup vs Hierarchy
2023-11-03AI & Education: Education when AI tools are smarter than us - Discussion with Kuang Wen (Part 2)
2023-11-03AI & Education: RAG Question-Answer, Test Question Generator, Autograder by Kuang Wen! (Part 1)
2023-10-31A Roadmap for AI: Past, Present and Future (Part 1)
2023-10-28Tutorial #10: StrictJSON v2 (StrictText): Handle any output - quotation marks or backslash!
2023-10-24ChatDev: Can LLM Agents really replace a software company?
2023-10-17LLMs and Robotics: An Overview by Daniel Tan!
2023-10-17LLM Q&A #1: Prompting vs Fine-Tuning, More vs Fewer Sources for RAG, Prompting vs LLMs as a System
2023-10-10LLMs as a System of Multiple Expert Agents to solve the ARC Challenge (Detailed Walkthrough)
2023-09-26Everything about LLM Agents - Chain of Thought, Reflection, Tool Use, Memory, Multi-Agent Framework
2023-09-19Moving Beyond Probabilities: Memory as World Modelling
2023-09-05Symbolic Regression: Doing What LLMs cannot - Deriving Arbitrary Mathematical Relations!
2023-08-29LLM Agents as a System (Prelim Findings Sharing): An Attempt to solve a 2-player 2D Escape Room!
2023-08-23LLM as Pattern Machines(Part 2) - Goal Directed Decision Transformers, 10-Year Plan for Intelligence
2023-08-18Tutorial #9: Evolution Game v2: ChatGPT (Text) and Dall-E (Image) API Integration!
2023-08-17Tutorial #8: Create a Web Scraper using ChatGPT and Selenium!
2023-08-17Tutorial #7: Create a Chatbot with Gradio and ChatGPT!
2023-08-15LLMs as General Pattern Machines: Use Arbitrary Tokens to Pattern Match?