LLM-Modulo: Using Critics and Verifiers to Improve Grounding of a Plan - Explanation + Improvements

Subscribers:
5,330
Published on ● Video Link: https://www.youtube.com/watch?v=_vZOIuV-nns



Duration: 0:00
438 views
13


LLMs are really bad at self-reflection / self-critique.

This is especially so if they do not have the domain knowledge required, or the problem is of the wrong abstraction space, or if rule-based accuracy is needed.

LLM Modulo architecture with verifiers/critics can mitigate a lot of issues after the output is generated.

The next question is, why not mitigate at the start, using agentic systems with rule-based components?

How can we create LLM agentic systems and incorporate robustness with verifiers/critics/rule-based grounding?

~~~

Slides: https://github.com/tanchongmin/Tensor...
Jupyter Notebook: https://github.com/tanchongmin/Tensor...


Paper: https://arxiv.org/abs/2402.01817

~~~

0:00 Introduction
3:03 One-shot and zero-shot BlocksWorld planning
7:54 ReAct only works for short trajectories
13:06 Planning
14:29 Reflection as a way of refining earlier experience
17:51 LLMs cannot self-critique well
23:12 Strawberry Example
29:10 The hard problem of planning
35:48 Simple schema for LLM Modulo
38:46 LLM Modulo Architecture
52:17 Comparison to ReAct
52:53 Critics
59:09 LLM Modulo for Travel Planning
1:03:21 My thoughts
1:06:50 My extension ideas for LLM Modulo
1:11:35 Critic/Verifiers in TaskGen
1:14:20 Discussion
1:27:27 Conclusion

~~~

AI and ML enthusiast. Likes to think about the essences behind breakthroughs of AI and explain it in a simple and relatable way. Also, I am an avid game creator.

Discord:   / discord  
LinkedIn:   / chong-min-tan-94652288  
Online AI blog: https://delvingintotech.wordpress.com/
Twitter:   / johntanchongmin  
Try out my games here: https://simmer.io/@chongmin




Other Videos By John Tan Chong Min


2024-10-21Tian Yu X John: Discussing Practical Gen AI Tips for Image Prompting
2024-10-08Jiafei Duan: Uncovering the 'Right' Representations for Multimodal LLMs for Robotics
2024-09-27TaskGen Tutorial 6: Conversation Wrapper
2024-09-26TaskGen Tutorial 5: External Functions & CodeGen
2024-09-24TaskGen Tutorial 4: Hierarchical Agents
2024-09-23TaskGen Tutorial 3: Memory
2024-09-19TaskGen Tutorial 2: Shared Variables and Global Context
2024-09-16Beyond Strawberry: gpt-o1 - Is LLM alone sufficient for reasoning?
2024-09-11TaskGen Tutorial 1: Agents and Equipped Functions
2024-09-11TaskGen Tutorial 0: StrictJSON
2024-09-10LLM-Modulo: Using Critics and Verifiers to Improve Grounding of a Plan - Explanation + Improvements
2024-09-06TaskGen: Co-create the best open-sourced LLM Agentic Framework together!
2024-08-21AriGraph (Part 2) - Knowledge Graph Construction and Retrieval Details
2024-08-13alphaXiv - Share Ideas, Build Collective Understanding, Interact with ANY open sourced paper authors
2024-07-30AriGraph: Learning Knowledge Graph World Models with Episodic Memory for LLM Agents
2024-07-23NeoPlanner - Continually Learning Planning Agent for Large Environments guided by LLMs
2024-07-17Intelligence = Sampling + Filtering
2024-07-12Michael Hodel: Reverse Engineering the Abstraction and Reasoning Corpus
2024-07-02TaskGen Conversational Class v2: JARVIS, Psychology Counsellor, Sherlock Holmes Shop Assistant
2024-06-04CodeAct: Code As Action Space of LLM Agents - Pros and Cons
2024-05-28TaskGen Conversation with Dynamic Memory - Math Quizbot, Escape Room Solver, Psychology Counsellor