How Visual ChatGPT works + Toolformer/Wolfram Alpha. LLMs with Tools/APIs/Plugins is the way ahead!
ChatGPT has been giving us very human-like responses, and has proven to be a user friendly interface to talk to. What if we can bring it to the next level by integrating it with plugins/tools/APIs? These tools can enable ChatGPT to communicate with applications, retrieve data from the web, buy your favourite meal from your favourite store, play chess using AlphaZero, send an email to your boss and many more.
We will go through how tools can be integrated in a few-shot manner using the Toolformer Input-Output example approach, in a zero-shot manner using the Visual ChatGPT/OpenAI plugin description approach. We also illustrate some failure cases (based on my own experimentation) of failing to call the right tool, or calling the tool with the wrong input. These failure cases reflect some of the weaknesses of ChatGPT to understand complex rules, such as finding words with no vowels, or comparing numbers against a certain threshold.
Overall, ChatGPT with plugins/tools/APIs is very promising, and the tools help to mitigate some of the flaws of ChatGPT. If infused with my "Learning, Fast & Slow"-style architecture, it can even be adaptive to the environment! There is a lot of promise for this approach, and I believe we are not too far from creating Jarvis from Iron Man.
~~~~~
Slides: https://github.com/tanchongmin/TensorFlow-Implementations/blob/main/Paper_Reviews/LLMs%20as%20API%20Interface.pdf
Related videos:
Part 2: https://www.youtube.com/watch?v=wLjJ34ygZVc
How ChatGPT works: https://www.youtube.com/watch?v=wA8rjKueB3Q
Learning, Fast and Slow (Adaptive learning): https://www.youtube.com/watch?v=Hr9zW7Usb7I
References:
Toolformer Paper: https://arxiv.org/abs/2302.04761
Visual ChatGPT Code + Paper: https://github.com/microsoft/visual-chatgpt
Wolfram Alpha Plugin announcement from Stephen Wolfram: https://writings.stephenwolfram.com/2023/03/chatgpt-gets-its-wolfram-superpowers/
Emergence properties of LLM paper: https://arxiv.org/abs/2206.07682
ReAct paper: https://arxiv.org/abs/2210.03629
LangChain documentation: https://python.langchain.com/en/latest/
OpenAI API Key: https://platform.openai.com/account/api-keys
OpenAI Plugin Documentation: https://platform.openai.com/docs/plugins/introduction
~~~~~
0:00 Introduction
6:38 LLMs vs Tools
11:15 What’s possible
35:03 What APIs can be interfaced
39:49 Neuro-symbolic AI
41:50 Make LLMs adaptive - Learning, Fast and Slow
44:30 Toolformer
52:38 Visual ChatGPT
58:53 ReAct for Chain of Thought prompting
1:04:53 Overall workflow of API calling
1:20:53 Code walkthrough
1:29:00 Live Demo of Visual ChatGPT
1:43:50 Limitations of LLMs + Tools
1:48:50 Failure cases for calling tools
1:53:13 Conclusion
~~~~~~
AI and ML enthusiast. Likes to think about the essences behind breakthroughs of AI and explain it in a simple and relatable way. Also, I am an avid game creator.
Discord: https://discord.gg/fXCZCPYs
LinkedIn: https://www.linkedin.com/in/chong-min-tan-94652288/
Online AI blog: https://delvingintotech.wordpress.com/.
Twitter: https://twitter.com/johntanchongmin
Try out my games here: https://simmer.io/@chongmin
Other Videos By John Tan Chong Min
Other Statistics
Wolfram Statistics For John Tan Chong Min
There are 1,842 views in 1 video for Wolfram. About an hours worth of Wolfram videos were uploaded to his channel, less than 0.61% of the total video content that John Tan Chong Min has uploaded to YouTube.