Learning Part-Whole Structure by Chunking - More Efficient than Deep Learning!!!

Subscribers:
5,330
Published on ● Video Link: https://www.youtube.com/watch?v=Y-goByHfsoo



Category:
Discussion
Duration: 2:33:52
813 views
13


I met Shuchen at NeurIPS 2022, and she is doing really interesting work on chunking at her lab, Max Planck Institute for Biological Cybernetics! Happy to invite her to give a talk about her work at our discussion session:

The world is complex - how do we learn a compact representation of the world? Shuchen will introduce the Hierarchical Chunking Model (HCM), a cognitive-inspired algorithm that learns nested hierarchies of chunks from sequences. Hierarchical chunking may be the key to creating part-whole structures that we use to perceive the world and applies to data domains, including text, image, audio, and other temporal sequences.

We had a great discussion, and ultimately, the consensus is that chunking may be a phenomenon that is online (that is, learned through experience), and we can use chunking to get the representation needed for semantic meaning later. This is much like Byte Pair Encoding generating tokens, which are then projected in an embedding space in the Transformer Architecture.

Paper: https://openreview.net/pdf?id=LceHl9wKmoQ

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

0:00 Introduction
4:11 Chunking is a widespread learning phenomenon
21:36 Hierarchical Generative Model
24:35 How to generate from hierarchical structure
25:28 Hierarchical Chunking Model
37:16 Link to Digital Signal Processing
44:14 Concerns with Exponential Increase in Chunk Length (no concern - only finitely many chunks possible)
55:07 HCM is better than RNN when sequence contains hierarchy
1:07:32 Forgetting is important - Pre-learned chunks can lead to worse performance if environment is different
1:11:26 HCM is similar to human chunk learning
1:15:15 Learning part-whole relationships
1:27:33 Summary
1:29:20 Start of discussion
1:37:35 Link to Hinton’s Part Whole Hierarchy and Link to LeCun’s Self-Supervised Learning
1:40:16 Link to Byte-Pair Encoding
1:43:38 Chunking as representation for more efficient deep learning models
1:51:24 What is the most basic unit for chunking?
2:02:10 Detailed discussion on linkage to Byte-Pair Encoding
2:14:13 Online Adaptive HCM!!
2:32:15 Concluding Remarks

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~


AI and ML enthusiast. Likes to think about the essences behind breakthroughs of AI and explain it in a simple and relatable way. Also, I am an avid game creator.

Discord: https://discord.gg/fXCZCPYs
LinkedIn: https://www.linkedin.com/in/chong-min-tan-94652288/
Online AI blog: https://delvingintotech.wordpress.com/.
Twitter: https://twitter.com/johntanchongmin
Try out my games here: https://simmer.io/@chongmin




Other Videos By John Tan Chong Min


2023-04-18The future is neuro-symbolic: Expressiveness of ChatGPT and generalizability of symbols (SymbolicAI)
2023-04-17Can GPT4 solve the Abstraction and Reasoning Corpus (ARC) Challenge Zero-Shot?
2023-04-12GPT4: Zero-shot Classification without any examples + Fine-tune with reflection
2023-04-11OpenAI Vector Embeddings - Talk to any book or document; Retrieval-Augmented Generation!
2023-04-11Tutorial #2: OpenAI Vector Embeddings and Pinecone for Retrieval-Augmented Generation
2023-04-04Creating JARVIS: ChatGPT + APIs - HuggingGPT, Memory-Augmented Context, Meta GPT structures
2023-04-02Is GPT4 capable of self-improving? Are we heading for AGI or AI doom?
2023-03-28How Visual ChatGPT works + Toolformer/Wolfram Alpha. LLMs with Tools/APIs/Plugins is the way ahead!
2023-03-21Tokenize any input, even continuous vectors! - Residual Vector Quantization - VALL-E (Part 2)
2023-03-07Using Transformers to mimic anyone's voice! - VALL-E (Part 1)
2023-02-28Learning Part-Whole Structure by Chunking - More Efficient than Deep Learning!!!
2023-02-21High-level planning with large language models - SayCan
2023-02-13Learning, Fast and Slow: Towards Fast and Adaptable Agents in Changing Environments
2023-02-07Using Logic Gates as Neurons - Deep Differentiable Logic Gate Networks!
2023-01-31Learn from External Memory, not just Weights: Large-Scale Retrieval for Reinforcement Learning
2023-01-17How ChatGPT works - From Transformers to Reinforcement Learning with Human Feedback (RLHF)
2023-01-09HyperTree Proof Search - Automated Theorem Proving with AlphaZero and Transformers!
2022-12-23CodinGame Fall Challenge 2022: A First Look (managed to get to Silver!)
2022-12-21Can ChatGPT solve CodinGame/Google Kickstart problems?
2022-12-19Reinforcement Learning Fast and Slow: Goal-Directed and Memory Retrieval Mechanism!
2022-12-12A New Framework of Memory for Learning (Part 1)