🧠Compressing for AGI
Jack Rae
相关解读:https://zhuanlan.zhihu.com/p/619511222
原视频:https://www.youtube.com/watch?v=dO4TPJkeaaU&t=161s
Theme of talk
Think deeply about the training objective of foundation models
What are we doing , why dose it make sense, and what are the limitations?
Takeaways
seek the minimum description length to solve perception
Generative models are lossless compressors
Large language models are state-of-the-art lossless text compressors(?!)
Current limitation of thr apporach
Minimum Description Length
...and why it relates to perception
We want to deepest understanding of our observations
ones that generalize
Last updated