No one really knows how generative AI works. Here’s how researchers working on AI interpretability are figuring it out.
Chapters
0:00 What we don’t know about AI
1:42 What we do know about AI
3:15 Emergent capabilities
5:12 Building ≠ understanding
6:23 Interpretability research - golden gate claude
7:32 Interpretability research - circuits
8:31 Who cares?
10:11 Sorry this wasn’t a skit :/
Papers Cited
Emergent Abilities of Large Language Models: arxiv.org/pdf/2206.07682
Attention is All You Need: arxiv.org/pdf/1706.03762
Emergent Abilities of Large Language Models: arxiv.org/pdf/2206.07682
Are Emergent Capabilities of LLMs a Mirage: arxiv.org/pdf/2304.15004
Anthropic Interpretability Research
Golden Gate Claude: www.anthropic.com/news/golden-gate-claude
Tracing thoughts: www.anthropic.com/news/tracing-thoughts-language-m…
Videos explaining LLMs / transformers
• Attention in transformers, step-by-step | DL6
• Large Language Models explained briefly
Merch
For developer mugs, hoodies and more, check out: alberta.nyc/
Socials
instagram.com/alberta.tech
tiktok.com/@alberta.nyc
x.com/albertadevs
#ai #largelanguagemodels #llms #airesearch #anthropic #softwareengineer
This video is not sponsored.
FTC Disclosure: This video contains affiliate links. If you purchase through one of these links, I may earn a small commission at no additional cost to you.
コメント