Loading...
「ツール」は右上に移動しました。
利用したサーバー: wtserver2
27699いいね 361793回再生

No one actually knows why AI works

No one really knows how generative AI works. Here’s how researchers working on AI interpretability are figuring it out.


Chapters
0:00 What we don’t know about AI
1:42 What we do know about AI
3:15 Emergent capabilities
5:12 Building ≠ understanding
6:23 Interpretability research - golden gate claude
7:32 Interpretability research - circuits
8:31 Who cares?
10:11 Sorry this wasn’t a skit :/

Papers Cited
Emergent Abilities of Large Language Models: arxiv.org/pdf/2206.07682
Attention is All You Need: arxiv.org/pdf/1706.03762
Emergent Abilities of Large Language Models: arxiv.org/pdf/2206.07682
Are Emergent Capabilities of LLMs a Mirage: arxiv.org/pdf/2304.15004

Anthropic Interpretability Research
Golden Gate Claude: www.anthropic.com/news/golden-gate-claude
Tracing thoughts: www.anthropic.com/news/tracing-thoughts-language-m…

Videos explaining LLMs / transformers
   • Attention in transformers, step-by-step | DL6  
   • Large Language Models explained briefly  

Merch
For developer mugs, hoodies and more, check out: alberta.nyc/

Socials
instagram.com/alberta.tech
tiktok.com/@alberta.nyc
x.com/albertadevs

#ai #largelanguagemodels #llms #airesearch #anthropic #softwareengineer

This video is not sponsored.
FTC Disclosure: This video contains affiliate links. If you purchase through one of these links, I may earn a small commission at no additional cost to you.

コメント