[{"data":1,"prerenderedAt":79},["ShallowReactive",2],{"term-t\u002Ftransformer":3,"related-t\u002Ftransformer":60},{"id":4,"title":5,"acronym":6,"body":7,"category":40,"description":41,"difficulty":42,"extension":43,"letter":44,"meta":45,"navigation":46,"path":47,"related":48,"seo":54,"sitemap":55,"stem":58,"subcategory":6,"__hash__":59},"terms\u002Fterms\u002Ft\u002Ftransformer.md","Transformer",null,{"type":8,"value":9,"toc":33},"minimark",[10,15,19,23,26,30],[11,12,14],"h2",{"id":13},"eli5-the-vibe-check","ELI5 — The Vibe Check",[16,17,18],"p",{},"The Transformer is THE architecture behind all modern AI. ChatGPT, Claude, Midjourney, Whisper — all transformers under the hood. The key innovation? The attention mechanism that lets the model look at all parts of the input at once, instead of one word at a time. The 2017 paper 'Attention Is All You Need' is probably the most impactful paper in the history of AI.",[11,20,22],{"id":21},"real-talk","Real Talk",[16,24,25],{},"The Transformer is a neural network architecture introduced in 'Attention Is All You Need' (Vaswani et al., 2017). It uses self-attention mechanisms to process entire sequences in parallel, replacing recurrent and convolutional architectures. Key components include multi-head self-attention, positional encoding, layer normalization, and feed-forward networks. It's the foundation of all modern LLMs and many vision and audio models.",[11,27,29],{"id":28},"when-youll-hear-this","When You'll Hear This",[16,31,32],{},"\"Every modern LLM is based on the Transformer architecture.\" \u002F \"Transformers parallelized sequence processing and changed everything.\"",{"title":34,"searchDepth":35,"depth":35,"links":36},"",2,[37,38,39],{"id":13,"depth":35,"text":14},{"id":21,"depth":35,"text":22},{"id":28,"depth":35,"text":29},"ai","The Transformer is THE architecture behind all modern AI. ChatGPT, Claude, Midjourney, Whisper — all transformers under the hood. The key innovation?","intermediate","md","t",{},true,"\u002Fterms\u002Ft\u002Ftransformer",[49,50,51,52,53],"Attention Mechanism","Self-Attention","Neural Network","LLM","Deep Learning",{"title":5,"description":41},{"changefreq":56,"priority":57},"weekly",0.7,"terms\u002Ft\u002Ftransformer","jni3WDC1hvA2Ix6rl1EZJpNrM-5FuJvRVMA9Kzew5qQ",[61,65,68,73,76],{"title":49,"path":62,"acronym":6,"category":40,"difficulty":63,"description":64},"\u002Fterms\u002Fa\u002Fattention-mechanism","advanced","The attention mechanism is how AI decides what to focus on — like when you're reading a long email and your eyes jump to the part that mentions your name.",{"title":53,"path":66,"acronym":6,"category":40,"difficulty":42,"description":67},"\u002Fterms\u002Fd\u002Fdeep-learning","Deep Learning is Machine Learning that's been hitting the gym.",{"title":52,"path":69,"acronym":70,"category":40,"difficulty":71,"description":72},"\u002Fterms\u002Fl\u002Fllm","Large Language Model","beginner","An LLM is a humongous AI that read basically the entire internet and learned to predict what words come next, really really well.",{"title":51,"path":74,"acronym":6,"category":40,"difficulty":42,"description":75},"\u002Fterms\u002Fn\u002Fneural-network","A neural network is a system loosely inspired by the human brain — lots of little math nodes connected together, passing numbers to each other.",{"title":50,"path":77,"acronym":6,"category":40,"difficulty":63,"description":78},"\u002Fterms\u002Fs\u002Fself-attention","Self-attention is how a model looks at a sentence and figures out which words are most important to each other.",1776518319858]