Learn. Build. Grow.
26:12
Scaling Laws Explained Simply (Model, Data, Compute + GPU)
Build AI with Sandeep
Shared 1 week ago
32 views
14:00
Pre vs Post Layer Normalization
Shared 3 weeks ago
105 views
19:02
ReLU vs GELU vs Swish (Simple Explanation)
Shared 1 month ago
37 views
23:55
BIG Mistake in Adam | Adam vs AdamW
62 views
44:21
OPTIMIZERS EXPLAINED
Shared 2 months ago
43 views
26:54
The Secret That Almost Killed AI | Why Transformers Failed in 2018
58 views
30:09
From Transformers to LLMs: How AI Evolved from 2017 to 2023 | Build AI with Sandeep
127 views
27:36
Magic of Cross Entropy Loss | Must watch
54 views
19:14
Why Transformer Decoder Uses Linear + Softmax? (No Confusion Anymore)
Shared 4 months ago
15:02
Cross Attention Made Easy | Decoder Learns from Encoder
26 views
10:12
Masked Self-Attention Explained
169 views
19:42
The Untold Secrets of FFN in Transformers
Shared 5 months ago
165 views
14:28
LayerNorm: The Unknown Hero of Transformers
28:22
Self Attention, Multi-Head Attention & Skip Connections Explained Simply and Visually | Transformers
1:00:30
Transformer Encoder Explained with Visuals | Attention, Embedding, PE, Residual Connections
23:08
How LLMs Actually Work: The Secret Behind ChatGPT, PaLM & GenAI Explained Simply!
107 views
20:20
Intro to Gen AI
88 views