AI and Football. Trying to stay young through constant learning.

Bengaluru
Joined April 2021
What are even the replies to these kinda posts? New brain-rot impression farming strat?
Guess the Legend Level : Hardest
"R2 you use because your boss only knows this one" 😭😭😭
- MSE you use when you don’t have outliers - RMSE you use when you want to interpret the above better - MAE you use when you have positive/zero/negative values and outliers - MAPE you use when you only have positive values and emphasize interpretability - RMSLE you use for positive values with nonnormal distribution - wMAPE you use when you want MAPE but have large vs small values - sMAPE you use when you want MAPE but have zero/negative values - R2 you use because your boss only knows this one
Casado = Vini? Jokes aside, not an yellow, not a red.
Sensei Karpathy just dropped the code for nanochat! ~8k lines of code (mostly handwritten!) - Rust trainer for tokenizer - FineWeb GPT pretraining - Smoltalk SFT - RL - GRPO on GSM8K - HumanEval, ARC, MMLU, GSM8K Eval - KV cache and prefill/decode inferenec
Excited to release new repo: nanochat! (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single, dependency-minimal codebase. You boot up a cloud GPU box, run a single script and in as little as 4 hours later you can talk to your own LLM in a ChatGPT-like web UI. It weighs ~8,000 lines of imo quite clean code to: - Train the tokenizer using a new Rust implementation - Pretrain a Transformer LLM on FineWeb, evaluate CORE score across a number of metrics - Midtrain on user-assistant conversations from SmolTalk, multiple choice questions, tool use. - SFT, evaluate the chat model on world knowledge multiple choice (ARC-E/C, MMLU), math (GSM8K), code (HumanEval) - RL the model optionally on GSM8K with "GRPO" - Efficient inference the model in an Engine with KV cache, simple prefill/decode, tool use (Python interpreter in a lightweight sandbox), talk to it over CLI or ChatGPT-like WebUI. - Write a single markdown report card, summarizing and gamifying the whole thing. Even for as low as ~$100 in cost (~4 hours on an 8XH100 node), you can train a little ChatGPT clone that you can kind of talk to, and which can write stories/poems, answer simple questions. About ~12 hours surpasses GPT-2 CORE metric. As you further scale up towards ~$1000 (~41.6 hours of training), it quickly becomes a lot more coherent and can solve simple math/code problems and take multiple choice tests. E.g. a depth 30 model trained for 24 hours (this is about equal to FLOPs of GPT-3 Small 125M and 1/1000th of GPT-3) gets into 40s on MMLU and 70s on ARC-Easy, 20s on GSM8K, etc. My goal is to get the full "strong baseline" stack into one cohesive, minimal, readable, hackable, maximally forkable repo. nanochat will be the capstone project of LLM101n (which is still being developed). I think it also has potential to grow into a research harness, or a benchmark, similar to nanoGPT before it. It is by no means finished, tuned or optimized (actually I think there's likely quite a bit of low-hanging fruit), but I think it's at a place where the overall skeleton is ok enough that it can go up on GitHub where all the parts of it can be improved. Link to repo and a detailed walkthrough of the nanochat speedrun is in the reply.
omg. A Macbook is great for beginner devs. I get it!! Now STFU!
> Games without Balde & Lamine playing 90 mins - amazing results! > Olmo and Lewa are THE combo - need to see more playing time together. > Pedri looked slightly tired, but FDJ - what an amazing player. > Araujo and Andreas - new pairing, but never looked out of touch. Onwards!
Frenkie is almost playing like an English box-to-box MF at times in this game, allowing Pedri to stay deeper. Never seen him go this forward in a game in quite a while.
I now recognise what @RobyPoto says - Olmo needs Lewa. Ferran needs Fermin - 100%.
• Ferran's off the ball movement is really really improving. • Gerard has been really under-appreciated - filling in for Balde and putting in solid performances. • The camoes from Rashford and Ronny cameo - really invigorating. Both play like they have a point to prove.
Understand this - No one can ever be the "best" midfielder. Not Pedri, Modric, Iniesta, Zidane, Xavi, Pirlo, Makelele, Scholes. There can be: - Favourite (fans') or - Perfectly suited (to team style) There are different player profiles, team structures, and opp play styles.
With all the magic Frenkie de Jong and Pedri produced yesterday, some accounts will just point to the 0 G/A stat. Unreal idiocracy.
Overall: • Great defensive performance - weather the storm at the start of both halves and slow the game down. • Right decision to switch from central progression dependence to wing play. • Amazing result without 2 of our best starters in Lamine and Balde.
• Rashy - absolutely fantastic, amazing moments. • Kounde, Joan - rock solid. 100% throughout 90 mins. • Pedri and FDJ - magical as usual. Held their own against a heavily physical English team.
• Rashy - absolutely fantastic, amazing moments. • Kounde, Joan - rock solid. 100% throughout 90 mins. • Pedri and FDJ - magical as usual. Held their own against a heavily physical English team.
Pedri is the kind of player every team needs, but none of the fans would openly accept it.
1
Marcus Rashford, what the holy fuck.
The GPU Glossary at @modal is one of the best resources for starters! Love the theme and Amazing work. @ekzhang1 @charles_irl - Amazing work
1
1
Atleast it knows...
Senior AI architect: I used ChatGPT to figure out a few solutions for this problem statement...
So United IS the problem...
Andre Onana in his first match with 10-man Trabzonspor against Fenerbahçe: - 29 shots faced 😳 - 8 saves 🧤 - 4 diving saves 🫸 - 1.63 goals prevented ❌ - 8.7 rating (@FotMob - MOTM) 📈