Pinned Tweet
First sneak peak at what I am building. A mix of Deep Research and Multi-agent systems. The goal is to have your own research lab that can go off for a long time and really do a deep dive on any topic you want. At least that’s the goal, will start small and document as I go along
I no longer ask people what they do, I ask them what they are building
1
1
2
Open-source model regulation incoming?
This is an important one, I think. AI progress and recommendations: openai.com/index/ai-progress…
Zakrea retweeted
Great to see GLM-4.6 powering Cerebras Code. This is exactly why we open weights: so teams can combine their own infra and ideas with GLM capabilities, and bring more choices to developers worldwide. Huge welcome to all partners building on GLM. Let’s grow the ecosystem together.
Cerebras Code just got an UPGRADE. It's now powered by GLM 4.6 Pro Plans ($50): 300k ▶️ 1M TPM @  24M Tokens/day Max Plans ($200): 400k ▶️ 1.5M TPM @ 120M Tokens/day Fastest GLM provider on the planet at 1000 tokens/s and at 131K context. Get yours before we run out 👇
Zakrea retweeted
I think people still don’t realize how absolutely insane it is that Kimi trained their model with only $4.6 million - and what that actually means. It’s mind-blowing and an incredible achievement by the team.
Zakrea retweeted
Which humanoid robot is actually the best? Which is the most capable across the board? I understand people may not fully realize it, but the reality is it's the Unitree G1. It's the Unitree G1... not because it's the best humanoid. It's b/c it ships To people who build.
22
6
4
160
I love Kimi K2 Thinking! Honestly, the best AI model I have used to date. Thank you @Kimi_Moonshot.
Ye I do not have this long Kimi
Zakrea retweeted
Kimi-K2 Thinking crushes all frontier models except GPT-5-high on the Artificial Intelligence Index Open-Source models closed the gap!
We need more of these open-source TTS and STT models. Thank you StepFun!
🚀 Step-Audio-EditX is now open source!! ✨ Zero-Shot TTS with high timbre similarity ✨ Iterative editing of dozens of audio emotion and speaking style ✨ Fine-grained control over paralinguistic features Whether for audio editing, interactive design, or personalized scenarios, it unlocks unprecedented audio expression for you! 🌟GitHub: github.com/stepfun-ai/Step-A… 📑 arXiv:arxiv.org/abs/2511.03601 🔥 Demo Page:stepaudiollm.github.io/step-… 🎮 HF playground: huggingface.co/spaces/stepfu…
Zakrea retweeted
🚀 Step-Audio-EditX is now open source!! ✨ Zero-Shot TTS with high timbre similarity ✨ Iterative editing of dozens of audio emotion and speaking style ✨ Fine-grained control over paralinguistic features Whether for audio editing, interactive design, or personalized scenarios, it unlocks unprecedented audio expression for you! 🌟GitHub: github.com/stepfun-ai/Step-A… 📑 arXiv:arxiv.org/abs/2511.03601 🔥 Demo Page:stepaudiollm.github.io/step-… 🎮 HF playground: huggingface.co/spaces/stepfu…
Zakrea retweeted
Unsurprisingly, Kimi K2 Thinking is already number one trending on HF. The AI frontier is open-source!
46
142
17
1,528
Zakrea retweeted
The new Kimi model was trained with only $4.6M !!!?? Like 1/100 of their US counterparts? Can that be true!? Remember: > In 1969, NASA’s Apollo mission landed people on the moon with a computer that had just 4KB of RAM. And as @crystalsssup said "Creativity loves constraints"
If you ever wonder how Chinese frontier models like Kimi, DeepSeek, and Qwen are trained on far fewer (and nerfed) Nvidia GPUs than US models. In 1969, NASA’s Apollo mission landed people on the moon with a computer that had just 4KB of RAM. Creativity loves constraints.
Zakrea retweeted
Kimi K2 Thinking just landed on Chutes 🪂 256K context, agentic reasoning, and true multi-step tool use, now powered by decentralized inference at scale. Try it on Chutes today: chutes.ai/app/chute/8d008c10… #KimiK2 #Chutes #OpenSource
2
24
2
286
There is a lot of pressure on DeepSeek to release V4 but I won’t be surprised if we get another V3 with DSA scaled to 512k or 1M.
Zakrea retweeted
The training code for Hunyuan World 1.1 (WorldMirror) is released now!🔥🔥🔥 This release provides researchers and developers the full stack for customization and fine-tuning: 📷 Your video to 3D worlds in 1 second. 🪄ANY input (image, video, 3D prior) to ANY output (3DGS, depth, cameras, normal, points) Start customizing your model and exploring the demo! Online demo: huggingface.co/spaces/tencen… Code: github.com/Tencent-Hunyuan/H…
I don't get all this anti-Logan noise around here. #TeamLogan Also, if he says it's fake, it's fake.
Zakrea retweeted
what did you say? what are you thinking? what will you see?
I spy with my little eye
K2, K2-Thinking, and …
1
70% on SWE bench verified 30% terminal bench those are two intuitive thresholds for "actually useful and not frustrating" coding assistant. Kimi k2 thinking got 71.3% on SWE-Bench Verified 47.1% on Terminal-Bench
12
18
499
Great sport!
This tweet is unavailable
1
Zakrea retweeted
It's SOTA, not only open weights SOTA :)
MoonshotAI has released Kimi K2 Thinking, a new reasoning variant of Kimi K2 that achieves #1 in the Tau2 Bench Telecom agentic benchmark and is potentially the new leading open weights model Kimi K2 Thinking is one of the largest open weights models ever, at 1T total parameters with 32B active. K2 Thinking is the first reasoning model release within @Kimi_Moonshot's Kimi K2 model family, following non-reasoning Kimi K2 Instruct models released previously in July and September 2025. Key takeaways: ➤ Strong performance on agentic tasks: Kimi K2 Thinking achieves 93% in 𝜏²-Bench Telecom, an agentic tool use benchmark where the model acts as a customer service agent. This is the highest score we have independently measured. Tool use in long horizon agentic contexts was a strength of Kimi K2 Instruct and it appears this new Thinking variant makes substantial gains ➤ Reasoning variant of Kimi K2 Instruct: The model, as per its naming, is a reasoning variant of Kimi K2 Instruct. The model has the same architecture and same number of parameters (though different precision) as Kimi K2 Instruct and like K2 Instruct only supports text as an input (and output) modality ➤ 1T parameters but INT4 instead of FP8: Unlike Moonshot’s prior Kimi K2 Instruct releases that used FP8 precision, this model has been released natively in INT4 precision. Moonshot used quantization aware training in the post-training phase to achieve this. The impact of this is that K2 Thinking is only ~594GB, compared to just over 1TB for K2 Instruct and K2 Instruct 0905 - which translates into efficiency gains for inference and training. A potential reason for INT4 is that pre-Blackwell NVIDIA GPUs do not have support for FP4, making INT4 more suitable for achieving efficiency gains on earlier hardware. Our full set of Artificial Analysis Intelligence Index benchmarks are in progress and we will provide an update as soon as they are complete.
64
103
6
1,936