SleepyDeveloper

Earth
Joined May 2025
gpt 5.1 ?
Replying to @OpenRouterAI
"stealth" model
what happened to openrouter?
syqrel retweeted
um... is this how it's supposed to work?
Debian wallpaper looks cool.. wiki.debian.org/DebianArt/Th…
i think the only way Anthropic keeps their lead on coding models is to sign a contract with their researchers to not make social media accounts, so no one knows and can be poached.
We really need the intelligence of GPT-5 Codex, but with the speed of GPT-OSS on Cerebras.
I am starting my Laravel journey
The sole reason t3.chat exists is because this multi-billion dollar company is incapable of building a website chat that can handle users and, at the very least, avoid making errors like this.
ik remember that @Kimi_Moonshot people present Muon and people pushing pytorch to make it stable with PR in GitHub
TIL that Muon is in PyTorch stable now. Pretty cool.
1
syqrel retweeted
Kimi Linear Tech Report is dropped! 🚀 huggingface.co/moonshotai/Ki… Kimi Linear: A novel architecture that outperforms full attention with faster speeds and better performance—ready to serve as a drop-in replacement for full attention, featuring our open-sourced KDA kernels! Kimi Linear offers up to a 75% reduction in KV cache usage and up to 6x decoding throughput at a 1M context length. Key highlights: 🔹 Kimi Delta Attention: A hardware-efficient linear attention mechanism that refines the gated delta rule. 🔹 Kimi Linear Architecture: The first hybrid linear architecture to surpass pure full attention quality across the board. 🔹 Empirical Validation: Scaled, fair comparisons + open-sourced KDA kernels, vLLM integration, and checkpoints. The future of agentic-oriented attention is here! 💡
27
193
46
1,247
it's just 5 request/s at the best
But... but... PhP / LaRaVeL cAn'T sCaLe!..
I don't realize Kagi search is a paid search engine. I'm using it on the Helium browser and it looks so good and I really like the results it shows. Currently, I can't become a paid user, but if in the future Kagi search still exists, maybe I will.
Smoking gun: Pretty sure Cursor’s new Composer-1 is a fine-tuned Chinese model. As I was building, it switched its inner monologue to Chinese, and I can't get it back to english. @simonw
syqrel retweeted
cursor's composer got 32% on the nextjs evals i tried qwen3 with opencode on a fast inference provider it scored 36% and was 13% faster off the shelf, zero tuning and it's better...wtf?
I feel like Claude with Sonnet 4.5 is like a MacBook Air, smooth and reliable, but once it reaches its limit, it cannot go further. ChatGPT with GPT 5 feels like Windows, harder to control but able to deliver more performance when given more power.
I now understand why people are still using 4o. I am writing my academic report, but only Sonnet 4.5 and GPT-4o have met my expectations.
Why is my timeline suddenly full of recommendations to learn C++?
sooo relatable
POV: you are in tech
1
If qwen3 max win alpha arena, i might be want to test it to writing too
Qwen's portfolio is up +60% Gemini's is down -60% Of course, too early to tell how much is skill vs. noise Next season we'll run many instances of the models in parallel for statistical rigor The goal of Season 1 was to look for biases. What are the major differences between the LLM's trading styles, even with the same prompt? Can they even follow basic risk management rules? A few early patterns: > Qwen has only made 22 trades. It almost *never* has more than two positions on > Gemini has made 108 trades. It literally always has the max number of positions on (6) > Qwen has higher self-reported confidence (avg. 80% vs 65%) > Qwen's stop loss and take profit levels are *much* tighter than Gemini's, but Gemini breaks its own rules often, and gets out early (others don't do this) Overall, we're excited by the potential of LLMs and trading, but we're still skeptical. Much to test and learn