Born again Christian! BS | MS @NDSU iloveneurons.com Building @SpeechSage Passion for biomedical arenas Love a good debate, but only if there is purpose

SF (Sioux Falls)
Joined November 2021
As great as science is, the humans performing it are going to be slightly biased
Nearly all good ML papers are at least implicitly position papers, providing substantive support for a meaningful position. Otherwise, what’s the point?
Noah Vandal retweeted
Introducing Nested Learning: A new ML paradigm for continual learning that views models as nested optimization problems to enhance long context processing. Our proof-of-concept model, Hope, shows improved performance in language modeling. Learn more: goo.gle/47LJrzI @GoogleAI
i am starting to wonder if the biggest predictor of positive outcomes is the ability to endure delayed gratification
1
5
did anyone else use Wombo in 2022? I remember thinking it was so cool, even if it was very abstract
1
that is a first. I have never gotten good clock results with any of the other AI image gen models
🚨 Nano Banana 2 completely smashed both the clock AND full wine glass tests in ONE IMAGE. "11:15 on the clock and a wine glass filled to the top" Holy. Shit.
1
when you are sending me invoices, please do not use exclamation marks. i am not happy to be receiving it. you shouldn't be either.
2
kimi k2 is a good example that you really don't need float. maybe for a few activations
Elon: we train in int Shareholders: confused what int is Elon: int operation is fundamentally better than float Shareholders: more confused, what is float Elon: we can do float but int is simpler with logic gates, int is power efficient, we do int Shareholders: still silent and confused, what the hell is a logic gate Elon: some of it is made in TSMC Texas Shareholders: everybody applauds Elon: man I love chips, it is a good chip sir. You know what, we gonna make a gigantic chip fab ourselves
i really love codex
1
2
if stores could date, @fleet_farm and @culvers would definitely be a couple
if you dont understand rank in the context of SVD, i would highly recommend learning it
How is memorized data stored in a model? We disentangle MLP weights in LMs and ViTs into rank-1 components based on their curvature in the loss, and find representational signatures of both generalizing structure and memorized training data
John Hopfield was 49 when he published his first neuroscience article
5
success will never fill the void in your heart
2
2
This will be the jevons principle in action. All these new use cases will need stewards. As has happened before, we will increase what is done, due to the ai we use. Jobs will still exist. Thinking we will be limited to the jobs we currently do is short sighted thinking
In 5 years from now, probably 95% of the tokens used by AI agents will be used on tasks that humans never did before. I just met with about 30 enterprises across 2 days and a dinner, and some of the most interesting use-cases that keep coming up for AI agents are on bringing automated work to areas that the companies would not have been able to apply labor to before. Most of the world hasn’t quite caught on to this point yet. We imagine AI as dropping into today’s workflows and just taking what we already do and making it more efficient by 20% or something. Yet most companies realize that most of the time they’re doing far less than they could because of the cost or limited capacity of talent. This shows up in different ways across every industry. In real estate it’s ideas like being able to read and analyze every lease agreement for every trend and business opportunity possible. In life sciences it’s being able to rapidly do drug discovery or improve quality by looking through errors in data. In financial services it’s being able to look through all past deals and figure out better future monetization. In legal it’s being able to execute on contracts or legal work for previously unprofitable segments or projects. And these are just the Box AI use cases that deal with documents and content. The same is going to be true in coding, where companies tackle software projects they wouldn’t have done before. Security of all systems and events they couldn’t get to. And so on. If you are working on AI Agents right now, the big opportunity is to bring enterprises “work” for problems that they couldn’t do before because it was nearly impossible to afford or scale. And if you’re deploying AI agents in an enterprise, consider what things you’d do more of (or differently) if the cost and speed of labor became 100X cheaper and faster. This is going to get you the real upside of automation.
1
1
i bet also those crashes are over 50% with human (mis) intervention
Tesla has revealed that owners using FSD experience just one crash per 4.92 million miles vs one crash per 700k miles for the US average.
1
happy for cerebras, and glad they are an option, but i do think it is funny whenever they report @GroqInc speeds, it is OBVIOUSLY the p95 or whatever latency, and normally it is 2-3x faster haha. gotta love competition
Cerebras beats Nvidia H100 but can it beat Blackwell? Blackwell inference endpoints are finally out and it’s fast. It runs GPT-OSS-120B at ~700 tokens/s, leapfrogging H100 and Groq. Cerebras clocked in at 3,000 TPS - still #1. Looking forward to Rubin!
sometimes these plots don't make sense. why is it more attractive to have been released earlier?
at first i thought hinton was wearing a fleet farm jacket
1
I have. It seen a good implementation of thinking models with voice agents yet!
Voice AI Meetup on Wednesday Nov 18th (in-person in SF, or live stream). Themes this time: - coordinating between multiple agents - incorporating thinking models into voice agents - using sandboxes for long-running sub-agents - dynamic user interfaces I'm always excited about the monthly meetups. But I'm *especially* excited about this one. I've been hacking on all these things lately both with enterprise partners and experimentally. We have a great group of people to talk about these "next next" emerging patterns. I'm also showing a demo of some stuff I've been working on for the past few weekends.
Noah Vandal retweeted
ts ai conf launch #4 you may have read my book: Principles of Building AI Agents well, we wrote a sequel: Patterns for Building AI Agents about bridging the gap where AI teams stumble between prototype and production if you like and share this post we will send you a copy