Founder @VocalityHealth (YC W25) | PhD in AI/ML | Prev founder @fulltrackai (4m users) | Sharing insights on building AI products and startup journeys.

Seattle WA
Joined March 2011
Excited to share our new paper: Constrained Diffusion Implicit Models! We use diffusion models to solve noisy inverse problems like inpainting, sparse-recovery, and colorization. 10-50x faster than previous methods! Paper: arxiv.org/pdf/2411.00359 Demo: huggingface.co/spaces/vivjay…
2
30
1
153
You can’t reduce something by 156%. You know we’re peak bubble when a launch video doesn’t even follow basic math sense. (Unless the model has negative sycophancy which is even more worrying).
I am excited to announce that Arcarae has $2.5M in funding and I am finally hiring. Arcarae’s mission is to help humanity remember and unlock the power each individual holds within themself so they can bring into reality their unique, authentic expression of self without fear or compromise. Our research endeavors are designed to support this mission via computationally modeling higher-order cognition and subjective internal world models. Specifically, we are building the computational models of the other side of intelligence that everyone has neglected:  Intuition. Our evolution is anchored in our current product, an immersive universe for self-discovery, and MIRROR, our AI research implementing cognitive inner-monologue in LLMs, reducing sycophancy by 21% on avg. & up to 156% vs. SOTA models. This marks Arcarae’s transition from a solo endeavor into a full-fledged consumer product and AI research company. I am seeking three very specific people to join me on this mission and help scale Arcarae to its next phase. I am hiring one researcher, one marketer, and one engineer as my founding team. These are far from normal roles; the application and hiring process even reflects this. If you think you are one of these three people, please apply right away. There is much to be done <3 And with that being said, Welcome to the era of Arcarae. And as always, I am excited for what’s to come <3
2
Jeff Bezos and I have the same phone. He can’t buy a better one than me. The same as going to happen to every other aspect of the lives across normal people and billionaires.
if ai drives true hyperdeflation, then most current power law outcomes start to collapse rapidly into a compressed distribution where billionaires & average folks converge in lived experience. private jets lose their edge if autonomous supersonic ridesharing exists, exclusive education vanishes if every tutor is gpt-x, hedge fund alpha dissolves if everyone has real time strategy agents, etc. this would be an incredibly fascinating human / societal experiment.
1
4
Nobody I know expects to get any money from Social Security. It’s just a tax.
The Social Security trust fund is expected to run out of cash in eight years, per NPR.
3
Vivek Jayaram retweeted
Excited to share our beta release of Music Arena, a live evaluation platform for state-of-the-art AI music generation models! 🎧 Listen to the latest models and 🗳️ vote for your favorite ⚔️ music-arena.org ⭐️ github.com/gclef-cmu/music-a… 📜 arxiv.org/abs/2507.20900
Unpopular opinion, if you can find people in India to do quality work for your company, you absolutely should. Costs 1/5th as much as US engineers, meaning you can provide a livelihood to 5x as many people, and do 5x the work with the same budget.
This tweet is unavailable
2
Awesome job to @chrisdonahuey and the rest of the magenta team.
Enter SpectroStream, released quietly by Google as part of Magenta Realtime, and it’s insanely good. Open source Cleaner sound Change bitrate at inference Works well on music and speech
1
1
Most audio models (Suno, Udio, ElevenLabs, etc) rely on this trick: ➡️ Encode audio into ~25 frames/sec ➡️ Each frame gets 4–16 discrete tokens ➡️ Train an LLM on these tokens like it’s text But there’s a tradeoff… Fewer tokens = easier modeling, but worse sound
We tokenize audio using something called Residual Vector Quantization (RVQ). RVQ turns a continuous audio signal into a small set of discrete codes, turning a WAV file into language-model-friendly building blocks.
Why does this matter? LLMs operate on discrete tokens (words, bytes, etc). But audio is continuous, often at 48kHz. That’s 48,000 values per second. Way too much for a language model to handle. So how do we fix that?
Google quietly released SpectroStream, a powerful new open-source audio tokenizer and it blows Encodec out of the water. The audio reconstruction quality is shockingly good, and I've already switched to it for my projects. Here's why this is so great for the audio generative modeling community:
1
4
0
I’m trying to learn how to promote B2C apps on TikTok. Anyone have any good tutorials or guides? AI influencers or pay for human influencers?
1
3
The Windsurf situation shows that it’s never been a better time to be a founder and a worse time to be an employee: Founders get stock, employees get options at ridiculous FMV with 30 days to exercise. Founders are selling secondaries earlier than ever (series A or 😱 even at Seed). Employees are waiting longer for IPOs or exits. Founders will leave for acquihire opportunities when it suits them.
2
YC companies today:
2
Vivek Jayaram retweeted
"you don’t seem to understand, i have a PhD in ML, i was meant to pretrain language model" "wrap the fucking API"
32
180
15
4,128
The clearest sign that we haven’t reached AGI is that AI can’t autonomously make money
New Anthropic Research: Project Vend. We had Claude run a small shop in our office lunchroom. Here’s how it went.
2
One of my favorite intuitions about gradient descent is that local minima are exponentially less common in high dimensions. That’s why neural nets don’t get stuck during training.
In machine learning, we take gradient descent for granted. We rarely question why it works. What's usually told is the mountain-climbing analogue: to find the valley, step towards the steepest descent. But why does this work so well? Read on.
1
In the long run, the only software companies that will exist will be foundational model research companies and undifferentiated sales shops.
1