We integrated LLMs into our aftermarket domain search tool while keeping sub-20 ms latency. How? Instead of querying an LLM on every keystroke, we used LLMs offline to teach a 22.7M-parameter embedding model. Learn how we built it: instantdomainsearch.com/blog…

Sep 8, 2025 · 5:12 PM UTC

2
3
12
Replying to @instantai
dayum insta .now!
We're actually working on updating our supported TLD list right now, including .now 🥳 Thanks as always!
2
1
6
Replying to @instantai
Brilliant approach! Using LLMs offline to train embeddings is such a clever way to get the intelligence without the latency hit. Curious about the training data pipeline - did you find semantic search worked better than keyword matching for domain suggestions?
1