The company AAI, founded a year ago, is still in stealth mode, but today we are doing a "partial unstealthing"..
A year and a half ago I gave a lecture at Reichman University about the latest status of large language models (LLMs). During the Q&A, Gil Kalai - a renowned mathematician - asked whether I see this technology one day reaching the level of a "great mathematician" or great scientist in general. My immediate instinct was negative.
Later I chatted with my partner in science and technology,
@shai_s_shwartz, to reflect on this question. First, can we actually prove that the trajectory of LLMs - with CoT, Tree-of-throughs, and today O1 - is subject to a performance ceiling of some sort? It is not obvious at all because these systems keep on improving.
Second, if there is a fundamental ceiling, then what would be the required leap to overcome this? We concluded that if we can answer both questions we can build something that could surpass anything we have done in the past.
A few months later, three of my doctoral students graduated:
@YoavLevine,
@or_sharir and Noam Wies. Also, Gal Biniamini’s doctoral student Nati Linial also graduated and all six of us founded AAI to pursue those two questions. A year later, we have a pretty good grasp of the two questions, and the paper we released today provides a conceptual framework for an AI that can become a "great scientist” and, along the way, prove why current technology would not get there.
The team has also been working on the implementation of this framework in the area of "Algorithmic expertise" using CodeForces as a proving ground. The results are very promising and will be shared in due course..
The link to paper:
bit.ly/3Bn2V1n
The link to blog:
bit.ly/3ZhGZwz