A paper co-authored by Prof. Alex Lew has been selected as one of four "Outstanding Papers" at this year's Conference on Language Modeling (COLM 2025), held in Montreal in October.
Large language models such as ChaptGPT have proven to be able to produce remarkably intelligent results, but the energy and monetary costs associated with running these massive algorithms is sky high.
Sapient Intelligence's prototype with only 27 million parameters outperformed systems from OpenAI, Anthropic, and DeepSeek on complex tasks including advanced Sudoku puzzles, maze-solving, and the ARC ...
Large language models (LLMs) leverage unsupervised learning to capture statistical patterns within vast amounts of text data. At the core of these models lies the Transformer architecture, which ...
Effective compression is about finding patterns to make data smaller without losing information. When an algorithm or model can accurately guess the next piece of data in a sequence, it shows it’s ...
They had to throw away most of what it produced but there was gold among the garbage. Google DeepMind has used a large language model to crack a famous unsolved problem in pure mathematics. In a paper ...
Results that may be inaccessible to you are currently showing.
Hide inaccessible results