Artificial intelligence
fromAxios
3 days agoOpenAI releases "Spud" GPT-5.5 model
GPT-5.5 enhances autonomous task handling and efficiency in various fields, marking a significant advancement in AI capabilities.
Next-word pretraining creates statistical pressure toward hallucination, even with idealized error-free data. Facts lacking repeated support in training data yield unavoidable errors, while recurring regularities do not.
Denise Dresser, Chief Revenue Officer at OpenAI, emphasizes the practical applicability. 'Infosys's deep expertise in large-scale software transformation enables enterprises to deploy Codex across areas like legacy code modernization, code review automation, vulnerability detection, and application development.'
Meta is working on two proprietary frontier models: Avocado, a large language model, and Mango, a multimedia file generator. The open-source variants are expected to be made available at a later date.
Today, OpenAI announced GPT-5.3-Codex, a new version of its frontier coding model that will be available via the command line, IDE extension, web interface, and the new macOS desktop app. (No API access yet, but it's coming.) GPT-5.3-Codex outperforms GPT-5.2-Codex and GPT-5.2 in SWE-Bench Pro, Terminal-Bench 2.0, and other benchmarks, according to the company's testing. There are already a few headlines out there saying "Codex built itself," but let's reality-check that, as that's an overstatement.
OpenAI's GPT-5.2 Pro does better at solving sophisticated math problems than older versions of the company's top large language model, according to a new study by Epoch AI, a non-profit research institute.