Journal De Bruxelles - OpenAI releases reasoning AI with eye on safety, accuracy

NYSE - LSE
NGG 1.4% 58.5 $
CMSD 0% 23.56 $
BCC -0.21% 122.75 $
SCS -4.94% 11.74 $
GSK 0.51% 33.6 $
RELX -0.68% 45.47 $
RYCEF -0.14% 7.27 $
RIO -0.15% 58.64 $
RBGPF 100% 59.96 $
CMSC 0.08% 23.86 $
AZN 1.39% 65.35 $
VOD 0.12% 8.39 $
BCE 0.22% 23.16 $
BTI 0.31% 36.24 $
BP 0.66% 28.6 $
JRI 0.91% 12.06 $
OpenAI releases reasoning AI with eye on safety, accuracy
OpenAI releases reasoning AI with eye on safety, accuracy / Photo: Jason Redmond - AFP/File

OpenAI releases reasoning AI with eye on safety, accuracy

ChatGPT creator OpenAI on Thursday released a new series of artificial intelligence models designed to spend more time thinking -- in hopes that generative AI chatbots provide more accurate and beneficial responses.

Text size:

The new models, known as OpenAI o1-Preview, are designed to tackle complex tasks and solve more challenging problems in science, coding and mathematics -- something that earlier models have been criticized for failing to provide consistently.

Unlike their predecessors, these models have been trained to refine their thinking processes, try different methods and recognize mistakes before they deploy a final answer.

The new release comes as OpenAI is raising funds that could see it valued around $150 billion, which would make it one of the world's most valuable private companies, according to US media.

Investors include Microsoft and Nvidia, and could also include a $7 billion investment from MGX, a United Arab Emirates-backed investment fund, The Information reported.

OpenAI CEO Sam Altman hailed the models as "a new paradigm: AI that can do general-purpose complex reasoning."

However, he cautioned that the technology "is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it."

OpenAI's push to improve "thinking" in its model is a response to the persistent problem of "hallucinations" in AI chatbots.

This refers to their tendency to generate persuasive but incorrect content that has somewhat cooled the excitement over ChatGPT-style AI features among business customers

"We have noticed that this model hallucinates less," OpenAI researcher Jerry Tworek told The Verge.

But "we can't say we solved hallucinations," he added.

The Microsoft-backed company said that in tests, the models performed comparably to PhD students on difficult tasks in physics, chemistry and biology.

They also excelled in mathematics and coding, achieving an 83 percent success rate on a qualifying exam for the International Mathematics Olympiad, compared to 13 percent for GPT-4o, its most advanced general use model.

OpenAI said that the new reasoning capabilities could be used for healthcare researchers to annotate cell sequencing data, physicists to generate complex formulas, or computer developers to build and execute multistep designs.

The company also said that the models survived rigorous jailbreaking tests and could better withstand attempts to circumvent its guardrails.

OpenAI said its strengthened safety measures also included recent agreements with the US and UK AI Safety Institutes, which were granted early access to the models for evaluation and testing.

R.Vandevelde--JdB