Thursday | 12 December 2024 | Reg No- 06
বাংলা
   
Thursday | 12 December 2024 | Epaper
BREAKING: 4 killed in 3-vehicle collission in Gazipur      WhatsApp, Instagram restored after Meta outages      Four more die, 444 hospitalised with dengue      India doesn't want democracy in neighbouring countries      ADB approves $600m loan for Bangladesh's economic reforms      Govt bans foreign travel of officials      Labour unrest in Ashulia demanding 15pc increment       

OpenAI releases reasoning AI with eye on safety, accuracy

Published : Friday, 13 September, 2024 at 1:35 PM  Count : 9717

ChatGPT creator OpenAI on Thursday released a new series of artificial intelligence models designed to spend more time thinking -- in hopes that generative AI chatbots provide more accurate and beneficial responses.

The new models, known as OpenAI o1-Preview, are designed to tackle complex tasks and solve more challenging problems in science, coding and mathematics -- something that earlier models have been criticized for failing to provide consistently.
Unlike their predecessors, these models have been trained to refine their thinking processes, try different methods and recognize mistakes before they deploy a final answer, AFP reports.

The new release comes as OpenAI is raising funds that could see it valued around $150 billion, which would make it one of the world's most valuable private companies, according to US media.

Investors include Microsoft and Nvidia, and could also include a $7 billion investment from MGX, a United Arab Emirates-backed investment fund, The Information reported.

OpenAI CEO Sam Altman hailed the models as "a new paradigm: AI that can do general-purpose complex reasoning."

However, he cautioned that the technology "is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it."

OpenAI's push to improve "thinking" in its model is a response to the persistent problem of "hallucinations" in AI chatbots.

This refers to their tendency to generate persuasive but incorrect content that has somewhat cooled the excitement over ChatGPT-style AI features among business customers

"We have noticed that this model hallucinates less," OpenAI researcher Jerry Tworek told The Verge.

But "we can't say we solved hallucinations," he added.

The Microsoft-backed company said that in tests, the models performed comparably to PhD students on difficult tasks in physics, chemistry and biology.

They also excelled in mathematics and coding, achieving an 83 percent success rate on a qualifying exam for the International Mathematics Olympiad, compared to 13 percent for GPT-4o, its most advanced general use model.

OpenAI said that the new reasoning capabilities could be used for healthcare researchers to annotate cell sequencing data, physicists to generate complex formulas, or computer developers to build and execute multistep designs.

The company also said that the models survived rigorous jailbreaking tests and could better withstand attempts to circumvent its guardrails.

OpenAI said its strengthened safety measures also included recent agreements with the US and UK AI Safety Institutes, which were granted early access to the models for evaluation and testing.

SR
Related topic   Subject:  OpenAI   AI   Safety   ChatGPT  


LATEST NEWS
MOST READ
Also read
Editor : Iqbal Sobhan Chowdhury
Published by the Editor on behalf of the Observer Ltd. from Globe Printers, 24/A, New Eskaton Road, Ramna, Dhaka.
Editorial, News and Commercial Offices : Aziz Bhaban (2nd floor), 93, Motijheel C/A, Dhaka-1000.
Phone: PABX- 41053001-06; Online: 41053014; Advertisement: 41053012.
E-mail: [email protected], news©dailyobserverbd.com, advertisement©dailyobserverbd.com, For Online Edition: mailobserverbd©gmail.com
🔝
close