Skip to main content

GPT-5 to take AI forward in these two important ways

Breaking Down Barriers to AI Innovation with Reid Hoffman & Kevin Scott

We could soon see generative AI systems capable of passing Ph.D. exams thanks to more “durable” memory and more robust reasoning operations, Microsoft CTO Kevin Scott revealed when he took to the stage with Reid Hoffman during a Berggruen Salon in Los Angeles earlier this week.

Recommended Videos

“It’s sort of weird right now that you have these interactions with agents and the memory is entirely episodic,” he lamented. “You have a transaction, you do a thing. It’s useful or not for whatever task you were doing, and then it forgets all about it.” The AI system isn’t learning from or even remembering previous interactions with the user, he continued. “There’s no way for you to refer back to a thing you were trying to get [the AI] to solve in the past.”

However, Scott is optimistic that,”we’re seeing technically all of the things fall in place to have really durable memories with the systems.” With more persistent memory, future AI systems will be able to respond more naturally and more accurately over the span of multiple conversations rather than being limited to the current session.

OpenAI announced in February that it was beginning to test a new persistent memory system, rolling it out to select free and Plus subscription users. Enabling the feature allows the AI to recall user tone, voice, and format preferences between conversations as well as make suggestions in new projects based on details the user mentioned in previous chats.

Scott was also buoyant about improving the “fragility” found in the reasoning of many AI systems today. “It can’t solve very complicated math problems,” he explained. “It has to bail out to other systems to do very complicated things.”

“Reasoning, I think, gets a lot better,” he continued. He compares GPT-4 and the current generation of models to high schoolers passing their AP exams. However, the next generation of AIs “could be the thing that could pass your qualified exam.”

To date, generative AI systems have outperformed their flesh-and-blood counterparts on a variety of exam and task formats. Last November, for example, GPT-4 passed the Multistate Professional Responsibility Exam (MPRE), better known as the bar exam, with 76% correct — that’s six points higher than the nation average for humans.

Scott was quick to point out, however, that training generative AIs to pass Ph.D. exams “probably sounds like a bigger deal than it actually is… the real test will be what we choose to do with it.”

Scott was especially excited to see the barriers to entry falling away so quickly. He noted that when he got into machine learning two decades ago, his work required graduate-level knowledge, stacks upon stacks of “very daunting, complicated, technical papers to figure out how to do what I wanted to do,” and around six months of coding. That same task today, he said, “a high school student could do in a Saturday morning.”

These lowered barriers to entry will likely accelerate the democratization of AI, Scott concluded. Finding solutions to the myriad social, environmental, and technological crises facing humanity are not — and cannot — be the sole responsibility of “just the people at tech companies in Silicon Valley or just people who graduated with Ph.D.s from top-five universities,” he said. “We have 8 billion people in the world who also have some idea about what it is that they want to do with powerful tools, if they just have access to them.”

Andrew Tarantola
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
Radiohead’s Thom Yorke among thousands of artists who issue AI protest
Thom Yorke on stage.

Leading actors, authors, musicians, and novelists are among 11,500 artists to have put their name to a statement calling for a halt to the unlicensed use of creative works to train generative AI tools like OpenAI’s ChatGPT, describing it as a “threat” to the livelihoods of creators.

The open letter, comprising just 29 words, says: “The unlicensed use of creative works for training generative AI is a major, unjust threat to the livelihoods of the people behind those works, and must not be permitted.”

Read more
The best AI chatbots to try: ChatGPT, Gemini, and more
Bing Chat shown on a laptop.

The idea of chatbots has been around since the early days of the internet. But even compared to popular voice assistants like Siri, the generated chatbots of the modern era are far more powerful.

Yes, you can converse with them in natural language. But these AI chatbots can generate text of all kinds, from poetry to code, and the results really are exciting. ChatGPT remains in the spotlight, but as interest continues to grow, more rivals are popping up to challenge it.
OpenAI ChatGPT and ChatGPT Plus

Read more
Perplexity’s two new features take it beyond just a chatbot
An abstract image showing floating squares used for a Perplexity blog post.

Perplexity AI, makers of the popular chatbot by the same name, announced Thursday that it is rolling out a pair of new features that promise to give users more flexibility over the sorts of sources they employ: Internal Knowledge Search and Spaces.

"Today, we're launching Perplexity for Internal Search: one tool to search over both the web and your team's files with multi-step reasoning and code execution," Perplexity AI CEO Aravind Srinivas wrote on X (formerly Twitter). Previously, users were able to upload personal files for the AI to chew through and respond upon, the same way they could with Gemini, ChatGPT, or Copilot. With Internal Search, Perplexity will now dig through both those personal documents and the internet to infer its response.

Read more