Connect with us

Technologies

AI Gets Smarter, Safer, More Visual With GPT-4 Update, OpenAI Says

If you subscribe to ChatGPT Plus, you can try it out now.

The hottest AI technology foundation got a big upgrade Tuesday with OpenAI’s GPT-4 release now available in the premium version of the ChatGPT chatbot.

GPT-4 can generate much longer strings of text and respond when people feed it images, and it’s designed to do a better job avoiding artificial intelligence pitfalls visible in the earlier GPT-3.5, OpenAI said Tuesday. For example, when taking bar exams that attorneys must pass to practice law, GPT-4 ranks in the top 10% of scores compared with the bottom 10% for GPT-3.5, the AI research company said.

GPT stands for Generative Pretrained Transformer, a reference to the fact that it can generate text on its own — now up to 25,000 words with GPT-4 — and that it uses an AI technology called transformers that Google pioneered. It’s a type of AI called a large language model, or LLM, that’s trained on vast swaths of data harvested from the internet, learning mathematically to spot patterns and reproduce styles. Human overseers rate results to steer GPT in the right direction, and GPT-4 has more of this feedback.

OpenAI has made GPT available to developers for years, but ChatGPT, which debuted in November, offered an easy interface ordinary folks can use. That yielded an explosion of interest, experimentation and worry about the downsides of the technology. It can do everything from generating programming code and answering exam questions to writing poetry and supplying basic facts. It’s remarkable if not always reliable.

ChatGPT is free, but it can falter when demand is high. In January, OpenAI began offering ChatGPT Plus for $20 per month with assured availability and, now, the GPT-4 foundation. Developers can sign up on a waiting list to get their own access to GPT-4.

GPT-4 advancements

«In a casual conversation, the distinction between GPT-3.5 and GPT-4 can be subtle. The difference comes out when the complexity of the task reaches a sufficient threshold,» OpenAI said. «GPT-4 is more reliable, creative and able to handle much more nuanced instructions than GPT-3.5.»

Another major advance in GPT-4 is the ability to accept input data that includes text and photos. OpenAI’s example is asking the chatbot to explain a joke showing a bulky decades-old computer cable plugged into a modern iPhone’s tiny Lightning port. This feature also helps GPT take tests that aren’t just textual, but it isn’t yet available in ChatGPT Plus.

Another is better performance avoiding AI problems like hallucinations — incorrectly fabricated responses, often offered with just as much seeming authority as answers the AI gets right. GPT-4 also is better at thwarting attempts to get it to say the wrong thing: «GPT-4 scores 40% higher than our latest GPT-3.5 on our internal adversarial factuality evaluations,» OpenAI said.

GPT-4 also adds new «steerability» options. Users of large language models today often must engage in elaborate «prompt engineering,» learning how to embed specific cues in their prompts to get the right sort of responses. GPT-4 adds a system command option that lets users set a specific tone or style, for example programming code or a Socratic tutor: «You are a tutor that always responds in the Socratic style. You never give the student the answer, but always try to ask just the right question to help them learn to think for themselves.»

«Stochastic parrots» and other problems

OpenAI acknowledges significant shortcomings that persist with GPT-4, though it also touts progress avoiding them.

«It can sometimes make simple reasoning errors … or be overly gullible in accepting obvious false statements from a user. And sometimes it can fail at hard problems the same way humans do, such as introducing security vulnerabilities into code it produces,» OpenAI said. In addition, «GPT-4 can also be confidently wrong in its predictions, not taking care to double-check work when it’s likely to make a mistake.»

Large language models can deliver impressive results, seeming to understand huge amounts of subject matter and to converse in human-sounding if somewhat stilted language. Fundamentally, though, LLM AIs don’t really know anything. They’re just able to string words together in statistically very refined ways.

This statistical but fundamentally somewhat hollow approach to knowledge led researchers, including former Google AI researchers Emily Bender and Timnit Gebru, to warn of the «dangers of stochastic parrots» that come with large language models. Language model AIs tend to encode biases, stereotypes and negative sentiment present in training data, and researchers and other people using these models tend «to mistake … performance gains for actual natural language understanding.»

OpenAI Chief Executive Sam Altman acknowledges problems, but he’s pleased overall with the progress shown with GPT-4. «It is more creative than previous models, it hallucinates significantly less, and it is less biased. It can pass a bar exam and score a 5 on several AP exams,» Altman tweeted Tuesday.

One worry about AI is that students will use it to cheat, for example when answering essay questions. It’s a real risk, though some educators actively embrace LLMs as a tool, like search engines and Wikipedia. Plagiarism detection companies are adapting to AI by training their own detection models. One such company, Crossplag, said Wednesday that after testing about 50 documents that GPT-4 generated, «our accuracy rate was above 98.5%.»

OpenAI, Microsoft and Nvidia partnership

OpenAI got a big boost when Microsoft said in February it’s using GPT technology in its Bing search engine, including a chat features similar to ChatGPT. On Tuesday, Microsoft said it’s using GPT-4 for the Bing work. Together, OpenAI and Microsoft pose a major search threat to Google, but Google has its own large language model technology too, including a chatbot called Bard that Google is testing privately.

Also on Tuesday, Google announced it’ll begin limited testing of its own AI technology to boost writing Gmail emails and Google Docs word processing documents. «With your collaborative AI partner you can continue to refine and edit, getting more suggestions as needed,» Google said.

That phrasing mirrors Microsoft’s «co-pilot» positioning of AI technology. Calling it an aid to human-led work is a common stance, given the problems of the technology and the necessity for careful human oversight. 

Microsoft uses GPT technology both to evaluate the searches people type into Bing and, in some cases, to offer more elaborate, conversational responses. The results can be much more informative than those of earlier search engines, but the more conversational interface that can be invoked as an option has had problems that make it look unhinged.

To train GPT, OpenAI used Microsoft’s Azure cloud computing service, including thousands of Nvidia’s A100 graphics processing units, or GPUs, yoked together. Azure now can use Nvidia’s new H100 processors, which include specific circuitry to accelerate AI transformer calculations.

AI chatbots everywhere

Another large language model developer, Anthropic, also unveiled an AI chatbot called Claude on Tuesday. The company, which counts Google as an investor, opened a waiting list for Claude.

«Claude is capable of a wide variety of conversational and text processing tasks while maintaining a high degree of reliability and predictability,» Anthropic said in a blog post. «Claude can help with use cases including summarization, search, creative and collaborative writing, Q&A, coding and more.»

It’s one of a growing crowd. Chinese search and tech giant Baidu is working on a chatbot called Ernie Bot. Meta, parent of Facebook and Instagram, consolidated its AI operations into a bigger team and plans to build more generative AI into its products. Even Snapchat is getting in on the game with a GPT-based chatbot called My AI.

Expect more refinements in the future.

«We have had the initial training of GPT-4 done for quite awhile, but it’s taken us a long time and a lot of work to feel ready to release it,» Altman tweeted. «We hope you enjoy it and we really appreciate feedback on its shortcomings.»

Editors’ note: CNET is using an AI engine to create some personalfinance explainers that are edited and fact-checked by our editors. Formore, see this post.

Technologies

How to Get Verizon’s New Internet Plan for Just $25 Per Month

Continue Reading

Technologies

This $20K Humanoid Robot Promises to Tidy Your Home. But There Are Strings Attached

The new Neo robot from 1X is designed to do chores. It’ll need help from you — and from folks behind the curtain.

It stands 5 feet, 6 inches tall, weighs about as much as a golden retriever and costs near the price of a brand-new budget car. 

This is Neo, the humanoid robot. It’s billed as a personal assistant you can talk to and eventually rely on to take care of everyday tasks, such as loading the dishwasher and folding laundry. 

Neo doesn’t work cheap. It’ll cost you $20,000. And even then, you’ll still have to train this new home bot, and possibly need a remote assist as well.

If that sounds enticing, preorders are now open (for a mere $200 down). You’ll be signing up as an early adopter for what Neo’s maker, a California-based company called 1X, is calling a «consumer-ready humanoid.» That’s opposed to other humanoids under development from the likes of Tesla and Figure, which are, for the moment at least, more focused on factory environments. 

Neo is a whole order of magnitude different from robot vacuums like those from Roomba, Eufy and Ecovacs, and embodies a long-running sci-fi fantasy of robot maids and butlers doing chores and picking up after us. If this is the future, read on for more of what’s in store.


Don’t miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source.


What the Neo robot can do around the house

The pitch from 1X is that Neo can do all manner of household chores: fold laundry, run a vacuum, tidy shelves, bring in the groceries. It can open doors, climb stairs and even act as a home entertainment system.

Neo appears to move smoothly, with a soft, almost human-like gait, thanks to 1X’s tendon-driven motor system that gives it gentle motion and impressive strength. The company says it can lift up to 154 pounds and carry 55 pounds, but it is quieter than a refrigerator. It’s covered in soft materials and neutral colors, making it look less intimidating than metallic prototypes from other companies.

The company says Neo has a 4-hour runtime. Its hands are IP68-rated, meaning they’re submersible in water. It can connect via Wi-Fi, Bluetooth and 5G. For conversation, it has a built-in LLM, the same sort of AI technology that powers ChatGPT and Gemini.

The primary way to control the Neo robot will be by speaking to it, just as if it were a person in your home.  

Still, Neo’s usefulness today depends heavily on how you define useful. The Wall Street Journal’s Joanna Stern got an up-close look at Neo at 1X’s headquarters and found that, at least for now, it’s largely teleoperated, meaning a human often operates it remotely using a virtual-reality headset and controllers. 

«I didn’t see Neo do anything autonomously, although the company did share a video of Neo opening a door on its own,» Stern wrote last week. 

1X CEO Bernt Børnich told her that Neo will do most things autonomously in 2026, though he also acknowledged that the quality «may lag at first.»

The company’s FAQ says that for any chore request Neo doesn’t know how to accomplish, «you can schedule a 1X Expert to guide it» to help the robot «learn while getting the job done.»

What you need to know about Neo and privacy

Part of what early adopters are signing up for is to let Neo learn from their environment so that future versions can operate more independently. 

That learning process raises privacy and trust questions. The robot uses a mix of visual, audio and contextual intelligence — meaning it can see, hear and remember interactions with users throughout their homes. 

«If you buy this product, it is because you’re OK with that social contract,» Børnich told the Journal. «It’s less about Neo instantly doing your chores and more about you helping Neo learn to do them safely and effectively.»

Neo’s reliance on human operation behind the scenes prompted a response from John Carmack, a computer industry luminary known for his work with VR systems and the lead programmer of classic video games including Doom and Quake. 

«Companies selling the dream of autonomous household humanoid robots today would be better off embracing reality and selling ‘remote operated household help’,» he wrote in a post on the X social network (formerly Twitter) on Monday.

1X says it’s taking steps to protect your privacy: Neo listens only when it recognizes it’s being addressed, and its cameras will blur out humans. You can restrict Neo from entering or viewing specific areas of your home, and the robot will never be teleoperated without owner approval, the company says. 

But inviting an AI-equipped humanoid to observe your home life isn’t a small step.

The first units will ship to customers in the US in 2026. There is a $499 monthly subscription alternative to the $20,000 full-purchase price, though that will be available at an unspecified later date. A broader international rollout is promised for 2027.

Neo’s got a long road ahead of it to live up to the expectations set by Rosie the Robot in The Jetsons way back when. But this is no Hanna-Barbera cartoon. What we’re seeing now is a much more tangible harbinger of change.

Continue Reading

Technologies

I Wish Nintendo’s New Switch 2 Zelda Game Was an Actual Zelda Game

Hyrule Warriors: Age of Imprisonment has great graphics, a great story and Zelda is actually in it. But the gameplay makes me wish for another true Zelda title instead.

I’ve never been a Hyrule Warriors fan. Keep that in mind when I say that Nintendo’s new Switch 2-exclusive Zelda-universe game has impressed me in several ways, but the gameplay isn’t one of them. Still, this Zelda spinoff has succeeded in showing off the Switch 2’s graphics power. Now can we have a true Switch 2 exclusive Zelda game next?

The upgraded graphics in Tears of the Kingdom and Breath of the Wild has made the Switch 2 a great way to play recent Zelda games, which had stretched the Switch’s capabilities to the limit before. And they’re both well worth revisiting, because they’re engrossing, enchanting, weird, epic wonders. Hyrule Warriors: Age of Imprisonment, another in the Koei-Tecmo developed spinoff series of Zelda-themed games, is a prequel to Tears of the Kingdom. It’s the story of Zelda traveling back in time to ancient Hyrule, and the origins of Ganondorf’s evil. I’m here for that, but a lot of hack and slash battles are in my way. 

A handful of hours in, I can say that the production values are wonderful. The voices and characters and worlds feel authentically Zelda. I feel like I’m getting a new chapter in the story I’d already been following. The Switch 2’s graphics show off smooth animation, too, even when battles can span hundreds of enemies.

But the game’s central style, which is endless slashing fights through hordes of enemies, gets boring for me. That’s what Hyrule Warriors is about, but the game so far feels more repetitive than strategic. And I just keep button-mashing to get to the next story chapter. For anyone who’s played Hyrule Warriors: Age of Calamity, expect more of the same, for the most part.

I do like that the big map includes parts in the depths and in the sky, mirroring the tri-level appeal of Tears of the Kingdom. But Age of Calamity isn’t a free-wandering game. Missions open up around the map, each one opening a contained map to battle through. Along the way, you unlock an impressive roster of Hyrule characters you can control.

As a Switch 2 exclusive to tempt Nintendo fans to make the console upgrade, it feels like a half success. I admire the production values, and I want to keep playing just to see where the story goes. But as a purchase, it’s a distant third to Donkey Kong Bananza and Mario Kart World.

Hyrule Warriors fans, you probably know what you’re probably in for, and will likely get this game regardless. Serious Zelda fans, you may enjoy it just for the story elements alone. 

As for me? I think I’ll play some more, but I’m already sort of tuning the game out a bit. I want more exploration, more puzzles, more curiosity. This game’s not about that. But it does show me how good a true next-gen Zelda could be on the Switch 2, whenever Nintendo decides to make that happen.

Continue Reading

Trending

Copyright © Verum World Media