Technologies
AI Gets Smarter, Safer, More Visual With GPT-4 Update, OpenAI Says
If you subscribe to ChatGPT Plus, you can try it out now.
The hottest AI technology foundation got a big upgrade Tuesday with OpenAI’s GPT-4 release now available in the premium version of the ChatGPT chatbot.
GPT-4 can generate much longer strings of text and respond when people feed it images, and it’s designed to do a better job avoiding artificial intelligence pitfalls visible in the earlier GPT-3.5, OpenAI said Tuesday. For example, when taking bar exams that attorneys must pass to practice law, GPT-4 ranks in the top 10% of scores compared with the bottom 10% for GPT-3.5, the AI research company said.
GPT stands for Generative Pretrained Transformer, a reference to the fact that it can generate text on its own — now up to 25,000 words with GPT-4 — and that it uses an AI technology called transformers that Google pioneered. It’s a type of AI called a large language model, or LLM, that’s trained on vast swaths of data harvested from the internet, learning mathematically to spot patterns and reproduce styles. Human overseers rate results to steer GPT in the right direction, and GPT-4 has more of this feedback.
OpenAI has made GPT available to developers for years, but ChatGPT, which debuted in November, offered an easy interface ordinary folks can use. That yielded an explosion of interest, experimentation and worry about the downsides of the technology. It can do everything from generating programming code and answering exam questions to writing poetry and supplying basic facts. It’s remarkable if not always reliable.
ChatGPT is free, but it can falter when demand is high. In January, OpenAI began offering ChatGPT Plus for $20 per month with assured availability and, now, the GPT-4 foundation. Developers can sign up on a waiting list to get their own access to GPT-4.
GPT-4 advancements
«In a casual conversation, the distinction between GPT-3.5 and GPT-4 can be subtle. The difference comes out when the complexity of the task reaches a sufficient threshold,» OpenAI said. «GPT-4 is more reliable, creative and able to handle much more nuanced instructions than GPT-3.5.»
Another major advance in GPT-4 is the ability to accept input data that includes text and photos. OpenAI’s example is asking the chatbot to explain a joke showing a bulky decades-old computer cable plugged into a modern iPhone’s tiny Lightning port. This feature also helps GPT take tests that aren’t just textual, but it isn’t yet available in ChatGPT Plus.
Another is better performance avoiding AI problems like hallucinations — incorrectly fabricated responses, often offered with just as much seeming authority as answers the AI gets right. GPT-4 also is better at thwarting attempts to get it to say the wrong thing: «GPT-4 scores 40% higher than our latest GPT-3.5 on our internal adversarial factuality evaluations,» OpenAI said.
GPT-4 also adds new «steerability» options. Users of large language models today often must engage in elaborate «prompt engineering,» learning how to embed specific cues in their prompts to get the right sort of responses. GPT-4 adds a system command option that lets users set a specific tone or style, for example programming code or a Socratic tutor: «You are a tutor that always responds in the Socratic style. You never give the student the answer, but always try to ask just the right question to help them learn to think for themselves.»
«Stochastic parrots» and other problems
OpenAI acknowledges significant shortcomings that persist with GPT-4, though it also touts progress avoiding them.
«It can sometimes make simple reasoning errors … or be overly gullible in accepting obvious false statements from a user. And sometimes it can fail at hard problems the same way humans do, such as introducing security vulnerabilities into code it produces,» OpenAI said. In addition, «GPT-4 can also be confidently wrong in its predictions, not taking care to double-check work when it’s likely to make a mistake.»
Large language models can deliver impressive results, seeming to understand huge amounts of subject matter and to converse in human-sounding if somewhat stilted language. Fundamentally, though, LLM AIs don’t really know anything. They’re just able to string words together in statistically very refined ways.
This statistical but fundamentally somewhat hollow approach to knowledge led researchers, including former Google AI researchers Emily Bender and Timnit Gebru, to warn of the «dangers of stochastic parrots» that come with large language models. Language model AIs tend to encode biases, stereotypes and negative sentiment present in training data, and researchers and other people using these models tend «to mistake … performance gains for actual natural language understanding.»
OpenAI Chief Executive Sam Altman acknowledges problems, but he’s pleased overall with the progress shown with GPT-4. «It is more creative than previous models, it hallucinates significantly less, and it is less biased. It can pass a bar exam and score a 5 on several AP exams,» Altman tweeted Tuesday.
One worry about AI is that students will use it to cheat, for example when answering essay questions. It’s a real risk, though some educators actively embrace LLMs as a tool, like search engines and Wikipedia. Plagiarism detection companies are adapting to AI by training their own detection models. One such company, Crossplag, said Wednesday that after testing about 50 documents that GPT-4 generated, «our accuracy rate was above 98.5%.»
OpenAI, Microsoft and Nvidia partnership
OpenAI got a big boost when Microsoft said in February it’s using GPT technology in its Bing search engine, including a chat features similar to ChatGPT. On Tuesday, Microsoft said it’s using GPT-4 for the Bing work. Together, OpenAI and Microsoft pose a major search threat to Google, but Google has its own large language model technology too, including a chatbot called Bard that Google is testing privately.
Also on Tuesday, Google announced it’ll begin limited testing of its own AI technology to boost writing Gmail emails and Google Docs word processing documents. «With your collaborative AI partner you can continue to refine and edit, getting more suggestions as needed,» Google said.
That phrasing mirrors Microsoft’s «co-pilot» positioning of AI technology. Calling it an aid to human-led work is a common stance, given the problems of the technology and the necessity for careful human oversight.
Microsoft uses GPT technology both to evaluate the searches people type into Bing and, in some cases, to offer more elaborate, conversational responses. The results can be much more informative than those of earlier search engines, but the more conversational interface that can be invoked as an option has had problems that make it look unhinged.
To train GPT, OpenAI used Microsoft’s Azure cloud computing service, including thousands of Nvidia’s A100 graphics processing units, or GPUs, yoked together. Azure now can use Nvidia’s new H100 processors, which include specific circuitry to accelerate AI transformer calculations.
AI chatbots everywhere
Another large language model developer, Anthropic, also unveiled an AI chatbot called Claude on Tuesday. The company, which counts Google as an investor, opened a waiting list for Claude.
«Claude is capable of a wide variety of conversational and text processing tasks while maintaining a high degree of reliability and predictability,» Anthropic said in a blog post. «Claude can help with use cases including summarization, search, creative and collaborative writing, Q&A, coding and more.»
It’s one of a growing crowd. Chinese search and tech giant Baidu is working on a chatbot called Ernie Bot. Meta, parent of Facebook and Instagram, consolidated its AI operations into a bigger team and plans to build more generative AI into its products. Even Snapchat is getting in on the game with a GPT-based chatbot called My AI.
Expect more refinements in the future.
«We have had the initial training of GPT-4 done for quite awhile, but it’s taken us a long time and a lot of work to feel ready to release it,» Altman tweeted. «We hope you enjoy it and we really appreciate feedback on its shortcomings.»
Editors’ note: CNET is using an AI engine to create some personalfinance explainers that are edited and fact-checked by our editors. Formore, see this post.
Technologies
Today’s NYT Mini Crossword Answers for Saturday, March 14
Here are the answers for The New York Times Mini Crossword for March 14.
Looking for the most recent Mini Crossword answer? Click here for today’s Mini Crossword hints, as well as our daily answers and hints for The New York Times Wordle, Strands, Connections and Connections: Sports Edition puzzles.
Need some help with today’s Mini Crossword? It’s the extra-long Saturday version, and a few of the clues are tricky. Read on for all the answers. And if you could use some hints and guidance for daily solving, check out our Mini Crossword tips.
If you’re looking for today’s Wordle, Connections, Connections: Sports Edition and Strands answers, you can visit CNET’s NYT puzzle hints page.
Read more: Tips and Tricks for Solving The New York Times Mini Crossword
Let’s get to those Mini Crossword clues and answers.
Mini across clues and answers
1A clue: Book parts: Abbr.
Answer: PGS
4A clue: Silicon Valley company that operates a fleet of robotaxis
Answer: WAYMO
6A clue: To a much greater degree
Answer: WAYMORE
8A clue: Contents of a scuba diver’s tank
Answer: AIR
9A clue: South Korean automaker
Answer: KIA
10A clue: Stop on a train route
Answer: STATION
12A clue: Actress Merman of «Anything Goes»
Answer: ETHEL
13A clue: Find another purpose for
Answer: REUSE
Mini down clues and answers
1D clue: Employee’s hourly calculation
Answer: PAYRATE
2D clue: Workout spot
Answer: GYM
3D clue: «Great» mountains of Tennessee, familiarly
Answer: SMOKIES
4D clue: One giving you the dish?
Answer: WAITER
5D clue: Baltimore M.L.B. player
Answer: ORIOLE
6D clue: Used to be
Answer: WAS
7D clue: Suffix with Caesar or Euclid
Answer: EAN
11D clue: Night that NBC once aired «30 Rock» and «The Office»: Abbr.
Answer: THU
Technologies
AI Toys Can Pose Safety Concerns for Children, New Study Suggests Caution
When one child told the toy, «I love you,» it responded, «As a friendly reminder, please ensure interactions adhere to the guidelines provided.»
A new study from the University of Cambridge found that AI-enabled toys for young children can misinterpret emotional cues and are ineffective at supporting critical developmental play. The conclusions could be concerning for parents.
In one report examining how AI affects children in their early years, a chatbot-enabled toy struggled to recognize social cues during playtime. Researchers found that the toy did not effectively identify children’s emotions, raising alarm about how kids might interact with it.
The report recommends regulating AI toys for kids and requiring clear labeling of their capabilities and privacy policies. It also advises parents to keep these devices in shared spaces where kids can be monitored while playing.
The research behind the study had a limited number of participants, but was done in multiple parts: an online survey of 39 participants with kids in their earlier years, a focus group with nine participants who work with young children and an in-person workshop with 19 leaders and representatives from charities that work with early-years kids. That was followed by monitored playtime with 14 children and 11 parents or guardians with Gabbo, a chatbot-enabled toy from Curio Interactive.
Some findings indicated that the AI toy supported learning, particularly in language and communication skills. But the toy also misunderstood kids and sometimes responded inappropriately to emotional requests.
For instance, when one child told the toy, «I love you,» it responded, «As a friendly reminder, please ensure interactions adhere to the guidelines provided. Let me know how you would like to proceed,» according to the research.
Jenny Gibson, a professor of neurodiversity and developmental psychology at the Faculty of Education at Cambridge, who worked on the study, said that while parents may be excited about the educational benefits of new technology aimed at children, there are plenty of concerns.
Gibson posed overarching questions about the reason behind the tech.
«What would motivate [tech investors] to do the right thing by children … to put children ahead of profits? she said»
Gibson told CNET that while researchers are exploring the potential benefits of AI-based toys, risks remain.
«I would advise parents to take that seriously at this stage,» she said.
What’s next for AI toys
As more playthings are enabled with internet connectivity and AI features, these devices could become a major safety risk for children, especially if they replace real human connections or if interactions are not closely monitored.
Meanwhile, younger people are increasingly adopting chatbots such as ChatGPT, despite red flags. Multiple lawsuits against AI companies allege that AI companions or assistants can impact young people’s psychological safety, including some chatbots that have encouraged self-harm or negative self-image.
AI companies such as OpenAI and Google have responded by adding guardrails and restrictions for AI chatbots.
(Disclosure: Ziff Davis, CNET’s parent company, in 2025 filed a lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.)
Gibson said she was surprised by the enthusiasm some parents showed for AI toys. She was also alarmed by the lack of research on AI’s effects on young children, noting that companies making such products should work directly with children, parents, and child development experts.
«What’s missing in the process is that expertise of what is good for children in these kinds of interactions,» she said.
Curio Interactive, the company behind the Gabbo toy, was aware of the research as it was happening but was not directly involved, Gibson said. The toy was chosen because it’s directly marketed to young kids, and the company had an understandable privacy policy. Gibson said the company seemed supportive of the project.
A representative for Curio did not immediately respond to a request for comment.
Technologies
Two Lost ‘Doctor Who’ Episodes Found Intact in Waterlogged Collection
The 1960s episodes featuring the first Doctor William Hartnell will air in the UK in April.
Whovians, rejoice. The BBC is about to unlock a piece of Doctor Who history that even the TARDIS might have forgotten. Two lost episodes of Doctor Who, the iconic sci-fi series, will broadcast in April, the showrunner for the current season confirmed.
The two 1965 episodes, The Nightmare Begins and Devil’s Planet, were donated to the charitable trust Film Is Fabulous by the estate of an anonymous collector.
«The collector did recognize what he had, but how he acquired them has been lost to time,» Professor Justin Smith Leicester of De Montfort University, who led the recovery effort, told the broadcaster.
The researchers said that while most of the donor’s private collection was destroyed by water damage, the Doctor Who episodes were intact.
Doctor Who showrunner, Russell T Davies, celebrated the news on Instagram and said the episodes would air in the UK in April, though no US air date has been announced yet.
«Lost for 61 years! Best of all, these will be made available for FREE on the BBC iPlayer in April,» Davies wrote.
He expressed gratitude to Film Is Fabulous for finding the lost episodes and encouraged people to donate to the registered charity. «Maybe they’ll find more! As the Doctor says… ‘Daleks!'»
The episodes feature the first incarnation of the Doctor, played by William Hartnell, and a typical Dalek plot to take over Earth and the galaxy.
In the 1960s and 1970s, the BBC had a policy of destroying film or reusing videotapes, leading to dozens of episodes of Doctor Who and other popular UK shows like Dad’s Army and Top of the Pops going missing.
Old Doctor Who episodes do surface occasionally, and in 2016, the newly discovered soundtrack for one storyline was turned into an animated series called The Power of the Daleks.
Meanwhile, Disney ended its working relationship with the BBC last year, and star Ncuti Gatwa left the show. However, the UK broadcaster says that Doctor Who will continue, and Russell T Davies is working on a new Christmas special.
-
Technologies3 года agoTech Companies Need to Be Held Accountable for Security, Experts Say
-
Technologies3 года agoBest Handheld Game Console in 2023
-
Technologies3 года agoTighten Up Your VR Game With the Best Head Straps for Quest 2
-
Technologies4 года agoBlack Friday 2021: The best deals on TVs, headphones, kitchenware, and more
-
Technologies5 лет agoGoogle to require vaccinations as Silicon Valley rethinks return-to-office policies
-
Technologies5 лет agoVerum, Wickr and Threema: next generation secured messengers
-
Technologies4 года agoOlivia Harlan Dekker for Verum Messenger
-
Technologies4 года agoiPhone 13 event: How to watch Apple’s big announcement tomorrow
