Connect with us

Technologies

AI Is Bad at Sudoku. It’s Even Worse at Showing Its Work

Researchers did more than ask chatbots to play games. They tested whether AI models could describe their thinking. The results were troubling.

Chatbots are genuinely impressive when you watch them do things they’re good at, like writing a basic email or creating weird, futuristic-looking images. But ask generative AI to solve one of those puzzles in the back of a newspaper, and things can quickly go off the rails.

That’s what researchers at the University of Colorado at Boulder found when they challenged large language models to solve sudoku. And not even the standard 9×9 puzzles. An easier 6×6 puzzle was often beyond the capabilities of an LLM without outside help (in this case, specific puzzle-solving tools).

A more important finding came when the models were asked to show their work. For the most part, they couldn’t. Sometimes they lied. Sometimes they explained things in ways that made no sense. Sometimes they hallucinated and started talking about the weather.

If gen AI tools can’t explain their decisions accurately or transparently, that should cause us to be cautious as we give these things more control over our lives and decisions, said Ashutosh Trivedi, a computer science professor at the University of Colorado at Boulder and one of the authors of the paper published in July in the Findings of the Association for Computational Linguistics.

«We would really like those explanations to be transparent and be reflective of why AI made that decision, and not AI trying to manipulate the human by providing an explanation that a human might like,» Trivedi said.


Don’t miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source.


The paper is part of a growing body of research into the behavior of large language models. Other recent studies have found, for example, that models hallucinate in part because their training procedures incentivize them to produce results a user will like, rather than what is accurate, or that people who use LLMs to help them write essays are less likely to remember what they wrote. As gen AI becomes more and more a part of our daily lives, the implications of how this technology works and how we behave when using it become hugely important.

When you make a decision, you can try to justify it, or at least explain how you arrived at it. An AI model may not be able to accurately or transparently do the same. Would you trust it?

Why LLMs struggle with sudoku

We’ve seen AI models fail at basic games and puzzles before. OpenAI’s ChatGPT (among others) has been totally crushed at chess by the computer opponent in a 1979 Atari game. A recent research paper from Apple found that models can struggle with other puzzles, like the Tower of Hanoi.

It has to do with the way LLMs work and fill in gaps in information. These models try to complete those gaps based on what happens in similar cases in their training data or other things they’ve seen in the past. With a sudoku, the question is one of logic. The AI might try to fill each gap in order, based on what seems like a reasonable answer, but to solve it properly, it instead has to look at the entire picture and find a logical order that changes from puzzle to puzzle. 

Read more: 29 Ways You Can Make Gen AI Work for You, According to Our Experts

Chatbots are bad at chess for a similar reason. They find logical next moves but don’t necessarily think three, four or five moves ahead — the fundamental skill needed to play chess well. Chatbots also sometimes tend to move chess pieces in ways that don’t really follow the rules or put pieces in meaningless jeopardy. 

You might expect LLMs to be able to solve sudoku because they’re computers and the puzzle consists of numbers, but the puzzles themselves are not really mathematical; they’re symbolic. «Sudoku is famous for being a puzzle with numbers that could be done with anything that is not numbers,» said Fabio Somenzi, a professor at CU and one of the research paper’s authors.

I used a sample prompt from the researchers’ paper and gave it to ChatGPT. The tool showed its work, and repeatedly told me it had the answer before showing a puzzle that didn’t work, then going back and correcting it. It was like the bot was turning in a presentation that kept getting last-second edits: This is the final answer. No, actually, never mind, this is the final answer. It got the answer eventually, through trial and error. But trial and error isn’t a practical way for a person to solve a sudoku in the newspaper. That’s way too much erasing and ruins the fun.

AI struggles to show its work

The Colorado researchers didn’t just want to see if the bots could solve puzzles. They asked for explanations of how the bots worked through them. Things did not go well.

Testing OpenAI’s o1-preview reasoning model, the researchers saw that the explanations — even for correctly solved puzzles — didn’t accurately explain or justify their moves and got basic terms wrong. 

«One thing they’re good at is providing explanations that seem reasonable,» said Maria Pacheco, an assistant professor of computer science at CU. «They align to humans, so they learn to speak like we like it, but whether they’re faithful to what the actual steps need to be to solve the thing is where we’re struggling a little bit.»

Sometimes, the explanations were completely irrelevant. Since the paper’s work was finished, the researchers have continued to test new models released. Somenzi said that when he and Trivedi were running OpenAI’s o4 reasoning model through the same tests, at one point, it seemed to give up entirely. 

«The next question that we asked, the answer was the weather forecast for Denver,» he said.

(Disclosure: Ziff Davis, CNET’s parent company, in April filed a lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.)

Explaining yourself is an important skill

When you solve a puzzle, you’re almost certainly able to walk someone else through your thinking. The fact that these LLMs failed so spectacularly at that basic job isn’t a trivial problem. With AI companies constantly talking about «AI agents» that can take actions on your behalf, being able to explain yourself is essential.

Consider the types of jobs being given to AI now, or planned for in the near future: driving, doing taxes, deciding business strategies and translating important documents. Imagine what would happen if you, a person, did one of those things and something went wrong.

«When humans have to put their face in front of their decisions, they better be able to explain what led to that decision,» Somenzi said.

It isn’t just a matter of getting a reasonable-sounding answer. It needs to be accurate. One day, an AI’s explanation of itself might have to hold up in court, but how can its testimony be taken seriously if it’s known to lie? You wouldn’t trust a person who failed to explain themselves, and you also wouldn’t trust someone you found was saying what you wanted to hear instead of the truth. 

«Having an explanation is very close to manipulation if it is done for the wrong reason,» Trivedi said. «We have to be very careful with respect to the transparency of these explanations.»

Technologies

Today’s NYT Mini Crossword Answers for Wednesday, Oct. 22

Here are the answers for The New York Times Mini Crossword for Oct. 22.

Looking for the most recent Mini Crossword answer? Click here for today’s Mini Crossword hints, as well as our daily answers and hints for The New York Times Wordle, Strands, Connections and Connections: Sports Edition puzzles.


Need some help with today’s Mini Crossword? It’s one of those with absolutely no empty spaces, just a grid of letters, which means if you correctly answer all the Across answers, you’ve solved the Down answers, too. Need help? Read on. And if you could use some hints and guidance for daily solving, check out our Mini Crossword tips.

If you’re looking for today’s Wordle, Connections, Connections: Sports Edition and Strands answers, you can visit CNET’s NYT puzzle hints page.

Read more: Tips and Tricks for Solving The New York Times Mini Crossword

Let’s get to those Mini Crossword clues and answers.

Mini across clues and answers

1A clue: Roomful of students
Answer: CLASS

6A clue: Something to bring in a brown paper bag
Answer: LUNCH

7A clue: __ Harbor, sightseeing area of Baltimore
Answer: INNER

8A clue: Where many Stephen King novels are set
Answer: MAINE

9A clue: Beagle or bulldog
Answer: BREED

Mini down clues and answers

1D clue: Go bouldering, e.g.
Answer: CLIMB

2D clue: ___ New Year
Answer: LUNAR

3D clue: Redhead of musical/movie fame
Answer: ANNIE

4D clue: Something an actor might steal
Answer: SCENE

5D clue: Tear to pieces
Answer: SHRED

Continue Reading

Technologies

These Small Tweaks Can Give Your Old Android a Big Speed Boost

Instead of buying a new phone, try clearing some space, updating your software and changing a few battery settings.

If your Android is a few years old and starting to feel sluggish, it doesn’t mean you have to rush out and buy the newest flagship model. Thanks to longer software support from brands like Google and Samsung, older models can still run smoothly, as long as you give them a little attention. 

Before you start shopping for a replacement, try a few simple adjustments. You might be surprised by how much faster your phone feels once you clear out unused apps, optimize battery use and turn off background drains.

Whether you use a Samsung Galaxy, Motorola or OnePlus phone, chances are you can still improve battery life and overall speed without buying something new. Just remember that Android settings vary slightly from brand to brand, so the menus may look a little different depending on your phone.


Don’t miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source.


Settings to improve your battery life

Living with a phone that has poor battery life can be infuriating, but there are some steps you can take to maximize each charge right from the very beginning:

1. Turn off auto screen brightness or adaptive brightness and set the brightness level slider to under 50%

The brighter your screen, the more battery power it uses. 

To get to the setting, pull down the shortcut menu from the top of the screen and adjust the slider, if it’s there. Some phones may have a toggle for auto brightness in the shortcut panel; otherwise, you need to open the settings app and search for «brightness» to find the setting and turn it off.

2. Use Adaptive Battery and Battery Optimization

These features focus on learning how you use your phone, including which apps you use and when, and then optimizing the apps and the amount of battery they use. 

Some Android phones have a dedicated Battery section in the Settings app, while other phones (looking at you, Samsung) bury these settings. It’s a little different for each phone. I recommend opening your settings and searching for «battery» to find the right screen. Your phone may also have an adaptive charging setting that can monitor how quickly your phone battery charges overnight to preserve its health.

Why you should use dark mode more often

Another way to improve battery life while also helping save your eyes is to use Android’s dedicated dark mode. Any Android phone running Android 10 or newer will have a dedicated dark mode option. 

According to Google, dark mode not only reduces the strain that smartphone displays cause on our eyes but also improves battery life because it takes less power to display dark backgrounds on OLED displays (used in most flagship phones) than a white background. 

Depending on which version of Android your phone is running, and what company made your phone, you may have to dig around the settings app to find a dark mode. If your phone runs Android 10 or newer, you’ll be able to turn on system-wide dark mode. If it runs Android 9, don’t despair. Plenty of apps have their own dark mode option in the settings that you can use, whether or not you have Android 10. 

To turn it on dark mode, open the Settings app and search for Dark Mode, Dark Theme or even Night Mode (as Samsung likes to call it). I suggest using dark mode all the time, but if you’re not sure, you can always set dark mode to automatically turn on based on a schedule, say from 7 p.m. to 7 a.m. every day, or allow it to automatically switch based on your location at sunset and sunrise. 

Keep your home screen free of clutter

Planning to hit up the Google Play Store for a bunch of new Android apps? Be prepared for a lot of icon clutter on your home screen, which is where shortcuts land every time you install something.

If you don’t want that, there’s a simple way out of this: Long-press on an empty area of your home screen and tap Settings. Find the option labeled something along the lines of Add icon to Home Screen or Add new apps to Home Screen and turn it off. 

Presto! No more icons on the home screen when you install new apps. You can still add shortcuts by dragging an app’s icon out of the app drawer, but they won’t appear on your home screen unless you want them to.

Read more: Best Android Phones You Can Buy in 2024

Set up Do Not Disturb so that you can better focus

If your phone routinely spends the night on your nightstand, you probably don’t want it beeping or buzzing every time there’s a call, message or Facebook alert — especially when you’re trying to sleep. Android offers a Do Not Disturb mode that will keep the phone more or less silent during designated hours. On some phones, this is referred to as the Downtime setting or even Quiet Time.

Head to Settings > Sounds (or Notifications), then look for Do Not Disturb or a similar name. If you can’t find it, search for it using the built-in search feature in your settings.

Using the feature, you can set up a range of hours when you want to turn off the digital noise. Don’t worry, any notifications you get while Do Not Disturb is turned on will still be waiting for you when you wake up. Also, you can typically make an exception that allows repeat callers and favorite contacts’ calls to go through. Turn that on. If someone is calling you in an emergency, odds are they are going to keep trying.

Always be prepared in case you lose your phone or it’s stolen

Is there anything worse than a lost or stolen phone? Only the knowledge that you could have tracked it down if you had turned on Google’s Find My Device feature.

To prepare for a successful recovery, here’s what you need to do: Open the Settings app and then search for Find My Device. It’s usually in the Security section of the Settings app.

If you have a Samsung device, you can use Samsung’s Find My Mobile service, which is found in Settings > Biometrics and security > Find My Mobile

Once that’s enabled, you can head to android.com/find from any PC or mobile device and sign in to your account. Samsung users can visit findmymobile.samsung.com to find a lost phone. 

If you have trouble setting any of this up, be sure to read our complete guide to finding a lost Android phone.

Assuming your phone is on and online, you should be able to see its location on a map. From there, you can make it ring, lock it, set a lock screen note to tell whoever has it how to get it back to you, or, worst-case scenario, remotely wipe the whole thing.

And always keep your phone up to date

As obvious as it may seem, a simple software update could fix bugs and other issues slowing down your Android device. 

Before you download and install the latest software update, make sure your device is connected to Wi-Fi, or else this won’t work.

Now, open the Settings application and type in Update. You’ll then either see Software update or System update — choose either one. Then just download the software, wait for a few minutes and install it when it’s ready. Your Android device will reboot and install the latest software update available.

There’s a lot more to learn about a new phone. Here are the best ways to boost your cell signal, and here’s a flagship phone head-to-head comparison. Plus, check out CNET’s list of the best cases for your Samsung phone. More of an Apple fan? We have tips for boosting your iPhone’s performance, too.

Continue Reading

Technologies

I’m Finally Using the iPhone 17 Pro’s Camera Control, Thanks to These iOS 26 Settings

In just a month, I’ve already used Camera Control on my iPhone 17 Pro Max more than I did in a whole year with the iPhone 16 Pro.

I was keen on using the Camera Control button when it first debuted on the iPhone 16 Pro. But in over a year of use, it caused more accidental swipes and presses than its intended use cases to take photos and adjust camera settings. I was frustrated with the experience and hoped that Apple would remove it from the iPhone 17 lineup. Instead, the Cupertino, California-based company made its touch-sensitive capacitive control surface more customizable with iOS 26. And I’m happy to report that it helped!

I’ve been using the iPhone 17 Pro since launch and spent 5 to 10 minutes customizing the Camera Control to my liking. The result? Minimized accidental swipes and more conscious usability.

I transformed my Camera Control experience by changing a few iOS 26 settings

When setting up a new iOS 26-supported iPhone, Apple includes a toggle (now turned off by default) called Light press to adjust Zoom, Exposure and more. This is what used to cause a lot of fake input earlier. I’m glad it is turned off by default.

Apple now also lets you customize the Camera Control further from the Settings menu. I tweaked settings there to personalize my shortcuts, functionalities and more.

For example, I’ve set the Camera Control to launch a Code Scanner on Double Click without requiring the screen to be on. This allows me to scan and pay at payment kiosks (my most frequently used mode of payment) without needing to open the payment app and then tap on a menu to scan a code. If I enter the Code Scanner without Face ID, it requires authentication before making the payment, so it is still as secure as ever.

Earlier, I had set an Action Button shortcut to open Google Pay, but I realized I still need a one-press solution to turn the phone to silent mode. Adding a Code Scanner shortcut to Camera Control frees the Action Button to be my Silent Switch again. Moreover, Code Scanner lets you select from multiple apps to pay a vendor, which could be useful for people who use multiple payment apps.

Secondly, I have turned off the Swipe gesture and selected only three controls that I use most often. Now, when I open the Camera app, I can lightly press on the Camera Control button and then swipe between my selected controls. It doesn’t register swipes from the get-go. This has reduced fake touches and my frustrating experience with the swipe gesture.

To further streamline my controls, I chose Exposure, Styles and Tone, and left out Depth, Zoom and Cameras. This way, I have access to hidden viewfinder settings with a single press-and-swipe gesture at my fingertips. 

I also turned off the Clean Preview toggle, so I can still switch between cameras with a single tap, and switched on the Lock Focus and Exposure toggle for a light press-and-hold gesture.

Customizing these settings helped me personalize Camera Control and use it more often. Now, it appeals to me with the settings I need and the way I need them, instead of being an overcrowded mess. And you can personalize your Camera Control, too. Here’s how:

Change Camera Control launch functionality

You can use Camera Control as another Action Button to launch an app of your choice. The only requirement is that the app should have access to the camera.

  1. Go to Settings > Camera > Camera Control.
  2. Under Launch Camera, select the app you need.
  3. Go back and select Single Click or Double Click to open the said app.

I rely on Double Click so I don’t accidentally trigger an app when taking out the iPhone from my pocket. In my opinion, it is the safer and more convenient choice.

Under the same Launch Camera menu, you can also choose if you want the screen to be on or off when opening the app. I have turned it off to save the extra step of scanning my face to access the said app.

Choose the Controls that you want to appear on Camera Control

Apple allows you to choose from six controls, namely, Exposure, Depth, Zoom, Cameras, Styles and Tone. I have chosen three because the other three are available as on-screen toggles in the viewfinder.

  1. Go to Settings > Camera > Camera Control.
  2. Under Controls, make sure Camera Adjustments is turned on.
  3. Tap on Customize.
  4. Under Gesture, turn on Light Press and toggle off Swipe.
  5. Under Controls, choose the functionalities you need.
  6. Now, turn off the Clean Preview toggle if you require the viewfinder toggles to remain accessible.

You can further adjust the Camera Control pressure by going into Settings> Camera > Camera Control > Accessibility.

Turn on Lock Exposure and Focus with Camera Control

This setting will help you lock the exposure and focus without needing to press and hold on the viewfinder. It can be beneficial when you need consistent settings, especially when moving the camera from one subject to another.

  1. Go to Settings > Camera > Camera Control.
  2. Swipe down to Lock Exposure and Focus.
  3. Tap on the toggle to turn it on.

For me, Camera Control was a hot mess when it debuted last year because I was either using on-screen controls or the new button. That’s why room for more personalization and customizability has been a game-changer. I realized I could access on-screen toggles while adding hidden settings to one-tap access. On my iPhone 17 Pro, I now use the Camera Control to open my payments app, adjust Exposure and Styles as well as trigger Visual Intelligence when needed.

Continue Reading

Trending

Exit mobile version