Technologies
Anker Recalls 535 Power Bank Over Fire Risk
An Anker 535 Power Bank was «the most probable cause» of a recent Maryland house fire, according to a local fire chief.

Anker recently recalled some of its 535 Power Bank products, citing possible overheating and fire risk as a safety hazard, according to a notice on the company’s website.
The China-based electronics manufacturer said on its website that it discovered a «manufacturing condition» that impacted a small number of its 535 Power Banks — specifically PowerCore 20K A1366 models — and issued a voluntary recall on Jan. 18. The 535 Power Bank is a portable charger typically used for devices like laptops, tablets and smartphones during travel.
The UK’s Office for Product Safety and Standards posted a recall notice on Feb. 1 about the Anker 535 Power Banks, stating that some of the products’ battery cells may have been damaged during the production process, which could cause a power bank to overheat.
In the US, an Anker 535 Power Bank was recently cited as «the most probable cause» of a Maryland house fire, as «there were no other electrical devices or anything else in the suitcase» where the fire appeared to start that would have caused the blaze, a local fire chief said.
Anker’s website says it’s working closely with local government agencies to «ensure a comprehensive and safe recall process.»
If you purchased an Anker 535 Power Bank, you can determine whether yours may have been affected by checking the bottom back of the battery and verifying the make and model. If it reads «535 Power Bank (PowerCore 20K) Model: A1366» you should stop using the device and return it to Anker.
Here’s how to safely dispose of 535 Power Bank and request a refund from Anker. The company says to:
- Stop using the device immediately.
- Store the 535 Power Bank in a safe location.
- Don’t throw the battery into the trash or recycling bin.
- Dispose of your device only at a facility that accepts Lithium batteries.
- FIll out this recall form and request a refund for your 535 Power Bank.
You can contact the company via email at support@anker.com or 800-988-7973 in the US.
In a statement to CNET, Anker said it has «stopped all sales of the affected product and has begun the process of informing customers and communicating with our retail channels,» as well as filing notice with the United States Consumer Products Safety Commission and cooperating with the recall effort.
Technologies
Today’s NYT Mini Crossword Answers for Tuesday, Oct. 14
Here are the answers for The New York Times Mini Crossword for Oct. 14.

Looking for the most recent Mini Crossword answer? Click here for today’s Mini Crossword hints, as well as our daily answers and hints for The New York Times Wordle, Strands, Connections and Connections: Sports Edition puzzles.
Today’s Mini Crossword has an odd vertical shape, with an extra Across clue, and only four Down clues. The clues are not terribly difficult, but one or two could be tricky. Read on if you need the answers. And if you could use some hints and guidance for daily solving, check out our Mini Crossword tips.
If you’re looking for today’s Wordle, Connections, Connections: Sports Edition and Strands answers, you can visit CNET’s NYT puzzle hints page.
Read more: Tips and Tricks for Solving The New York Times Mini Crossword
Let’s get to those Mini Crossword clues and answers.
Mini across clues and answers
1A clue: Smokes, informally
Answer: CIGS
5A clue: «Don’t have ___, man!» (Bart Simpson catchphrase)
Answer: ACOW
6A clue: What the vehicle in «lane one» of this crossword is winning?
Answer: RACE
7A clue: Pitt of Hollywood
Answer: BRAD
8A clue: «Yeah, whatever»
Answer: SURE
9A clue: Rd. crossers
Answer: STS
Mini down clues and answers
1D clue: Things to «load» before a marathon
Answer: CARBS
2D clue: Mythical figure who inspired the idiom «fly too close to the sun»
Answer: ICARUS
3D clue: Zoomer around a small track
Answer: GOCART
4D clue: Neighbors of Norwegians
Answer: SWEDES
Technologies
Watch SpaceX’s Starship Flight Test 11
Technologies
New California Law Wants Companion Chatbots to Tell Kids to Take Breaks
Gov. Gavin Newsom signed the new requirements on AI companions into law on Monday.

AI companion chatbots will have to remind users in California that they’re not human under a new law signed Monday by Gov. Gavin Newsom.
The law, SB 243, also requires companion chatbot companies to maintain protocols for identifying and addressing cases in which users express suicidal ideation or self-harm. For users under 18, chatbots will have to provide a notification at least every three hours that reminds users to take a break and that the bot is not human.
It’s one of several bills Newsom has signed in recent weeks dealing with social media, artificial intelligence and other consumer technology issues. Another bill signed Monday, AB 56, requires warning labels on social media platforms, similar to those required for tobacco products. Last week, Newsom signed measures requiring internet browsers to make it easy for people to tell websites they don’t want them to sell their data and banning loud advertisements on streaming platforms.
AI companion chatbots have drawn particular scrutiny from lawmakers and regulators in recent months. The Federal Trade Commission launched an investigation into several companies in response to complaints by consumer groups and parents that the bots were harming children’s mental health. OpenAI introduced new parental controls and other guardrails in its popular ChatGPT platform after the company was sued by parents who allege ChatGPT contributed to their teen son’s suicide.
«We’ve seen some truly horrific and tragic examples of young people harmed by unregulated tech, and we won’t stand by while companies continue without necessary limits and accountability,» Newsom said in a statement.
Don’t miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source.
One AI companion developer, Replika, told CNET that it already has protocols to detect self-harm as required by the new law, and that it is working with regulators and others to comply with requirements and protect consumers.
«As one of the pioneers in AI companionship, we recognize our profound responsibility to lead on safety,» Replika’s Minju Song said in an emailed statement. Song said Replika uses content-filtering systems, community guidelines and safety systems that refer users to crisis resources when needed.
Read more: Using AI as a Therapist? Why Professionals Say You Should Think Again
A Character.ai spokesperson said the company «welcomes working with regulators and lawmakers as they develop regulations and legislation for this emerging space, and will comply with laws, including SB 243.» OpenAI spokesperson Jamie Radice called the bill a «meaningful move forward» for AI safety. «By setting clear guardrails, California is helping shape a more responsible approach to AI development and deployment across the country,» Radice said in an email.
One bill Newsom has yet to sign, AB 1064, would go further by prohibiting developers from making companion chatbots available to children unless the AI companion is «not foreseeably capable of» encouraging harmful activities or engaging in sexually explicit interactions, among other things.
-
Technologies3 года ago
Tech Companies Need to Be Held Accountable for Security, Experts Say
-
Technologies3 года ago
Best Handheld Game Console in 2023
-
Technologies3 года ago
Tighten Up Your VR Game With the Best Head Straps for Quest 2
-
Technologies4 года ago
Verum, Wickr and Threema: next generation secured messengers
-
Technologies4 года ago
Google to require vaccinations as Silicon Valley rethinks return-to-office policies
-
Technologies4 года ago
Black Friday 2021: The best deals on TVs, headphones, kitchenware, and more
-
Technologies4 года ago
Olivia Harlan Dekker for Verum Messenger
-
Technologies4 года ago
iPhone 13 event: How to watch Apple’s big announcement tomorrow