Connect with us

Technologies

Scientists Develop ‘Cellular Glue’ That Could Heal Wounds, Regrow Nerves

One day, these special synthetic molecules could also help mitigate the organ shortage crisis.

Researchers from the University of California, San Francisco announced a fascinating innovation on Monday. They call it «cellular glue» and say it could one day open doors to massive medical achievements, like building organs in a lab for transplantation and reconstructing nerves that’ve been damaged beyond the reach of standard surgical repair.

Basically, the team engineered a set of synthetic molecules that can be manipulated to coax cells within the human body to bond with one another. Together, these molecules constitute the so-called «cellular glue» and act like adhesive molecules naturally found in and around cells that involuntarily dictate the way our tissues, nerves and organs are structured and anchored together.

Only in this case scientists can voluntarily control them.

«The properties of a tissue, like your skin for example, are determined in large part by how the different cells are organized within it,» Adam Stevens, a researcher at UCSF’s Cell Design Institute and first author of a paper in the journal Nature, said in a statement. «We’re devising ways to control this organization of cells, which is central to being able to synthesize tissues with the properties we want them to have.»

Doctors could eventually use the sticky material as a viable mechanism to mend patients’ wounds, regrow nerves otherwise deemed destroyed and potentially even work toward regenerating diseased lungs, livers and other vital organs.

That last bit could lend a hand in alleviating the crisis of donor organs rapidly running out of supply. According to the Health Resources and Services Administration, 17 people in the US die each day while on the waitlist for an organ transplant, yet every 10 minutes, another person is added to that list.

«Our work reveals a flexible molecular adhesion code that determines which cells will interact, and in what way,» Stevens said. «Now that we are starting to understand it, we can harness this code to direct how cells assemble into tissues and organs.»

Ikea cells

Right after babies are born (and even when they’re still in the womb) their cells essentially find it easy to reconnect with one another when a bond is lost. This is primarily because kids are still growing, so their cells are still actively coming together. But as a consequence, that’s also why their scratches and scrapes tend to heal quite quickly.

In other words, think of children’s cell molecules as having lots of clear-cut instructions on how to put themselves together to make tissues, organs and nerves. They’re like sentient little pieces of Ikea furniture with the store’s building booklet in hand.

As people get older, however, those biological Ikea instructions get put in the attic, the team explains. That’s because, for the most part, the body is pretty solidified — and this is sometimes a problem. For instance, when someone’s liver gets really damaged, their liver cell molecules may need to refer back to those Ikea instructions but can’t find them.

But that’s where «cellular glue» molecules come in. These rescuers can essentially be primed with those Ikea instructions before being sent into the body, so their blueprint is fresh. Scientists can load them up with information on which cell molecules to bond with and even how strongly to bond with them.

Then, these glue molecules can guide relevant cells toward one another, helping along the healing and regeneration processes.

«In a solid organ, like a lung or a liver, many of the cells will be bonded quite tightly,» explains a UCSF description of the new invention. «But in the immune system, weaker bonds enable the cells to flow through blood vessels or crawl between the tightly bound cells of skin or organ tissues to reach a pathogen or a wound.»

To make this kind of customization possible, the researchers added two important components to their cellular glue. First, part of the molecule acts as a receptor. It remains on the outside of the cell and determines which other cells the molecule is allowed to interact with. Second, there’s the bond-strength-tuner. This section exists within the cell. Mix and match those two traits and, the team says, you can create an array of cell adhesion molecules prepped to bond in various ways.

«We were able to engineer cells in a manner that allows us to control which cells they interact with, and also to control the nature of that interaction,» Wendell Lim, director of UCSF’s Cell Design Institute and senior author of the paper, said in a statement.

In fact, the team says the range of potential molecules is wide enough that they could inform the academic stage of medical studies, too. Researchers could make mock tissues, for example, to deepen understanding of the human body as a whole.

Or as Stevens put it, «These tools could be really transformative.»

Technologies

Tariffs Explained: I Have Everything You Need to Know as Walmart, Subaru Hike Prices

Continue Reading

Technologies

Google I/O Announcements: The Latest AI Upgrades Coming to Gemini, XR and More

From its new Project Aura XR glasses to Chrome’s wants-to-be-more-helpful AI mode, Gemini Live and new Flow generative video tool, Google puts AI everywhere.

As you’d expect, this year’s Google I/O developer’s conference focused almost exclusively on AI — where the company’s Gemini AI platform stands, where it’s going and how much it’s going to cost you now for its new AI Ultra subscription plan (spoiler: $250 per month). Meanwhile, a new Flow app expands the company’s video-generation toolset, and its Android XR glasses make their debut. 

Plus, all AI usage and performance numbers are up! (Given that a new 42.5-exaflop Ironwood Tensor processing unit is coming to Google Cloud later this year, they’ll continue to rise.) 

Google’s Project Aura, a developer kit for Android XR that includes new AR glasses from Xreal, is the company’s next step in the company’s roadmap toward glasses-based, AI-driven extended reality. CNET’s Scott Stein goes in-depth in an exclusive interview with Shahram Izadi, Google’s VP and GM for Android XR about that future. And headset-based Project Moohan, developed in conjunction with Samsung, is now available, and Google’s working with Samsung to extend beyond headsets. 

For a play-by-play of the event, you can read the archive of our live blog.

Google already held a separate event for Android, where it launched Android 16, debuting its new Material 3 Expressive interface, updates to security and an update on Gemini integration and features. 

A lot of the whizzy new AI features are only available via one of its subscription levels. AI Pro is just a rebranding of Google’s $20-per-month Gemini Advanced plan (adding some new features), but Google AI Ultra is a pricier new option — $250 per month, with half off the first three months for the moment — that provides access to the latest, spiffiest and least usage-limited of all its tools and models,  as well as a prototype for managing AI agents and the 30 terabytes of storage you’re going to need to store it all. They’re both available today.

Google also wants to make your automation sound smarter with Personalized Smart Replies, which makes your generated answers sound more like you, as well as plowing through pieces of information on your device to provide relevant information. It’ll be in Gmail this summer for subscribers. Eventually, it’ll be everywhere. 

Also, it includes lots of better models, better coding tools and other details on developer-friendly things you expect from a developer conference. The announcement included its conversational Gemini Live, formerly part of Project Astra, its interactive, agentic, voice AI, kitchen sink AI app. (As Managing Editor Patrick Holland says, «Astra is a rehearsal of features that, when they’re ready for the spotlight, get added to Gemini Live.») And for researchers, NotebookLM incorporates Gemini Live to improve its… everything.

It’s available now in the US. 

Chrome AI Mode

People (that is, those over 18) who pony up for the subscriptions, plus users on the Chrome Beta, Dev and Canary tracks, will be able to try out the company’s expanded Gemini integration with Chrome — summary, research and agentic chat based on the contents of your screen, somewhat like Gemini Live does for phones (which, by the way, is available for free on Android and iOS as of today). But the Chrome version is more suited to the type of things you do at a computer rather than a phone. (Microsoft already does this with Copilot in its own Edge browser.)

Eventually, Google plans for Gemini in Chrome to be capable of synthesizing using multiple tabs and voice navigation. 

The company is also expanding how you can interact with its AI Overviews in Google Search as part of AI Mode, with interactions with AI Overviews and more agentic shopping help. It’s a new tab with search, or on the search bar, and it’s available now. It includes deeper searches, Personal Context — which uses all the information it knows about you, and that’s a lot — to make suggestions and customize replies.

The company detailed its new AI Mode for shopping, which has an improved conversational shopping experience, a checkout that monitors for the best pricing, and an updated «try on» interface that lets you upload a photo of yourself rather than modeling it on a generic body. 

Google plans to launch it soon, though the updated «try on» feature is now available in the US via Search Labs.

Google Beam

Formerly known as Project Starline, Google Beam is the updated version of the company’s 3D videoconferencing, now with AI. It uses a six-camera array to capture all angles of you, which the AI then stitches together, uses head tracking to follow your movements, and sends at up to 60 frames per second.

The platform uses a light field display that doesn’t require wearing any special equipment, but that technology also tends to be sensitive to off-angle viewing. HP is an old hand in the large-scale scanning biz, including 3D scanning, so the partnership with Google isn’t a big surprise. 

Flow and other generative creative tools

Google Flow is a new tool that builds on Imagen 4 and Veo 3 to perform tasks like creating AI video clips and stitching them into longer sequences, or extending them, with a single prompt while keeping them consistent from scene to scene. It also provides editing tools like camera controls. It’s available as part of Gemini AI Ultra. 

Imagen 4 image generation is more detailed, with improved tonality and better text and typography. And it’s faster. Meanwhile, Veo 3, also available today, has a better understanding of physics and native audio generation — sound effects, background sounds and dialogue.

Of course, all this is available under the AI Pro plan. Google’s Synth ID gen AI detection tool is also available today.

Continue Reading

Technologies

WWDC 2025 Set for June 9. What to Expect From iOS 19 and How to Watch

The tech giant has published the lineup for its summer software event, including details of how to stream the keynote.

Apple has revealed the lineup for this year’s WWDC, its annual developer conference, which is scheduled to take place from June 9 to June 13, with a mix of online and in-person sessions. The company has scheduled a special event for the first day at Apple Park, where it will stream its keynote and platforms state of the union address.

The June 9 keynote is scheduled to take place at 10 a.m. PT, and will detail «groundbreaking updates coming to Apple platforms,» according to a press release. During the keynote, we expect to see the company announce its plans for iOS 19 and other software updates across iPad, Mac, Apple Watch and Apple TV.

The platforms state of the union will take place on the same day at 1 p.m. PT, offering «a deeper dive into the new tools that will further empower Apple developers, including advances across iOS, iPadOS, macOS, tvOS, visionOS, and watchOS.»

You can watch both events online via apple.com, the Apple TV app, and the Apple YouTube channel.

As always, iOS 19 likely won’t arrive later in the year (probably after the annual September iPhone event), but WWDC will provide us with our first glimpse of Apple’s plans for our devices. In the meantime, check out our iOS 19 wish list of features that CNET staff hope will be coming with this next big software update.

What is WWDC?

WWDC is a conference for the developers who create the apps we use across our Apple devices, from our iPhones to our MacBooks, to learn about new tools and features they can take advantage of that year. However, the event also garners significant attention from beyond the developer community, since it tells us all we need to know about what to expect from the next generation of software updates that will come to our devices.

The online event is free to attend for all developers, and anyone who is curious can tune into Apple’s keynote. But those who want to attend the June 9 event at Apple Park in person need to apply online. In the days following, students and developers will attend sessions where they can engage more deeply with the changes to Apple’s platforms, which will allow them to ensure their apps are fully compatible with and that they take advantage of those changes.

«We’re excited to mark another incredible year of WWDC with our global developer community,» said Susan Prescott, Apple’s vice president of worldwide developer relations. «We can’t wait to share the latest tools and technologies that will empower developers and help them continue to innovate.»

What to Expect at WWDC 2025

If rumors are to be believed, Apple’s operating systems are set for a dramatic design overhaul this year. According to Bloomberg, iOS 19, iPadOS 19 and MacOS 16 will all feature new icons and menus, which will also change how our devices and controlled and navigated. The aim is reportedly to create a more cohesive style across platforms that will be reminiscent of VisionOS, the operating system designed for the company’s mixed-reality headset.

But that’s not all — Apple will likely use WWDC to outline the progress its made in the field of AI. «I’m looking forward to seeing more details about Apple Intelligence and the new features that will become available,» said Francisco Jeronimo, VP of devices at IDC. «This is critical for Apple, as Google Gemini and the Android vendors have taken significant steps forward with their AI offerings.»

Last year at WWDC 2024, Apple announced iOS 18 and Apple Intelligence — as well as updates to Siri that have yet to materialize.

Continue Reading

Trending

Copyright © Verum World Media