Skip to main content

Autonomous wheelchair could change the lives of motor-impaired users

autonomous wheelchair tech dsc00148
Image used with permission by copyright holder
There are people developing autonomous cars, boats, planes and, heck, even garbage-munching aquatic drones — so why not wheelchairs, too?

That’s the mission of Brenna Argall, an assistant professor of Rehabilitation Robotics at Northwestern University and research scientist at the Rehabilitation Institute of Chicago. Argall and her team are developing autonomous wheelchairs designed for people with severe disabilities, such as significant motor impairments or the physical inability to operate a traditional control mechanism like a joystick.

“By making assistive machine easier to control, because they can autonomously control themselves, our goal is to make people with severe motor impairments more able and independent,” Argall told Digital Trends.

Argall first became interested in robotics as a math undergraduate at Carnegie Mellon University. She later went on to receive a Ph.D. in the subject, but it was only when she became faculty at Northwestern five years ago, that she decided to combine it with an interest in medicine and healthcare.

screen-shot-2016-09-15-at-06-52-21
Image used with permission by copyright holder

“It took a bit of time to figure out my lab’s positioning,” she said. “Robotics autonomy doesn’t have much representation within the field of rehabilitation robotics, but there clearly is a lot of low-hanging fruit. The key is to be judicious in selecting which problem to tackle, so that it not only is feasible technically but also is addressing user need in a way that is likely to be adopted by the user. And that is where all the current projects in — and core research direction of — my lab come from: leveraging robotics autonomy in order to enable human autonomy.”

Unlike a self-driving car, which will be used in broadly the same by all of its users, Argall’s work is challenging because every user will have different requirements. “Each person is unique in their personal preferences and abilities, and we need to be responsive to this,” she said. As a result, her team is working with modular software and hardware that, in the former case, lets users opt in and out of various autonomous assistance modules, and in the latter, add additional sensors for their wheelchair depending on requirements.

Argall also said that finances will vary from person to person — particularly since autonomy add-ons “will not be covered by insurance anytime soon.” As a result, the team prioritizes cheap, off-the-shelf components, with the goal of making the necessary modifications to existing wheelchairs (rather than manufacturing entirely new ones) for the cost of a laptop computer.

There’s still a way to go, but Argall said, “the basic functionality and technology are there.” So how far are we from an autonomous wheelchair that could carry out the kind of self-driving travel and obstacle avoidance users would need? “My estimate is five years,” she concluded. “Perhaps faster, for a version … that provides just basic obstacle avoidance, or if we decided to go full steam ahead on commercialization, which we at the moment are not.”

As cool as self-driving cars undoubtedly are, it’s hard to think of too many more transformative uses of AI and smart sensors than projects like this one.

Luke Dormehl
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Digital Trends’ Top Tech of CES 2023 Awards
Best of CES 2023 Awards Our Top Tech from the Show Feature

Let there be no doubt: CES isn’t just alive in 2023; it’s thriving. Take one glance at the taxi gridlock outside the Las Vegas Convention Center and it’s evident that two quiet COVID years didn’t kill the world’s desire for an overcrowded in-person tech extravaganza -- they just built up a ravenous demand.

From VR to AI, eVTOLs and QD-OLED, the acronyms were flying and fresh technologies populated every corner of the show floor, and even the parking lot. So naturally, we poked, prodded, and tried on everything we could. They weren’t all revolutionary. But they didn’t have to be. We’ve watched enough waves of “game-changing” technologies that never quite arrive to know that sometimes it’s the little tweaks that really count.

Read more
Digital Trends’ Tech For Change CES 2023 Awards
Digital Trends CES 2023 Tech For Change Award Winners Feature

CES is more than just a neon-drenched show-and-tell session for the world’s biggest tech manufacturers. More and more, it’s also a place where companies showcase innovations that could truly make the world a better place — and at CES 2023, this type of tech was on full display. We saw everything from accessibility-minded PS5 controllers to pedal-powered smart desks. But of all the amazing innovations on display this year, these three impressed us the most:

Samsung's Relumino Mode
Across the globe, roughly 300 million people suffer from moderate to severe vision loss, and generally speaking, most TVs don’t take that into account. So in an effort to make television more accessible and enjoyable for those millions of people suffering from impaired vision, Samsung is adding a new picture mode to many of its new TVs.
[CES 2023] Relumino Mode: Innovation for every need | Samsung
Relumino Mode, as it’s called, works by adding a bunch of different visual filters to the picture simultaneously. Outlines of people and objects on screen are highlighted, the contrast and brightness of the overall picture are cranked up, and extra sharpness is applied to everything. The resulting video would likely look strange to people with normal vision, but for folks with low vision, it should look clearer and closer to "normal" than it otherwise would.
Excitingly, since Relumino Mode is ultimately just a clever software trick, this technology could theoretically be pushed out via a software update and installed on millions of existing Samsung TVs -- not just new and recently purchased ones.

Read more
AI turned Breaking Bad into an anime — and it’s terrifying
Split image of Breaking Bad anime characters.

These days, it seems like there's nothing AI programs can't do. Thanks to advancements in artificial intelligence, deepfakes have done digital "face-offs" with Hollywood celebrities in films and TV shows, VFX artists can de-age actors almost instantly, and ChatGPT has learned how to write big-budget screenplays in the blink of an eye. Pretty soon, AI will probably decide who wins at the Oscars.

Within the past year, AI has also been used to generate beautiful works of art in seconds, creating a viral new trend and causing a boon for fan artists everywhere. TikTok user @cyborgism recently broke the internet by posting a clip featuring many AI-generated pictures of Breaking Bad. The theme here is that the characters are depicted as anime characters straight out of the 1980s, and the result is concerning to say the least. Depending on your viewpoint, Breaking Bad AI (my unofficial name for it) shows how technology can either threaten the integrity of original works of art or nurture artistic expression.
What if AI created Breaking Bad as a 1980s anime?
Playing over Metro Boomin's rap remix of the famous "I am the one who knocks" monologue, the video features images of the cast that range from shockingly realistic to full-on exaggerated. The clip currently has over 65,000 likes on TikTok alone, and many other users have shared their thoughts on the art. One user wrote, "Regardless of the repercussions on the entertainment industry, I can't wait for AI to be advanced enough to animate the whole show like this."

Read more