Skip to main content

Apple may introduce augmented reality functionality into the iPhone’s camera

Report: Apple has up to 1,000 engineers working on AR product for the iPhone

apple encryption court order news logo
Image used with permission by copyright holder
Google isn’t the only company looking to cash in on the augmented reality craze. Business Insider reports that Apple plans to integrate augmented reality tech directly into the iPhone’s camera app.

The new feature would reportedly take the form of computer vision. Apple’s work-in-progress camera app will be able to identify objects in-frame and recognize faces, according to the report. And it will be made available to developers in the form of a software development kit.

The most recent news we’ve heard about Apple’s augmented reality plans suggest that Apple may have as many as 1,000 engineers working on an AR-related product in Israel — a product that will make its way to the next iPhone, at least according to analyst Steven Milunovich and his team at UBS. The report, from Business Insider, also highlights that Apple has recently made a number of AR-related purchases, including PrimeSense, a 3D sensing company in Tel Aviv, and RealFace, a facial recognition company also based in Tel Aviv.

The Cupertino, California-based company is far from the first to enter the AR arena. Google Goggles, Google’s eponymous object recognition app, recognizes landmarks, barcodes, books, and works of art, and parses text of labels and signage using optical character recognition. And Amazon’s Flow app can decode QR codes, scan business cards, and recognize tens of millions of books, DVDs, and packaged products.

Apple’s system sounds nearly as ambitious: an app that can identify objects that users point the iPhone’s camera at in real time. It will rely on machine learning, a type of artificial intelligence that “learns” and improves over time, and a database of 3D objects that Apple will either license or build itself.

Beyond those basics, the project’s implications aren’t clear. Google’s Project Tango, an augmented reality platform, leverages sensors to measure the depth of surroundings, but the iPhone lacks the hardware necessary to perform that sort of tracking. Apple Insider speculates that Apple’s brand of machine-learning-powered object tracking could be used for spatial mapping, and that facial recognition, meanwhile, could be used to apply Snapchat-style filters to people’s faces.

Spearheading the project is a team comprised of employees from recent Apple acquisitions. The iPhone maker purchased PrimeSense, the Israeli company behind the motion-tracking hardware in Microsoft’s Kinect sensor, in 2013. It bought Metaio in February 2014, FaceShift in 2015, and Flyby Media in January 2016, all of which specialize in virtual reality and AR technologies. And it hired a senior optics manufacturing engineer specializing in heads-up displays, camera systems, and image sensors in September.

The project builds on another reportedly in development at Apple’s Cupertino headquarters: AR glasses. According to Business Insider, Apple is developing a compact pair of AR eyewear that connect wirelessly with an iPhone and show images and other information in the wearer’s field of vision.

If anything is for certain, it’s that an updated camera app will debut far ahead of a headset. Bloomberg reports that Apple has begun talking with suppliers about the glasses and has ordered “small quantities” of displays for testing. The publication pegs 2018 as the product’s earliest possible release window.

Until then, Snapchat’s Spectacles will have to do.

Updated on 03-02-2017 by Christian de Looper: Added new Business Insider report saying that Apple had as many as 1,000 engineers working on AR.

Kyle Wiggers
Former Digital Trends Contributor
Kyle Wiggers is a writer, Web designer, and podcaster with an acute interest in all things tech. When not reviewing gadgets…
iOS 18 makes an 11-year-old iPhone feature exciting again
Someone holding an iPhone 14, showing the Lock Screen.

Following the Worldwide Developers Conference (WWDC 2024) keynote, developers are starting to dig into the first iOS 18 developer beta. Though this beta lacks Apple Intelligence and many of the other features demoed on Monday, it offers a surprising new take on an old iOS feature: the flashlight.

The built-in flashlight feature has been available on the iPhone since iOS 7, which was released in 2013. It hasn't changed much at all since then, which makes sense, given its basic function. Interestingly, it has received a significant update in iOS 18.

Read more
iOS 18 is official. Here’s how it’s going to change your iPhone forever
Screenshots of new features in iOS 18.

It’s been a long time coming, but it’s finally here: iOS 18 has just been announced at Apple’s Worldwide Developers Conference (WWDC) 2024 keynote. And, like the rumors have been saying, this is a very big, juicy update for your iPhone.

There's been a lot of anticipation for iOS 18. Rumors and leaks for the new update have been particularly intense this year, more so than iOS 17 rumors last year. So, was the hype worth it? Here's what's coming to your iPhone with iOS 18.
Home screen changes

Read more
iOS 18 may have a fun surprise for longtime iPhone fans
Apple iPhone 6S

If you’ve been a longtime iPhone user, you’ve probably seen every default wallpaper iOS has to offer — and you’ve likely had a few favorites that you miss. The good news is that, according to the latest news about Worldwide Developers Conference (WWDC) 2024, Apple is set to bring back many old-school wallpapers dating back to the early days of the Mac and iPhone.

“Apple’s devices are getting new wallpaper packs, including Mac versions that reference old-school icons and slogans,” said Bloomberg’s Mark Gurman in a report. “The iPhone wallpapers will have options that look similar to early ones on the phone.”

Read more