Apple Live Text takes on Google Lens, can read your photos
Live Text can pull information out of Apple Photos, much like Google Lens does.
Apple is pulling more information out of photos: its new Live Text feature coming to iOS 15, announced at its virtual WWDC developer conference, looks like a variation on Google's clever computer vision smarts baked into Lens.
The feature will recognize text in photos or through Apple's Camera app, and will recognize seven languages to start. The computer vision-based tool will be able to search for text found in photos, much like Google Lens already does.
While Live Text isn't quite augmented reality as defined by technologies like Apple's ARKit, Live Text could very well be a key tool for upcoming AR software (or hardware). The ability to recognize and pull information on the fly looks like a pretty necessary trick for glasses, although how it plays out in iOS 15 remains to be seen.