Google Lens is an application announced by Google during Google I/O 2017, designed to bring up relevant information using visual analysis.


When directing the phone's camera at an object, Google Lens will attempt to identify the object and show relevant search results and information. For example, when pointing the device's camera at a Wi-Fi label containing the network name and password, it will automatically connect to the Wi-Fi source that has been scanned. The lens is also integrated with Google Photos and Google Assistant apps. The service is similar to Google Goggles, a previous app that functioned similarly but with lesser capability. Lens uses more advanced deep learning routines, similar to other apps like Bixby Vision (integrated with the newer Samsung smartphones) and Image Analysis Toolset (available on Google Play); artificial neural networks are used to detect and identify objects, landmarks and to improve optical character recognition (OCR) accuracy.


Google officially launched Google Lens on October 4, with app previews pre-installed into the Google Pixel 2. In November 2017, the feature began rolling out into the Google Assistant for Pixel and Pixel 2 phones. A preview of Lens has also been implemented into the Google Photos app for Pixel phones. On March 5, 2018, Google officially released Google Lens to Google Photos on non-Pixel phones. Support for Lens in the iOS version of Google Photos was made on March 15, 2018. Beginning in May 2018, Google Lens was made available within Google Assistant on OnePlus devices, as well as being integrated into camera apps of various Android phones. A standalone Google Lens app was made available on Google Play in June 2018. Device support is limited, although it is not clear which devices are not supported or why. It requires Android Marshmallow or newer.

Community content is available under CC-BY-SA unless otherwise noted.