Apple has announced that visual search will be coming to the iPhone, thanks to its suite of AI capabilities known as Apple Intelligence. This new feature, which can be accessed through the Camera Control button on the iPhone 16 and 16 Plus models, will allow users to launch what Apple calls "visual intelligence" – a reverse image search combined with some text recognition.
What is Visual Intelligence?
Visual intelligence is essentially a powerful tool that uses artificial intelligence (AI) to identify objects, scenes, and even text within an image. When you point the camera at something, visual intelligence will analyze the image and provide relevant information about it. For example, if you use visual intelligence to search for a restaurant, it’ll pull up details such as:
- Restaurant hours
- Ratings
- Options to check out the menu or make a reservation
Or, if you come across a flier for an event, you can use visual intelligence to quickly add the title, time, date, and location to your calendar.
How Does Visual Intelligence Work?
Visual intelligence works by using machine learning algorithms to analyze images. These algorithms are trained on vast amounts of data, allowing them to recognize patterns and objects within images. When you use visual intelligence, it will source information from Google Search, ensuring that the results are accurate and up-to-date.
But What About My Privacy?
Apple has assured users that their services will never store your images. This means that when you use visual intelligence, your photos will not be stored on Apple’s servers or used for any other purpose than to provide you with relevant information.
A Partnership with OpenAI
In addition to using Google Search as a source of information, Apple has also partnered with OpenAI to integrate ChatGPT into the Camera Control button. This means that users will have the option to send their query off to ChatGPT instead of Google Search.
For example, if you’re stuck on a homework assignment and want some help, you can use visual intelligence to search for answers or examples related to the topic. You’ll be able to receive detailed responses from ChatGPT, helping you complete your task.
Availability and Rollout
Visual intelligence will launch along with other Apple Intelligence features in beta in October for U.S. English language users. Users in other countries can expect to get it in December and early 2025. This means that you’ll have plenty of time to try out visual intelligence and give feedback to help improve the feature.
Conclusion
Visual search is coming to the iPhone, powered by Apple Intelligence. With its powerful capabilities and seamless integration with Google Search and ChatGPT, this new feature has the potential to revolutionize how we use our smartphones. Whether you’re trying to find information about a restaurant or need help with your homework, visual intelligence will be there to assist you.