This page lists a set of Vision API samples. Samples are organized by language and mobile platform.
Product Search examples
Search for matching products with Cloud Vision Product Search
Using Cloud Vision Product Search you can create a product set (catalog) with corresponding reference images of select product categories. You can then use the service to take a new image of a product and search for matching products in your product set. See the official documentation and tutorial for more information.
Label tagging using Kubernetes
Making text in images searchable
This sample uses
TEXT_DETECTION Vision API requests to build an
inverted index from the stemmed words found in the images, and stores that index
in a Redis database. The example uses the
nltk (Natural Language Toolkit) library for
finding stopwords and doing stemming. The resulting index can be queried to find
images that match a given set of words, and to list text that was found in each
Mobile platform examples
Vision and more with ML Kit for Firebase
These sample apps show how you can easily use the Cloud Vision label detection, landmark detection, and text recognition APIs from your mobile apps with ML Kit for Firebase. ML Kit also provides APIs to perform face detection, barcode scanning, inference using custom ML models, and more, all on the device, without requiring a network call.
Image detection using Android device photos
This simple single-activity sample shows you how to make a call to the Vision API with an image picked from your device's gallery.
Image detection using iOS device photos
The Swift and Objective-C versions of this app use the Vision API to run label and face detection on an image from the device's photo library. The resulting labels and face metadata from the API response are displayed in the UI.
Check out the Swift or Objective-C READMEs for specific getting started instructions.