Sample applications

This page lists a set of Vision API samples. Samples are organized by language and mobile platform.

Product Search examples

Using Cloud Vision Product Search you can create a product set (catalog) with corresponding reference images of select product categories. You can then use the service to take a new image of a product and search for matching products in your product set. See the official documentation and tutorial for more information.

Language examples

Label tagging using Kubernetes

Awwvision is a Kubernetes and Cloud Vision API sample that uses the Vision API to classify (label) images from Reddit's /r/aww subreddit, and display the labeled results in a web application.

Documentation and Python code

Making text in images searchable

This sample uses TEXT_DETECTION Vision API requests to build an inverted index from the stemmed words found in the images, and stores that index in a Redis database. The example uses the nltk (Natural Language Toolkit) library for finding stopwords and doing stemming. The resulting index can be queried to find images that match a given set of words, and to list text that was found in each matching image.

Documentation and Python code

Documentation and Node.js code

Mobile platform examples

Vision and more with ML Kit for Firebase

These sample apps show how you can easily use the Cloud Vision label detection, landmark detection, and text recognition APIs from your mobile apps with ML Kit for Firebase. ML Kit also provides APIs to perform face detection, barcode scanning, inference using custom ML models, and more, all on the device, without requiring a network call.



Image detection using Android device photos

This simple single-activity sample shows you how to make a call to the Vision API with an image picked from your device's gallery.


Android code

Image detection using iOS device photos

The Swift and Objective-C versions of this app use the Vision API to run label and face detection on an image from the device's photo library. The resulting labels and face metadata from the API response are displayed in the UI.

Check out the Swift or Objective-C READMEs for specific getting started instructions.

Documentation (Objective-C)

Documentation (Swift)

iOS Sample code