# Image Captioning AI Blueprint

# Inference

Image captioning uses neural networks to generate textual descriptions for images. It utilizes both computer vision and natural language processing (NLP) to generate these descriptions.

# Purpose

Use this inference blueprint to immediately generate descriptive captions for images. To use this pretrained caption-generator model, create a ready-to-use API-endpoint that can be quickly integrated with your data and application.

This blueprint deploys an OFA-pretrained image captioning model that returns a caption for an image provided. The user can provide single or multiple images and the blueprint returns captions for the images in the same order as the JSON response(s).

# Instructions


The minimum resource recommendations to run this blueprint are 3.5 CPU and 14.5 GB RAM or AWS xlarge memory on-demand instance.


This blueprint’s performance can benefit from using GPU as its compute.

Complete the following steps to deploy this caption-generator endpoint:

  1. Click the Use Blueprint button. Use Blueprint
  2. In the dialog, select the relevant compute to deploy the API endpoint and click the Start button.
  3. The cnvrg software redirects to your endpoint. Complete one or both of the following options:
    • Use the Try it Live section with any image to check your model. Try it Live
    • Use the bottom integration panel to integrate your API with your code by copying in the code snippet. Integration

An API endpoint that generates textual captions for images has now been deployed. For information on this blueprint's software version and release details, click here.

Refer to the following blueprint related to this inference blueprint:

Last Updated: 1/17/2023, 10:52:15 PM