Skip to content

Latest commit

 

History

History
41 lines (27 loc) · 1.54 KB

File metadata and controls

41 lines (27 loc) · 1.54 KB

Visual Question Answering and Image Captioning using BLIP and OpenVINO

BLIP is a pre-training framework for unified vision-language understanding and generation, which achieves state-of-the-art results on a wide range of vision-language tasks. This tutorial considers ways to use BLIP for visual question answering and image captioning.

The complete pipeline of this demo is shown below:

Image Captioning

The following image shows an example of the input image and generated caption:

Visual Question Answering

The following image shows an example of the input image, question and answer generated by model

Notebook Contents

The tutorial consists of the following parts:

  1. Instantiate a BLIP model.
  2. Convert the BLIP model to OpenVINO IR.
  3. Run visual question answering and image captioning with OpenVINO.

Installation Instructions

If you have not installed all required dependencies, follow the Installation Guide.