@@ -220,7 +220,8 @@ EncodedImage VisionEncoderPhi3V::encode(const ov::Tensor& image, const ov::AnyMa
220
220
encoder.set_input_tensor (pixel_values);
221
221
ov::Tensor res{ov::element::f32, encoder.get_output_tensor ().get_shape ()};
222
222
encoder.set_output_tensor (res);
223
- encoder.infer ();
223
+ encoder.start_async ();
224
+ encoder.wait ();
224
225
return {std::move (res), image_size};
225
226
}
226
227
@@ -365,7 +366,8 @@ ov::Tensor reshape_hd_patches_2x2merge(const ov::Tensor& image_features, size_t
365
366
hd_feature_transformer.set_input_tensor (1 , height);
366
367
ov::Tensor width{ov::element::i32, {}, &w_crop};
367
368
hd_feature_transformer.set_input_tensor (2 , width);
368
- hd_feature_transformer.infer ();
369
+ hd_feature_transformer.start_async ();
370
+ hd_feature_transformer.wait ();
369
371
return hd_feature_transformer.get_output_tensor ();
370
372
}
371
373
@@ -432,7 +434,8 @@ ov::Tensor hd_feature_transform(const EncodedImage& image_features, InferRequest
432
434
ov::Tensor sub_image_features_hd_newline = add_image_newline (sub_image_features_hd, sub_GN); // [1,h_crop*12*(w_crop*12+1), 4096]
433
435
ov::Tensor image_embeddings = concatenate_2d (sub_image_features_hd_newline, glb_GN, global_image_features_hd_newline); // [1,l,4096]
434
436
vision_projection.set_input_tensor (image_embeddings);
435
- vision_projection.infer ();
437
+ vision_projection.start_async ();
438
+ vision_projection.wait ();
436
439
ov::Tensor out = vision_projection.get_output_tensor ();
437
440
ov::Tensor res{out.get_element_type (), out.get_shape ()};
438
441
out.copy_to (res);
0 commit comments