@@ -104,46 +104,40 @@ void init_vlm_pipeline(py::module_& m) {
104
104
.def (py::init ([](
105
105
const std::string& model_path,
106
106
const std::string& device,
107
- const std::map<std::string, py::object>& config,
108
- const ov::Core& core
107
+ const std::map<std::string, py::object>& config
109
108
) {
110
109
ScopedVar env_manager (utils::ov_tokenizers_module_path ());
111
- return std::make_unique<ov::genai::VLMPipeline>(model_path, device, utils::properties_to_any_map (config), core );
110
+ return std::make_unique<ov::genai::VLMPipeline>(model_path, device, utils::properties_to_any_map (config));
112
111
}),
113
112
py::arg (" model_path" ), " folder with openvino_model.xml and openvino_tokenizer[detokenizer].xml files" ,
114
113
py::arg (" device" ) = " CPU" , " device on which inference will be done" ,
115
- py::arg (" config" ) = ov::AnyMap ({}), " openvino.properties map" ,
116
- py::arg (" core" ) = ov::Core (), " openvino.Core object" ,
114
+ py::arg (" config" ) = ov::AnyMap ({}), " openvino.properties map"
117
115
R"(
118
116
VLMPipeline class constructor.
119
117
model_path (str): Path to the model file.
120
118
device (str): Device to run the model on (e.g., CPU, GPU). Default is 'CPU'.
121
119
Add {"scheduler_config": ov_genai.SchedulerConfig} to config properties to create continuous batching pipeline.
122
- core (openvino.Core): openvino.Core instance.
123
120
)" )
124
121
125
122
.def (py::init ([](
126
123
const std::string& model_path,
127
124
const ov::genai::Tokenizer& tokenizer,
128
125
const std::string& device,
129
- const std::map<std::string, py::object>& config,
130
- const ov::Core& core
126
+ const std::map<std::string, py::object>& config
131
127
) {
132
128
ScopedVar env_manager (utils::ov_tokenizers_module_path ());
133
- return std::make_unique<ov::genai::VLMPipeline>(model_path, tokenizer, device, utils::properties_to_any_map (config), core );
129
+ return std::make_unique<ov::genai::VLMPipeline>(model_path, tokenizer, device, utils::properties_to_any_map (config));
134
130
}),
135
131
py::arg (" model_path" ),
136
132
py::arg (" tokenizer" ),
137
133
py::arg (" device" ) = " CPU" ,
138
- py::arg (" config" ) = ov::AnyMap ({}), " openvino.properties map" ,
139
- py::arg (" core" ) = ov::Core (), " openvino.Core object" ,
134
+ py::arg (" config" ) = ov::AnyMap ({}), " openvino.properties map"
140
135
R"(
141
136
VLMPipeline class constructor for manualy created openvino_genai.Tokenizer.
142
137
model_path (str): Path to the model file.
143
138
tokenizer (openvino_genai.Tokenizer): tokenizer object.
144
139
device (str): Device to run the model on (e.g., CPU, GPU). Default is 'CPU'.
145
140
Add {"scheduler_config": ov_genai.SchedulerConfig} to config properties to create continuous batching pipeline.
146
- core (openvino.Core): openvino.Core instance.
147
141
)" )
148
142
.def (" start_chat" , &ov::genai::VLMPipeline::start_chat)
149
143
.def (" finish_chat" , &ov::genai::VLMPipeline::finish_chat)
0 commit comments