Skip to content

Commit 415b8b2

Browse files
committed
code style
1 parent 2dbf413 commit 415b8b2

File tree

5 files changed

+11
-10
lines changed

5 files changed

+11
-10
lines changed

optimum/exporters/openvino/__main__.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -213,7 +213,7 @@ def main_export(
213213
subfolder=subfolder,
214214
revision=revision,
215215
cache_dir=cache_dir,
216-
token=token
216+
token=token,
217217
)
218218
if library_name == "sentence_transformers":
219219
logger.warning(

optimum/exporters/openvino/convert.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -631,7 +631,7 @@ def export_from_model(
631631
task = TasksManager.map_from_synonym(task)
632632
else:
633633
try:
634-
task = TasksManager._infer_task_from_model_or_model_class(model=model, library_name=library_name)
634+
task = TasksManager._infer_task_from_model_or_model_class(model=model)
635635
except (ValueError, KeyError) as e:
636636
raise RuntimeError(
637637
f"The model task could not be automatically inferred in `export_from_model`. Please provide the argument `task` with the relevant task from {', '.join(TasksManager.get_all_tasks())}. Detailed error: {e}"

optimum/exporters/openvino/model_patcher.py

+1-4
Original file line numberDiff line numberDiff line change
@@ -3944,10 +3944,7 @@ def __exit__(self, exc_type, exc_value, traceback):
39443944

39453945
class JanusVisionGenEmbeddingModelPatcher(ModelPatcher):
39463946
def __init__(
3947-
self,
3948-
config: "OnnxConfig",
3949-
model: Union["PreTrainedModel", "TFPreTrainedModel"],
3950-
model_kwargs: Dict[str, Any]
3947+
self, config: "OnnxConfig", model: Union["PreTrainedModel", "TFPreTrainedModel"], model_kwargs: Dict[str, Any]
39513948
):
39523949
model.__orig_forward = model.forward
39533950
model.forward = model.prepare_gen_img_embeds

optimum/intel/openvino/modeling_visual_language.py

+7-3
Original file line numberDiff line numberDiff line change
@@ -719,7 +719,7 @@ def _submodel_names(self):
719719
for part in self.additional_parts:
720720
if part == "lm_head" and getattr(self, part + "_model", None) is not None:
721721
model_names.append(part + "_model")
722-
continue
722+
continue
723723
if getattr(self, part, None) is not None:
724724
model_names.append(part + "_model")
725725
return model_names
@@ -2475,7 +2475,7 @@ def generate_image(
24752475
image_token_num_per_image: int = 576,
24762476
img_size: int = 384,
24772477
patch_size: int = 16,
2478-
generator=None
2478+
generator=None,
24792479
):
24802480
from PIL import Image
24812481

@@ -2524,7 +2524,11 @@ def generate_image(
25242524
logits = logit_uncond + cfg_weight * (logit_cond - logit_uncond)
25252525
probs = torch.softmax(logits / temperature, dim=-1)
25262526

2527-
next_token = torch.multinomial(probs, num_samples=1) if generator is None else torch.multinomial(probs, num_samples=1, generator=generator)
2527+
next_token = (
2528+
torch.multinomial(probs, num_samples=1)
2529+
if generator is None
2530+
else torch.multinomial(probs, num_samples=1, generator=generator)
2531+
)
25282532
generated_tokens[:, i] = next_token.squeeze(dim=-1)
25292533

25302534
next_token = torch.cat([next_token.unsqueeze(dim=1), next_token.unsqueeze(dim=1)], dim=1).view(-1)

tests/openvino/utils_tests.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -170,7 +170,7 @@
170170
"st-bert": "sentence-transformers/all-MiniLM-L6-v2",
171171
"st-mpnet": "sentence-transformers/all-mpnet-base-v2",
172172
"sana": "katuni4ka/tiny-random-sana",
173-
"janus": "katuni4ka/tiny-random-janus"
173+
"janus": "katuni4ka/tiny-random-janus",
174174
}
175175

176176

0 commit comments

Comments
 (0)