Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Deprecate OVTrainer #1167

Merged
merged 2 commits into from
Feb 19, 2025

Conversation

nikita-savelyevv
Copy link
Collaborator

What does this PR do?

Deprecate OVTrainer and related logic as stated in #890.

Before submitting

  • This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • Did you make sure to update the documentation with your changes?
  • Did you write any new necessary tests?

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

@echarlaix echarlaix requested a review from AlexKoff88 February 18, 2025 13:55
@echarlaix
Copy link
Collaborator

Works for me, @AlexKoff88 feel free to merge

@AlexKoff88
Copy link
Collaborator

@echarlaix, we decided to deprecate and remove it due to low interest. However, we are working on QAT for LLMs which could be contributed as an example by inheriting and implementing the HF Trainer interface instead of bringing it to the core functionality of Optimum-Intel.

@AlexKoff88
Copy link
Collaborator

@ljaljushkin, please take a look as well.

@echarlaix
Copy link
Collaborator

@echarlaix, we decided to deprecate and remove it due to low interest. However, we are working on QAT for LLMs which could be contributed as an example by inheriting and implementing the HF Trainer interface instead of bringing it to the core functionality of Optimum-Intel.

Why deprecating the OVTrainer in this context, what would be the benefit of having in the examples and not in optimum-intel directly as currently the case ?

@AlexKoff88
Copy link
Collaborator

@echarlaix, we decided to deprecate and remove it due to low interest. However, we are working on QAT for LLMs which could be contributed as an example by inheriting and implementing the HF Trainer interface instead of bringing it to the core functionality of Optimum-Intel.

Why deprecating the OVTrainer in this context, what would be the benefit of having in the examples and not in optimum-intel directly as currently the case ?

The main reason is to get flexibility in this flow:

  • In the example, we will show how to implement QAT Trainer for LLMs with NNCF so that it is PyTorch model in and out
  • User can modify the Trainer from the example according to the model training pipeline it has. NNCF part remains the same.
  • User decides how to use optimized model: go to OV IR, use torch.compile, executorch, export to ONNX for ORT OV EP, etc.

@AlexKoff88 AlexKoff88 merged commit c8c6beb into huggingface:main Feb 19, 2025
16 of 20 checks passed
AlexanderDokuchaev added a commit to openvinotoolkit/nncf that referenced this pull request Feb 20, 2025
shumaari pushed a commit to shumaari/nncf that referenced this pull request Mar 8, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants