Replies: 4 comments 6 replies
-
I put the guidelines in this discussion item |
Beta Was this translation helpful? Give feedback.
-
Notice that we already support HF tokenizers in https://github.com/pytorch-labs/tokenizers natively. |
Beta Was this translation helpful? Give feedback.
-
Thanks for putting this together. From my experiences working with product teams, I think this will be a massive shift in usability for common GenAI/LLM use cases. On the documentation front, I'd like to put the HF integrate front and center. I'm thinking we can create/update our top-level LLM doc, put a callout in the getting started section, and otherwise try to funnel people to HF optimum-executorch as much as possible. I'll reach out later this week and we can put together a doc update PR. Thanks! |
Beta Was this translation helpful? Give feedback.
-
Hey, so I have been working on some similar stuff, mainly the issues with multi-method export w/ regards to encoder/decoder type models where you run into graph breaks. see #8030 I got distracted most of February, but so far, I have been able to export encoder/decoder seq2seq models from hugging face, with a paired down generation pipeline, the high-level goal being that for a given model class (seq2seq -encoder/decoder), users should just be able to provide the model/tokenizer link and everything will just export. Also, one should minimize the amount of C++ code written that duplicates python model processing code. AKA do not write logit processing in c++, the only thing in C++ should be the bare minimum to handle the graph breaks from data/shape-dependent control flow. I still need to get some changes merged in on the hugging face side:
I have fixes for both of these but need to sit down and get feedback from huggingface/ figure out how I should integrate these things in. Additionally, I honestly add more unit tests/ validate against more models, as my main development test case has been a Bart variant OPUS translation model. I would appreciate feedback/ how best to integrate this between both repos -> been mostly working on the Executorch side so far (wanted to validate that things worked, especially since you seem to be highly familiar with both repos. Ideally |
Beta Was this translation helpful? Give feedback.
-
Goals:
UX to Enable:
With native HF models and within HF ecosystem:
Hugging Face/ExecuTorch-Community
This continues the partnership from 2024. In 2025, we aim to strengthen it with a focus on ExecuTorch's general availability and improving UX in the open-source community.
Hugging Face Transformers
Detailed tasks will be filed and tracked in huggingface/transformers: huggingface/transformers#32253
Hugging Face Optimum-ExecuTorch
Detailed tasks will be filed and tracked in huggingface/optimum-executorch
ExecuTorch Recipe API
ExecuTorch will host a limited number of canonical recipes covering all major backends as the default, out-of-the-box solution for most use cases.
Each recipe will adhere to the following contract:
Details about this section will be posted separately by @tarun292
Documentation/Navigation
In ExecuTorch repo
In HF repo
This section will expand the work from @GregoryComer here #8178 and #8676 to provide a seamless navigation between Hugging Face and ExecuTorch.
Cross-repo Integration
Runtime
Benchmark & Deployment to Mobile
Beta Was this translation helpful? Give feedback.
All reactions