-
Notifications
You must be signed in to change notification settings - Fork 79
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[HF][fix] Allow kwargs
in ModelParser.run()
#882
Conversation
Lol literally a 1-line fix. I tested by removing the model name explicitly from HF model parsers to use the default, but this errored since we're setting it to "" instead of None. <img width="610" alt="Screenshot 2024-01-11 at 03 02 00" src="https://github.com/lastmile-ai/aiconfig/assets/151060367/1a045dbb-ad1c-46db-ba1c-32b05d923484"> ## Test Plan Before ``` model_name="" ``` I got an error saying that no model "" exists, forgot to take screenshot but it was real pls believe me After ``` model_name=None No model was supplied, defaulted to sshleifer/distilbart-cnn-12-6 and revision a4f8f3e (https://huggingface.co/sshleifer/distilbart-cnn-12-6). Using a pipeline without specifying a model name and revision in production is not recommended. ``` https://github.com/lastmile-ai/aiconfig/assets/151060367/db6b35d4-9e03-4ff5-b035-c77263978fd4
See #877 for context, and #880 for why that original fix needed to be reverted Just a quick hack to get unblocked. I tried originally to make the ASR and Image2Text ParameterizedModel but that caused other errors. ## Test Plan Before https://github.com/lastmile-ai/aiconfig/assets/151060367/a23a5d5c-d9a2-415b-8a6e-9826da56e985 After https://github.com/lastmile-ai/aiconfig/assets/151060367/f29580e9-5cf6-43c5-b848-bb525eb368f2
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We can fix on top but this is a critical API we are adding this to.
[ez][fix][HF]Explicitly set hf_model to None if it's empty string Lol literally a 1-line fix. I tested by removing the model name explicitly from HF model parsers to use the default, but this errored since we're setting it to "" instead of None. <img width="610" alt="Screenshot 2024-01-11 at 03 02 00" src="https://github.com/lastmile-ai/aiconfig/assets/151060367/1a045dbb-ad1c-46db-ba1c-32b05d923484"> ## Test Plan Before ``` model_name="" ``` I got an error saying that no model "" exists, forgot to take screenshot but it was real pls believe me After ``` model_name=None No model was supplied, defaulted to sshleifer/distilbart-cnn-12-6 and revision a4f8f3e (https://huggingface.co/sshleifer/distilbart-cnn-12-6). Using a pipeline without specifying a model name and revision in production is not recommended. ``` https://github.com/lastmile-ai/aiconfig/assets/151060367/db6b35d4-9e03-4ff5-b035-c77263978fd4 --- Stack created with [Sapling](https://sapling-scm.com). Best reviewed with [ReviewStack](https://reviewstack.dev/lastmile-ai/aiconfig/pull/879). * #885 * #883 * #882 * #881 * __->__ #879
The problem is this currently breaks though so we need this. This is btw already a problem that we've flagged and have Sudhanshu working on it in #664 (I need to reply back to him) |
#881) Revert "[HF][fix] Use ParameterizedModelParser instead of ModelParser" Reverts #877 This caused a bug because image and audio inputs can't be parametrized. See #882 for a better fix instead --- Stack created with [Sapling](https://sapling-scm.com). Best reviewed with [ReviewStack](https://reviewstack.dev/lastmile-ai/aiconfig/pull/881). * #885 * #883 * #882 * __->__ #881 * #879
I looked at the code and I understand what the issue has been -- until now all our model parsers have been ParameterizedModelParser instances -- that one accepts kwargs in |
…ms, get_output_text in right format (#883) [HF][fix] Small nit fixes to ASR: Remove "model" from completion params, get_output_text in right format Getting an error saying that "model" param is defined twice, which it is because we're passing it through completion params and again through the explicit model param. This along with the change in #879 will ensure that the right model of `None` is passed if the value is empty string. I also fixed typo in the aiconfig.json file ## Test Plan I also had to change the data field from `"./hi.mp3"` to `"cookbooks/Gradio/hi.mp3"` becuase I can only run the config from the high level `aiconfig` directory due to this line when running the editor: https://github.com/lastmile-ai/aiconfig/blob/b632094618c8f96b186b1849d0036e1313e02c5a/python/src/aiconfig/scripts/aiconfig_cli.py#L126 After https://github.com/lastmile-ai/aiconfig/assets/151060367/7387706f-deea-414b-b429-732c78322615 --- Stack created with [Sapling](https://sapling-scm.com). Best reviewed with [ReviewStack](https://reviewstack.dev/lastmile-ai/aiconfig/pull/883). * #885 * __->__ #883 * #882 * #881 * #879
…Text2Speech (#885) Use facebook/mms-tts-eng instead of suno/bark in Gradio cookbook for Text2Speech Bark is ~5GB and took my laptop 30mins to download for the first time. The Facebook MMS (https://huggingface.co/facebook/mms-tts-eng?fbclid=IwAR22U5VSaZ_V7QW4TJe77CM7f4fsTHjqCMjgLKV31wUl66ww1vXYSrEl-MA) is only ~100MB which is way better and at least for cookbook I feel much more worth it to quickly test things out. Yes the output quality is way more memey and "voice AI" stereotype, but I feel the tradeoff is worth it --- Stack created with [Sapling](https://sapling-scm.com). Best reviewed with [ReviewStack](https://reviewstack.dev/lastmile-ai/aiconfig/pull/885). * __->__ #885 * #883 * #882 * #881 * #879
[HF][fix] Allow
kwargs
inModelParser.run()
See #877 for context, and #880 for why that original fix needed to be reverted
Just a quick hack to get unblocked. I tried originally to make the ASR and Image2Text ParameterizedModel but that caused other errors.
Test Plan
Before
Screen.Recording.2024-01-11.at.02.00.14.mov
After
Screen.Recording.2024-01-11.at.05.01.47.mov
Stack created with Sapling. Best reviewed with ReviewStack.
kwargs
inModelParser.run()
#882