You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
use_cache is something that you can specify at model load time as a kwarg, or at generation time as a kwarg. We can probably make some adjustments to make this more automatic. Thanks!
Environment
llmfoundry:latest
To reproduce
Steps to reproduce the behavior:
llm-foundry/scripts/inference/convert_composer_to_hf.py
-> When generating, the model throws an exception that
use_cache
must be enabled in the HF config.Expected behavior
Model generates output texts.
Manually editing the HF config and enabling the cache did the trick for me.
The text was updated successfully, but these errors were encountered: