You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
ValueError: Conversion of a LlamaForCausalLM does not support these configuration attributes: {'use_cache': False, '_name_or_path': 'Unbabel/TowerInstruct-7B-v0.2'}
I use transformers.__version__ == '4.51.2' and penzai.__version__ == '0.2.5'. While use_cache is documented here, the _name_or_path does not seem to be documented.
The text was updated successfully, but these errors were encountered:
There's a check for unrecognized configuration arguments in llama_from_huggingface_model because it is otherwise pretty difficult to make sure that the converted model has the same behavior as the original one. But this might be a false positive:
use_cache seems like it controls whether the HF model returns cached keys and values, but this isn't relevant for the Penzai model since Penzai handles the KV cache differently.
_name_or_path seems like it's probably metadata that doesn't get used when the model runs.
If you're feeling adventurous you could try adding these attributes to
and see if the resulting Penzai model produces the same outputs as the original hf_model. If so I'd be happy to include that change in the next Penzai release.
When I run
the second line fails with
I use
transformers.__version__ == '4.51.2'
andpenzai.__version__ == '0.2.5'
. Whileuse_cache
is documented here, the_name_or_path
does not seem to be documented.The text was updated successfully, but these errors were encountered: