Update config.json
#13
by
AntonV
HF Staff
- opened
See https://github.com/huggingface/transformers/blob/0fe03afeb82e1a435a75704d1f434c47e49a0bbb/tests/models/ernie4_5_moe/test_modeling_ernie4_5_moe.py#L172 for the correct class, sorry about that
Didn't use the auto class in the test
Such as JetMoeModel, Qwen2MoeModel, we consider that 'Moe' is more suitable.
Fair, let me change this on the transformers side then. I'll close the PRs on the other repos :D
AntonV
changed pull request status to
closed