mirror of
https://github.com/bigscience-workshop/petals
synced 2024-10-31 09:20:41 +00:00
Update src/petals/models/mixtral/model.py
Co-authored-by: Max Ryabinin <mryabinin0@gmail.com>
This commit is contained in:
parent
866927d88c
commit
5fe8fa4a45
@ -24,7 +24,7 @@ logger = get_logger(__name__)
|
||||
|
||||
|
||||
class DistributedMixtralModel(DefaultRevisionMixin, FromPretrainedMixin, PTuneMixin, MixtralModel):
|
||||
"""MixtalModel, but all transformer layers are hosted by the swarm"""
|
||||
"""MixtralModel, but all transformer layers are hosted by the swarm"""
|
||||
|
||||
_keys_to_ignore_on_load_missing = PTuneMixin._keys_to_ignore_on_load_missing
|
||||
_keys_to_ignore_on_load_unexpected = [r"^model\.layers\."]
|
||||
|
Loading…
Reference in New Issue
Block a user