You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Fix checkpoint compatibility for upsampling_layers (#314)
* Fix checkpoint compatibility for upsampling_layers
Older checkpoints may have 'upsampling_layers' parameters without the
'head.' prefix, causing warnings when loading models after head
refactoring. This fix remaps these keys during checkpoint loading to
ensure backwards compatibility.
Fixes warning: "Found keys that are not in the model state dict but in
the checkpoint: ['upsampling_layers.1.weight', 'upsampling_layers.1.bias', ...]"
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <[email protected]>
* Add fallback mechanism for checkpoint loading with weights_only=False
- Implement try/catch blocks around torch.load() calls in three files
- First attempts standard loading, falls back to weights_only=False on failure
- Provides clear warning messages when fallback is used
- Resolves pickle deserialization errors with older checkpoints
- Maintains security by attempting safer method first
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <[email protected]>
* add more transformer tests
---------
Co-authored-by: Muad Abd El Hay <[email protected]>
Co-authored-by: Claude <[email protected]>
Co-authored-by: Muad Abd El Hay <[email protected]>
Co-authored-by: themattinthehatt <[email protected]>
0 commit comments