You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* Add AIM-v2 encoders from https://github.com/apple/ml-aim, see on Hub: https://huggingface.co/models?search=timm%20aimv2
18
+
* Add PaliGemma2 encoders from https://github.com/google-research/big_vision to existing PaliGemma, see on Hub: https://huggingface.co/models?search=timm%20pali2
* Fix existing RmsProp layer to match standard formulation, use PT 2.5 impl when possible. Move old impl to `SimpleNorm` layer, it's LN w/o centering or bias. There were only two `timm` models using it, and they have been updated.
21
+
* Allow override of `cache_dir` arg for model creation
22
+
* Pass through `trust_remote_code` for HF datasets wrapper
23
+
*`inception_next_atto` model added by creator
24
+
* Adan optimizer caution, and Lamb decoupled weighgt decay options
25
+
* Some feature_info metadata fixed by https://github.com/brianhou0208
26
+
* All OpenCLIP and JAX (CLIP, SigLIP, Pali, etc) model weights that used load time remapping were given their own HF Hub instances so that they work with `hf-hub:` based loading, and thus will work with new Transformers `TimmWrapperModel`
27
+
15
28
## Nov 28, 2024
16
29
* More optimizers
17
30
* Add MARS optimizer (https://arxiv.org/abs/2411.10438, https://github.com/AGI-Arena/MARS)
@@ -248,7 +261,7 @@ Add a set of new very well trained ResNet & ResNet-V2 18/34 (basic block) weight
248
261
### April 11, 2024
249
262
* Prepping for a long overdue 1.0 release, things have been stable for a while now.
250
263
* Significant feature that's been missing for a while, `features_only=True` support for ViT models with flat hidden states or non-std module layouts (so far covering `'vit_*', 'twins_*', 'deit*', 'beit*', 'mvitv2*', 'eva*', 'samvit_*', 'flexivit*'`)
251
-
* Above feature support achieved through a new `forward_intermediates()` API that can be used with a feature wrapping module or direclty.
264
+
* Above feature support achieved through a new `forward_intermediates()` API that can be used with a feature wrapping module or directly.
@@ -486,7 +499,7 @@ Included optimizers available via `timm.optim.create_optimizer_v2` factory metho
486
499
*`madgrad` an implementation of MADGRAD adapted from https://github.com/facebookresearch/madgrad - https://arxiv.org/abs/2101.11075
487
500
*`mars` MARS optimizer from https://github.com/AGI-Arena/MARS - https://arxiv.org/abs/2411.10438
488
501
*`nadam` an implementation of Adam w/ Nesterov momentum
489
-
*`nadamw` an impementation of AdamW (Adam w/ decoupled weight-decay) w/ Nesterov momentum. A simplified impl based on https://github.com/mlcommons/algorithmic-efficiency
502
+
*`nadamw` an implementation of AdamW (Adam w/ decoupled weight-decay) w/ Nesterov momentum. A simplified impl based on https://github.com/mlcommons/algorithmic-efficiency
490
503
*`novograd` by [Masashi Kimura](https://github.com/convergence-lab/novograd) - https://arxiv.org/abs/1905.11286
491
504
*`radam` by [Liyuan Liu](https://github.com/LiyuanLucasLiu/RAdam) - https://arxiv.org/abs/1908.03265
492
505
*`rmsprop_tf` adapted from PyTorch RMSProp by myself. Reproduces much improved Tensorflow RMSProp behaviour
Copy file name to clipboardExpand all lines: UPGRADING.md
+3-3
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,10 @@
1
1
# Upgrading from previous versions
2
2
3
-
I generally try to maintain code interface and especially model weight compability across many `timm` versions. Sometimes there are exceptions.
3
+
I generally try to maintain code interface and especially model weight compatibility across many `timm` versions. Sometimes there are exceptions.
4
4
5
5
## Checkpoint remapping
6
6
7
-
Pretrained weight remapping is handled by `checkpoint_filter_fn` in a model implementation module. This remaps old pretrained checkpoints to new, and also 3rd party (original) checkpoints to `timm` format if the model was modified when brough into `timm`.
7
+
Pretrained weight remapping is handled by `checkpoint_filter_fn` in a model implementation module. This remaps old pretrained checkpoints to new, and also 3rd party (original) checkpoints to `timm` format if the model was modified when brought into `timm`.
8
8
9
9
The `checkpoint_filter_fn` is automatically called when loading pretrained weights via `pretrained=True`, but they can be called manually if you call the fn directly with the current model instance and old state dict.
10
10
@@ -19,6 +19,6 @@ Many changes were made since the 0.6.x stable releases. They were previewed in 0
19
19
* The pretrained_tag is the specific weight variant (different head) for the architecture.
20
20
* Using only `architecture` defaults to the first weights in the default_cfgs for that model architecture.
21
21
* In adding pretrained tags, many model names that existed to differentiate were renamed to use the tag (ex: `vit_base_patch16_224_in21k` -> `vit_base_patch16_224.augreg_in21k`). There are deprecation mappings for these.
22
-
* A number of models had their checkpoints remaped to match architecture changes needed to better support `features_only=True`, there are `checkpoint_filter_fn` methods in any model module that was remapped. These can be passed to `timm.models.load_checkpoint(..., filter_fn=timm.models.swin_transformer_v2.checkpoint_filter_fn)` to remap your existing checkpoint.
22
+
* A number of models had their checkpoints remapped to match architecture changes needed to better support `features_only=True`, there are `checkpoint_filter_fn` methods in any model module that was remapped. These can be passed to `timm.models.load_checkpoint(..., filter_fn=timm.models.swin_transformer_v2.checkpoint_filter_fn)` to remap your existing checkpoint.
23
23
* The Hugging Face Hub (https://huggingface.co/timm) is now the primary source for `timm` weights. Model cards include link to papers, original source, license.
24
24
* Previous 0.6.x can be cloned from [0.6.x](https://github.com/rwightman/pytorch-image-models/tree/0.6.x) branch or installed via pip with version.
0 commit comments