Releases: keras-team/keras-hub
Releases · keras-team/keras-hub
v0.21.1
Summary:
- Comprehensive docstrings to QwencausalLM, resolve integration test issues for Keras-IO, and coverage tracking for Keras-Hub.
What's Changed
- Add QwencausalLM docstrings, coverage tracking, keras-io integration fix by @laxmareddyp in #2284
- Version bump to 0.21.1 by @laxmareddyp in #2285
Full Changelog: v0.21.0...v0.21.1
v0.21.0
Summary
-
New Models.
- Xception: Added Xception architecture for image classification tasks.
- Qwen: Added Qwen2.5 large language models and presets of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters.
- Qwen MoE: Added transformer-based Mixture of Experts (MoE) decoder-only language model with a base variant having 2.7B activated parameters during runtime.
- Mixtral: Added Mixtral LLM, a pretrained generative Sparse Mixture of Experts with pre-trained and instruction tuned models having 7 billion activated parameters.
- Moonshine: Added Moonshine, a speech recognition task model.
- CSPNet: Added Cross Stage Partial Network (CSPNet) classification task model.
- Llama3: Added support for Llama 3.1 and 3.2.
-
Added sharded weight support to KerasPresetSaver and KerasPresetLoader, defaulting to a 10GB maximum shard size.
What's Changed
- Fix Roformer export symbol by @abheesht17 in #2199
- Bump up master version to 0.21 by @abheesht17 in #2204
- reenable test by @mattdangerw in #2188
- Add xception model by @mattdangerw in #2179
- Make image converter built by @mattdangerw in #2206
- Qwen - Fix Preset Loader + Add Causal LM Test by @kanpuriyanawab in #2193
- Update Qwen conversion script by @laxmareddyp in #2207
- Revert "Do not export Qwen for release" by @sachinprasadhs in #2208
- Fixes compute_output_shape for PaliGemmaVitEncoder and Gemma3VisionEncoderBlock by @JyotinderSingh in #2210
- Python 3.12 fix by @mattdangerw in #2211
- Small Gemma3 doc-string edits by @abheesht17 in #2214
- Llama3.1 by @pctablet505 in #2132
- Update gemma3_causal_lm_preprocessor.py by @pctablet505 in #2217
- fix: apply
weights_only = True
by @b8zhong in #2215 - Fix the keras_hub package for typecheckers and IDEs by @mattdangerw in #2222
- Add utility to map COCO IDs to class names by @mattdangerw in #2219
- Set GPU timeouts to 2 hours by @mattdangerw in #2226
- Fix nightly by @mattdangerw in #2227
- Another fix for nightly builds by @mattdangerw in #2229
- Cast a few more input to tensors in SD3 by @mattdangerw in #2234
- Fix up package build scripts again by @mattdangerw in #2230
- Add qwen presets by @laxmareddyp in #2241
- script for converting retinanet weights from trochvision by @sineeli in #2233
- Sharded weights support by @james77777778 in #2218
- Add Qwen Moe by @kanpuriyanawab in #2163
- Add Mixtral by @kanpuriyanawab in #2196
- Made label data optional for inference and adopted other required changes by @laxmareddyp in #2183
- Fix the layer names by @kanpuriyanawab in #2247
- Add new CSPNet preset and add manual padding. by @sachinprasadhs in #2212
- Update the int8 quant logic in
ReversibleEmbedding
by @james77777778 in #2250 - Add Moonshine to KerasHub by @harshaljanjani in #2093
- Add Kaggle handle for moonshine presets by @laxmareddyp in #2253
- Update requirements-jax-cuda.txt by @pctablet505 in #2252
- Add Mixtral,Qwen-MoE presets and Update conversion script. by @laxmareddyp in #2248
- fix flash attention test by @divyashreepathihalli in #2263
- Fix JAX bugs for qwen moe & mixtral by @kanpuriyanawab in #2258
- Create pull_request_template.md by @sachinprasadhs in #2262
- Update preset versions for sharded models by @laxmareddyp in #2264
- Add AudioToText and AudioToTextPreprocessor class stubs to enable auto class functionality by @harshaljanjani in #2265
- register moonshine presets by @sachinprasadhs in #2267
- Version bump 0.21.0.dev1 by @laxmareddyp in #2273
- Version bump to 0.21.0 by @laxmareddyp in #2275
New Contributors
- @JyotinderSingh made their first contribution in #2210
- @pctablet505 made their first contribution in #2132
- @b8zhong made their first contribution in #2215
Full Changelog: v0.20.0...v0.21.0
v0.20.0
What's Changed
- Install TF Text on non-Windows only by @abheesht17 in #2115
- Add SigLIP by @james77777778 in #2113
- Fix
PaliGemmaVitEncoder
output shape by @abheesht17 in #2123 - Cspnet architecture. by @sachinprasadhs in #2091
- Update our master version to be a dev release by @mattdangerw in #2131
- Add top 3 HF Presets for Mobilenet by @pkgoogle in #2105
- Add SigLIP2 by @james77777778 in #2127
- update Gemma attention for TPU by @divyashreepathihalli in #2130
- Update dev version rule for nightly by @SamanehSaadat in #2139
- Fix dtype bug in image converter by @abheesht17 in #2147
- Add instruction in .md for manual pre-commit run by @abheesht17 in #2148
- Add Qwen 2.5 by @shivance in #2088
- Updated CONTRIBUTING.md (Fixes issue #2153) by @villurignanesh in #2156
- Update kaggle preset paths for SigLip model by @laxmareddyp in #2164
- Routine Kaggle HF sync by @divyashreepathihalli in #2165
- Enable LoRA target names arg by @divyashreepathihalli in #2166
- Update retinanet_presets.py by @sineeli in #2157
- Add Gemma3 by @abheesht17 in #2152
- Add precommit to the common requirements file by @mattdangerw in #2173
- Add back a format script for compat by @mattdangerw in #2174
- Add a TextToImagePreprocessor base class by @mattdangerw in #2181
- Bump the python group with 2 updates by @dependabot in #2185
- implement of roformerv2 by @pass-lin in #2145
- Move sliding window attn before FA block for Gemma by @abheesht17 in #2187
- Update gating condition to include check for supporting GPUs for flash attention by @divyashreepathihalli in #2184
- Revert "Fix dtype bug in image converter (#2147)" by @mattdangerw in #2180
- Add vision for Gemma3 by @abheesht17 in #2170
- Do not export Qwen for release by @abheesht17 in #2198
- Version bump to 0.20.0.dev1 by @abheesht17 in #2200
- Version bump to 0.20.0 by @abheesht17 in #2202
New Contributors
- @villurignanesh made their first contribution in #2156
Full Changelog: v0.19.3...v0.20.0
v0.20.0.dev1
What's Changed
- Version bump to 0.20.0.dev1 by @abheesht17 in #2200
Full Changelog: v0.20.0.dev0...v0.20.0.dev1
v0.20.0.dev0
What's Changed
- Install TF Text on non-Windows only by @abheesht17 in #2115
- Add SigLIP by @james77777778 in #2113
- Fix
PaliGemmaVitEncoder
output shape by @abheesht17 in #2123 - Cspnet architecture. by @sachinprasadhs in #2091
- Update our master version to be a dev release by @mattdangerw in #2131
- Add top 3 HF Presets for Mobilenet by @pkgoogle in #2105
- Add SigLIP2 by @james77777778 in #2127
- update Gemma attention for TPU by @divyashreepathihalli in #2130
- Update dev version rule for nightly by @SamanehSaadat in #2139
- Fix dtype bug in image converter by @abheesht17 in #2147
- Add instruction in .md for manual pre-commit run by @abheesht17 in #2148
- Add Qwen 2.5 by @shivance in #2088
- Updated CONTRIBUTING.md (Fixes issue #2153) by @villurignanesh in #2156
- Update kaggle preset paths for SigLip model by @laxmareddyp in #2164
- Routine Kaggle HF sync by @divyashreepathihalli in #2165
- Enable LoRA target names arg by @divyashreepathihalli in #2166
- Update retinanet_presets.py by @sineeli in #2157
- Add Gemma3 by @abheesht17 in #2152
- Add precommit to the common requirements file by @mattdangerw in #2173
- Add back a format script for compat by @mattdangerw in #2174
- Add a TextToImagePreprocessor base class by @mattdangerw in #2181
- Bump the python group with 2 updates by @dependabot in #2185
- implement of roformerv2 by @pass-lin in #2145
- Move sliding window attn before FA block for Gemma by @abheesht17 in #2187
- Update gating condition to include check for supporting GPUs for flash attention by @divyashreepathihalli in #2184
- Revert "Fix dtype bug in image converter (#2147)" by @mattdangerw in #2180
- Add vision for Gemma3 by @abheesht17 in #2170
- Do not export Qwen for release by @abheesht17 in #2198
New Contributors
- @villurignanesh made their first contribution in #2156
Full Changelog: v0.19.0.dev0...v0.20.0.dev0
v0.19.3
What's Changed
- Version bump to 0.19.3.dev0 by @abheesht17 in #2168
- Version bump to 0.19.3 by @abheesht17 in #2169
Full Changelog: v0.19.2...v0.19.3
v0.19.3.dev0
What's Changed
- Version bump to 0.19.3.dev0 by @abheesht17 in #2168
Full Changelog: v0.19.2...v0.19.3.dev0
v0.19.2
v0.19.1
What's Changed
- TF Text dependency is installed only on non-Windows system only
Full Changelog: v0.19.0.dev0...v0.19.1
v0.19.0.dev0
Summary
- Flash Attention was enabled for KearsHub models
- New models added: SD3.5medium version, VIT, MobileNet and BASNet
- New model presets for: MobileNet, EfficientNet, BasNet, VIT, etc
- Updated formatting to Ruff
- Added download support for ModelScope
- Bounding box utilities moved to Keras
- Minor fixes
What's Changed
- Bump version number to 0.19 by @mattdangerw in #2008
- Remove .DS_Store files by @mattdangerw in #2011
- Add supported tasks types to metadata by @mattdangerw in #1997
- Add a numeric check to the PaliGemma2 conversion script by @james77777778 in #2012
- Adds efficientnet2 presets by @pkgoogle in #1983
- Update README.md by @mattdangerw in #2017
- [ViT] Vision Transformer (ViT) backbone, layers, and image classifier by @sineeli in #1989
- Enable Flash Attention for SD3 MMDiT by @james77777778 in #2014
- Use Ruff for formatting by @mattdangerw in #2019
- update devcontainer as per ruff by @sineeli in #2020
- Add ViT Presets by @sineeli in #2021
- Add BASNet to keras hub by @laxmareddyp in #1984
- Some routine cleanup while writing some new tools for checkpoint admin by @mattdangerw in #2023
- Fix broken preset links; presets without versions by @mattdangerw in #2024
- New tools for model preset admin by @mattdangerw in #2025
- Avoid hard error if tf is not installed by @mattdangerw in #2028
- Try requiring a miniumum version of keras by @mattdangerw in #2029
- Update metadata by @mattdangerw in #2026
- Add SD 3.5 medium by @james77777778 in #2033
- Run HF sync by @mattdangerw in #2030
- Improve error messages for tokenizer trainers by @mattdangerw in #2037
- Update formatting for latest Ruff version by @mattdangerw in #2041
- BASNet Kaggle presets path update by @laxmareddyp in #2052
- Update asserts to avoid deprecated methods by @mattdangerw in #2053
- Add
pad_to_aspect_ratio
flag to ImageConverter by @sineeli in #2045 - Use Flash Attention if available by @james77777778 in #2058
- os.make_dirs is not a thing; os.makedirs is by @mattdangerw in #2061
- Update README.md by @mattdangerw in #2063
- Update auto-assignment.js by @balanprasanth in #2057
- Remove
mask = None
by @abheesht17 in #2067 - Bump the python group with 2 updates by @dependabot in #2066
- Make gemma inputs int32 same as other models by @mattdangerw in #2069
- Vit bug by @sineeli in #2070
- Update auto-assignment.js by @balanprasanth in #2065
- Fix Pytorch GPU test by @divyashreepathihalli in #2087
- Remove bounding box utils and refactor retinanet by @sineeli in #2039
- Add download support for modelscople. by @pass-lin in #2032
- Port MobileNet by @pkgoogle in #2049
- Add VGG preset test cases by @laxmareddyp in #2090
- Made y optional when we use for inference by @laxmareddyp in #2092
- fix mobilenet tests by @divyashreepathihalli in #2096
- Basnet preset rename in kaggle by @laxmareddyp in #2097
- Fix tensorflow dep in
requirements.txt
by @abheesht17 in #2100 - test and preset fixes for mobilenet by @pkgoogle in #2098
- Keep TF Text as optional dependency by @abheesht17 in #2103
- Add PaliGemma 2 mix checkpoints by @bebechien in #2106
- Add
query_proj
,value_proj
to target names forenable_lora
by @abheesht17 in #2107 - Enable Flash attention in Gemma by @divyashreepathihalli in #2064
- Update gemma_attention.py by @divyashreepathihalli in #2109
- fix attention mask dtype by @divyashreepathihalli in #2110
- Fixing imports in init file and preset name update by @laxmareddyp in #2108
- Add pre-commit hooks by @abheesht17 in #2111
- batch mismatch hotfix by @pkgoogle in #2112
New Contributors
- @laxmareddyp made their first contribution in #1984
- @balanprasanth made their first contribution in #2057
- @pass-lin made their first contribution in #2032
- @bebechien made their first contribution in #2106
Full Changelog: v0.18.1...v0.19.0.dev0