Skip to content

v0.21.0

Compare
Choose a tag to compare
@laxmareddyp laxmareddyp released this 28 May 19:07
· 2 commits to r0.21 since this release
933efe6

Summary

  • New Models.

    • Xception: Added Xception architecture for image classification tasks.
    • Qwen: Added Qwen2.5 large language models and presets of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters.
    • Qwen MoE: Added transformer-based Mixture of Experts (MoE) decoder-only language model with a base variant having 2.7B activated parameters during runtime.
    • Mixtral: Added Mixtral LLM, a pretrained generative Sparse Mixture of Experts with pre-trained and instruction tuned models having 7 billion activated parameters.
    • Moonshine: Added Moonshine, a speech recognition task model.
    • CSPNet: Added Cross Stage Partial Network (CSPNet) classification task model.
    • Llama3: Added support for Llama 3.1 and 3.2.
  • Added sharded weight support to KerasPresetSaver and KerasPresetLoader, defaulting to a 10GB maximum shard size.

What's Changed

New Contributors

Full Changelog: v0.20.0...v0.21.0