Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Releases: foundation-model-stack/foundation-model-stack

v1.7.0

16 Feb 20:40
9124614

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.6.0...v1.7.0

v1.6.0

14 Jan 16:39
057f620

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.5.0...v1.6.0

v1.5.0

31 Oct 16:42
c9bc7ee

Choose a tag to compare

What's Changed

  • Add Roberta for classification to FMS by @ani300 in #466
  • Add BERT support to FMS by @ani300 in #467
  • updated transformers to 4.55.4 and fixed bug where fms expected cache to be None by @JRosenkranz in #474
  • Granite 2b & 3b expand K Q V and Dense to head dim 128 to compile for AIU by @rzbhatti in #472
  • Chunked prefill support for paged attention by @ani300 in #463
  • head_dim expansion managed at get_model() level by @rzbhatti in #476

Full Changelog: v1.4.0...v1.5.0

v1.4.0

27 Sep 00:26
71eb2ea

Choose a tag to compare

What's Changed

Full Changelog: v1.3.0...v1.4.0

v1.3.0

26 Sep 23:14
fc7fd8e

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.2.1...v1.3.0

v1.2.1

07 Aug 14:09

Choose a tag to compare

What's Changed

  • Change how quantized layers are sharded during Tensor Parallel to support FP8 and other cases by @ani300 in #457

Full Changelog: v1.2.0...v1.2.1

v1.2.0

05 Aug 00:10

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.1.0...v1.2.0

v1.1.0

13 Jun 19:46

Choose a tag to compare

What's Changed

  • [tokenizer] enable new methods/attributes for tokenizer that are commonly used by @kcirred in #409
  • [hf_adapter] added GenerationMixin for transformers version compatibility by @kcirred in #418
  • fix get_signature so that it no longer ignores some optional params by @JRosenkranz in #413
  • Enable Granite HF Adapter by @kcirred in #402
  • Don't set cache_dir in snapshot_download by @tjohnson31415 in #417
  • Delay the eos-based loop break to at least get one timing in by @ani300 in #422
  • fixed errors in llama and gpt_bigcode docstring by @Zephyr271828 in #390
  • paged attention implementation using attn_kwargs by @JRosenkranz in #411

New Contributors

Full Changelog: v1.0.0...v1.1.0

v1.0.0

16 May 18:53

Choose a tag to compare

What's Changed

Bug Fixes

  • fixed bug where when hf_configured set as architecture in get_model, weights get downloaded by @JRosenkranz in #373
  • Fix issue where CICD fails occasionally on a test case by @JRosenkranz in #378
  • Fixed mypy failure in CI by @JRosenkranz in #381
  • fix some bugs in local/test training by @nairbv in #393
  • Fix performance when use_contiguous_cache=True by @ani300 in #389
  • Fix extra graph generation during generate with contiguous cache by @ani300 in #397
  • fixed bug where expectation tests were being skipped by @JRosenkranz in #410

Changes

  • Change HF default checkpoint from bin to safetensors by @ani300 in #377
  • Support for Granite GPTQ model weight adaptation from HF by @JRosenkranz in #375
  • Roberta Question-Answering by @andrea-fasoli in #379
  • Add code of conduct by @spzala in #374
  • Select linear layer based on module_name by @andrea-fasoli in #369
  • Update dependencies for FMS to a more modern stack by @ani300 in #380
  • Bamba Model Support by @JRosenkranz in #372
  • Added option to specify model name in the model consistency test suite expectation file path by @JRosenkranz in #383
  • Fixed issues inferring Roberta QA as well as added features for encoder-only testing by @JRosenkranz in #386
  • Update TP and MoE kernels with modern pytorch constructs by @ani300 in #382
  • Add fixes for full dynamic with masks by @ani300 in #387
  • Add support for fms_mo-based INT8 Granite by @andrea-fasoli in #391
  • Add a carve-out for Bamba SSM layers in contiguous check by @ani300 in #394
  • Flexible model inputs in generate by @JRosenkranz in #388
  • added optional_params/input_ids option to consistency expectation testing by @JRosenkranz in #398
  • mistralai/Mistral-7B-Instruct-v0.3 model ported to fms by @rzbhatti in #395
  • Add handling of token_type_ids to RoBERTa by @andrea-fasoli in #399
  • Raise exception in case of shape mismatch during ckpt loading by @andrea-fasoli in #400
  • Add Rope implementations and corrections for llama 3 and llama 3.1 by @ani300 in #385
  • Inject custom attention op into MultiHeadAttention by @JRosenkranz in #408

New Contributors

Full Changelog: v0.0.8...v1.0.0

v0.0.8

12 Dec 18:57

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.0.7...v0.0.8

Morty Proxy This is a proxified and sanitized view of the page, visit original site.