Commit History

Author SHA1 Message Date
  Wang, Yi d72edb3b0d fix lm head overriden issue, move it from checkpoint in-loop loading to out loop (#4206) 1 year ago
  Reza Yazdani 468882fb68 Add the policy to run llama model from the official repo (#4313) 1 year ago
  Lev Kurilenko cc3a7c9cba Fix Meta Tensor checkpoint load for BLOOM models (#3885) 1 year ago
  Lev Kurilenko 194053bd58 Hybrid Engine Fix Llama (#3505) 1 year ago
  Michael Wyatt b361c72761 Update DeepSpeed copyright license to Apache 2.0 (#3111) 1 year ago
  Jeff Rasley 91d63e0228 update formatter version and style settings (#3098) 1 year ago
  Lev Kurilenko 3798e60519 Fix Meta Tensor checkpoint load for OPT models (#2990) 1 year ago
  Heyang Qin dc01cee5ca using container when loading inference checkpoints (#2875) 1 year ago
  Jeff Rasley da84e60d98 add missing license info to top of all source code (#2889) 1 year ago
  Lev Kurilenko fd1449c766 Port Reza's INT8-quantization fix to container architecture (#2725) 1 year ago
  Ma, Guokai 98cc35b6a8 Abstract accelerator (step 3) (#2677) 1 year ago
  Ammar Ahmad Awan 867da307d0 Inference Refactor (replace_with_policy, model_implementations) (#2554) 1 year ago
  Lev Kurilenko 503706ac44 Remove GatheredParameters context from replace_with_policy (#2591) 1 year ago
  Reza Yazdani 35b350b28c Fix quantized-inference & Add generic support of checkpoint loading (#2547) 1 year ago
  lokoppakmsft f2710bbe1d Make data contiguous before the inplace reshape-copy_ function (#2489) 1 year ago
  Reza Yazdani afdc72879f Ds-inference Int8 support through ZeroQuant technology (#2217) 2 years ago
  Reza Yazdani 556f005152 Fix random token-generation issue + MP-checkpoint loading/saving (#2132) 2 years ago
  Alex Hedges 316c4a43e0 Add flake8 to pre-commit checks (#2051) 2 years ago
  Reza Yazdani aa88137b8d Add Inference support for running the BigScience-BLOOM Architecture (#2083) 2 years ago