Wang, Yi
|
d72edb3b0d
fix lm head overriden issue, move it from checkpoint in-loop loading to out loop (#4206)
|
1 年之前 |
Reza Yazdani
|
468882fb68
Add the policy to run llama model from the official repo (#4313)
|
1 年之前 |
Lev Kurilenko
|
cc3a7c9cba
Fix Meta Tensor checkpoint load for BLOOM models (#3885)
|
1 年之前 |
Lev Kurilenko
|
194053bd58
Hybrid Engine Fix Llama (#3505)
|
1 年之前 |
Michael Wyatt
|
b361c72761
Update DeepSpeed copyright license to Apache 2.0 (#3111)
|
1 年之前 |
Jeff Rasley
|
91d63e0228
update formatter version and style settings (#3098)
|
1 年之前 |
Lev Kurilenko
|
3798e60519
Fix Meta Tensor checkpoint load for OPT models (#2990)
|
1 年之前 |
Heyang Qin
|
dc01cee5ca
using container when loading inference checkpoints (#2875)
|
1 年之前 |
Jeff Rasley
|
da84e60d98
add missing license info to top of all source code (#2889)
|
1 年之前 |
Lev Kurilenko
|
fd1449c766
Port Reza's INT8-quantization fix to container architecture (#2725)
|
1 年之前 |
Ma, Guokai
|
98cc35b6a8
Abstract accelerator (step 3) (#2677)
|
1 年之前 |
Ammar Ahmad Awan
|
867da307d0
Inference Refactor (replace_with_policy, model_implementations) (#2554)
|
1 年之前 |
Lev Kurilenko
|
503706ac44
Remove GatheredParameters context from replace_with_policy (#2591)
|
1 年之前 |
Reza Yazdani
|
35b350b28c
Fix quantized-inference & Add generic support of checkpoint loading (#2547)
|
1 年之前 |
lokoppakmsft
|
f2710bbe1d
Make data contiguous before the inplace reshape-copy_ function (#2489)
|
1 年之前 |
Reza Yazdani
|
afdc72879f
Ds-inference Int8 support through ZeroQuant technology (#2217)
|
2 年之前 |
Reza Yazdani
|
556f005152
Fix random token-generation issue + MP-checkpoint loading/saving (#2132)
|
2 年之前 |
Alex Hedges
|
316c4a43e0
Add flake8 to pre-commit checks (#2051)
|
2 年之前 |
Reza Yazdani
|
aa88137b8d
Add Inference support for running the BigScience-BLOOM Architecture (#2083)
|
2 年之前 |