.. |
adagrad
|
69af361167
CPUAdam fp16 and bf16 support (#5409)
|
5 months ago |
adam
|
69af361167
CPUAdam fp16 and bf16 support (#5409)
|
5 months ago |
aio
|
b361c72761
Update DeepSpeed copyright license to Apache 2.0 (#3111)
|
1 year ago |
deepspeed4science
|
1d1a20c5a1
Fix the openfold training. (#4657)
|
11 months ago |
fp_quantizer
|
5e6c9b9311
OptimizedLinear implementation (#5355)
|
6 months ago |
lamb
|
b361c72761
Update DeepSpeed copyright license to Apache 2.0 (#3111)
|
1 year ago |
lion
|
69af361167
CPUAdam fp16 and bf16 support (#5409)
|
5 months ago |
quantizer
|
b361c72761
Update DeepSpeed copyright license to Apache 2.0 (#3111)
|
1 year ago |
random_ltd
|
c8d3f5eb19
fix typo in comments with deepspeed/ (#3537)
|
1 year ago |
sparse_attention
|
c8d3f5eb19
fix typo in comments with deepspeed/ (#3537)
|
1 year ago |
transformer
|
f4efef21b8
[INF] DSAttention allow input_mask to have false as value (#5546)
|
5 months ago |
__init__.py
|
c08e69f212
Make op builder detection adapt to accelerator change (#5206)
|
7 months ago |
csrc
|
31f46feee2
DeepSpeed JIT op + PyPI support (#496)
|
4 years ago |
op_builder
|
31f46feee2
DeepSpeed JIT op + PyPI support (#496)
|
4 years ago |