Search

Your search keyword '"Rajbhandari, Samyam"' showing total 41 results

Search Constraints

Start Over You searched for: Author "Rajbhandari, Samyam" Remove constraint Author: "Rajbhandari, Samyam"
41 results on '"Rajbhandari, Samyam"'

Search Results

1. SwiftKV: Fast Prefill-Optimized Inference with Knowledge-Preserving Model Transformation

2. DeepSpeed Ulysses: System Optimizations for Enabling Training of Extreme Long Sequence Transformer Models

3. DeepSpeed-VisualChat: Multi-Round Multi-Image Interleave Chat via Multi-Modal Causal Attention

4. DeepSpeed-Chat: Easy, Fast and Affordable RLHF Training of ChatGPT-like Models at All Scales

5. ZeRO++: Extremely Efficient Collective Communication for Giant Model Training

6. A Hybrid Tensor-Expert-Data Parallelism Approach to Optimize Mixture-of-Experts Training

7. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

8. DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale

9. Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model

10. DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to Power Next-Generation AI Scale

11. Scalable and Efficient MoE Training for Multitask Multilingual Models

12. ZeRO-Infinity: Breaking the GPU Memory Wall for Extreme Scale Deep Learning

13. 1-bit LAMB: Communication Efficient Large-Scale Large-Batch Training with LAMB's Convergence Speed

14. 1-bit Adam: Communication Efficient Large-Scale Training with Adam's Convergence Speed

15. ZeRO-Offload: Democratizing Billion-Scale Model Training

16. APMSqueeze: A Communication Efficient Adam-Preconditioned Momentum SGD Algorithm

17. ZeRO: Memory Optimizations Toward Training Trillion Parameter Models

18. AntMan: Sparse Low-Rank Compression to Accelerate RNN inference

19. Learning Intrinsic Sparse Structures within Long Short-Term Memory

20. DeepSpeed-FastGen: High-throughput Text Generation for LLMs via MII and DeepSpeed-Inference

23. A Hybrid Tensor-Expert-Data Parallelism Approach to Optimize Mixture-of-Experts Training

27. ZeRO-infinity

29. DeepSpeed

40. International Conference on Computational Science, ICCS 2012.

41. Locality Optimizations for Regular and Irregular Applications

Catalog

Books, media, physical & digital resources