Search

Your search keyword '"Park, Jongsoo"' showing total 289 results

Search Constraints

Start Over You searched for: Author "Park, Jongsoo" Remove constraint Author: "Park, Jongsoo"
289 results on '"Park, Jongsoo"'

Search Results

1. Context Parallelism for Scalable Million-Token Inference

2. The Llama 3 Herd of Models

3. Wukong: Towards a Scaling Law for Large-Scale Recommendation

4. Disaggregated Multi-Tower: Topology-aware Modeling Technique for Efficient Large-Scale Recommendation

7. MTrainS: Improving DLRM training efficiency using heterogeneous memories

8. With Shared Microexponents, A Little Shifting Goes a Long Way

9. RecD: Deduplication for End-to-End Deep Learning Recommendation Model Training Infrastructure

10. DHEN: A Deep and Hierarchical Ensemble Network for Large-Scale Click-Through Rate Prediction

15. First-Generation Inference Accelerator Deployment at Facebook

16. Low-Precision Hardware Architectures Meet Recommendation Model Inference at Scale

17. Alternate Model Growth and Pruning for Efficient Training of Recommendation Systems

18. Software-Hardware Co-design for Fast and Scalable Training of Deep Learning Recommendation Models

19. Efficient Soft-Error Detection for Low-precision Deep Learning Recommendation Models

20. FBGEMM: Enabling High-Performance Low-Precision Deep Learning Inference

21. Mixed-Precision Embedding Using a Cache

22. Adaptive Dense-to-Sparse Paradigm for Pruning Online Recommendation System with Non-Stationary Data

23. Post-Training 4-bit Quantization on Embedding Tables

24. Deep Learning Recommendation Model for Personalization and Recommendation Systems

25. A Study of BFLOAT16 for Deep Learning Training

26. Spatial-Winograd Pruning Enabling Sparse Winograd Convolution

28. Deep Learning Inference in Facebook Data Centers: Characterization, Performance Optimizations and Hardware Implications

29. On Periodic Functions as Regularizers for Quantization of Neural Networks

30. Glow: Graph Lowering Compiler Techniques for Neural Networks

33. Two-step approach to scheduling quantum circuits

34. Enabling Sparse Winograd Convolution by Native Pruning

38. Faster CNNs with Direct Sparse Convolutions and Guided Pruning

49. With Shared Microexponents, A Little Shifting Goes a Long Way

50. Parallel Efficient Sparse Matrix-Matrix Multiplication on Multicore Platforms

Catalog

Books, media, physical & digital resources