Liang Luo
Title
Cited by
Cited by
Year
Incbricks: Toward in-network computation with an in-network cache
M Liu, L Luo, J Nelson, L Ceze, A Krishnamurthy, K Atreya
Proceedings of the Twenty-Second International Conference on Architectural …, 2017
882017
Parameter hub: a rack-scale parameter server for distributed deep neural network training
L Luo, J Nelson, L Ceze, A Phanishayee, A Krishnamurthy
Proceedings of the ACM Symposium on Cloud Computing, 41-54, 2018
602018
Laser: Light, accurate sharing detection and repair
L Luo, A Sriraman, B Fugate, S Hu, G Pokam, CJ Newburn, J Devietti
2016 IEEE International Symposium on High Performance Computer Architecture …, 2016
262016
Troubleshooting transiently-recurring errors in production systems with blame-proportional logging
L Luo, S Nath, LR Sivalingam, M Musuvathi, L Ceze
2018 {USENIX} Annual Technical Conference ({USENIX}{ATC} 18), 321-334, 2018
102018
Parameter box: High performance parameter servers for efficient distributed deep neural network training
L Luo, J Nelson, L Ceze, A Phanishayee, A Krishnamurthy
SysML 2018, 2018
92018
Motivating in-network aggregation for distributed deep neural network training
L Luo, M Liu, J Nelson, L Ceze, A Phanishayee, A Krishnamurthy
Workshop on Approximate Computing Across the Stack, 2017
92017
PLink: Discovering and Exploiting Locality for Accelerated Distributed Training on the public Cloud.
L Luo, P West, J Nelson, A Krishnamurthy, L Ceze
Proceedings of the 3rd MLSys Conference, 2020, 2020
72020
PLink: Discovering and Exploiting Datacenter Network Locality for Efficient Cloud-based Distributed Training
L Luo, P West, A Krishnamurthy, L Ceze, J Nelson
MLSys, 2020
52020
PLink: Efficient Cloud-based Training with Topology-aware Dynamic Hierarchical Aggregation
L Luo, P West, J Nelson, A Krishnamurthy, L Ceze
Proceedings of the 3rd MLSys Conference, 2020
42020
High-performance, distributed training of large-scale deep learning recommendation models
D Mudigere, Y Hao, J Huang, A Tulloch, S Sridharan, X Liu, M Ozdal, ...
arXiv preprint arXiv:2104.05158, 2021
12021
Cloud Collectives: Towards Cloud-aware Collectives forML Workloads with Rank Reordering
L Luo, J Nelson, A Krishnamurthy, L Ceze
arXiv preprint arXiv:2105.14088, 2021
2021
Accelerating SpMM Kernel with Cache-First Edge Sampling for GNN Inference
CY Lin, L Luo, L Ceze
arXiv preprint arXiv:2104.10716, 2021
2021
Srift: Swift and Thrift Cloud-Based Distributed Training
L Luo, P West, A Krishnamurthy, L Ceze
arXiv preprint arXiv:2011.14243, 2020
2020
Towards More Efficient Communication for Distributed Learning Systems
L Luo
PQDT-Global, 2020
2020
The system can't perform the operation now. Try again later.
Articles 1–14