Yuxin Wang

Orcid: 0000-0001-7680-1419

Affiliations:
  • Hong Kong Baptist University, Department of Computer Science, Hong Kong


According to our database1, Yuxin Wang authored at least 24 papers between 2019 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
Reasoning Language Model Inference Serving Unveiled: An Empirical Study.
CoRR, October, 2025

ElasWave: An Elastic-Native System for Scalable Hybrid-Parallel Training.
CoRR, October, 2025

DreamDDP: Accelerating Data Parallel Distributed LLM Training with Layer-wise Scheduled Partial Synchronization.
CoRR, February, 2025

2024
ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference.
CoRR, 2024

FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression.
CoRR, 2024

Towards Efficient and Reliable LLM Serving: A Real-World Workload Study.
CoRR, 2024

Peering the Edge: Enabling Low-Latency Interdomain Edge Communication via Collaborative Transmission.
Proceedings of the 32nd IEEE/ACM International Symposium on Quality of Service, 2024

Bandwidth-Aware and Overlap-Weighted Compression for Communication-Efficient Federated Learning.
Proceedings of the 53rd International Conference on Parallel Processing, 2024

2023
Reliable and Efficient In-Memory Fault Tolerance of Large Language Model Pretraining.
CoRR, 2023

FusionAI: Decentralized Training and Deploying LLMs with Massive Consumer-Level GPUs.
CoRR, 2023

FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training.
CoRR, 2023

Beyond a Centralized Verifier: Scaling Data Plane Checking via Distributed, On-Device Verification.
Proceedings of the ACM SIGCOMM 2023 Conference, 2023

More Than Capacity: Performance-oriented Evolution of Pangu in Alibaba.
Proceedings of the 21st USENIX Conference on File and Storage Technologies, 2023

NAS-LID: Efficient Neural Architecture Search with Local Intrinsic Dimension.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

2022
Energy-Efficient Online Scheduling of Transformer Inference Services on GPU Servers.
IEEE Trans. Green Commun. Netw., 2022

Lamda: The Last Mile of the Datacenter Network Does matter.
CoRR, 2022

Toward low-latency end-to-end communication in 5G using interdomain edge peering.
Proceedings of the NAI '22: Proceedings of the ACM SIGCOMM Workshop on Network-Application Integration, 2022

Network can check itself: scaling data plane checking via distributed, on-device verification.
Proceedings of the 21st ACM Workshop on Hot Topics in Networks, 2022

2020
A Survey of Deep Learning Techniques for Neural Machine Translation.
CoRR, 2020

Energy-efficient Inference Service of Transformer-based Deep Learning Models on GPUs.
Proceedings of the 2020 International Conferences on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, 2020

Benchmarking the Performance and Energy Efficiency of AI Accelerators for AI Training.
Proceedings of the 20th IEEE/ACM International Symposium on Cluster, 2020

2019
A Distributed Synchronous SGD Algorithm with Global Top-k Sparsification for Low Bandwidth Networks.
Proceedings of the 39th IEEE International Conference on Distributed Computing Systems, 2019

The Impact of GPU DVFS on the Energy and Performance of Deep Learning: an Empirical Study.
Proceedings of the Tenth ACM International Conference on Future Energy Systems, 2019

Computer-Aided Clinical Skin Disease Diagnosis Using CNN and Object Detection Models.
Proceedings of the 2019 IEEE International Conference on Big Data (IEEE BigData), 2019


  Loading...