Follow
Yu Emma Wang
Title
Cited by
Cited by
Year
Glam: Efficient scaling of language models with mixture-of-experts
N Du, Y Huang, AM Dai, S Tong, D Lepikhin, Y Xu, M Krikun, Y Zhou, ...
International Conference on Machine Learning, 5547-5569, 2022
3472022
A Systematic Methodology for Analysis of Deep Learning Hardware and Software Platforms
YE Wang, GY Wei, D Brooks
the 3rd MLSys Conference, 2020
64*2020
Exploiting parallelism opportunities with deep learning frameworks
YE Wang, CJ Wu, X Wang, K Hazelwood, D Brooks
ACM Transactions on Architecture and Code Optimization (TACO) 18 (1), 1-23, 2020
312020
Predicting New Workload or CPU Performance by Analyzing Public Datasets
YE Wang, V Lee, GY Wei, D Brooks
ACM Transactions on Architecture and Code Optimization 15 (4), 53, 2019
312019
A flexible approach to autotuning multi-pass machine learning compilers
PM Phothilimthana, A Sabne, N Sarda, KS Murthy, Y Zhou, ...
2021 30th International Conference on Parallel Architectures and Compilation …, 2021
242021
Exploring the limits of Concurrency in ML Training on Google TPUs
S Kumar, Y Wang, C Young, J Bradbury, N Kumar, D Chen, A Swing
Proceedings of Machine Learning and Systems 3, 81-92, 2021
162021
A Learning Algorithm for Bayesian Networks and Its Efficient Implementation on GPUs
Y Wang, W Qian, S Zhang, X Liang, B Yuan
IEEE Transactions on Parallel and Distributed Systems 27 (1), 17-30, 2016
102016
Demystifying Bayesian Inference Workloads
YE Wang, Y Zhu, GG Ko, B Reagen, GY Wei, D Brooks
IEEE International Symposium on Performance Analysis of Systems and Software, 2019
92019
Yazhou Zu, Yuanzhong Xu, and Andy Swing. Exploring the limits of concurrency in ml training on google TPUs
S Kumar, J Bradbury, C Young, YE Wang, A Levskaya, B Hechtman, ...
arXiv preprint arXiv:2011.03641 1 (1), 1.2, 2020
82020
AutoDistill: An end-to-end framework to explore and distill hardware-efficient language models
X Zhang, Z Zhou, D Chen, YE Wang
arXiv preprint arXiv:2201.08539, 2022
72022
Exploring hardware profile-guided green datacenter scheduling
W Tang, Y Wang, H Liu, T Zhang, C Li, X Liang
2015 44th International Conference on Parallel Processing, 11-20, 2015
52015
Mixed precision post training quantization of neural networks with sensitivity guided search
CJS Schaefer, E Guo, C Stanton, X Zhang, T Jablin, N Lambert-Shirzad, ...
arXiv preprint arXiv:2302.01382, 2023
12023
Hadamard Domain Training with Integers for Class Incremental Quantized Learning
M Schiemer, CJS Schaefer, JP Vap, MJ Horeni, YE Wang, J Ye, S Joshi
arXiv preprint arXiv:2310.03675, 2023
2023
Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training Quantization
CJS Schaefer, N Lambert-Shirzad, X Zhang, C Chou, T Jablin, J Li, E Guo, ...
arXiv preprint arXiv:2306.04879, 2023
2023
Sparsely Activated Language Models are Efficient In-Context Learners
A Yu, A Dai, C Cui, DD Lepikhin, E Wang, K Meier-Hellstern, K Webster, ...
2022
Performance Analysis for Machine Learning Applications
YE Wang
Harvard University, 2019
2019
The system can't perform the operation now. Try again later.
Articles 1–16