Shaden Smith

Orcid: 0000-0003-4072-9990

According to our database1, Shaden Smith authored at least 26 papers between 2012 and 2023.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
HEAT: A Highly Efficient and Affordable Training System for Collaborative Filtering Based Recommendation on CPUs.
Proceedings of the 37th International Conference on Supercomputing, 2023

2022
Scalable Label Propagation for Multi-Relational Learning on the Tensor Product of Graphs.
IEEE Trans. Knowl. Data Eng., 2022

Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model.
CoRR, 2022

DeepSpeed- Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale.
Proceedings of the SC22: International Conference for High Performance Computing, 2022

2021
ZeRO-infinity: breaking the GPU memory wall for extreme scale deep learning.
Proceedings of the International Conference for High Performance Computing, 2021

High Performance Streaming Tensor Decomposition.
Proceedings of the 35th IEEE International Parallel and Distributed Processing Symposium, 2021

2020
PIUMA: Programmable Integrated Unified Memory Architecture.
CoRR, 2020

Tensaurus: A Versatile Accelerator for Mixed Sparse-Dense Tensor Computations.
Proceedings of the IEEE International Symposium on High Performance Computer Architecture, 2020

2018
HPC formulations of optimization algorithms for tensor completion.
Parallel Comput., 2018

Scalable Label Propagation for Multi-relational Learning on Tensor Product Graph.
CoRR, 2018

Streaming Tensor Factorization for Infinite Data Sources.
Proceedings of the 2018 SIAM International Conference on Data Mining, 2018

Blocking Optimization Techniques for Sparse Tensor Computation.
Proceedings of the 2018 IEEE International Parallel and Distributed Processing Symposium, 2018

Scalability and Distribution of Collaborative Recommenders.
Proceedings of the Collaborative Recommendations, 2018

2017
Bridging the Gap between HPC and Big Data frameworks.
Proc. VLDB Endow., 2017

Sparse Tensor Factorization on Many-Core Processors with High-Bandwidth Memory.
Proceedings of the 2017 IEEE International Parallel and Distributed Processing Symposium, 2017

Constrained Tensor Factorization with Accelerated AO-ADMM.
Proceedings of the 46th International Conference on Parallel Processing, 2017

Exploring optimizations on shared-memory platforms for parallel triangle counting algorithms.
Proceedings of the 2017 IEEE High Performance Extreme Computing Conference, 2017

Truss decomposition on shared-memory parallel systems.
Proceedings of the 2017 IEEE High Performance Extreme Computing Conference, 2017

Accelerating the Tucker Decomposition with Compressed Sparse Tensors.
Proceedings of the Euro-Par 2017: Parallel Processing - 23rd International Conference on Parallel and Distributed Computing, Santiago de Compostela, Spain, August 28, 2017

2016
An exploration of optimization algorithms for high performance tensor completion.
Proceedings of the International Conference for High Performance Computing, 2016

A Medium-Grained Algorithm for Sparse Tensor Factorization.
Proceedings of the 2016 IEEE International Parallel and Distributed Processing Symposium, 2016

2015
Tensor-matrix products with a compressed sparse tensor.
Proceedings of the 5th Workshop on Irregular Applications - Architectures and Algorithms, 2015

SPLATT: Efficient and Parallel Sparse Tensor-Matrix Multiplication.
Proceedings of the 2015 IEEE International Parallel and Distributed Processing Symposium, 2015

2014
Memory-efficient parallel computation of tensor and matrix products for big tensor decomposition.
Proceedings of the 48th Asilomar Conference on Signals, Systems and Computers, 2014

Big Data Frequent Pattern Mining.
Proceedings of the Frequent Pattern Mining, 2014

2012
Weighted-Sequence Problem: ASP vs CASP and Declarative vs Problem-Oriented Solving.
Proceedings of the Practical Aspects of Declarative Languages, 2012


  Loading...