Tsuguchika Tabaru

Orcid: 0000-0001-6568-1968

According to our database1, Tsuguchika Tabaru authored at least 20 papers between 2015 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
NEEBS: Nonexpert large-scale environment building system for deep neural network.
Concurr. Comput. Pract. Exp., 2023

A novel structured sparse fully connected layer in convolutional neural networks.
Concurr. Comput. Pract. Exp., 2023

Optimization of NumPy Transcendental Functions for Arm SVE.
Proceedings of the HPC Asia 2023 Workshops, 2023

2022
Accelerating LiNGAM Causal Discovery with Massive Parallel Execution on Supercomputer Fugaku.
IEICE Trans. Inf. Syst., December, 2022

Structured Pruning with Automatic Pruning Rate Derivation for Image Processing Neural Networks.
Proceedings of the ISMSI 2022: 6th International Conference on Intelligent Systems, Metaheuristics & Swarm Intelligence, Seoul, Republic of Korea, April 9, 2022

Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks.
Proceedings of the 26th International Conference on Pattern Recognition, 2022

Regularizing Data for Improving Execution Time of NLP Model.
Proceedings of the Thirty-Fifth International Florida Artificial Intelligence Research Society Conference, 2022

2021
MLPerf HPC: A Holistic Benchmark Suite for Scientific Machine Learning on HPC Systems.
CoRR, 2021


Efficient and Large Scale Pre-training Techniques for Japanese Natural Language Processing.
Proceedings of the Ninth International Symposium on Computing and Networking, 2021

The 16, 384-node Parallelism of 3D-CNN Training on An Arm CPU based Supercomputer.
Proceedings of the 28th IEEE International Conference on High Performance Computing, 2021

2020
Efficient convolution pooling on the GPU.
J. Parallel Distributed Comput., 2020

An Efficient Multicore CPU Implementation for Convolution-Pooling Computation in CNNs.
Proceedings of the 2020 IEEE International Parallel and Distributed Processing Symposium Workshops, 2020

Huffman Coding with Gap Arrays for GPU Acceleration.
Proceedings of the ICPP 2020: 49th International Conference on Parallel Processing, 2020

An Efficient Technique for Large Mini-batch Challenge of DNNs Training on Large Scale Cluster.
Proceedings of the HPDC '20: The 29th International Symposium on High-Performance Parallel and Distributed Computing, 2020

2019
Yet Another Accelerated SGD: ResNet-50 Training on ImageNet in 74.7 seconds.
CoRR, 2019

Efficient cuDNN-Compatible Convolution-Pooling on the GPU.
Proceedings of the Parallel Processing and Applied Mathematics, 2019

Structured Sparse Fully-Connected Layers in the CNNs and Its GPU Acceleration.
Proceedings of the Seventh International Symposium on Computing and Networking Workshops, 2019

2017
Fast algorithm using summed area tables with unified layer performing convolution and average pooling.
Proceedings of the 27th IEEE International Workshop on Machine Learning for Signal Processing, 2017

2015
Design of a Shared Memory mechanism for efficient paralell processing in PostgreSQL.
Proceedings of the 6th International Conference on Information, 2015


  Loading...