Alert button
Picture for Geonhwa Jeong

Geonhwa Jeong

Alert button

Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition

Add code
Bookmark button
Alert button
Mar 12, 2024
Geonhwa Jeong, Po-An Tsai, Abhimanyu R. Bambhaniya, Stephen W. Keckler, Tushar Krishna

Figure 1 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Figure 2 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Figure 3 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Figure 4 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Viaarxiv icon

GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM

Add code
Bookmark button
Alert button
Mar 11, 2024
Hao Kang, Qingru Zhang, Souvik Kundu, Geonhwa Jeong, Zaoxing Liu, Tushar Krishna, Tuo Zhao

Figure 1 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Figure 2 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Figure 3 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Figure 4 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Viaarxiv icon

Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference

Add code
Bookmark button
Alert button
Mar 08, 2024
Akshat Ramachandran, Zishen Wan, Geonhwa Jeong, John Gustafson, Tushar Krishna

Figure 1 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Figure 2 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Figure 3 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Figure 4 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Viaarxiv icon

VEGETA: Vertically-Integrated Extensions for Sparse/Dense GEMM Tile Acceleration on CPUs

Add code
Bookmark button
Alert button
Feb 23, 2023
Geonhwa Jeong, Sana Damani, Abhimanyu Rajeshkumar Bambhaniya, Eric Qin, Christopher J. Hughes, Sreenivas Subramoney, Hyesoon Kim, Tushar Krishna

Figure 1 for VEGETA: Vertically-Integrated Extensions for Sparse/Dense GEMM Tile Acceleration on CPUs
Figure 2 for VEGETA: Vertically-Integrated Extensions for Sparse/Dense GEMM Tile Acceleration on CPUs
Figure 3 for VEGETA: Vertically-Integrated Extensions for Sparse/Dense GEMM Tile Acceleration on CPUs
Figure 4 for VEGETA: Vertically-Integrated Extensions for Sparse/Dense GEMM Tile Acceleration on CPUs
Viaarxiv icon

RASA: Efficient Register-Aware Systolic Array Matrix Engine for CPU

Add code
Bookmark button
Alert button
Oct 05, 2021
Geonhwa Jeong, Eric Qin, Ananda Samajdar, Christopher J. Hughes, Sreenivas Subramoney, Hyesoon Kim, Tushar Krishna

Figure 1 for RASA: Efficient Register-Aware Systolic Array Matrix Engine for CPU
Figure 2 for RASA: Efficient Register-Aware Systolic Array Matrix Engine for CPU
Figure 3 for RASA: Efficient Register-Aware Systolic Array Matrix Engine for CPU
Figure 4 for RASA: Efficient Register-Aware Systolic Array Matrix Engine for CPU
Viaarxiv icon

Union: A Unified HW-SW Co-Design Ecosystem in MLIR for Evaluating Tensor Operations on Spatial Accelerators

Add code
Bookmark button
Alert button
Sep 17, 2021
Geonhwa Jeong, Gokcen Kestor, Prasanth Chatarasi, Angshuman Parashar, Po-An Tsai, Sivasankaran Rajamanickam, Roberto Gioiosa, Tushar Krishna

Figure 1 for Union: A Unified HW-SW Co-Design Ecosystem in MLIR for Evaluating Tensor Operations on Spatial Accelerators
Figure 2 for Union: A Unified HW-SW Co-Design Ecosystem in MLIR for Evaluating Tensor Operations on Spatial Accelerators
Figure 3 for Union: A Unified HW-SW Co-Design Ecosystem in MLIR for Evaluating Tensor Operations on Spatial Accelerators
Figure 4 for Union: A Unified HW-SW Co-Design Ecosystem in MLIR for Evaluating Tensor Operations on Spatial Accelerators
Viaarxiv icon

Evaluating Spatial Accelerator Architectures with Tiled Matrix-Matrix Multiplication

Add code
Bookmark button
Alert button
Jun 19, 2021
Gordon E. Moon, Hyoukjun Kwon, Geonhwa Jeong, Prasanth Chatarasi, Sivasankaran Rajamanickam, Tushar Krishna

Figure 1 for Evaluating Spatial Accelerator Architectures with Tiled Matrix-Matrix Multiplication
Figure 2 for Evaluating Spatial Accelerator Architectures with Tiled Matrix-Matrix Multiplication
Figure 3 for Evaluating Spatial Accelerator Architectures with Tiled Matrix-Matrix Multiplication
Figure 4 for Evaluating Spatial Accelerator Architectures with Tiled Matrix-Matrix Multiplication
Viaarxiv icon

ConfuciuX: Autonomous Hardware Resource Assignment for DNN Accelerators using Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 04, 2020
Sheng-Chun Kao, Geonhwa Jeong, Tushar Krishna

Figure 1 for ConfuciuX: Autonomous Hardware Resource Assignment for DNN Accelerators using Reinforcement Learning
Figure 2 for ConfuciuX: Autonomous Hardware Resource Assignment for DNN Accelerators using Reinforcement Learning
Figure 3 for ConfuciuX: Autonomous Hardware Resource Assignment for DNN Accelerators using Reinforcement Learning
Figure 4 for ConfuciuX: Autonomous Hardware Resource Assignment for DNN Accelerators using Reinforcement Learning
Viaarxiv icon