Picture for Bum Jun Kim

Bum Jun Kim

The Disappearance of Timestep Embedding in Modern Time-Dependent Neural Networks

May 23, 2024
Viaarxiv icon

Configuring Data Augmentations to Reduce Variance Shift in Positional Embedding of Vision Transformers

Add code
May 23, 2024
Viaarxiv icon

Scale Equalization for Multi-Level Feature Fusion

Feb 02, 2024
Viaarxiv icon

Analysis of NaN Divergence in Training Monocular Depth Estimation Model

Add code
Nov 07, 2023
Figure 1 for Analysis of NaN Divergence in Training Monocular Depth Estimation Model
Figure 2 for Analysis of NaN Divergence in Training Monocular Depth Estimation Model
Figure 3 for Analysis of NaN Divergence in Training Monocular Depth Estimation Model
Figure 4 for Analysis of NaN Divergence in Training Monocular Depth Estimation Model
Viaarxiv icon

Resolution-Aware Design of Atrous Rates for Semantic Segmentation Networks

Add code
Jul 26, 2023
Figure 1 for Resolution-Aware Design of Atrous Rates for Semantic Segmentation Networks
Figure 2 for Resolution-Aware Design of Atrous Rates for Semantic Segmentation Networks
Figure 3 for Resolution-Aware Design of Atrous Rates for Semantic Segmentation Networks
Figure 4 for Resolution-Aware Design of Atrous Rates for Semantic Segmentation Networks
Viaarxiv icon

Understanding Gaussian Attention Bias of Vision Transformers Using Effective Receptive Fields

May 08, 2023
Figure 1 for Understanding Gaussian Attention Bias of Vision Transformers Using Effective Receptive Fields
Figure 2 for Understanding Gaussian Attention Bias of Vision Transformers Using Effective Receptive Fields
Figure 3 for Understanding Gaussian Attention Bias of Vision Transformers Using Effective Receptive Fields
Figure 4 for Understanding Gaussian Attention Bias of Vision Transformers Using Effective Receptive Fields
Viaarxiv icon

How to Use Dropout Correctly on Residual Networks with Batch Normalization

Feb 13, 2023
Figure 1 for How to Use Dropout Correctly on Residual Networks with Batch Normalization
Figure 2 for How to Use Dropout Correctly on Residual Networks with Batch Normalization
Figure 3 for How to Use Dropout Correctly on Residual Networks with Batch Normalization
Figure 4 for How to Use Dropout Correctly on Residual Networks with Batch Normalization
Viaarxiv icon

On the Ideal Number of Groups for Isometric Gradient Propagation

Feb 07, 2023
Figure 1 for On the Ideal Number of Groups for Isometric Gradient Propagation
Figure 2 for On the Ideal Number of Groups for Isometric Gradient Propagation
Figure 3 for On the Ideal Number of Groups for Isometric Gradient Propagation
Figure 4 for On the Ideal Number of Groups for Isometric Gradient Propagation
Viaarxiv icon

Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks

May 15, 2022
Figure 1 for Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks
Figure 2 for Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks
Figure 3 for Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks
Figure 4 for Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks
Viaarxiv icon

Improved Robustness of Vision Transformer via PreLayerNorm in Patch Embedding

Nov 16, 2021
Figure 1 for Improved Robustness of Vision Transformer via PreLayerNorm in Patch Embedding
Figure 2 for Improved Robustness of Vision Transformer via PreLayerNorm in Patch Embedding
Figure 3 for Improved Robustness of Vision Transformer via PreLayerNorm in Patch Embedding
Figure 4 for Improved Robustness of Vision Transformer via PreLayerNorm in Patch Embedding
Viaarxiv icon