I am an undergraduate student, majoring in computer science and engineering at the National Institute of Technology Rourkela, India. I am broadly interested in deep learning and machine learning research, with a focus on multi-modal learning and interpretability.

Currently, I am working on continual learning and test-time training for Vision Transformers at the Visual Computing Lab (VCL), IISc Bangalore, under the guidance of Dr. Anirban Chakraborty.

Previously, in the summer of 2024, I worked with Dr. Konda Reddy Mopuri at IIT Hyderabad on Vision Transformers and explainability; in the summer of 2025, I worked with Prof. Vineeth Balasubramanian on faithful Concept Bottleneck Models (CBMs) for medical imaging; and in the subsequent semester, I worked with Dr. Prasenjit Dey on diffusion-based sketch-to-face synthesis.

For more details, drop me an email.

News & Honors

Publications

CapsoNet: A CNN-Transformer Ensemble for Multi-Class Abnormality Detection in Video Capsule Endoscopy
Arnav Samal, Ranya Batsyas
arXiv preprint | Oct, 2024
pdf | abstract | code

Selected Projects

SketchWarp
Developed a self-supervised learning framework in PyTorch for dense photo-to-sketch correspondences, enabling automatic image-to-sketch warping. Designed and implemented training and evaluation pipelines inspired by the “Learning Dense Correspondences between Photos and Sketches” paper.
code | paper

NeurIPS Ariel Data Challenge 2024
Developed a pipeline for predicting spectral values in the NeurIPS Ariel Data Challenge 2024 using time-series calibration, spatial aggregation, and gradient-based phase detection. Ranked 257/1,152 by applying Nelder-Mead optimization and cubic polynomial fitting to model planetary transits from raw sensor data.
code | kaggle

Paper Implementations
Implemented significant AI and machine learning research papers, including transformers (such as GPT variants, BERT, ViTs) as well as LoRA and neural style transfer. I actively implement new papers and continuously update this repository.
code

Measuring Patch Importance in ViT's (Vanilla & Attention Rollout)
Analyzed patch importance in Vision Transformers using attention scores of the [CLS] token across MHSA mechanims in all blocks, visualizing the distribution of top-k patch tokens. Implemented Attention Rollout to propagate attention through layers, creating interpretable visualizations of information flow and enhancing understanding of self-attention mechanisms.
code

IISc Bangalore
2026
RespAI Lab
2025
IIT Hyderabad
S2024 & S2025
NIT Rourkela
2022 - Present