Shivank Garg

I am Shivank Garg, a final-year undergraduate at the Indian Institute of Technology Roorkee. I have previously worked at Adobe MDSR Labs on video generation models and at Sony AI, and I have also collaborated with professors, Ph.D. scholars, and master's students from leading institutions such as Stanford, CMU, Mila, and Microsoft. My current research focuses on building self-improving AI systems, with broad interests in LLM and VLM alignment. In particular, I explore post-training RL-based techniques for generative models, AI agents, and diffusion models, along with their practical applications.

Feel free to reach out if you would like to collaborate on any research project or discuss any of my research.

Email  /  CV  /  Github  /  Google Scholar

Shivank Garg profile photo

Research

I'm interested in LLM and VLM Alignment, post-training RL based techniques for generative models, and diffusion models and their applications . Some papers are highlighted.

ipo Do Biased Models Have Biased Thoughts?
Shivank Garg, Swati Rajwal, Reem Abdel-Salam, Abdelrahman Zayed
COLM, 2025
bibtex / arXiv

We study fairness in Large Language Models (LLMs) by examining chain-of-thought prompting. Our results show that biased outputs do not always stem from biased reasoning, distinguishing model behavior from human decision-making.

ipo IPO: Your Language Model is Secretly a Preference Classifier
Shivank Garg, Ayush Singh, Shweta Singh, Paras Chopra
ACL Main, 2025
bibtex / arXiv

We introduce Implicit Preference Optimization (IPO), a novel framework that enhances Large Language Models (LLMs) through self-improvement. IPO leverages the model's intrinsic capability to generate continuous reward signals, which are then used to create preference pairs for fine-tuning with Direct Preference Optimization (DPO), eliminating the need for external preference models.

unmasking Unmasking the Veil: An Investigation into Concept Ablation for Privacy and Copyright Protection in Images
Shivank Garg, Manyana Tiwari
TMLR
bibtex / OpenReview

We introduce a technique for concept ablation, specifically targeting the removal of copyrighted logos and trademarks from images to protect intellectual property. We demonstrate the limitations of existing methods, particularly their vulnerability to jailbreaking via paraphrased prompts, and propose a more robust solution.

hint Give me a hint: Can LLMs take a hint to solve math problems?
Vansh Agrawal, Pratham Singla, Amitoj Singh Miglani, Shivank Garg, Ayush Mangal
NeurIPS MathAI Workshop, 2024
bibtex / arXiv

We investigate the mathematical reasoning capabilities of LLMs. Systematically evaluating their performance on complex math problems when provided with hints, and further assessed their robustness under various adversarial conditions to probe the limits of their problem-solving abilities.

attn_shift Attention Shift: Steering AI Away from Unsafe Content
Shivank Garg, Manyana Tiwari
NeurIPS RBFM Workshop 2024
bibtex / arXiv

We introduced Attention Shift, an inference-time technique for content safety in text-to-image diffusion models utilising attention map manipulation of diffusion models to generate safe images from diffusion models while maintaining semantic consistency.

position Positional Embeddings in Transformer Models: Evolution from Text to Vision Domains
Abhinav Kumar, Adesh Gupta, Shivank Garg, Mansi Gupta
ICLR BlogPost, 2025
bibtex / OpenReview

A survey on different positional embeddings used in Vision Transformers and their generalisation to variable resolution images.

lora LoRA-Mini: Adaptation Matrices Decomposition and Selective Training
Ayush Singh, Rajdeep Aher, Shivank Garg
AAAI CoLORAI Workshop, 2025
bibtex / arXiv

We introduce LoRA-Mini, an efficient parameter-efficient fine-tuning (PEFT) method. By decomposing the standard LoRA adaptation matrices and applying selective training of these matrices, LoRA-Mini significantly reduces the number of trainable parameters and lowers memory consumption during fine-tuning.

planning Adaptive Urban Planning: A Hybrid Framework for Balanced City Development
Pratham Singla, Ayush Singh, Adesh Gupta, Shivank Garg
AAAI Workshop on Urban Planning, 2025
bibtex / arXiv

We introduce a novel hybrid framework for adaptive urban planning utilising LLMs and genetic algorithms. The agentic system is designed to generate sustainable and balanced city layouts by optimizing for multiple socio-economic and environmental factors.

snowy Snowy Scenes, Clear Detections: A Robust Model for Traffic Light Detection in Adverse Weather Conditions
Shivank Garg, Abhishek Baghel, Amit Agarwal, Durga Toshniwal
KDD UG Consortium, 2024
bibtex / arXiv

We present a data-centric approach to improve the robustness of traffic light detection models in adverse weather. We introduce a novel data curation and augmentation strategy that simulates challenging conditions like snow and fog, demonstrating a significant improvement in model generalization and performance on out-of-distribution data.

lira Confidence is all you need for MI attacks (student abstract)
Abhishek Sinha, Himanshi Tibrewal, Mansi Gupta, Nikhar Waghela, Shivank Garg
AAAI Student Abstract, 2024
bibtex / AAAI

We introduce a novel confidence-based Membership Inference Attack (MIA). Our method, LIRA, utilizes a model's prediction confidence to infer whether a data point was included in its training set.


The webpage is template is stolen from Jon Barron's website.