Research Interests

Adversarial Machine Learning

Investigating vulnerabilities in LLMs and Reinforcement Learning agents against intentional perturbations.

Trustworthy & Safe AI

Building alignment protocols and safety guardrails to ensure AI systems behave reliably in edge cases.

NLP Robustness

Analyzing the linguistic limits of foundational models and improving their generalization in medical and legal domains.

Publications

arXiv Pre-print2025

Robustness in Large Language Models via Adversarial Alignment

Piyush Pant, Xiao Zhang, John Doe

arXiv Pre-print2024

Advancing Robustness in Large Language Models via Adversarial Alignment

Piyush Pant, Xiao Zhang, John Doe