Research Interests
Adversarial Machine Learning
Investigating vulnerabilities in LLMs and Reinforcement Learning agents against intentional perturbations.
Trustworthy & Safe AI
Building alignment protocols and safety guardrails to ensure AI systems behave reliably in edge cases.
NLP Robustness
Analyzing the linguistic limits of foundational models and improving their generalization in medical and legal domains.
Publications
arXiv Pre-print2025
Robustness in Large Language Models via Adversarial Alignment
Piyush Pant, Xiao Zhang, John Doe
arXiv Pre-print2024
Advancing Robustness in Large Language Models via Adversarial Alignment
Piyush Pant, Xiao Zhang, John Doe