AI Safety

Published:

Time: 2025 - Present
Location: Penn State University
Advisor: Prof. Rui Zhang
Skills: Group Relative Policy Optimization (GRPO) Reinforcement Learning , Uncertainty quantification (UQ), Retrieval-Augmented Generation (RAG)

Currently, I am developing a safety framework that utilizes model internal signals to detect and mitigate hallucinations in Large Language Models. By leveraging Group Relative Policy Optimization (GRPO) and Reinforcement Learning, the system dynamically adjusts model outputs to align with factual correctness, aiming to create more robust and safe AI decision making agents.