arxiv:2510.21285
Yingz
KigYzi
AI & ML interests
None yet
Recent Activity
authored
a paper
about 16 hours ago
When Models Outthink Their Safety: Mitigating Self-Jailbreak in Large
Reasoning Models with Chain-of-Guardrails
upvoted
a
paper
about 17 hours ago
When Models Outthink Their Safety: Mitigating Self-Jailbreak in Large
Reasoning Models with Chain-of-Guardrails
liked
a model
8 days ago
jiawei-ucas/Qwen-2.5-7B-ConsistentChat
Organizations
None yet