Skylar Zhai is an undergraduate researcher in Computer Science at the University of Minnesota Twin Cities, advised by Prof. Dongyeop Kang in the Minnesota NLP Group. He also feels fortunate to collaborate with Linxin Song, whose work centers on LLM/VLM evaluation and synthetic data.
Research Interests
My research broadly lies in natural language processing and trustworthy AI. Specifically, I am interested in the following questions:
- How can we build trustworthy AI systems, including models that are honest, calibrated, and safe to deploy in the real world?
- How can we use reinforcement learning to unlock and extend the capabilities of foundation models?
- How can language models meaningfully support human writing, and how can we use them to characterize the linguistic and structural features of scientific research at scale?
π» Experience
- present, University of Minnesota Twin Cities, Minnesota NLP Group
- Advisor: Dongyeop Kang
- Research Focus: Trustworthy LLM, AI4Writing
π₯ News
- 2026.04: Β π Abstain-R1 accepted at ACL 2026 Findings.
- 2026.04: Β π OS-Blind (The Blind Spot of Agent Safety) released on arXiv.
- 2025.11: Β π FAST-CAD accepted as an Oral at AAAI 2026. Congrats to Tommy Sha!
- 2025.10: Β βοΈ Attended ICCV 2025.
- 2025.06: Β π One paper accepted at ICCV 2025 on test-time adaptation.
- 2025.05: Β π Submitted one paper about video understanding to NeurIPS 2025.
- 2025.03: Β π One paper accepted at ICME 2025 on test-time adaptation.
- 2025.03: Β π One paper accepted at ICLR 2025 FM-Wild Workshop on test-time adaptation.
π Publications
Trustworthy LLM / MLLM (Abstention & Agent Safety)
Abstain-R1: Calibrated Abstention and Post-Refusal Clarification via Verifiable RL
Reinforcement fine-tuning sharpens LLM reasoning, but also pushes models to guess on unanswerable queries. We propose a clarification-aware RLVR reward that jointly optimizes explicit abstention and semantically aligned post-refusal clarification, treating βwhat is missingβ as a first-class training target. The resulting 3B Abstain-R1 model substantially improves abstention and clarification on unanswerable queries while preserving strong performance on answerable ones, matching larger systems like DeepSeek-R1 across Abstain-Test, Abstain-QA, and SelfAware.
Role: First Author.
Supported by: Lambda and CloudRift research grants ($1k compute credits each).
Initiated as a course project for CSCI 5541 (Natural Language Processing) at UMN.
Project Page Β |Β Code Β |Β Dataset
Featured on π€ Hugging Face Daily Papers (Apr 15, 2026)
Computer-use agents can autonomously drive real digital environments, and when misled they can automate harm at scale. Existing safety evaluations focus on explicit misuse or prompt injection; we study a subtler setting where user instructions are entirely benign and harm emerges from the task context or execution. We introduce OS-Blind, a 300-task human-crafted benchmark spanning 12 categories, 8 applications, and 2 threat clusters. Most CUAs exceed 90% attack success rate; even safety-aligned Claude 4.5 Sonnet reaches 73.0%, climbing to 92.7% in multi-agent deployments, showing that alignment mostly fires in the first few steps and rarely re-engages during execution.
Role: Co-first Author.
Test-Time Adaptation for Vision-Language Models
Multi-Cache enhanced Prototype Learning for Test-Time Generalization of Vision-Language Models
Project Page Β |Β Code
We observed that cache-based test-time adaptation performance is positively correlated with intra-class compactness. To address the unreliability of low-entropy samples under distribution shifts, we propose MCP, which uses an entropy cache for prototype initialization, an align cache to fuse visual and textual information and tighten intra-class distributions, and a negative cache to calibrate high-entropy predictions. We further extend this into the MCP++ framework by introducing cross-modal prototype alignment and residual learning, achieving state-of-the-art generalization on 15 downstream tasks.
Role: Co-first Author.
Mitigating Cache Noise in Test-Time Adaptation for Large Vision-Language Models
Also accepted at ICLR 2025 FM-Wild Workshop
We analyzed the root causes of the performance gap between zero-shot and few-shot TTA, identifying noisy cache labels as a critical bottleneck. We then propose the CRG framework, which maintains positive and negative visual prototypes alongside text prototypes, employs learnable residuals to align modalities, and leverages Gaussian Discriminant Analysis to dynamically model class distributions and suppress noisy samples. Finally, by jointly minimizing prediction entropy and maximizing inter-prototype distances, CRG achieves superior robustness and generalization across 13 benchmarks..
Role: First Author.
Medical AI
FAST-CAD: A Fairness-Aware Framework for Non-Contact Stroke Diagnosis
Project Page Β |Β MIT Tech Review
Stroke is an acute cerebrovascular disease, so we propose FAST-CAD, a DAT + Group-DRO framework that jointly enforces demographic-invariant representations and worst-group robustness for non-contact stroke diagnosis. Built on a 12-subgroup multimodal dataset, it couples adversarial domain discrimination with self-supervised encoders and optimizes worst-group risk, delivering 91.2% AUC and tight fairness bounds backed by domain adaptation and minimax theory.
Role: Collaborating Author.
π¨ Project
βCan We Make Feature-3DGS Faster, Better, and Smaller?β
We accelerate and shrink Feature-3DGS with semantic-aware Gaussian pruning and consistency loss, keeping fine details while boosting FPS and mIoU across Replica and Gopher/LindHall scenes.
π Honors and Awards
-
2025.05: Β πππ Selected for the Spring 2025 Deanβs List, University of Minnesota.
-
2023.10: Β πππ Achieved a silver (π₯) and a bronze (π₯) medal at the ICPC Asia Regional Contest.
π Educations
- present - 2027.6 (expected), Bachelor of Arts in Computer Science, University of Minnesota Twin Cities
π€ Academic Service
- Conferences: Reviewer for ICME 2025/2026, AAAI 2026, ICASSP 2026
- Journals: Reviewer for IEEE Transactions on Industrial Informatics (TII), IEEE Transactions on Neural Networks and Learning Systems (TNNLS).