AIR-ML
Home
Research
News
Team
Project
Publication
Contact
Article
SplineFlow: Flow Matching for Dynamical Systems with B-Spline Interpolants
We introduce SplineFlow, a B-spline based flow matching algorithm for modeling complex dynamical systems.
Santanu Rathod
,
Pietro Liò
,
Xiao Zhang
PDF
Cite
Code
ArXiv
Jailbreaking Attacks vs. Content Safety Filters: How Far Are We in the LLM Safety Arms Race?
We present the first systematic evaluation of jailbreak attacks targeting LLM safety alignment, assessing their success across the full inference pipeline, including both input and output filtering stages.
Yuan Xin
,
Dingfan Chen
,
Linyi Yang
,
Michael Backes
,
Xiao Zhang
PDF
Cite
ArXiv
GREAT: Generalizable Backdoor Attacks in RLHF via Emotion-Aware Trigger Synthesis
we develop a novel framework for crafting generalizable backdoors in RLHF through emotion-aware trigger synthesis
Subrat Kishore Dutta
,
Yuelin Xu
,
Piyush Pant
,
Xiao Zhang
PDF
Cite
ArXiv
ContextFlow: Context-Aware Flow Matching For Trajectory Inference From Spatial Omics Data
We propose ContextFlow, a novel context-aware flow matching framework that incorporates prior knowledge to guide the inference of structural tissue dynamics from spatially resolved omics data.
Santanu Rathod
,
Francesco Ceccarelli
,
Sean B. Holden
,
Pietro Liò
,
Xiao Zhang
,
Jovan Tanevski
PDF
Cite
Code
ArXiv
Controllable Adversarial Makeup for Privacy via Text-Guided Diffusion
We introduce MASQUE, a diffusion-based framework that generates localized adversarial makeup guided by user-defined text prompts.
Youngjin Kwon
,
Xiao Zhang
PDF
Cite
Code
ArXiv
Generalizable Targeted Data Poisoning against Varying Physical Objects
We take the first step toward understanding the real-world threats of TDP by studying its generalizability across varying physical conditions.
Zhizhen Chen
,
Zhengyu Zhao
,
Subrat Kishore Dutta
,
Chenhao Lin
,
Chao Shen
,
Xiao Zhang
PDF
Cite
Code
ArXiv
Safe in Isolation, Dangerous Together: Agent-Driven Multi-Turn Decomposition Jailbreaks on LLMs
We propose a multi-agent, multi-turn jailbreak strategy that systematically bypasses LLM safety mechanisms by decomposing harmful queries into seemingly benign sub-tasks.
Devansh Srivastav
,
Xiao Zhang
PDF
Cite
Code
Source Document
DiffCAP: Diffusion-based Cumulative Adversarial Purification for Vision Language Models
This paper introduces DiffCAP, a novel diffusion-based purification strategy that can effectively neutralize adversarial corruptions in VLMs.
Jia Fu
,
Yongtao Wu
,
Yihang Chen
,
Kunyu Peng
,
Xiao Zhang
,
Volkan Cevher
,
Sepideh Pashami
,
Anders Holst
PDF
Cite
ArXiv
Predicting Time-varying Flux and Balance in Metabolic Systems using Structured Neural ODE Processes
We propose a structured neural ODE process model to estimate flux and balance samples using gene-expression time-series data
Santanu Rathod
,
Pietro Liò
,
Xiao Zhang
PDF
Cite
ArXiv
OpenReview
Invisibility Cloak: Disappearance under Human Pose Estimation via Backdoor Attacks
We propose a general framework to craft invisibility cloak for human pose estimation models
Minxing Zhang
,
Michael Backes
,
Xiao Zhang
PDF
Cite
ArXiv
»
Cite
×