AIware 2025
Wed 19 - Thu 20 November 2025
co-located with ASE 2025

This program is tentative and subject to change.

Wed 19 Nov 2025 10:00 - 10:08 at Grand Hall 1 - AIware & Security

Safety alignment is critical for AI-powered systems. While recent LLM-powered guardrail approaches such as LlamaGuard achieve high detection accuracy of unsafe inputs written in English (e.g., ``How to create a bomb?''), they struggle with multilingual unsafe inputs. This limitation leaves LLM systems vulnerable to unsafe and jailbreak prompts written in low-resource languages such as those in Southeast Asia. This paper introduces SEALGuard, a multilingual guardrail designed to improve the safety alignment across diverse languages. It aims to address the multilingual safety alignment gap of existing guardrails and ensure effective filtering of unsafe and jailbreak prompts in AI-powered systems. We adapt a general-purpose multilingual language model into a multilingual guardrail using low-rank adaptation (LoRA). We construct SEALSBench, a large-scale multilingual safety alignment dataset containing over 260,000 prompts in ten languages, including safe, unsafe, and jailbreak cases. We evaluate SEALGuard against state-of-the-art guardrails such as LlamaGuard on this benchmark. Our findings show that multilingual unsafe and jailbreak prompts substantially degrade the performance of the state-of-the-art LlamaGuard, which experiences a drop in Defense Success Rate (DSR) by 9% and 18%, respectively, compared to its performance on English-only prompts. In contrast, SEALGuard outperforms existing guardrails in detecting multilingual unsafe and jailbreak prompts, improving DSR by 48% over LlamaGuard and achieving the best DSR, precision, and F1-score. Our ablation study further reveals the contributions of adaptation strategies and model size to the overall performance of SEALGuard. We release our pre-trained model and benchmark at https://github.com/awsm-research/SEALGuard to support further research.

Preprint (Multilingual_Guardrails.pdf)6.88MiB
SEALGuard Slide Deck (AIWare2025_SEALGuard_SlideDeck.pdf)6.2MiB

This program is tentative and subject to change.

Wed 19 Nov

Displayed time zone: Seoul change

09:20 - 10:30
AIware & SecurityMain Track at Grand Hall 1
09:20
8m
Talk
CHASE: LLM Agents for Dissecting Malicious PyPI Packages
Main Track
Takaaki Toda Waseda University, Tatsuya Mori Waseda University
File Attached
09:28
8m
Talk
CFCEval: Evaluating Security Aspects in Code Generated by Large Language Models
Main Track
Cheng Cheng Concordia University, Jinqiu Yang Concordia University
Pre-print
09:36
8m
Talk
Security in the Wild: An Empirical Analysis of LLM-Powered Applications and Local Inference Frameworks
Main Track
Julia Gomez-Rangel Texas A&M University - Corpus Christi, Young Lee Texas A & M University - San Antonio, Bozhen Liu Texas A&M University - Corpus Christi
Pre-print
09:44
8m
Talk
How Quantization Impacts Privacy Risk on LLMs for Code?
Main Track
Md Nazmul Haque North Carolina State University, Hua yang North Carolina State University, Zhou Yang University of Alberta, Alberta Machine Intelligence Institute , Bowen Xu North Carolina State University
Pre-print
09:52
8m
Talk
Securing the Multi-Chain Ecosystem: A Unified, Agent-Based Framework for Vulnerability Repair in Solidity and Move
Main Track
Rabimba Karanjai University of Houston, Lei Xu Kent State University, Weidong Shi University of Houston
10:00
8m
Talk
SEALGuard: Safeguarding the Multilingual Conversations in Southeast Asian Languages for AI-Powered Software
Main Track
Wenliang Shan Monash University, Michael Fu The University of Melbourne, Rui Yang Monash University and Transurban, Kla Tantithamthavorn Monash University and Atlassian
Pre-print File Attached
10:10
20m
Live Q&A
Joint Q&A and Discussion #AISecurity
Main Track