MMAISB 2025

The 1st International Workshop for Multi-Modal AI Safety Benchmark (MMAISB)

February 9 (Sunday), 2025, Nexus Resort & Spa Karambunai, Kota Kinabalu, Malaysia

In conjunction with the IEEE BigComp 2025 - 12th IEEE International Conference on Big Data and Smart Computing

Generative AI techniques have opened new business opportunities that can change and impact our ordinary lives. However, evaluating and measuring the safety and faithfulness of the AI-generated contents remains generally under-explored. The workshop of this year introduces some research efforts to construct benchmark datasets for evaluating safety and faithfulness of generative AI, currently undergoing in South Korea by TTA (Telecommunications Technology Association) with the collaboration of researchers and practitioners from KAIST, University of Seoul, Keimyung University, SelectStar, and Kakao. Our ultimate goal is to establish a methodology for constructing benchmark datasets, evaluation metrics, protocols and annotations to evaluate the safety and faithfulness of multimodal generative AI.


Organizing Committee

TBA


Keynote

TBA

Speaker


Program

※ This tentative program may be updated.

14:00 - 17:00, February 9 (Sunday), 2025

  • Welcoming Address
  • Multi-Modal Risk Detection: Combining Video and Image Processing for Safety Analysis
  • Dongkun Lee, Ho-Jin Choi (KAIST, Korea)
  • From Text to Speech: Building Multimodal Datasets for AI Safety Assessment
  • Jonghwan Hyeon, Ho-Jin Choi (KAIST, Korea)
  • A Recipe for Multilingual and Multi-Modal Red-Teaming
  • Young-Jun Lee, Ho-Jin Choi (KAIST, Korea)
  • An Analysis of Unsafe Responses Across Large Language Models
  • Eojin Joo, Ho-Jin Choi (KAIST, Korea)
  • Defining and Monitoring Potential Risks of Large Action Models and AI Agents
  • Yechan Hwang, Ho-Jin Choi (KAIST, Korea)
  • Enhancing Keyphrase Generation in AI Risk Evaluation Datasets through Prompt Engineering
  • Chae-Gyun Lim, Jeongyun Han, Ho-Jin Choi (KAIST, Korea)
  • A Quality Assurance Framework for Multimodal Assessment Datasets on AI Risk Factors
  • Seung-Ho Han, Jeongyun Han, Ho-Jin Choi (KAIST, Korea)
  • Exploring Korean Socio-Cultural Factors for Database Development and AI Risk Testing
  • Soohyun Cho (Keimyung University, Korea)
  • Coffee Break
  • Between Assistance and Reliance: Assessing the Fine Line in AI Dependence
  • Dasom Choi, Hyunseung Lim, HwaJung Hong (KAIST, Korea)
  • AI for the Vulnerable: Incorporating User Context in LLM Safety
  • Juhoon Lee, Yoojin Hong, Jeanne Choi, Yubin Choi, Joseph Seering (KAIST, Korea)
  • Exploring LMM Risk Assessment Datasets Across Diverse Modalities
  • Minhyeong An, Joyce Jiyoung Whang (KAIST, Korea)
  • An Empirical Study on Assessing LMM Privacy Risks and Exploring Measures to Mitigate Privacy Concerns
  • Hyunsoo Lee, Uichin Lee (KAIST, Korea)
  • An Empirical Study on AI Safety Assessment Using a Large Language Model
  • Kyunghoon Kim, Myungsik Ha, Sojung Lee (Kakao Crop., Korea)
  • Trustworthy AI: Ensuring Quality and Safety in Large Language Model (LLM) Application
  • Chansu Lee, Bohyun Kim (Selectstar Corp., Korea)
  • Closing Remarks

Contact

All questions about submissions should be emailed to chairs Chae-Gyun Lim (rayote@kaist.ac.kr).