Skip to content

Latest commit

 

History

History
59 lines (45 loc) · 2.96 KB

README.md

File metadata and controls

59 lines (45 loc) · 2.96 KB

Multimodal Situational Safety

Kaiwen Zhou*1, Chengzhi Liu*1, Xuandong Zhao2, Anderson Compalas1, Dawn Song2, Xin Eric Wang†1

1University of California, Santa Cruz, 2University of California, Berkley

*Equal contribution

Teaser figure

Dataset Structure

The Dataset can be downloaded from Hugging Face.

Each entry in the Chat Task dataset contains the following fields:

  • safe_image_path: the file path to the safe image.
  • intent: The user's intent in the context of images.
  • unsafe_image: The description of unsafe image.
  • unsafe_image_path: the file path to the unsafe image.
  • Type: The multimodal situational safety category of the entry.
  • queries: The user's question in Chat Task.

Each entry in the Embodied Task dataset contains the following fields:

  • task: the specific embodied task.
  • category: The multimodal situational safety category of the entry.
  • safe_instruction/safe_instructions: The user's safe instructions and related variations.
  • unsafe_instruction/unsafe_instructions: The user's unsafe instructions and related variations.
  • safe: the file path to the safe image.
  • unsafe: the file path to the unsafe image.

Figure 1 Figure 2

Evaluation

You can evaluate different MLLMs by running our evaluation code inference.py and changing the "--mllm" parameter:

python inference.py --mllm gemini --data_root xxx --output_dir xxx

The deployment of the model can refer to models. For proprietary models, please set up your API key first.

Citation

@misc{zhou2024multimodalsituationalsafety,
      title={Multimodal Situational Safety}, 
      author={Kaiwen Zhou and Chengzhi Liu and Xuandong Zhao and Anderson Compalas and Dawn Song and Xin Eric Wang},
      year={2024},
      eprint={2410.06172},
      archivePrefix={arXiv},
      primaryClass={cs.AI},
      url={https://arxiv.org/abs/2410.06172}, 
}