Skip to content

Latest commit

 

History

History
15 lines (12 loc) · 442 Bytes

README.md

File metadata and controls

15 lines (12 loc) · 442 Bytes

MODEL SAFETY

Evaluate safety of local LLMs.

Build

Run

  • eval_model.py
python src\eval_model.py

This program reads unsafe prompts from the data directory and asks the llm to classify if they are safe or unsafe. The output is generated in the output directory. Use the -h (--help) flag for usage details.