Tag: safety
All the articles with the tag "safety".
-
LLMs Evals: A General Framework for Custom Evaluations
A general framework for building rule-based and model-graded evaluations for LLM-based applications.
-
Malicious LLM Prompt Detection in Python
Building a malicious prompt detector using traditional ML classifiers in sklearn, trained on Deepset's prompt-injection dataset.