SomeAI.org
  • Hot AI Tools
  • New AI Tools
  • AI Category
  • Free Submit
  • Find More AI Tools
SomeAI.org
SomeAI.org

Discover 10,000+ free AI tools instantly. No login required.

About

  • Blog

© 2025 • SomeAI.org All rights reserved.

  • Privacy Policy
  • Terms of Service
Home
Model Benchmarking
OR-Bench Leaderboard

OR-Bench Leaderboard

Measure over-refusal in LLMs using OR-Bench

You May Also Like

View All
🌎

Push Model From Web

Push a ML model to Hugging Face Hub

9
⚡

Goodharts Law On Benchmarks

Compare LLM performance across benchmarks

0
😻

2025 AI Timeline

Browse and filter machine learning models by category and modality

56
👀

Model Drops Tracker

Find recent high-liked Hugging Face models

33
🔥

OPEN-MOE-LLM-LEADERBOARD

Explore and submit models using the LLM Leaderboard

32
🏅

PTEB Leaderboard

Persian Text Embedding Benchmark

12
🥇

GIFT Eval

GIFT-Eval: A Benchmark for General Time Series Forecasting

64
🥇

Open Tw Llm Leaderboard

Browse and submit LLM evaluations

20
🧠

SolidityBench Leaderboard

SolidityBench Leaderboard

7
📊

DuckDB NSQL Leaderboard

View NSQL Scores for Models

7
🎨

SD-XL To Diffusers (fp16)

Convert a Stable Diffusion XL checkpoint to Diffusers and open a PR

5
🚀

stm32 model zoo app

Explore and manage STM32 ML models with the STM32AI Model Zoo dashboard

2

What is OR-Bench Leaderboard ?

OR-Bench Leaderboard is a tool designed to measure and benchmark over-refusal in Large Language Models (LLMs). It provides a standardized framework to evaluate when and how models refuse to answer questions or generate text. This leaderboard helps researchers and developers understand the limitations and safety mechanisms of LLMs by comparing their performance across different scenarios.

Features

• Model Comparison: Allows users to compare multiple models based on their refusal patterns.
• Refusal Trigger Evaluation: Tests models against a curated set of triggers to assess their refusal thresholds.
• Metric Aggregation: Provides aggregated metrics such as refusal rates and response patterns.
• Result Sharing: Enables sharing of benchmark results for community collaboration.
• Documented Methodology: Offers transparent documentation of evaluation methods and datasets.

How to use OR-Bench Leaderboard ?

  1. Visit the OR-Bench Leaderboard website: Access the platform to explore benchmarked models.
  2. Select a model: Choose from the list of available LLMs to view its performance.
  3. Input refusal triggers: Enter specific prompts or scenarios to test the model's refusal behavior.
  4. Run the benchmark: Execute the evaluation to generate results.
  5. Analyze metrics: Review refusal rates, response patterns, and other key metrics.
  6. Compare results: Use the leaderboard to compare results across different models.

Frequently Asked Questions

1. What is the purpose of OR-Bench Leaderboard?
The purpose of OR-Bench Leaderboard is to provide a standardized way to measure and compare over-refusal behaviors in LLMs, helping to identify models with balanced safety and utility.

2. Why is measuring over-refusal important?
Measuring over-refusal is important because it helps assess a model's ability to avoid harmful or inappropriate responses while still providing useful answers.

3. How can I interpret the results from OR-Bench Leaderboard?
Results show how often a model refuses to answer and under what conditions. Lower refusal rates may indicate a model that is more permissive, while higher rates suggest stricter safety mechanisms.

Recommended Category

View All
🧠

Text Analysis

✂️

Remove background from a picture

📋

Text Summarization

🎧

Enhance audio quality

🕺

Pose Estimation

📈

Predict stock market trends

👤

Face Recognition

🚫

Detect harmful or offensive content in images

🔍

Detect objects in an image

💻

Code Generation

🎎

Create an anime version of me

📄

Document Analysis

🤖

Create a customer service chatbot

😊

Sentiment Analysis

🔊

Add realistic sound to a video