Browse and evaluate language models
Explore GenAI model efficiency on ML.ENERGY leaderboard
View and submit machine learning model evaluations
Determine GPU requirements for large language models
Retrain models for new data at edge devices
Display model benchmark results
Pergel: A Unified Benchmark for Evaluating Turkish LLMs
Explore and submit models using the LLM Leaderboard
Evaluate open LLMs in the languages of LATAM and Spain.
Convert a Stable Diffusion XL checkpoint to Diffusers and open a PR
Track, rank and evaluate open LLMs and chatbots
Export Hugging Face models to ONNX
Evaluate model predictions with TruLens
The Hebrew LLM Leaderboard is a comprehensive platform designed for benchmarking and evaluating language models specifically tailored for the Hebrew language. It provides a centralized repository where users can explore, compare, and analyze the performance of various large language models (LLMs) on Hebrew datasets and tasks. This tool is invaluable for researchers, developers, and professionals looking to identify the most suitable models for their Hebrew-based NLP applications.
What is the purpose of the Hebrew LLM Leaderboard?
The Hebrew LLM Leaderboard aims to simplify the process of identifying and evaluating language models for Hebrew-specific tasks, helping users make informed decisions.
How are the models evaluated?
Models are evaluated using standardized datasets and tasks specific to the Hebrew language, ensuring consistent and comparable results.
Is the Hebrew LLM Leaderboard suitable for non-experts?
Yes, the platform is designed to be user-friendly, with clear visualizations and explanations, making it accessible to both experts and non-experts.