Explore and benchmark visual document retrieval models
Quantize a model for faster inference
Track, rank and evaluate open LLMs and chatbots
Evaluate adversarial robustness using generative models
Convert a Stable Diffusion XL checkpoint to Diffusers and open a PR
Download a TriplaneGaussian model checkpoint
Compare and rank LLMs using benchmark scores
Explore and manage STM32 ML models with the STM32AI Model Zoo dashboard
Visualize model performance on function calling tasks
Create and manage ML pipelines with ZenML Dashboard
Create and upload a Hugging Face model card
Teach, test, evaluate language models with MTEB Arena
Compare audio representation models using benchmark results
Vidore Leaderboard is a tool designed for exploring and benchmarking visual document retrieval models. It provides a platform to compare and evaluate the performance of different models in the domain of visual document retrieval, helping users understand their strengths and weaknesses.
• Comprehensive Model Database: Access a wide range of pre-trained models for visual document retrieval. • Customizable Benchmarking: Define custom benchmarks to evaluate models based on specific criteria. • Performance Metrics: Detailed metrics to assess model accuracy, efficiency, and robustness. • Visual Results: Interactive visualizations to compare model performance side-by-side. • Community Sharing: Share benchmark results and insights with the broader AI research community.
What is visual document retrieval?
Visual document retrieval involves systems that retrieve documents based on visual content, such as images or layouts, rather than text-based search.
How do I interpret the performance metrics?
Performance metrics are provided in an easy-to-understand format, with visual charts and numerical scores to help compare model effectiveness.
Can I use Vidore Leaderboard for non-public models?
Yes, Vidore Leaderboard supports benchmarking private models by uploading them through the platform or API.