Browse and submit LLM evaluations
Submit deepfake detection models for evaluation
Browse and submit evaluations for CaselawQA benchmarks
Push a ML model to Hugging Face Hub
Visualize model performance on function calling tasks
View and submit machine learning model evaluations
Determine GPU requirements for large language models
View NSQL Scores for Models
Submit models for evaluation and view leaderboard
Generate leaderboard comparing DNA models
Browse and evaluate ML tasks in MLIP Arena
Teach, test, evaluate language models with MTEB Arena
Export Hugging Face models to ONNX
The Open Medical-LLM Leaderboard is a comprehensive platform designed for benchmarking and comparing large language models (LLMs) specifically tailored for medical and healthcare applications. It provides a centralized hub where users can browse, evaluate, and submit their own model evaluations, fostering transparency and collaboration in the development of AI for medical use cases.
What types of medical applications are supported?
The Open Medical-LLM Leaderboard supports a wide range of medical applications, including clinical text analysis, medical question answering, and healthcare document summarization.
How do I submit my own model evaluation?
To submit your model evaluation, follow these steps:
Is the leaderboard open to non-experts?
Yes, the leaderboard is designed to be accessible to both experts and non-experts. Researchers, developers, and healthcare professionals can all benefit from the platform's resources and tools.