Explore and submit NER models
Label data for machine learning models
Multilingual metrics for the LMSys Arena Leaderboard
Browse and filter AI model evaluation results
Search and save datasets generated with a LLM in real time
Cluster data points using KMeans
Display and analyze PyTorch Image Models leaderboard
Explore and filter model evaluation results
Browse LLM benchmark results in various categories
Submit evaluations for speaker tagging and view leaderboard
Need to analyze data? Let a Llama-3.1 agent do it for you!
Leaderboard for text-to-video generation models
Analyze autism data and generate detailed reports
Clinical NER Leaderboard is a platform designed to evaluate and compare Named Entity Recognition (NER) models specifically within the clinical domain. It serves as a centralized hub for researchers and developers to explore, submit, and benchmark their NER models. The leaderboard provides transparency into model performance, fostering innovation and advancements in clinical NLP.
What types of models can I submit?
You can submit any NER model designed for clinical text processing, including rule-based, machine learning, and deep learning models.
How are models evaluated?
Models are evaluated using standardized metrics such as precision, recall, F1-score, and throughput on curated clinical datasets.
Can I access the datasets used for benchmarking?
Yes, the datasets used for benchmarking are available for download, allowing you to train and fine-tune your models effectively.