Explore and filter language model benchmark results
Analyze text using tuned lens and visualize predictions
Explore and interact with HuggingFace LLM APIs using Swagger UI
Identify named entities in text
Track, rank and evaluate open Arabic LLMs and chatbots
Predict song genres from lyrics
Upload a PDF or TXT, ask questions about it
Demo emotion detection
Find collocations for a word in specified part of speech
Determine emotion from text
Submit model predictions and view leaderboard results
Explore BERT model interactions
Generative Tasks Evaluation of Arabic LLMs
Open Ko-LLM Leaderboard is a web-based platform designed for exploring and filtering benchmark results of language models (LLMs). It focuses on providing a comprehensive overview of model performance, particularly for Korean language models, enabling users to compare and evaluate different models based on various metrics and criteria.
• Benchmark Summaries: Access detailed performance metrics of various language models. • Advanced Filtering: Filter models by parameters like model size, architecture, and training data. • Performance Metrics: View metrics such as perplexity, accuracy, and F1-score across different tasks. • Model Comparison: Compare multiple models side-by-side to identify strengths and weaknesses. • Regular Updates: Stay informed with the latest benchmark results as new models are released. • User-Friendly Interface: Intuitive design for easy navigation and finding relevant information.
What is the purpose of the Open Ko-LLM Leaderboard?
The leaderboard aims to provide a centralized platform for comparing and evaluating the performance of Korean language models across various tasks and metrics.
How often is the leaderboard updated?
The leaderboard is updated regularly as new models are released and benchmarked.
Can I use the leaderboard for model selection?
Yes, the leaderboard is designed to help users select models based on specific requirements by providing detailed performance metrics and comparisons.