MTEM Pruner
Multilingual Text Embedding Model Pruner
You May Also Like
View AllContextualBench-Leaderboard
View and submit language model evaluations
Merge Lora
Merge Lora adapters with a base model
stm32 model zoo app
Explore and manage STM32 ML models with the STM32AI Model Zoo dashboard
Submission Portal
Evaluate and submit AI model results for Frugal AI Challenge
OR-Bench Leaderboard
Evaluate LLM over-refusal rates with OR-Bench
Titanic Survival in Real Time
Calculate survival probability based on passenger details
Memorization Or Generation Of Big Code Model Leaderboard
Compare code model performance on benchmarks
Ilovehf
View RL Benchmark Reports
OR-Bench Leaderboard
Measure over-refusal in LLMs using OR-Bench
Vidore Leaderboard
Explore and benchmark visual document retrieval models
InspectorRAGet
Evaluate RAG systems with visual analytics
AICoverGen
Launch web-based model application
What is MTEM Pruner ?
MTEM Pruner is an advanced tool designed for pruning multilingual text embedding models. It allows users to .optimize large multilingual models by focusing on a specific language or set of languages. This makes the model more efficient and lightweight while maintaining high performance for the target language(s). MTEM Pruner is particularly useful for developers and researchers working on model benchmarking and fine-tuning.
Features
- Efficient Pruning: Reduces model size while preserving key performance metrics for the target language.
- Multilingual Support: Works with a wide range of languages, enabling customized pruning for specific use cases.
- Integration with Hugging Face: Supports popular libraries and frameworks for seamless implementation.
- Benchmarking Tools: Provides detailed metrics to evaluate model performance before and after pruning.
- User-Friendly Interface: Simplifies the pruning process with intuitive API and commands.
How to use MTEM Pruner ?
-
Install the Tool
Install MTEM Pruner using pip or directly from source:pip install mtem-pruner -
Import the Library
Load the required libraries and initialize the pruner:from mtem_pruner import MTEMPruner pruner = MTEMPruner() -
Load the Model
Load the pre-trained multilingual model you want to prune:model = AutoModel.from_pretrained("your_multilingual_model") -
Define Pruning Parameters
Specify the target language(s) and pruning settings:params = { "target_language": "en", "pruning_ratio": 0.5, "device": "cuda" } -
Perform Pruning
Apply the pruning process to the model:pruned_model = pruner.prune_model(model, **params) -
Export the Pruned Model
Save the pruned model for deployment or further use:pruned_model.save_pretrained("pruned_model_directory") -
Deploy the Model
Use the pruned model in your application, benefiting from reduced size and optimized performance.
Frequently Asked Questions
What models does MTEM Pruner support?
MTEM Pruner is compatible with most multilingual text embedding models, including popular ones like Multilingual BERT, DistilBERT, and XLM-RoBERTa.
Can I prune the model for more than one language?
Yes, MTEM Pruner allows you to define multiple target languages. Simply specify them in the target_language parameter as a list:
target_language: ["en", "es", "fr"]
How do I choose the optimal pruning ratio?
The pruning ratio depends on your specific needs. Start with a lower ratio (e.g., 0.3) and evaluate performance. Gradually increase the ratio while monitoring accuracy and model size to find the best balance for your use case.