Extract information from Indonesian receipts
Ask questions about PDF documents
Submit your Hugging Face username to check certification progress
Convert files to Markdown and extract metadata
Generate documentation for Hugging Face spaces
Parse PDF to extract trip data and metadata
Extract structured data from documents using images
Extract text and metadata from PDF files
Ask questions about a PDF file
Show evaluation results on a leaderboard
Upload documents and chat with a smart assistant based on them
Search ChatGPT-related repositories
This space contains 4 usecases in Law Domain.
Donut Base Finetuned Cord V2 is a specialized AI model designed for document analysis, particularly optimized for extracting information from Indonesian receipts. It leverages advanced computer vision and natural language processing techniques to accurately identify and extract key details such as vendor names, dates, totals, and item descriptions. This model is fine-tuned for Indonesian language support, making it highly effective for processing receipts in Indonesian.
• Indonesian language support: Optimized for receipts written in Indonesian.
• High accuracy: Advanced fine-tuning ensures precise extraction of relevant information.
• Multi-format handling: Works with scanned images, digital receipts, and even rotated or blurry documents.
• Comprehensive extraction: Capable of extracting vendor names, dates, totals, and item details.
• User-friendly integration: Easily integrates with existing document processing workflows.
pip install donut
from donut import Donut
model = Donut.load('donut-base-finetuned-cord-v2')
image_path = 'path_to_your_receipt_image.jpg'
result = model.infer(image_path)
What languages does Donut Base Finetuned Cord V2 support?
Donut Base Finetuned Cord V2 is primarily optimized for Indonesian receipts but can also process basic English text within documents.
Can the model handle blurry or rotated images?
Yes, the model is designed to handle rotated, blurry, or low-quality images. However, the accuracy may improve with clearer input images.
How do I access or use this model?
You can use the Donut library to load and run the model. Install the library using pip install donut and follow the usage instructions provided.