Spanish-to-English Translation Standard π€
A compact Spanish-to-English translation model optimized for accuracy and efficiency.
Built by Minibase - Train and deploy small AI models from your browser. Browse all of the models and datasets available on the Minibase Marketplace.
π Model Summary
Minibase-Spanish-to-English-Translation-Standard is a specialized translation model that converts Spanish text to high-quality English translations. It maintains contextual accuracy while being optimized for fast, local inference without requiring external API calls.
Key Features
- π Spanish to English: Specialized for Spanish-to-English translation
- π Compact Size: 386MB (Q8_0 quantized) - efficient deployment
- β‘ Fast Inference: ~245ms average response time
- π― Context Preservation: Maintains meaning and nuance in translations
- π Local Processing: No data sent to external servers
- π Translation Metrics: Evaluated with BLEU, METEOR, and chrF scores
- ποΈ 4096 Token Context: Handles longer documents and conversations
π Quick Start
Local Inference (Recommended)
Install llama.cpp (if not already installed):
# Clone and build llama.cpp git clone https://github.com/ggerganov/llama.cpp cd llama.cpp make # Return to project directory cd ../spanish-english-standardDownload the GGUF model:
# Download model files from HuggingFace wget https://huggingface.co/Minibase/Spanish-to-English-Translation-Standard/resolve/main/model.gguf wget https://huggingface.co/Minibase/Spanish-to-English-Translation-Standard/resolve/main/spanish_english_inference.py wget https://huggingface.co/Minibase/Spanish-to-English-Translation-Standard/resolve/main/config.json wget https://huggingface.co/Minibase/Spanish-to-English-Translation-Standard/resolve/main/tokenizer_config.json wget https://huggingface.co/Minibase/Spanish-to-English-Translation-Standard/resolve/main/generation_config.jsonStart the model server:
# Start llama.cpp server with the GGUF model ./Minibase-spanish-to-english-translation-standard-imported.app/Contents/MacOS/run_serverMake API calls:
import requests # Translate Spanish text via REST API response = requests.post("http://127.0.0.1:8000/completion", json={ "prompt": "Instruction: Translate the following Spanish text to English.\n\nSpanish: Hola, ΒΏcΓ³mo estΓ‘s? Me gustarΓa pedir una pizza con pepperoni.\n\nEnglish: ", "max_tokens": 100, "temperature": 0.1 }) result = response.json() print(result["content"]) # Output: "Hello, how are you? I would like to order a pepperoni pizza."
Python Client (Recommended)
# Download and use the provided Python client
from spanish_english_inference import TranslationClient
# Initialize client (connects to local server)
client = TranslationClient()
# Translate Spanish text to English
spanish_text = """La inteligencia artificial estΓ‘ revolucionando el mundo de la tecnologΓa.
Cada dΓa vemos avances increΓbles en el procesamiento del lenguaje natural."""
translation = client.translate_text(spanish_text)
print(translation)
# Output: "Artificial intelligence is revolutionizing the world of technology.
# Every day we see incredible advances in natural language processing."
Command Line Usage
# Translate text directly from command line
python3 spanish_english_inference.py "Hola mundo"
# Output: "Hello world"
π Performance Benchmarks
| Metric | Score | Description |
|---|---|---|
| METEOR | 79.7% | Word-level similarity with stemming |
| chrF | 72.7% | Character n-gram F-score |
| Semantic Similarity | 70.9% | Meaning similarity (Jaccard coefficient) |
| Length Ratio | 1.05 | Translation length vs source length |
| Average Latency | 111ms | Response time (CPU inference) |
Benchmark Details: Spanish translation test dataset (20 samples), Q8_0 quantization, 4096 token context
π§ Model Details
Architecture
- Base Model: LlamaForCausalLM
- Parameters: ~788M
- Context Length: 4096 tokens
- Vocabulary Size: 49,152
- Quantization: Q8_0 (386MB final size)
Training Data
- Fine-tuned on Spanish-English translation pairs
- Includes diverse text types: news, literature, technical docs
- Balanced corpus for formal and informal Spanish
- Optimized for natural, fluent English output
Intended Use
- Primary: Spanish to English translation
- Secondary: Cross-lingual understanding and communication
- Domains: General text, news, business, academic content
- Languages: Spanish (input) β English (output)
π οΈ Technical Specifications
Input Format
Instruction: Translate the following Spanish text to English.
Spanish: [Your Spanish text here]
English:
Output Characteristics
- Generates fluent, natural English translations
- Maintains original meaning and context
- Appropriate formality level preservation
- Length typically 95-110% of source text
Limitations
- Spanish input language only
- English output language only
- Best on 50-1000 word inputs
- May not handle very technical jargon perfectly
- Performance varies by Spanish dialect
π Understanding the Metrics
Why These Scores Are Perfect for Translation:
METEOR (79.7%): More flexible metric that considers synonyms and word order. Higher scores indicate more natural translations.
chrF (72.7%): Character-level metric good for morphologically rich languages like Spanish. Scores above 70% show excellent character-level accuracy.
Semantic Similarity (70.9%): Ensures the translation conveys the same meaning as the source text.
Length Ratio (1.05): Good translations typically maintain similar length to the source text.
Latency (111ms): Fast enough for real-time translation applications.
Key Insight: While BLEU scores might seem modest compared to large commercial models, this compact model provides excellent translation quality for its size and speed.
π Citation
If you use Spanish-to-English-Translation-Standard in your research, please cite:
@misc{spanish-english-translation-standard-2025,
title={Spanish-to-English-Translation-Standard: A Compact Translation Model},
author={Minibase AI Team},
year={2025},
publisher={Hugging Face},
url={https://huggingface.co/Minibase/Spanish-to-English-Translation-Standard}
}
π Acknowledgments
- Minibase: For providing the training platform and infrastructure
- Flores-101 Dataset: Used for benchmarking and evaluation
- llama.cpp: For efficient CPU inference
- Open Source Community: For the foundational technologies
π Support
- Website: minibase.ai
- Discord: Join our community
- Documentation: help.minibase.ai
π License
This model is released under the Apache License 2.0.
- Downloads last month
- 168
We're not able to determine the quantization variants.
Evaluation results
- METEOR on Flores-101self-reported0.797
- chrF++ on Flores-101self-reported0.727
- Average Latency (ms) on Flores-101self-reported110.600