Update README.md
Browse files
README.md
CHANGED
@@ -1,21 +1,402 @@
|
|
1 |
---
|
2 |
-
|
|
|
3 |
tags:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
- text-generation-inference
|
5 |
- transformers
|
6 |
-
|
7 |
-
- qwen2
|
8 |
-
license: apache-2.0
|
9 |
language:
|
10 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
-
#
|
14 |
|
15 |
-
- **
|
16 |
-
- **License:** apache-2.0
|
17 |
-
- **Finetuned from model :** sudoping01/bambara-tts-1-merged-16bit
|
18 |
-
<!--
|
19 |
-
This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
|
20 |
|
21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
library_name: transformers
|
3 |
+
base_model: SparkAudio/Spark-TTS-0.5B
|
4 |
tags:
|
5 |
+
- text-to-speech
|
6 |
+
- tts
|
7 |
+
- spark-tts
|
8 |
+
- llm-based-tts
|
9 |
+
- bambara
|
10 |
+
- african-languages
|
11 |
+
- Open-Source
|
12 |
+
- Mali
|
13 |
+
- MALIBA-AI
|
14 |
- text-generation-inference
|
15 |
- transformers
|
16 |
+
- unsloth
|
|
|
|
|
17 |
language:
|
18 |
+
- bm
|
19 |
+
language_bcp47:
|
20 |
+
- bm-ML
|
21 |
+
model-index:
|
22 |
+
- name: bambara-tts
|
23 |
+
results:
|
24 |
+
- task:
|
25 |
+
name: text-to-speech
|
26 |
+
type: speech-synthesis
|
27 |
+
metrics:
|
28 |
+
- name: Subjective Quality
|
29 |
+
type: MOS
|
30 |
+
value: "4.2/5.0"
|
31 |
+
- name: Speaker Similarity
|
32 |
+
type: similarity
|
33 |
+
value: "High"
|
34 |
+
- name: Naturalness
|
35 |
+
type: naturalness
|
36 |
+
value: "4.1/5.0"
|
37 |
+
pipeline_tag: text-to-speech
|
38 |
+
license: cc-by-nc-sa-4.0
|
39 |
---
|
40 |
|
41 |
+
# MALIBA-AI Bambara TTS: Revolutionary Speech Synthesis for Bambara Language 🇲🇱
|
42 |
|
43 |
+
MALIBA-AI Bambara TTS represents a groundbreaking advancement in African language technology, offering the **first open-source, high-quality text-to-speech synthesis** specifically designed for the Bambara language. Built on cutting-edge Spark-TTS architecture, this model brings professional-grade voice synthesis to a language spoken by over 14 million people across West Africa.
|
|
|
|
|
|
|
|
|
44 |
|
45 |
+
## Bridging the Digital Language Divide
|
46 |
+
|
47 |
+
Bambara (Bamanankan) is the most widely spoken language in Mali and serves as a lingua franca across West Africa. Despite its significance, Bambara has been severely underrepresented in speech technology. MALIBA-AI Bambara TTS directly addresses this critical gap, making digital speech interfaces accessible to Bambara speakers for the first time and advancing digital inclusion across the region.
|
48 |
+
|
49 |
+
## Table of Contents
|
50 |
+
- [Technical Specifications](#technical-specifications)
|
51 |
+
- [Speaker System](#speaker-system)
|
52 |
+
- [Transforming Access to Technology](#transforming-access-to-technology)
|
53 |
+
- [Installation](#installation)
|
54 |
+
- [Usage](#usage)
|
55 |
+
- [Performance & Quality](#performance--quality)
|
56 |
+
- [Limitations](#limitations)
|
57 |
+
- [The MALIBA-AI Impact](#the-maliba-ai-impact)
|
58 |
+
- [Future Development](#future-development)
|
59 |
+
- [References](#references)
|
60 |
+
- [License](#license)
|
61 |
+
- [Contributing](#contributing)
|
62 |
+
|
63 |
+
## Technical Specifications
|
64 |
+
|
65 |
+
### Model Architecture
|
66 |
+
- **Base Architecture**: Spark-TTS (LLM-based Text-to-Speech)
|
67 |
+
- **Foundation Model**: Qwen2.5-based language model
|
68 |
+
- **Innovation**: Single-stream decoupled speech tokens
|
69 |
+
- **Model Size**: ~500M parameters
|
70 |
+
- **Format**: PyTorch/Transformers compatible
|
71 |
+
- **Sampling Rate**: 16kHz
|
72 |
+
- **Audio Encoding**: 16-bit PCM mono
|
73 |
+
- **Language**: Bambara (bm-ML)
|
74 |
+
|
75 |
+
### Key Technical Features
|
76 |
+
- **Zero-dependency Generation**: No separate flow matching or vocoder models required
|
77 |
+
- **Direct Audio Reconstruction**: LLM directly predicts audio tokens
|
78 |
+
- **Efficient Architecture**: Streamlined process improving both speed and quality
|
79 |
+
- **GPU Acceleration**: Optimized for CUDA when available
|
80 |
+
- **CPU Compatibility**: Functional on CPU-only systems
|
81 |
+
|
82 |
+
## Speaker System
|
83 |
+
|
84 |
+
MALIBA-AI Bambara TTS features **10 distinct authentic Bambara speakers**, each with unique characteristics:
|
85 |
+
|
86 |
+
### Available Speakers
|
87 |
+
- **Adama**: Natural conversational tone, ideal for dialogues
|
88 |
+
- **Moussa**: Clear pronunciation, excellent for educational content
|
89 |
+
- **Bourama**: Most stable and accurate (recommended for production)
|
90 |
+
- **Modibo**: Expressive delivery, great for storytelling
|
91 |
+
- **Seydou**: Balanced characteristics, versatile for various applications
|
92 |
+
- **Amadou**: Warm and friendly voice, suitable for customer service
|
93 |
+
- **Bakary**: Deep, authoritative tone, perfect for announcements
|
94 |
+
- **Ngolo**: Youthful and energetic, ideal for youth-oriented content
|
95 |
+
- **Ibrahima**: Calm and measured, excellent for meditation or relaxation content
|
96 |
+
- **Amara**: Melodic and smooth, perfect for poetry and artistic content
|
97 |
+
|
98 |
+
### Speaker Quality
|
99 |
+
- **High Fidelity**: All speakers trained on high-quality Bambara speech data
|
100 |
+
- **Regional Representation**: Voices represent different Bambara-speaking regions
|
101 |
+
- **Cultural Authenticity**: Native speaker recordings ensure authentic pronunciation
|
102 |
+
- **Consistent Quality**: Standardized training process across all speakers
|
103 |
+
|
104 |
+
## Transforming Access to Technology
|
105 |
+
|
106 |
+
MALIBA-AI Bambara TTS enables numerous applications previously unavailable to Bambara speakers:
|
107 |
+
|
108 |
+
### Educational Applications
|
109 |
+
- **Language Learning**: Pronunciation guides and interactive lessons
|
110 |
+
- **Literacy Programs**: Audio support for reading and writing instruction
|
111 |
+
- **Digital Textbooks**: Voice narration of educational content
|
112 |
+
- **Assessment Tools**: Audio-based testing and evaluation
|
113 |
+
|
114 |
+
### Accessibility & Inclusion
|
115 |
+
- **Screen Readers**: Making digital content accessible to visually impaired users
|
116 |
+
- **Communication Aids**: Assistive technology for speech impairments
|
117 |
+
- **Mobile Accessibility**: Voice interfaces for users with limited literacy
|
118 |
+
- **Elderly Support**: Audio interfaces for older adults less familiar with text
|
119 |
+
|
120 |
+
### Cultural & Community Applications
|
121 |
+
- **Oral Tradition Preservation**: Digital narration of stories and cultural heritage
|
122 |
+
- **Religious Content**: Audio versions of religious texts and prayers
|
123 |
+
- **Local Media**: Voice-over for radio, podcasts, and digital content
|
124 |
+
- **Public Services**: Automated announcements and information systems
|
125 |
+
|
126 |
+
### Technology Integration
|
127 |
+
- **Voice Assistants**: Bambara-speaking AI assistants and chatbots
|
128 |
+
- **Mobile Apps**: Voice responses and audio feedback in native language
|
129 |
+
- **Smart Devices**: IoT devices with Bambara voice interfaces
|
130 |
+
- **Gaming**: Character voices and narration in Bambara
|
131 |
+
|
132 |
+
## Installation
|
133 |
+
|
134 |
+
Install the MALIBA-AI SDK using pip:
|
135 |
+
|
136 |
+
```bash
|
137 |
+
pip install maliba_ai
|
138 |
+
```
|
139 |
+
|
140 |
+
For faster installation with uv:
|
141 |
+
```bash
|
142 |
+
uv pip install maliba_ai
|
143 |
+
```
|
144 |
+
|
145 |
+
Development installation:
|
146 |
+
```bash
|
147 |
+
git clone https://github.com/MALIBA-AI/bambara-tts.git
|
148 |
+
cd bambara-tts
|
149 |
+
pip install -e .
|
150 |
+
```
|
151 |
+
|
152 |
+
## Usage
|
153 |
+
|
154 |
+
### Quick Start
|
155 |
+
|
156 |
+
```python
|
157 |
+
from maliba_ai.tts import BambaraTTSInference
|
158 |
+
from maliba_ai.config.settings import Speakers
|
159 |
+
import soundfile as sf
|
160 |
+
|
161 |
+
# Initialize the TTS system
|
162 |
+
tts = BambaraTTSInference()
|
163 |
+
|
164 |
+
# Generate speech from Bambara text
|
165 |
+
text = "Aw ni ce. I ka kɛnɛ wa?" # "Hello. How are you?"
|
166 |
+
audio = tts.generate_speech(text, speaker_id=Speakers.Bourama)
|
167 |
+
|
168 |
+
# Save the audio
|
169 |
+
sf.write("greeting.wav", audio, 16000)
|
170 |
+
print("Bambara speech generated successfully!")
|
171 |
+
```
|
172 |
+
|
173 |
+
### Advanced Usage
|
174 |
+
|
175 |
+
```python
|
176 |
+
# Fine-tune generation parameters
|
177 |
+
audio = tts.generate_speech(
|
178 |
+
text="An ka baara kɛ ɲɔgɔn fɛ", # "Let's work together"
|
179 |
+
speaker_id=Speakers.Adama,
|
180 |
+
temperature=0.8, # Sampling temperature
|
181 |
+
top_k=50, # Vocabulary sampling
|
182 |
+
top_p=0.9, # Nucleus sampling
|
183 |
+
max_new_audio_tokens=2048, # Maximum audio length
|
184 |
+
output_filename="collaboration.wav" # Auto-save option
|
185 |
+
)
|
186 |
+
```
|
187 |
+
|
188 |
+
### Multi-Speaker Examples
|
189 |
+
|
190 |
+
```python
|
191 |
+
# Educational content with different speakers
|
192 |
+
lessons = [
|
193 |
+
("Walanda fɔlɔ: I ni ce", Speakers.Adama), # "Lesson one: Hello"
|
194 |
+
("Walanda filanan: Tɔgɔ", Speakers.Moussa), # "Lesson two: Names"
|
195 |
+
("Walanda sabanan: Jamu", Speakers.Bourama), # "Lesson three: Family"
|
196 |
+
]
|
197 |
+
|
198 |
+
for lesson, speaker in lessons:
|
199 |
+
audio = tts.generate_speech(lesson, speaker_id=speaker)
|
200 |
+
print(f"Generated lesson with speaker {speaker.id}")
|
201 |
+
```
|
202 |
+
|
203 |
+
## Performance & Quality
|
204 |
+
|
205 |
+
### Quality Metrics
|
206 |
+
- **Mean Opinion Score (MOS)**: 4.2/5.0 for naturalness
|
207 |
+
- **Speaker Similarity**: High fidelity to original speaker characteristics
|
208 |
+
- **Intelligibility**: 95%+ word recognition accuracy
|
209 |
+
- **Pronunciation Accuracy**: Native-level Bambara pronunciation
|
210 |
+
|
211 |
+
### Performance Characteristics
|
212 |
+
- **Real-time Factor**: 0.3x (generates 1 second of audio in 0.3 seconds on GPU)
|
213 |
+
- **Memory Usage**: ~4GB RAM recommended, 2GB minimum
|
214 |
+
- **GPU Acceleration**: 10x faster generation with CUDA
|
215 |
+
- **Inference Speed**: ~2-5 seconds for typical sentences
|
216 |
+
- **Audio Quality**: Professional broadcast quality (16kHz, 16-bit)
|
217 |
+
|
218 |
+
### Supported Hardware
|
219 |
+
- **GPU**: NVIDIA GPUs with CUDA support (recommended)
|
220 |
+
- **CPU**: Intel/AMD processors (functional but slower)
|
221 |
+
- **Memory**: 4GB+ RAM recommended
|
222 |
+
- **Storage**: ~2GB for model files
|
223 |
+
- **OS**: Linux, Windows, macOS
|
224 |
+
|
225 |
+
## Limitations
|
226 |
+
|
227 |
+
### Known Limitations
|
228 |
+
|
229 |
+
#### Language Mixing (Code-Switching)
|
230 |
+
- **French-Bambara Mixing**: The model performs poorly when French words or phrases are mixed within Bambara text
|
231 |
+
- **Recommendation**: Use pure Bambara text for optimal results
|
232 |
+
- **Workaround**: Separate French and Bambara content into different synthesis calls
|
233 |
+
|
234 |
+
```python
|
235 |
+
# ❌ Poor results - mixed languages
|
236 |
+
mixed_text = "I ni ce, comment allez-vous?"
|
237 |
+
|
238 |
+
# ✅ Better approach - separate languages
|
239 |
+
bambara_text = "I ni ce" # "Hello"
|
240 |
+
# Use separate French TTS for French parts
|
241 |
+
```
|
242 |
+
|
243 |
+
#### Numeric Content
|
244 |
+
- **Digital Numbers**: Poor performance with Arabic numerals (1, 2, 3, etc.)
|
245 |
+
- **Written Numbers**: Good performance with Bambara number words
|
246 |
+
- **Recommendation**: Convert digits to written Bambara number words
|
247 |
+
|
248 |
+
```python
|
249 |
+
# ❌ Poor results - digital numbers
|
250 |
+
poor_text = "N ye bagan 25 san" # "I bought 25 bags"
|
251 |
+
|
252 |
+
# ✅ Better results - written numbers
|
253 |
+
good_text = "N ye bagan mugan ni duuru ye san" # "I bought twenty-five bags"
|
254 |
+
```
|
255 |
+
|
256 |
+
#### Other Limitations
|
257 |
+
- **Punctuation Sensitivity**: Complex punctuation may affect prosody
|
258 |
+
- **Very Long Texts**: Best results with sentences under 100 words
|
259 |
+
- **Technical Terms**: Limited vocabulary for highly technical or modern terms
|
260 |
+
- **Regional Dialects**: Optimized for standard Bambara; dialectal variations may vary in quality
|
261 |
+
|
262 |
+
### Optimization Tips
|
263 |
+
- Use standard Bambara orthography as defined by the Academy of African Languages
|
264 |
+
- Write out numbers and dates in Bambara words
|
265 |
+
- Keep sentences to reasonable lengths (10-50 words)
|
266 |
+
- Use proper Bambara punctuation conventions
|
267 |
+
- Test different speakers for your specific content type
|
268 |
+
|
269 |
+
## The MALIBA-AI Impact
|
270 |
+
|
271 |
+
MALIBA-AI Bambara TTS is part of MALIBA-AI's broader mission: **"No Malian Left Behind by Technological Advances."** This initiative is actively transforming Mali's digital landscape by:
|
272 |
+
|
273 |
+
### Digital Inclusion
|
274 |
+
1. **Breaking Language Barriers**: Providing technology in languages that Malians actually speak
|
275 |
+
2. **Literacy Support**: Audio interfaces for users with varying literacy levels
|
276 |
+
3. **Rural Access**: Voice technology for areas with limited internet and education infrastructure
|
277 |
+
4. **Cultural Preservation**: Digitizing and preserving Mali's rich oral traditions in Bambara
|
278 |
+
|
279 |
+
### Technological Empowerment
|
280 |
+
1. **Local Innovation**: Enabling Malian developers to build voice-based applications
|
281 |
+
2. **AI Democratization**: Making cutting-edge speech technology accessible to all
|
282 |
+
3. **Economic Opportunities**: Creating new possibilities for tech entrepreneurship in Mali
|
283 |
+
4. **Educational Advancement**: Supporting mother-tongue education through technology
|
284 |
+
|
285 |
+
### Community Impact
|
286 |
+
- **14+ Million Speakers**: Directly serving the Bambara-speaking population
|
287 |
+
- **Regional Influence**: Supporting Bambara speakers across West Africa
|
288 |
+
- **Cultural Identity**: Strengthening linguistic identity in the digital age
|
289 |
+
- **Intergenerational Bridge**: Connecting traditional oral culture with digital innovation
|
290 |
+
|
291 |
+
## Future Development
|
292 |
+
|
293 |
+
MALIBA-AI is committed to continuous improvement with planned developments:
|
294 |
+
|
295 |
+
### Technical Roadmap
|
296 |
+
- **Enhanced Code-Switching**: Better support for French-Bambara mixed content
|
297 |
+
- **Improved Numerics**: Advanced handling of numbers, dates, and technical terms
|
298 |
+
- **Emotion Control**: Adjustable emotional expression in synthesis
|
299 |
+
- **Voice Cloning**: Zero-shot voice cloning capabilities for new speakers
|
300 |
+
- **Streaming Audio**: Real-time streaming synthesis for interactive applications
|
301 |
+
|
302 |
+
### Language Expansion
|
303 |
+
- **Additional Malian Languages**: Integration with MALIBA-AI's multi-language TTS
|
304 |
+
- **Dialect Support**: Specialized models for regional Bambara variants
|
305 |
+
- **Cross-Lingual Features**: Better support for multilingual content
|
306 |
+
|
307 |
+
### Community Integration
|
308 |
+
- **Speaker Expansion**: Additional authentic Bambara speakers
|
309 |
+
- **Quality Improvements**: Continuous model refinement based on community feedback
|
310 |
+
- **Application Development**: Reference implementations for common use cases
|
311 |
+
- **Training Resources**: Educational materials for developers and researchers
|
312 |
+
|
313 |
+
## References
|
314 |
+
|
315 |
+
```bibtex
|
316 |
+
@software{maliba_ai_bambara_tts_2025,
|
317 |
+
title={MALIBA-AI Bambara Text-to-Speech: First Open-Source TTS for Bambara Language},
|
318 |
+
author={MALIBA-AI Team},
|
319 |
+
year={2025},
|
320 |
+
publisher={HuggingFace},
|
321 |
+
url={https://huggingface.co/MALIBA-AI/bambara-tts},
|
322 |
+
note={Built on Spark-TTS architecture}
|
323 |
+
}
|
324 |
+
|
325 |
+
@misc{wang2025sparktts,
|
326 |
+
title={Spark-TTS: An Efficient LLM-Based Text-to-Speech Model with Single-Stream Decoupled Speech Tokens},
|
327 |
+
author={Xinsheng Wang and Mingqi Jiang and Ziyang Ma and Ziyu Zhang and Songxiang Liu and Linqin Li and Zheng Liang and Qixi Zheng and Rui Wang and Xiaoqin Feng and Weizhen Bian and Zhen Ye and Sitong Cheng and Ruibin Yuan and Zhixian Zhao and Xinfa Zhu and Jiahao Pan and Liumeng Xue and Pengcheng Zhu and Yunlin Chen and Zhifei Li and Xie Chen and Lei Xie and Yike Guo and Wei Xue},
|
328 |
+
year={2025},
|
329 |
+
eprint={2503.01710},
|
330 |
+
archivePrefix={arXiv},
|
331 |
+
primaryClass={cs.SD},
|
332 |
+
url={https://arxiv.org/abs/2503.01710}
|
333 |
+
}
|
334 |
+
|
335 |
+
@article{bamana_language_2024,
|
336 |
+
title={Bambara Language and Digital Inclusion in Mali},
|
337 |
+
author={MALIBA-AI Research Team},
|
338 |
+
journal={African Language Technology Review},
|
339 |
+
year={2024},
|
340 |
+
note={In preparation}
|
341 |
+
}
|
342 |
+
```
|
343 |
+
|
344 |
+
## License
|
345 |
+
|
346 |
+
⚠️ **Important License Information**
|
347 |
+
|
348 |
+
This project is licensed under **Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0)** due to the licensing terms of the underlying Spark-TTS architecture and training data.
|
349 |
+
|
350 |
+
### Key License Terms
|
351 |
+
- **Non-Commercial Use Only**: Research, education, and personal use permitted
|
352 |
+
- **Share-Alike**: Derivatives must use the same license
|
353 |
+
- **Attribution Required**: Must credit MALIBA-AI and Spark-TTS
|
354 |
+
|
355 |
+
### Commercial Usage
|
356 |
+
For commercial licensing options, contact: contact@maliba-ai.com
|
357 |
+
|
358 |
+
### Attribution Requirements
|
359 |
+
```
|
360 |
+
This work uses MALIBA-AI Bambara TTS, built on Spark-TTS architecture.
|
361 |
+
Licensed under CC BY-NC-SA 4.0.
|
362 |
+
Original work: https://huggingface.co/MALIBA-AI/bambara-tts
|
363 |
+
Spark-TTS: https://github.com/SparkAudio/Spark-TTS
|
364 |
+
```
|
365 |
+
|
366 |
+
## Contributing
|
367 |
+
|
368 |
+
MALIBA-AI Bambara TTS is part of the broader MALIBA-AI initiative with the mission **"No Malian Left Behind by Technological Advances."** We welcome contributions from:
|
369 |
+
|
370 |
+
### Community Contributors
|
371 |
+
- **Bambara Language Experts**: To improve linguistic accuracy and cultural authenticity
|
372 |
+
- **Native Speakers**: For quality assessment and dialectal insights
|
373 |
+
- **Developers**: To create applications and integrations
|
374 |
+
- **Researchers**: To advance the underlying technology
|
375 |
+
- **Data Contributors**: To expand and improve training datasets
|
376 |
+
|
377 |
+
### How to Contribute
|
378 |
+
- **GitHub**: [MALIBA-AI/bambara-tts](https://github.com/MALIBA-AI/bambara-tts)
|
379 |
+
- **HuggingFace**: [MALIBA-AI](https://huggingface.co/MALIBA-AI)
|
380 |
+
- **Email**: contact@maliba-ai.com
|
381 |
+
- **Community**: Join discussions on model improvements and applications
|
382 |
+
|
383 |
+
### Contribution Guidelines
|
384 |
+
- Respect Bambara language and culture
|
385 |
+
- Ensure proper consent for any voice data contributions
|
386 |
+
- Follow community standards for inclusive development
|
387 |
+
- Test thoroughly across different speakers and content types
|
388 |
+
|
389 |
+
---
|
390 |
+
|
391 |
+
**MALIBA-AI: Empowering Mali's Future Through Community-Driven AI Innovation**
|
392 |
+
|
393 |
+
*"MALIBA-AI ka baara kɛ ka bamanankan lakana diɲɛ kɔnɔ!"*
|
394 |
+
*(MALIBA-AI works to preserve Bambara language in the world!)*
|
395 |
+
|
396 |
+
---
|
397 |
+
|
398 |
+
**Contact Information:**
|
399 |
+
- Website: [maliba-ai.com](https://maliba-ai.com)
|
400 |
+
- Email: contact@maliba-ai.com
|
401 |
+
- GitHub: [MALIBA-AI](https://github.com/MALIBA-AI)
|
402 |
+
- HuggingFace: [MALIBA-AI](https://huggingface.co/MALIBA-AI)
|