DingTalk and Tongyi Lab Launch Fun-ASR Speech Recognition Model
DingTalk and Tongyi Lab Introduce Fun-ASR: A Breakthrough in Speech Recognition
DingTalk, in collaboration with Tongyi Laboratory, has launched Fun-ASR, a next-generation speech recognition model designed to revolutionize enterprise communication. This advanced model excels in understanding specialized terminology across ten industries, including home decoration and animal husbandry, while offering customizable training for businesses.
Key Features of Fun-ASR
The Fun-ASR model leverages deep learning to accurately transcribe diverse voice signals, handling multiple languages and accents with ease. Its standout feature is the ability to contextually enhance transcription accuracy, making it ideal for enterprise environments where precision is critical.

Integration with DingTalk
Fun-ASR has been seamlessly integrated into several DingTalk functionalities:
- Meeting subtitles and simultaneous interpretation
- Smart meeting minutes
- Voice assistants
These integrations aim to provide businesses with a stable, scalable, and highly accurate speech recognition foundation, particularly useful in scenarios demanding deep contextual understanding.
Future Developments
While Fun-ASR already boasts impressive capabilities, its potential is far from exhausted. DingTalk and Tongyi Lab plan to focus on:
- Dialect recognition to cater to regional linguistic variations.
- Noise robustness for clearer transcription in noisy environments.
- Multilingual support to expand global usability.
- Enterprise deep customization for tailored business solutions.
Key Points:
- 🎤 Fun-ASR understands multi-industry terminology and supports custom training.
- 📊 Integrated into DingTalk’s meeting and voice assistant features.
- 🚀 Future upgrades will enhance dialect recognition and noise handling.





