Skip to main content

Volcano Engine Unveils BeanPod 1.6 and Seedance 1.0 Pro AI Models

At the FORCE Original Power Conference, Volcano Engine made waves in the AI industry with the official release of two groundbreaking products: the BeanPod Large Model 1.6 and its accompanying video generation model, Seedance 1.0 Pro. This launch represents a major leap forward in AI cloud-native services, offering enterprises powerful new tools for digital transformation.

Revolutionary Pricing Structure

The most immediate impact comes from BeanPod's innovative pricing model. ByteDance CEO Liang Ruobo revealed that version 1.6 introduces interval-based pricing tied to input length—a first in the industry. For the typical enterprise usage range of 0-32K inputs, costs plummet to just ¥0.8 per million tokens for input and ¥8 for output, slashing expenses by an impressive 63% compared to previous versions.

Seedance 1.0 Pro follows suit with equally competitive rates: generating a professional-grade 5-second 1080P video now costs merely ¥3.67, while text-to-video conversion runs at ¥0.015 per thousand tokens.

Image

Three Models, Unlimited Potential

The BeanPod 1.6 series comprises three specialized variants:

  • doubao-seed-1.6: China's first model supporting massive 256K context windows, excelling in deep thinking and multimodal comprehension
  • doubao-seed-1.6-thinking: Optimized for complex reasoning tasks requiring advanced cognitive processing
  • doubao-seed-1.6-flash: The speed demon of the family, delivering near-instant responses with visual understanding rivaling industry leaders

Together, these models enable businesses to implement sophisticated AI solutions across diverse scenarios with unprecedented efficiency.

Image

Smart Features Redefine Productivity

BeanPod's "simultaneous thinking and search" capability represents a paradigm shift—the model can now conduct information retrieval while independently analyzing problems. Imagine planning supplies for a Guangdong student in Beijing: BeanPod doesn't just list items but understands cultural preferences and climate needs.

The "Deep Research" function compresses days of analytical work into minutes, generating comprehensive reports in 5-30 minutes that would traditionally require extensive human effort.

Image

Seeing and Doing Like Never Before

Multimodal capabilities take center stage in version 1.6. The model natively processes complex real-world information across formats—text, images, and more—making it invaluable for applications from e-commerce reviews to autonomous vehicle training data annotation.

Perhaps most impressively, BeanPod now interacts with software interfaces like a human operator. Need hotels filtered by specific criteria or invoices organized automatically? The model handles these GUI operations with remarkable precision.

Image

Hollywood-Quality Video from Text

The debut of Seedance 1.0 Pro marks a watershed moment for creative industries. This video generation model produces seamless multi-shot narratives with professional-grade motion stability and aesthetic quality—all from simple text prompts or images.

Third-party evaluations confirm Seedance's dominance in both text-to-video and image-to-video categories, opening new possibilities for e-commerce marketing, film previsualization, and game asset creation.

Image

As enterprises worldwide race to implement AI solutions, Volcano Engine's latest offerings provide both the technological firepower and cost efficiency needed to stay competitive in an increasingly digital landscape.

Key Points

  1. BeanPod Large Model 1.6 reduces costs by up to 63% with innovative length-based pricing
  2. Three specialized model variants cater to different enterprise needs—from deep reasoning to lightning-fast responses
  3. Advanced features enable simultaneous research/analysis and complex GUI operations
  4. Seedance 1.0 Pro sets new standards for AI-generated video quality across multiple industries
  5. Multimodal capabilities allow seamless processing of real-world information across formats

Enjoyed this article?

Subscribe to our newsletter for the latest AI news, product reviews, and project recommendations delivered to your inbox weekly.

Weekly digestFree foreverUnsubscribe anytime

Related Articles

Alibaba's Fun-CineForge Brings Hollywood-Style AI Dubbing to Open Source
News

Alibaba's Fun-CineForge Brings Hollywood-Style AI Dubbing to Open Source

Alibaba's Tongyi Lab has open-sourced Fun-CineForge, a groundbreaking AI system that solves film dubbing's toughest challenges. Unlike traditional robotic voiceovers, this multimodal model masters lip sync, emotional expression, and voice adaptation - even handling complex scenes with multiple speakers. The release includes both the AI model and CineDub, the first large-scale Chinese TV dubbing dataset. Early demos show startlingly natural results when redubbing classics like 'Romance of the Three Kingdoms.'

March 17, 2026
AI dubbingmultimodal AIvoice synthesis
News

Mistral AI's Small4: A Triple-Threat Open Source Model Arrives

Mistral AI has unveiled its latest open-source marvel - the Small4 model. This isn't just another incremental update; it combines three powerful capabilities into one package: logical reasoning, multimodal processing, and coding assistance. With its efficient 128-expert architecture and configurable performance modes, developers now have a versatile tool that adapts to different needs while cutting computational costs.

March 17, 2026
AI modelsopen sourceMistral AI
IBM's Granite 4.0 Speech Model: Smaller Size, Bigger Performance
News

IBM's Granite 4.0 Speech Model: Smaller Size, Bigger Performance

IBM has unveiled Granite 4.0 1B Speech, a compact yet powerful multilingual speech recognition model designed for edge computing. Half the size of its predecessor, it delivers improved accuracy while supporting Japanese ASR and English-Chinese translation. The innovative two-stage architecture allows flexible deployment on resource-constrained devices, topping benchmarks with an impressive 5.52% word error rate.

March 16, 2026
IBMspeech recognitionedge computing
Alibaba's New AI Brings Movie Characters to Life with Perfect Lip Sync
News

Alibaba's New AI Brings Movie Characters to Life with Perfect Lip Sync

Alibaba's Tongyi Lab has unveiled Fun-CineForge, an open-source voice synthesis model that solves Hollywood's toughest AI challenge - making digital voices match actors' lips perfectly. The breakthrough technology handles complex scenes with multiple characters, camera cuts, and obscured faces while maintaining emotional authenticity. Alongside the model, researchers released CineDub, an innovative dataset creation method that slashes production costs.

March 16, 2026
voice synthesisAI in entertainmentmultimodal AI
xAI's Grok 4.20 Prioritizes Truth Over Speed in AI Race
News

xAI's Grok 4.20 Prioritizes Truth Over Speed in AI Race

While competitors chase raw performance, Elon Musk's xAI takes a different path with Grok 4.20 Beta. This new model sets industry records for truthfulness, boasting a 78% non-hallucination rate and the honesty to say 'I don't know' when uncertain. With three specialized API modes and competitive pricing starting at $2 per million tokens, Grok positions itself as the reliable choice for businesses tired of AI 'making up nonsense.'

March 13, 2026
xAIGrokAI reliability
OpenAI's Sora Video API Gets Major Upgrade with Character Consistency and Longer Clips
News

OpenAI's Sora Video API Gets Major Upgrade with Character Consistency and Longer Clips

OpenAI has rolled out significant improvements to its Sora video generation API, making life easier for content creators. The update introduces character consistency - no more random facial changes between scenes - and extends maximum video length to 20 seconds. Creators can now generate both landscape and portrait versions simultaneously, eliminating tedious reformatting. These changes promise to streamline production for ads, short films, and social media content.

March 13, 2026
OpenAIvideo generationAI tools