Comprehensively Upgrading SenseNova 5.5: China’s First Real-time Multimodal Model
Comprehensively Upgrading SenseNova 5.5: China’s First Real-time Multimodal Model
Blog Article
By gathering and processing data across modalities including audio, texts, images and videos, SenseNova 5o, China’s first real-time multimodal model provides a brand-new interactive AI experience.
Users are able to interact with SenseNova 5o, akin to having a conversation with an actual person. The interactive model is especially suitable for applications such as real-time conversation and speech recognition. It is highly adaptable and can manage multiple tasks within the same model, while adjusting its responses based on different contexts.
SenseNova 5.0, which was recently released in April, was China's first large model on par with the capabilities of GPT-4 Turbo. In quick succession two months later, the upgraded SenseNova 5.5 has registered a 30% improvement in overall performance compared to SenseNova 5.0. With significantly enhanced abilities in mathematical reasoning, English proficiency and following commands, SenseNova 5.5 interactivity and multiple core indicators are on par with GPT-4o.
SenseNova 5.5 adopts a hybrid cloud-edge collaborative expert architecture to maximize the “Cloud-to-Edge” synergy and reduce inference costs. The model training was based on over 10TB tokens of high-quality training data, including a large amount of synthetically-generated reasoning chain data, which help to enhance its reasoning capabilities.
To lower the barriers to entry for enterprise users in leveraging the robust capabilities of the SenseNova Large Model, SenseTime has recently launched the "Project $0 Go" scheme. This is a free and comprehensive onboarding bundle for all new enterprise users who are migrating from the OpenAI platform, including a 50 million tokens package and API migration consulting services.
reference link:https://www.sensetime.com/en/technology-index