AI在线 AI在线

DeepSeek Affiliate Company Releases Innovative Deployment Patent for Large Language Models, Promoting New Developments in AI Technology

Recently, DeepSeek's affiliated company —— Hangzhou DeepSeek Artificial Intelligence Basic Technology Research Co., Ltd., has publicly disclosed a patent titled "A Deployment Method and System for a Large Language Model." The release of this patent marks another significant advancement by the company in the field of artificial intelligence, especially in the deployment of large language models.

DeepSeek

Image source note: The image is AI-generated, and the image licensing service provider is Midjourney

According to the patent summary, this invention mainly involves core technologies in artificial intelligence. Its innovation lies in deploying two key stages of a large language model —— the pre-filling stage and the decoding stage —— on machines with high-performance computing capabilities and large memory. This distributed deployment method can effectively balance workload tasks and maximize hardware resource utilization. By reducing idle computing capacity, this method not only reduces overall latency but also significantly improves system throughput.

In the current development of AI technology, system scalability and fault tolerance have become particularly important. DeepSeek's patent achieves this by optimizing resource allocation, enhancing the system's adaptability to different workloads. This innovative deployment method suggests that future AI models will be more efficient and intelligent, and are expected to provide better support for various application scenarios.

Notably, DeepSeek-V3, one of the company's core products, features a powerful Mixture-of-Experts (MoE) language model with 671B parameters, and each token activates 37B parameters. This technological advancement will undoubtedly promote the popularization and application of AI technology, supporting the digital transformation of various industries.

Key points:

🌟 DeepSeek releases a new patent, innovating the method of deploying large language models, improving system performance.  

🚀 Distributed deployment maximizes hardware resource utilization, reducing overall latency.  

📈 Enhances system scalability and fault tolerance, supporting the development of future AI technology.

相关资讯

Mistral Launches New Agents API: Empowering Developers to Build Intelligent AI Agents

Mistral recently released its new Agents API, a framework designed specifically for developers to simplify the creation of AI agents that can perform various tasks such as running Python code, generating images, and conducting retrieval-augmented generation (RAG).The introduction of this API aims to provide a unified environment for large language models (LLMs) to interact with multiple tools and data sources in a structured and persistent manner.The Agents API is built on top of Mistral's language model and integrates multiple built-in connectors. These connectors enable agents to run Python code in a controlled environment, generate images through dedicated models, access real-time web searches, and utilize user-provided document libraries. One highlight is its persistent memory feature, which allows agents to maintain context across multiple interactions, supporting coherent and stateful conversations..
5/28/2025 11:01:20 AM
AI在线

Breaking Traditions! FUDOKI Model Makes Multi-Modal Generation and Understanding More Flexible and Efficient

In recent years, the field of artificial intelligence has undergone tremendous changes, particularly with large language models (LLMs) making significant progress in multi-modal tasks. These models demonstrate powerful potential in understanding and generating language, but most current multi-modal models still adopt autoregressive (AR) architectures, which limit their inference process to be relatively monotonous and lacking in flexibility. To address this limitation, a research team from The University of Hong Kong and Huawei Noah’s Ark Lab has proposed a novel model called FUDOKI.The core innovation of FUDOKI lies in its entirely new non-masked discrete flow matching architecture.
6/16/2025 9:49:02 AM
AI在线

Meta's Open-Source Strategy Now in Question? Report Says Senior Leaders Discuss Abandoning Behemoth Model in Favor of Closed Development

The New York Times, citing sources, reported that senior members of Meta's newly established super intelligent laboratory are discussing a major strategic shift, abandoning the company's powerful next-generation open-source artificial intelligence model Behemoth in favor of developing a closed model.It is reported that Meta has completed the training of the Behemoth model, but due to poor internal performance tests, its release was postponed, and the testing work has also been suspended after the new laboratory was launched.
7/15/2025 9:51:06 AM
AI在线
  • 1