中文
English

A 'small company' that generates revenue with DeepSeek, both painful and happy

2025-04-13

A 'small company' that generates revenue with DeepSeek, both painful and happy

Chinese Entrepreneur Magazine

Chinese Entrepreneur Magazine

follow with interest

This article is from WeChat official account: China Entrepreneur Magazine (ID: iceo com cn), written by Kong Yuexin, edited by Ma Jiying, and the title is from AI Generation


Article Summary

The DeepSeek big model triggers a revolution in the AI industry, driving a surge in traffic and technological upgrades for small and medium-sized enterprises. Its high cost performance API and open source strategy lower the threshold for enterprise access, stimulate innovation in C-end/B-end applications, and intensify industry competition. Model optimization makes it possible to deploy end-to-end AI, and hybrid computing solutions accelerate the landing of AI application scenarios. The market expects to usher in an application explosion period in 2025.


•    Model hot: DeepSeeker R1 full blood version triggers enterprise access wave, Tencent Yuanbao APP downloads top app store


•    Cost Revolution: API price is only 1/30 of GPT-4, theoretical profit margin reaches 545%, promoting industry cost reduction and efficiency improvement


•   ⚙️ Technological breakthrough: MoE architecture optimization reduces computing power requirements, end side chips can locally deploy 670B parameter models


•    Ecological Expansion: Over 200 Enterprises Connect to Open Source Models, Forming a New Paradigm of Cloud+End Side Hybrid Deployment


•    Traffic impact: Silicon based mobile traffic surges 40 times, server overload reflects market demand explosion


•    Application turning point: The cost of large models will decrease by 90% in 12 months, and it is expected to usher in the first year of AI intelligent agent explosion in 2025

After AI Infra announced its integration with DeepSeeker R1, many small and medium-sized enterprises will come to contact, hoping to obtain products deployed with the R1 model. Qingcheng Jizhi encountered a similar situation.




Is your DeepSeek a 'full blooded version'? "Tang Xiongchao, CEO of Qingcheng Jizhi, was once asked by a client.




Note: DeepSeek Full Blood Version is the top-level version of the DeepSeeker R1 model, with model parameters reaching 671B (671 billion), which is more than 20 times that of the regular version (14B/32B). The Full Blood Version supports local/API deployment and complex scientific research calculations, with a higher upper limit of capabilities and higher hardware requirements.




After receiving too many such inquiries, the Qingcheng Jizhi team decided to use engineering to solve this problem - launching a "full health version" identification mini program on the official website, and carefully selecting several distinguishable questions that users can use to ask. If the system answers correctly, it is basically the "full health version"; If you can't answer it, it may not be the 'full blooded version'.




After the mini program went online, its traffic exceeded the expectations of Qingcheng Jizhi.




In fact, the experience of Qingcheng Jizhi is just a microcosm of the recent AI industry. The entire AI industry should have had a very fulfilling month, "said an industry insider. The popularity of DeepSeek has put all practitioners in the AI industry in a state of "pain and happiness".




On the one hand, the emergence of DeepSeek has stimulated the awareness and demand of ordinary users to use AI tools, promoting the popularization of AI. DeepSeek has also become the fastest-growing AI application in history. According to the AI product list, DeepSeek had 157 million active users in February, which is close to 20% of ChatGPT's 749 million. The influx of too many users often puts DeepSeek chatbot in a "server busy" state.




On the other hand, DeepSeek's rapid iteration and open source have led the already "rolling" AI industry into a new round of "arms race", with many companies from the model layer to the application layer hardly taking a break during this year's Spring Festival. Numerous companies have announced their integration into DeepSeek, including B-end companies such as cloud service providers and chip manufacturers, as well as various C-end application companies. According to statistics from Zhenghe Island, over 200 companies have completed the integration and deployment of DeepSeek technology interfaces.




The companies that have joined have also received a wave of "sky high traffic" - Tencent Yuanbao APP has rapidly climbed in download volume after joining DeepSeek, and topped the free APP download ranking list of Apple App Store in China on March 3; As an AI infrastructure company, silicon-based mobile has the fastest access to DeepSeeker R1 on the entire network, with a 40 fold increase in traffic and a staggering 17.19 million visits in February.




The emergence of DeepSeeker R1 has further raised expectations from all parties for the accelerated development of AIGC applications. When Monica.im released its AI agent product Manus on March 6th, it once again sparked a frenzy of "invitation code buying".




Both major model manufacturers and companies in the upstream and downstream of the AI industry chain are eagerly waiting for the key path to the future AI world.




1、 How to access DeepSeek




As early as the release of the DeepSeeker V2 model in 2024, the industry had already paid attention to this company and its open-source models.




Guo Chenhui, the technical director of Meitu Design Studio, stated that in order to provide users with a better experience in Meitu's AI application scenarios, Meitu has also been paying attention to excellent large models at home and abroad based on its own research. When DeepSeeker V2 was released, Meitu's external AI team paid attention to the model and tried to collaborate with the DeepSeeker team. However, in order to seek stability, Meitu mainly called the DeepSeek model API through third-party AIInfra service providers at that time. In September 2024, Meitu Design Studio integrated the V2 model to assist with copywriting expansion. After the release of the V3 and R1 models, they also updated them one after another. When our product and business teams see some models that are suitable for integration, they will conduct performance evaluations, and those that are suitable may be introduced into our own application scenarios, "said Guo Chenhui.




DeepSeek officially provides two access methods: one is to call its API interface through some programming methods after the model runs; The second is for users to install an app on their mobile phone or open a chat window on the official website, and directly chat with it. Behind the chat window is calling the API.




However, due to the current high traffic of DeepSeek and the shortage of servers and manpower, DeepSeek's own API may experience issues such as timeouts. Guo Chenhui stated that Meitu's products have a large user base, and after promoting some features, traffic may increase by tens or even hundreds of times. In this case, the service guarantee capability of public cloud is relatively stronger.




Not only that, DeepSeek's model is relatively large, especially the "full blooded version" model, which has certain hardware requirements; Based on the consideration of cost-effectiveness, Meitu's business scenarios have significant peak and low peak effects, and cloud providers can smooth out the differences in the peak and low peak periods of API calls among different companies. If we deploy it ourselves, the utilization rate of resources during low peak periods may be relatively low, resulting in significant waste of resources, "said Guo Chenhui.




Therefore, the current way for Meitu to access the DeepSeek-R1 model is mainly to call the API of cloud vendors and deploy it privately on this basis.




Similar to Meitu, this chip technology that deploys end side chips has also been keeping an eye on various newly released large models, especially those that are more suitable for localized deployment on the end side. Zhou Jie, the general manager of the ecological strategy of this chip technology, stated that for some open source large models, especially SOTA models (State of the Art, the best performing model in a certain field or task), they will invest resources in corresponding heterogeneous adaptation as soon as possible. Therefore, after DeepSeek released V2 last year and R1 this year, this chip technology immediately attempted to adapt to these models.




In Zhou Jie's view, there are two main innovations of the DeepSeek-V2 model. Firstly, it effectively reduces the overhead of KV cache (an optimization technique used by Transformer models in autoregressive decoding) through MLA (Multi Head Latent Attention) architecture. This is because large language models have high requirements for memory bandwidth and capacity. Once KV cache can be reduced, it can greatly help computing platforms; The second is the MoE (Mixed Expert) model released by DeepSeek, which optimizes and transforms the traditional MoE architecture. This architecture allows a larger model with limited resources to be used.




At that time, this chip technology quickly adapted to the light version of the V2 model, which is a model of size 16B. Although the 16A parameter may seem large, in actual operation, it only activates the 2.4B parameter. We believe that this model is very suitable for running on the end side, and the P1 chip of this chip technology can also provide good support for models with a 2.4B parameter scale, "Zhou Jie told" Chinese Entrepreneur ".




Regarding how this chip technology "connects" to DeepSeek, Zhou Jie explained, "Users now use applications such as DeepSeek, which require a lot of computing power from the cloud. DeepSeek's own data center or cloud vendors provide some APIs for terminal applications to call. When users use the DeepSeek APP, they can call the AI capabilities in the cloud. However, some terminal scenarios may have high requirements for data privacy and other aspects. In this case, local computing is needed. After deployment on the terminal side, users can run DeepSeek and other models in the event of network disconnection




After meeting the basic requirements for running a large language model from the aspects of computing power and system, this chip technology can combine with the actual needs of customer projects, collaborate commercially with model vendors such as DeepSeek, fine tune and optimize the model, and implement specific projects.




After the launch of V2, Qingcheng Jizhi also attempted to integrate the model internally, but due to low market demand at that time, they did not promote its use. After the release of R1 this year, they felt it was a great opportunity and decided to integrate with DeepSeek and promote it to customers on a large scale.




Qingcheng Jizhi specializes in system software and provides inference services based on system software. Therefore, unlike some application companies that directly access DeepSeek's API, it provides customers with a dedicated DeepSeek API for application services. Our way of accessing is to download DeepSeek's open-source model and deploy the service on our computing system using system software, "said Tang Xiongchao.




Simply put, the R1 model is a file of several hundred gigabytes in size, but it cannot be directly used after downloading. It's just a file, not a usable service. What we need to do is to run this model and let it provide service interfaces to the outside world. Through the API service interface, users can talk to the model, "explained Tang Xiongchao.




Based on previous technical accumulation, Qingcheng Jizhi iterated the first version within a day after downloading the model file, and then optimized the R1 model structure. The official "full blood version" was officially announced and launched in just one week.




In Tang Xiongchao's opinion, the technical work has been relatively smooth, and after integrating with DeepSeek, more challenges come from the business or market side. Specifically, DeepSeek's traffic has brought the company a lot of consulting clients, but each client's needs are different. Including different computing power platforms, chip models, server specifications, etc., we need to make targeted optimizations based on different computing power foundations, "said Tang Xiongchao.




2、 Reducing API costs promotes the popularization of large models




After the release of the V2 model in May 2024, DeepSeek was awarded the title of "Pinduoduo in the AI industry" due to its ultimate cost-effectiveness, which led to a price war among major domestic manufacturers for large models.




The price war has reduced API costs. Taking Meitu's "AI Product Image" as an example, in Guo Chenhui's view, on the one hand, Meitu has a strong technological advantage in AI image processing, and the integration of DeepSeek model brings positive feedback on user experience and conversion. Moreover, the cost of calling the big language model API is very low, which complements Meitu's business scenarios very well. Therefore, Meitu will also increase its attention to the application of big language models.




On February 9th, DeepSeek stopped the 45 day discounted price trial period for the V3 model, and restored the API to its original price of 0.5 yuan per million input tokens (cache hit)/2 yuan per million output tokens (cache miss), and 8 yuan per million output tokens. The input price of one million tokens (cache hit) for R1 is 1 yuan, the input price of one million tokens (cache miss) is 4 yuan, and the output price is 16 yuan.




But according to OpenAI's official website, GPT-4o's input tokens cost $2.5 per million and output tokens cost $10 per million; The latest release of GPT-4.5's million input/output tokens is as high as $75/$150, which is 15 to 30 times higher than GPT-4o alone.




In Guo Chenhui's view, on the one hand, the cost of calling the DeepSeek model does not account for a high proportion of the overall cost of Meitu AI research and investment; On the other hand, DeepSeek remains in a relatively affordable price range even after restoring its original price, and Meitu's integration with DeepSeek has shown positive user conversion and feedback. Therefore, they will increase their investment in large language models.




Zhou Jie also believes that the API price of DeepSeek is many times lower than OpenAI, which greatly reduces the cost of buying tokens for enterprises and users. At the end model level, a 3B model may now be able to achieve the same level of performance as previous models with a scale of 7B or above, with relatively reduced costs such as memory.




This is a process of software hardware collaboration. Under the same hardware conditions, it is now equivalent to being able to achieve model effects with larger parameter scales, or to achieve the same model effects, the requirements for hardware have become lower, "said Zhou Jie.




In early March, after the five-day "DeepSeek Open Source Week", the DeepSeek team released key information such as the optimization technology details and cost profit margin of the model for the first time. According to DeepSeek's calculations, its cost profit margin can theoretically reach 545%.




The rapid reduction of the cost of large models and the improvement of capabilities have also brought about a high-speed growth of users in the to B and to C fields. Tang Xiongchao revealed that many small and medium-sized enterprises now proactively contact them, hoping to obtain products based on the R1 model.




3、 AI applications will accelerate the explosion




Robin Lee, the founder, chairman and CEO of Baidu, wrote in the article "Seizing the first year opportunity of AI agent explosion and promoting the rapid development of new quality productivity" that the reasoning cost of the big model is reduced by more than 90% every 12 months, far exceeding the "Moore's Law". With the iteration of big model technology and the sharp decline in costs, the application of artificial intelligence will explode.




At present, the AI market is in a stage of rapid growth. Tang Xiongchao believes that DeepSeek's theoretical profit margin is as high as 545%, which has a very positive significance and impact on the entire industry, and has educated the market on the importance of computing power system software.




In the past, people did not attach great importance to the ability of software. DeepSeek made people realize that spending money on software is not a waste of money, but to save money better Tang Xiongchao stated that in an educated market environment, the advantages of core system software can be more fully utilized; In the short term, DeepSeek's open source can also help reduce the commercial cost of product delivery for all parties involved.




As more and more enterprises join DeepSeek and provide feedback on its open source ecosystem, the development process of DeepSeek is also accelerating.




Guo Chenhui believes that this is also the biggest advantage of DeepSeek's open source ecosystem - while the connected enterprises create differentiated products in their respective application scenarios, the application scenarios can also promote the development of DeepSeek and other foundational models. The differentiated deployment of open source ecosystems by various companies not only accelerates AI innovation, but also helps to reduce the cost of large models in vertical segmentation fields, bringing greater imagination to AI applications, "said Guo Chenhui.




In Zhou Jie's view, in addition to the explosion of cloud applications, end-to-end AI applications will also achieve explosive development in 2025 under the promotion of DeepSeek.




The future AI is actually a hybrid artificial intelligence, not everything runs in the cloud, nor is everything run on the end side, because each has its own advantages. For example, the end side can only run relatively small-scale parameter models, but for certain tasks that require higher accuracy, cloud computing power is still needed. In order to ensure data security and privacy, it is necessary to use end side capabilities to achieve the effects of larger parameter scale models, which forms a hybrid deployment solution. "Zhou Jie said that this chip technology is also exploring applications with cloud vendors in this area.




The concept of "the first year of AI application" is no longer a new one, but as of now, AI industry practitioners and investors are still searching for more suitable landing scenarios for AI applications. In Zhou Jie's view, it's just a matter of time. "The development of a new ecosystem definitely takes some time, and everything won't suddenly improve. It requires continuous iteration of software and hardware. Currently, the chip side, model side, and other aspects have laid a solid foundation for the large-scale application of AI. More developers are needed to develop AI applications in the future to meet practical scenario requirements




This article is from WeChat official account: China Entrepreneur Magazine (ID: iceo com cn), written by Kong Yuexin and edited by Ma Jiying


This content is the author's independent viewpoint and does not represent the stance of Tiger Sniff. Reproduction without permission is not allowed. Please contact for authorization matters hezuo@huxiu.com

If you have any objections or complaints about this article, please contact tougao@huxiu.com

People who are changing and want to change the world are all on the Tiger Sniff app




用DeepSeek创收的“小公司”,痛并快乐着

本文来自微信公众号:中国企业家杂志 (ID:iceo-com-cn),作者:孔月昕,编辑:马吉英,题图来自:AI生成



文章摘要
DeepSeek大模型引发AI行业变革,推动中小企业流量激增与技术升级。其高性价比API和开源策略降低企业接入门槛,激发C端/B端应用创新,同时加剧行业竞争。模型优化使端侧AI部署成为可能,混合计算方案加速AI应用场景落地,市场预期2025年将迎来应用爆发期。

• 模型火爆:DeepSeek-R1满血版引发企业接入潮,腾讯元宝APP下载量登顶应用商店

• 成本革命:API价格仅为GPT-4的1/30,理论利润率达545%推动行业降本增效

• ⚙️技术突破:MoE架构优化降低算力需求,端侧芯片可本地部署670B参数模型

• 生态扩张:超200家企业接入开源模型,形成云端+端侧混合部署新范式

• 流量冲击:硅基流动访问量激增40倍,服务器过载反映市场需求井喷

• 应用拐点:大模型成本12个月降90%,2025年预计迎来AI智能体爆发元年

AI Infra公司在宣布接入DeepSeek-R1后,会有很多中小企业前来联系,希望获得部署了R1模型的产品,清程极智就遇到了类似情形。


“你们的DeepSeek是不是‘满血版’?”清程极智CEO汤雄超曾被客户这么问。


注:DeepSeek满血版即DeepSeek-R1模型的顶级版本,模型参数达671B(6710亿),是普通版(14B/32B)的20倍以上,满血版支持本地/API部署及复杂科研计算,能力上限更高,且对硬件要求也更高。


收到太多此类问询后,清程极智团队决定用工程解决这个问题——在官网上线一个“满血版”鉴别小程序,并精心挑选了几道比较有区分度的题目,用户可以用这些题目去提问,如果系统回答得对,基本上就是“满血版”;如果回答不出来,可能就不是“满血版”。


该小程序上线后,其访问量超出了清程极智的预期。


实际上,清程极智的经历只是近期AI行业的一个缩影。“整个AI行业这一个多月应该都过得很充实。”有业内人士表示。DeepSeek的火爆,让整个AI赛道的从业人员处于一种“痛并快乐着”的状态。


一方面,DeepSeek的出现激发了普通用户使用AI工具的意识和需求,推动了AI的普及。DeepSeek也成为有史以来增速最快的AI应用,据AI产品榜统计,DeepSeek 2月份活跃用户1.57亿,已接近ChatGPT 7.49亿的20%。过多用户的涌入,也让DeepSeek对话机器人常常陷入“服务器繁忙”的状态。


另一方面,DeepSeek的快速迭代和开源,让本来就“卷”的AI行业进入了新一轮的“军备竞赛”,从模型层至应用层的很多公司,在今年春节期间几乎没有休息。众多企业宣布接入DeepSeek,既有云服务提供商、芯片制造商等B端公司,也有各类C端应用公司。据正和岛统计,已有超200家企业完成DeepSeek技术接口的集成部署。


接入的企业也迎来了一波“泼天流量”——腾讯元宝APP在接入DeepSeek之后,下载量迅速攀升,并在3月3日登顶中国区苹果应用商店免费APP下载排行榜TOP1;作为AI基础设施公司,硅基流动在全网最快接入DeepSeek-R1,访问量激增40倍,2月访问量高达1719万人次。


DeepSeek-R1的出现,也进一步提高了各方对AIGC应用加速发展的期待。当3月6日Monica.im发布AI智能体产品Manus时,再次引发了一场“邀请码抢购”狂潮。


无论是各大模型厂商还是AI产业链上下游的公司,都在兴奋地等待通往未来AI世界的关键路径。


一、如何接入DeepSeek


早在2024年DeepSeek-V2模型发布时,业内已经关注到这家公司及旗下的开源模型。


美图设计室技术负责人郭晨晖表示,为了在美图的AI应用场景中给用户更好的效果体验,在自研的基础上,美图也一直对国内外优秀的大模型保持关注。DeepSeek-V2发布时,美图的外采AI团队就关注到了该模型,与DeepSeek团队接触尝试合作。不过为了寻求稳定性,美图当时主要通过第三方AIInfra服务商调用DeepSeek模型API。2024年9月,美图设计室接入了V2模型,辅助文案扩写,V3、R1模型发布后,他们也陆续进行了更新。“我们产品和业务团队看到一些适合结合的模型,就会去做效果评估,合适的可能就会引入到我们自己的应用场景里。”郭晨晖说。


DeepSeek官方提供了两个接入方法,一是模型跑起来后,通过一些编程方式去调用它的API接口;二是用户在手机上装一个APP或打开官网的聊天窗口,直接跟它对话,聊天窗口的背后就在调用API。


不过,由于目前DeepSeek的流量过高,又存在服务器、人手不足等情况,导致DeepSeek自己的API会出现超时等问题。郭晨晖表示,美图旗下产品有着大体量的用户基数,一些功能推广开来后流量可能会激增数十倍、上百倍,这种情况下,公有云的服务保障能力相对更强。


不仅如此,DeepSeek的模型比较大,尤其是“满血版”模型对硬件有一定要求;基于性价比层面的考虑,美图的业务场景存在很显著的(使用)高峰、低峰效应,云厂商可以抹平各家调用API高低峰期的差异。“如果我们自己进行部署,低峰期资源利用率可能比较低,会有比较大的资源浪费。”郭晨晖说。


因此,美图目前接入DeepSeek-R1模型的方式,主要是调用云厂商的API,在此基础上进行一定的私有化部署。


与美图类似,部署端侧芯片的此芯科技,也一直对新发布的各种大模型保持关注,尤其是比较适合在端侧进行本地化部署的模型。此芯科技生态战略总经理周杰表示,对于一些开源的大模型,尤其是SOTA模型(State of the Art,在某一领域或任务中表现最佳的模型),他们会第一时间投入资源进行相应的异构适配。因此在DeepSeek去年发布V2以及今年发布R1后,此芯科技都第一时间尝试适配这些模型。


在周杰看来,DeepSeek-V2模型的主要创新点有两个,一是通过MLA(多头潜在注意力)架构有效地降低了KV缓存(Transformer模型在自回归解码过程中使用的一种优化技术)的开销,因为大语言模型对于内存带宽和容量的要求很高,一旦能够降低KV缓存,可以给算力平台带来很大帮助;二是DeepSeek发布的MoE(混合专家)模型,对传统MoE架构进行了优化改造,这个架构可以让一个(参数)更大的模型在资源有限的情况下被使用。


当时,此芯科技很快适配了V2模型的light版本,即16B大小的模型。“虽然16B参数看起来也很大,但实际运行时,它只会激活2.4B参数。我们觉得这样的模型非常适合在端侧运行,此芯科技的P1芯片也可以给2.4B参数规模的模型提供比较好的支持。”周杰告诉《中国企业家》。


对于此芯科技如何“接入”DeepSeek,周杰解释道:“用户现在使用DeepSeek等应用,很多需要调用云端的算力,相当于DeepSeek自己的数据中心或云厂商,提供了一些API给终端侧应用调用,用户使用DeepSeek APP时,就可以调用云端的AI能力。但是部分端侧场景可能对数据隐私等方面有很高的要求,这种情况下就需要在本地进行运算,在端侧部署后,用户可以在断网的情况下运行DeepSeek等模型。”


从算力和系统层面满足了运行一个大语言模型的基本要求后,此芯科技就可以结合客户项目的实际需求,跟DeepSeek等模型厂商进行商业化合作,对模型进行微调优化,把具体项目落地。


V2推出后,清程极智内部也尝试接入该模型,但当时的市场需求较少,他们就没有推广使用。今年R1出来后,他们觉得这是一个非常好的机会,决定接入DeepSeek并大规模向客户推广。


清程极智是做系统软件的,对外基于系统软件提供推理服务,因此不是像部分应用公司那样直接接入DeepSeek的API,而是为客户提供一套专属的DeepSeek的API用于应用服务。“我们接入的方式是把DeepSeek的开源模型下载下来,在我们的算力系统上用系统软件把服务部署起来。”汤雄超说。


通俗来讲,R1模型是一个几百G大小的文件,但下载后无法直接使用。“它只是一个文件,不是一个可用的服务,我们要做的是把这个模型运行起来,让它去对外提供服务的接口。通过API的服务接口,用户就可以跟模型进行对话了。”汤雄超解释道。


基于前期技术积累,清程极智在把模型文件下载下来后,一天内就迭代出了第一个版本,随后针对R1模型结构进行了优化,正式“满血版”官宣上线只用了一周。


在汤雄超看来,技术环节的工作都比较顺利,接入DeepSeek后,更多的挑战来自于商务侧或市场侧。具体来说,DeepSeek的流量给公司带来了非常多来咨询的客户,但每个客户的需求都不太一样。“包括算力平台、芯片型号、服务器规格等都不一样,我们需要针对不同的算力等基础,做针对性的调优。”汤雄超说。


二、API成本降低推动大模型普及


在2024年5月发布V2模型后,因其极致性价比,DeepSeek获得“AI界拼多多”的称号,并带动国内大厂打起了大模型价格战。


价格战降低了API费用。以美图“AI商品图”为例,在郭晨晖看来,一方面,美图在AI图像处理上拥有强大技术优势,而DeepSeek模型的接入带来了用户体验和转化的正向反馈,且大语言模型API的调用成本占比很低,这与美图的业务场景形成了很好的优势互补,因此美图也会加大对大语言模型应用的关注。


2月9日,DeepSeek停止了V3模型为期45天的优惠价格体验期,API恢复原价,每百万输入tokens 0.5元(缓存命中)/2元(缓存未命中),每百万输出tokens 8元。R1的百万tokens输入价格(缓存命中)为1元,百万tokens输入价格(缓存未命中)为4元,输出价格为16元。


但OpenAI官网显示,GPT-4o的2.5美元/百万输入tokens,10美元/百万输出tokens;最新发布的GPT-4.5的百万输入/输出tokens更是高达75美元/150美元,仅较GPT-4o就上涨了15~30倍。


在郭晨晖看来,一方面DeepSeek模型调用费用在美图AI研投的整体成本占比不高;另一方面,DeepSeek恢复原价后依然处于比较便宜的价格区间,且美图接入DeepSeek后,在用户转化及反馈上是正向的,因此他们会加大在大语言模型方面的投入。


周杰也认为,DeepSeek的API价格比OpenAI低了很多倍,对于企业和用户而言,买token的费用大大降低。在端侧模型层面,现在一个3B的模型可能就能做到以前7B以上规模的模型效果,内存等成本也相对降低了。


“这是个软硬件协同的过程。同样的硬件条件下,现在相当于能实现以前更大参数规模的模型效果,或者要达到同样的模型效果,对硬件的要求变低了。”周杰说。


3月初,持续五天的“DeepSeek开源周”结束后,DeepSeek团队首次对外公布了模型的优化技术细节、成本利润率等关键信息。按DeepSeek测算,其成本利润率理论上能达到545%。


大模型成本的快速降低以及能力的提升,也带来了to B和to C领域用户的高速增长。汤雄超透露,现在有很多的中小企业会主动联系到他们,希望获得基于R1模型的产品。


三、AI应用将加速爆发


百度创始人、董事长兼CEO李彦宏在《紧抓AI智能体爆发元年机遇,推动新质生产力加快发展》一文中写道,大模型的推理成本,每12个月就降低90%以上,远超“摩尔定律”。随着大模型技术的迭代和成本的直线下降,人工智能应用将大爆发。


目前,AI市场处于高速增长阶段,汤雄超认为,DeepSeek的理论利润率高达545%,对于整个行业的意义和影响非常积极,给市场科普了算力系统软件的重要性。


“过去大家并不是非常重视软件的能力,DeepSeek让大家认识到,花钱买软件不是浪费钱,而是为了更好地省钱。”汤雄超表示,在受过教育的市场环境下,核心系统软件的优势能被更大地发挥出来;短期来看,DeepSeek的开源也能让各方降低产品交付的商业成本。


随着越来越多企业接入DeepSeek,在其开源生态上做“建设”反馈,DeepSeek的发展进程也在加速。


郭晨晖认为,这也是DeepSeek的开源生态最大的优势——接入的企业在各自应用场景上打造差异化能力产品的同时,应用场景也能推动DeepSeek等基座大模型的发展。“各家公司在开源生态的差异化部署不仅能加速AI的创新,大模型的低成本化也有助于大模型在垂直细分领域的可用性,给AI的应用带来更大的想象空间。”郭晨晖说。


在周杰看来,除了云端应用爆发外,在DeepSeek的推动下,端侧AI应用也会在2025年实现井喷式发展。


“未来的AI其实是一个混合式的人工智能,不是所有的东西都在云端跑,也不是所有东西都在端侧跑,因为各有各的优势。如端侧只能跑相对小规模参数的模型,但对于某些任务来说,对精度有更高要求,还是要用云端算力;而为了保证数据安全和隐私,就需要使用端侧能力实现以前更大参数规模的模型效果,这就形成一个混合式的部署方案。”周杰说,此芯科技也在跟云厂商进行这方面的应用探索。


“AI应用元年”已经不是一个新概念,但截至目前,AI行业从业者以及投资人,还在寻找更适合AI应用的落地场景。在周杰看来,这只是时间问题,“一个新生态的发展肯定需要一定时间,所有的东西不会突然完善,需要软件和硬件不断迭代。目前来看,芯片侧、模型侧等已经为AI的大规模应用打下了坚实的基础,后面需要更多的开发者来开发AI应用,满足实际的场景需求。”


本文来自微信公众号:中国企业家杂志 (ID:iceo-com-cn),作者:孔月昕,编辑:马吉英

本内容为作者独立观点,不代表虎嗅立场。未经允许不得转载,授权事宜请联系hezuo@huxiu.com
如对本稿件有异议或投诉,请联系tougao@huxiu.com
正在改变与想要改变世界的人,都在 虎嗅APP


Read0
share
Write a Review...
推荐文章
1  /  178