I recently came across market research data on AI Agent engineering applications, which I found quite interesting. Here are some key findings:



In terms of practical applications, customer service has become the most popular scenario, accounting for 26.5%. Next are research and data analysis (24.4%), followed by internal collaboration and productivity tools (17.7%). Code generation ranks fourth, making up only 9.8%. This may differ from many people's expectations.

However, the biggest obstacle remains quality issues. 32% of practitioners list it as the primary barrier, covering accuracy, relevance, consistency, and other dimensions—consistent with pain points from last year. Cost concerns, on the other hand, are less prominent this year.

Interestingly, 89% of companies have installed some form of observability tools on their Agents, with 62% achieving relatively fine-grained tracking. This indicates that everyone recognizes the importance of tracking multi-step reasoning and tool invocation capabilities; it’s no longer optional.

Regarding model selection, although OpenAI’s models are used by over two-thirds of organizations, multi-model ensembles are mainstream—more than 75% of organizations run multiple different models simultaneously. Interestingly, one-third of organizations are still investing in building their own model infrastructure, showing that beyond API convenience, on-premises deployment remains attractive to enterprises.

As for fine-tuning, it’s still somewhat niche. 57% of organizations do not perform fine-tuning at all; instead, most meet their needs through base models combined with prompt engineering and Retrieval-Augmented Generation (RAG).

This data mainly comes from B2B enterprise users and should reflect the current practical situation.
View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • 4
  • Repost
  • Share
Comment
0/400
BoredRiceBallvip
· 8h ago
Customer service scenarios have actually been predictable for a long time. After all, reducing costs and increasing efficiency are the main goals, and code generation only accounts for 9.8%. I'm a bit surprised... Quality issues remain a longstanding challenge. It seems the "hallucination" problem of LLMs still needs ongoing treatment. Has multi-model mixed usage become standard? If so, companies solely relying on OpenAI should start to panic. But one-third of those building their own infrastructure are truly wealthy, or is it that domestic vendors are finally becoming competitive? People who think RAG with prompt engineering can be easily fine-tuned, it seems the return on investment for fine-tuning is indeed average. Is the high usage rate of Agent's observability tools real, or are people just copying data from each other... 57% not doing fine-tuning suggests that directly deploying the base model is the best solution—it's simpler. Isn't this data a bit exaggerated? It feels too "idealized."
View OriginalReply0
TopBuyerForevervip
· 8h ago
Customer service accounts for the highest proportion? Alas, it still depends on RAG+ prompts. Fine-tuning feels increasingly useless. Quality issues have been bottlenecking for a year, and that's the real pain point. Cost is no longer a big deal. 89% have installed observability tools, it seems everyone knows—no one can handle a black-box Agent. OpenAI accounts for two-thirds of usage, but I believe multi-model combinations are the real way to go. After all, if one model has an issue. Some still invest in local deployment, but API convenience is overrated; data security is the real need.
View OriginalReply0
OvertimeSquidvip
· 8h ago
Wait, only 9.8% for code generation? I always thought that was the main course. Looks like I was too naive haha Customer service at 26.5% is higher than I expected, it feels like rescuing people from hell Quality issues are always the biggest headache. It seems they haven't been truly solved in the past couple of years. RAG and prompt optimization are just for show Speaking of which, 75% are playing with multi-model combinations. Doesn't that make OpenAI the standard? Feels all the same 89% for installing observability tools—are these numbers real? We don't even have this concept here
View OriginalReply0
LiquidityHuntervip
· 9h ago
Customer Service 26.5%? It shows that everyone is still using Agent to do the least technical work. Quality issues are always the biggest pitfall, but with OpenAI being so heavily invested, do other models really have no chance? RAG+ prompts are enough; fine-tuning is indeed somewhat overhyped. 89% are equipped with observability tools? That has already become standard. Are the one-third of self-built models truly valuable, or are they just trying to reverse bind to suppliers? I agree that multi-models are the mainstream; no one dares to rely solely on a single one.
View OriginalReply0
Trade Crypto Anywhere Anytime
qrCode
Scan to download Gate App
Community
English
  • 简体中文
  • English
  • Tiếng Việt
  • 繁體中文
  • Español
  • Русский
  • Français (Afrique)
  • Português (Portugal)
  • Bahasa Indonesia
  • 日本語
  • بالعربية
  • Українська
  • Português (Brasil)