Communications of the ACM - Artificial Intelligence 16小时前
Achieving Early Wins in Generative AI
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了大型语言模型(LLM)的结构化输出功能如何成为企业AI转型的关键驱动力。在企业急于从生成式AI中获得回报的背景下,结构化输出能够解决LLM生成自由格式文本与企业IT系统集成难题,提升数据质量和处理效率。文章详细阐述了结构化输出在提取非结构化数据洞察、优化AI数据管道、增强检索增强生成(RAG)的可靠性、实现实时监控以及简化自动化流程等方面的应用。尽管面临提示设计和成本等挑战,结构化输出仍是实现早期AI效益、推动组织AI能力升级的重要途径。

🌟 **解决LLM集成瓶颈,加速AI落地**:企业AI转型面临将LLM输出集成到现有IT系统的挑战,因为IT系统通常需要结构化数据。LLM的结构化输出功能允许模型生成严格符合预定义模式的数据,从而显著改善了LLM与企业IT系统的无缝集成,为AI应用落地提供了关键技术支持。

📊 **释放非结构化数据价值,驱动业务洞察**:企业80%以上的数据是非结构化的(如文本、图像)。LLM的结构化输出能够从这些海量数据中高效提取关键信息,并将其转化为结构化数据,便于企业利用现有的商业智能工具进行分析,从而提升KPI可见性,优化风险管理,并加速数据分析价值的交付。

⚙️ **优化AI数据管道,提升模型训练效率**:结构化输出使LLM能够成为AI模型训练数据管道的 integral part。例如,它可以从非结构化数据中提取特征、聚合数值,并生成可直接加载的合成数据,同时为文档添加结构化元数据,从而提高检索准确性,并为AI模型提供更可靠、更易于处理的训练数据。

💡 **强化RAG性能,减少模型“幻觉”**:在检索增强生成(RAG)技术中,LLM有时会因处理多份非结构化文档而产生“幻觉”。通过结构化输出将非结构化数据转化为图表、表格等结构化格式,能够为RAG提供更清晰、更准确的上下文信息,有效提升知识密集型推理能力,降低模型生成错误信息的概率。

🚀 **赋能自动化流程,提升IT系统可观测性**:结构化输出能够将半结构化和非结构化的IT系统日志转换为结构化日志,便于下游监控系统进行实时跟踪和分析,提升了老旧系统的可观测性,有助于故障排查和威胁检测。同时,它还能将自然语言请求转化为可执行的工作流定义,简化自动化流程,提高效率。

Agentic AI has become the center of gravity in AI transformation initiatives despite being complex and challenging to implement. Effective AI transformation requires fundamental changes to the organizational structure and processes. This needs long-term strategic planning and collaborative effort, but companies are under pressure to show results from Generative AI to keep the momentum sustained. A survey by Deloitte found that 55% of senior leaders want returns from Generative AI (GenAI) within two years and 34% are concerned that not achieving expected value from GenAI could slow down future adoption.1 A tactical approach that produces early results while progressing towards AI transformation is the current need. This post explores how Structured Outputs, a functionality in large language models (LLMs), can help.

Data quality and seamless integration between IT systems and GenAI platforms are essential when scaling IT infrastructure for AI transformation. LLMs are efficient at interpreting natural language texts and producing natural language outputs because they are pretrained for conversation. But the freeform output impedes their integration with enterprise IT systems, as IT systems expect structured data as input. Structured Outputs enables LLMs to generate output that strictly conforms to the schema specified by the developer. This would improve the integration between LLMs and existing IT systems.

Use Cases

The primary use of Structured Outputs is enabling GenAI’s integration with other IT systems, but there are different facets to this integration. I have listed five illustrative use cases for Structured Outputs below.

Insights from Unstructured Data

Surveys estimate that more than 80% of enterprise data are unstructured.2 The major sources of unstructured data are text and images from emails, websites, applications, social media feeds, logs, contract documents, and in recent times, conversational agents. Organizations struggle to process unstructured data even as they hold increasingly more value in them. Valuable data points on customer sentiment, operational efficiency, and risk and compliance issues lie hidden in plain sight within unstructured data.

Natural Language Processing (NLP) tools are the traditional choice for performing analysis on unstructured data, but they have limited capabilities. LLMs are more versatile in handling unstructured data, as they are trained on a large amount of unstructured Internet content. LLMs can extract key information from unstructured data and produce structured data adhering to schema specifications, using Structured Outputs. Extract, Transform, Load (ETL) tools can then load the structured data into databases for further analysis. Organizations can leverage their existing investments in business intelligence tools to analyze the structured data. This would improve the visibility of Key Performance Indicators and facilitate efficient risk management. According to a Gartner report, Generative AI would lead to 40% faster delivery of value for Data and Analytics programs by 2027.3

Feeding the AI Data Pipeline

Data engineering tools have started to integrate LLM Structured Outputs, simplifying the extraction of features and aggregation of values from unstructured data. MotherDuck, a serverless data warehousing platform, has introduced a SQL-based interface to LLMs. This functionality allows the querying, analysis, and transformation of unstructured data from within SQL using Structured Outputs.4 This example shows that Structured Outputs can enable LLMs to become an integral part of the data pipelines used to train AI models.

The presence of personal and sensitive information in the training data used for AI has been a concern for a long time. LLMs can mitigate this by generating synthetic data in ready-to-load form using Structured Outputs. Moreover, Structured Outputs can enrich unstructured documents with structured meta-data, improving retrieval accuracy in GenAI applications.5

Improving Reliability and Performance of RAG

Retrieval Augmented Generation (RAG) is a well-known technique that enables LLMs to generate content that is contextually relevant to the user’s query. RAG uses vector search to retrieve content related to the users’ query and feeds the content as context to an LLM, along with the user’s original query. The LLM responds to the query using the context, even if the query falls outside of the LLM’s training data distribution. This is supposed to reduce hallucination and make LLM responses grounded on facts. However, in practice RAG is vulnerable to hallucinations when LLMs deal with the task of comparing and summarizing data from multiple unstructured documents.

Effective techniques to mitigate hallucination in RAG often involve the transformation of unstructured data into structured data at the retrieval stage. For example, StructRAG is a RAG technique that aims to improve knowledge-intensive reasoning by building the context in various structured formats such as graphs and tables.6 Structured Outputs are a natural fit for building such context structures in RAG. Apart from this, Mistral has used Structured Outputs in its ‘LLM as a judge’ framework to evaluate the performance of RAG.7

Realtime Tracking and Observability

Web 2.0 has created a world where the consumer is also a producer. The new generation of Web users are ready to engage, providing original content, feedback, ideas, and alternatives. Companies already track customer sentiment using NLP tools, but LLMs expand the range of analysis that can be done on unstructured data. Structured Outputs in LLMs can extract and filter pointers of interest and alert stakeholders. It is even possible to graph and chart key indicators in real time.

Structured Outputs can create structured logs from semi-structured and unstructured IT systems logs and send them to downstream monitoring systems. The monitoring systems can track key parameters and raise alerts if needed. This can improve the observability of legacy systems which often lack structured logging, improving troubleshooting and threat detection.

Streamlining Automation

Processing unstructured data in current automation frameworks such as Robotic Process Automation is difficult. They often require human intervention when long text or images are involved. LLMs can fill these gaps in automation. They can accept text and images as input and feed structured data to automation workflows for further processing. This allows organizations to use traditional automation frameworks efficiently for use cases that do not need agentic automation.

Structured Outputs could also help to automate activities that are out of bounds for traditional automation frameworks. For example, ServiceNow uses Structured Outputs to automate the generation of workflow definitions from natural language user requests.8

Challenges

Structured Outputs are a recent innovation in LLMs, and have their fair share of challenges as LLMs continue to evolve. The performance of LLMs in Structured Outputs tasks is dependent on the quality of prompts. In the case of badly designed prompts or schema, LLMs may generate structurally correct but semantically flawed outputs. Performance may also vary across models.

Model cost is still a factor for consideration in GenAI solutions, even as McKinsey estimates them to be only 15% of the overall solution cost.9 Token count may increase in tasks that involve deeply nested schema,10 leading to increased latency and degraded performance, apart from spiraling costs. Focusing on proper prompt and schema design may offset these risks.

Structured Outputs may look like a straightforward functionality, but they require thorough planning and design. For maximum impact, it is important to implement Structured Outputs in line with the broader organizational strategy for AI transformation.

Conclusion

Data is digital gold when curated properly. LLM Structured Outputs can sift through unstructured data and output business critical structured data, which can help in AI transformation efforts. The ability to extract structured data from unstructured content may look plain and mundane, but this could unclog data pipelines, illuminate dashboards with new insights, streamline automation workflows, and improve reliability in IT and AI systems.

Structured Outputs use cases would fit within the augmentation tier of the Agentic AI Value Pyramid,11 leading to early wins in the AI transformation journey. However, successful AI transformation will depend on a clear organizational strategy and roadmap for full-scale adoption of agentic AI.

References

1. Deloitte, Now decides next: Generating a new future, January 2025, https://www.deloitte.com/content/dam/assets-zone3/us/en/docs/campaigns/2025/us-state-of-gen-ai-2024-q4.pdf

2. Harbert, T. Tapping the power of unstructured data, MIT Sloan School of Management, February 2021, https://mitsloan.mit.edu/ideas-made-to-matter/tapping-power-unstructured-data

3. Gartner, Gartner Predicts 80% of D&A Governance Initiatives Will Fail by 2027, February 2024, https://bit.ly/4kxNaW5

4. Krishnan, A. LLM-driven data pipelines with prompt() in MotherDuck and dbt,” December 2024. [Online]. Available: https://motherduck.com/blog/llm-data-pipelines-prompt-motherduck-dbt/

5. Celik, T. Advanced RAG: Automated Structured Metadata Enrichment, April 2025, https://haystack.deepset.ai/cookbook/metadata_enrichment

6. Li, Z., Chen, X., Yu, H., Lin, H., Lu, Y., Tang, Q., Huang, F., Han, X., Sun, L., and Li, Y. StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization,” October 2024, https://doi.org/10.48550/arXiv.2410.08815

7. Mistral. Evaluating RAG with LLM as a Judge, April 2025, https://mistral.ai/news/llm-as-rag-judge

8. Béchard, P. and Ayala, O.M. Reducing hallucination in structured outputs via Retrieval-Augmented Generation, April 2024, https://doi.org/10.48550/arXiv.2404.08189

9. McKinsey. Moving past gen AI’s honeymoon phase: Seven hard truths for CIOs to get from pilot to scale, May 2024, https://bit.ly/3IemJHD

10. Snowflake. AI_COMPLETE Structured Outputs, https://docs.snowflake.com/en/user-guide/snowflake-cortex/complete-structured-outputs

11. Sudalaimuthu, S. AI Agents: Automation is Not Enough, Communications, January 2025, https://cacm.acm.org/blogcacm/ai-agents-automation-is-not-enough/

Shanmugam Sudalaimuthu is a software architect with more than 20 years of experience building innovative solutions for Fortune 500 companies across diverse industries. He specializes in Generative AI and Cloud technologies, and holds a degree in Physics and a master’s degree in Computer Applications.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

LLM 结构化输出 AI转型 生成式AI 数据集成
相关文章