Contact

Contact HaxiTAG for enterprise services, consulting, and product trials.

Showing posts sorted by relevance for query openai. Sort by date Show all posts
Showing posts sorted by relevance for query openai. Sort by date Show all posts

Friday, January 23, 2026

From “Controlled Experiments” to “Replicable Scale”: How BNY’s Eliza Platform Turns Generative AI into a Bank-Grade Operating System

Opening: Context and Inflection Point

The Bank of New York Mellon (BNY) is not an institution that can afford to “experiment at leisure.” It operates at the infrastructural core of the global financial system—asset custody, clearing, and the movement and safeguarding of data and cash. As of the third quarter of 2025, the value of assets under custody and/or administration reached approximately USD 57.8 trillion. Any error, delay, or compliance lapse in its processes is therefore magnified into systemic risk. ([bny.com][1])

When ChatGPT ignited the wave of generative AI at the end of 2022, BNY did not confine its exploration to a small circle of engineers or innovation labs. Instead, it elevated the question to the level of how the enterprise itself should operate. If AI is destined to become the operating system of future technology, then within a systemically important financial institution it cannot exist as a peripheral tool. It must scale within clearly defined boundaries of governance, permissions, auditability, and accountability. ([OpenAI][2])

This marked the inflection point. BNY chose to build a centralized platform—Eliza—integrating model capabilities, governance mechanisms, and workforce enablement into a single, scalable system of work, developed in collaboration with frontier model providers such as OpenAI. ([OpenAI][2])

Problem Recognition and Internal Reflection: The Bottleneck Was Not Models, but Structural Imbalance

In large financial institutions, the main barrier to scaling AI is rarely compute or model availability. More often, it lies in three forms of structural imbalance:

  • Information silos and fragmented permissions: Data and knowledge across legal, compliance, business, and engineering functions fail to flow within a unified boundary, resulting in “usable data that cannot be used” and “available knowledge that cannot be found.”

  • Knowledge discontinuity and poor reuse: Point-solution proofs of concept generate prompts, agents, and best practices that are difficult to replicate across teams. Innovation is repeatedly reinvented rather than compounded.

  • Tension between risk review and experimentation speed: In high-risk industries, governance is often layered into approval stacks, slowing experimentation and deployment until both governance and innovation lose momentum.

BNY reached a clear conclusion: governance should not be the brake on AI at scale—it should be the accelerator. The prerequisite is to design governance into the system itself, rather than applying it as an after-the-fact patch. Both OpenAI’s case narrative and BNY’s official communications emphasize that Eliza’s defining characteristic is governance embedded at the system level. Prompts, agent development, model selection, and sharing all occur within a controlled environment, with use cases continuously reviewed through cross-functional mechanisms. ([OpenAI][2])

Strategic Inflection and the Introduction of an AI Platform: From “Using AI” to “Re-architecting Work”

BNY did not define generative AI as a point-efficiency tool. It positioned it as a system of work and a platform capability. This strategic stance is reflected in three concrete moves:

  1. Centralized AI Hub + Enterprise Platform Eliza
    A single entry point, a unified capability stack, and consistent governance and audit boundaries. ([OpenAI][2])

  2. From Use-Case Driven to Platform-Driven Adoption
    Every department is empowered to build first, with sharing and reuse enabling scale. Eliza now supports 125+ active use cases, with 20,000 employees actively building agents. ([OpenAI][2])

  3. Embedding “Deep Research” into the Decision Chain
    For complex tasks such as legal analysis, risk modeling, and scenario planning, multi-step reasoning is combined with internal and external data as a pre-decision thinking partner, working in tandem with agents to trigger follow-on actions. ([OpenAI][2])

Organizational Intelligence Re-architecture: From Departmental Coordination to Integrated Knowledge, Workflow, and Accountability

Eliza is not “another chat tool.” It represents a reconfiguration of how the organization operates. The transformation can be summarized along three linked pathways:

1. Departmental Coordination → Knowledge-Sharing Mechanisms

Within Eliza, BNY developed a mode of collaboration characterized by joint experimentation, shared prompts, reusable agents, and continuous iteration. Collaboration no longer means more meetings; it means faster collective validation and reuse. ([OpenAI][2])

2. Data Reuse → Formation of Intelligent Workflows

By unifying permissions, controls, and oversight at the platform level, Eliza allows “usable data” and “usable knowledge” to enter controlled workflows. This reduces redundant labor and gray processes while laying the foundation for scalable reuse. ([bny.com][3])

3. Decision Models → Model-Based Consensus

In high-risk environments, model outputs must be tied to accountability. BNY’s approach productizes governance itself: cross-functional review and visible, in-platform controls ensure that use cases evolve from the outset within a consistent risk and oversight framework. ([bny.com][3])

From HaxiTAG’s perspective, the abstraction is clear: the deliverable of AI transformation is not a single model, but a replicable intelligent work system. In product terms, this often corresponds to a composable platform architecture—such as YueLi Engine (knowledge computation and orchestration), EiKM (knowledge accumulation and reuse), and vertical systems like ESGtank—that connects knowledge, tools, workflows, and auditability within a unified boundary.

Performance and Quantified Impact: Proving That Scale Is More Than a Slogan

What makes BNY’s case persuasive is that early use cases were both measurable and repeatable:

  • Contract Review Assistant: For more than 3,000 supplier contracts per year, legal review time was reduced from four hours to one hour, a 75% reduction. ([OpenAI][2])

  • Platform Scale Metrics: With 125+ active use cases and 20,000 employees building agents, capability has expanded from a small group of experts to the organizational mainstream. ([bny.com][3])

  • Cultural and Capability Diffusion: Training programs and community-based initiatives encouraged employees to see themselves as problem solvers and agent builders, reinforced through cross-functional hackathons. ([OpenAI][2])

Together, these indicators point to a deeper outcome: AI’s value lies not merely in time savings, but in upgrading knowledge work from manual handling to controlled, autonomous workflows, thereby increasing organizational resilience and responsiveness.

Governance and Reflection: Balancing Technology and Ethics Through “Endogenous Governance”

In financial services, AI risks are tangible rather than theoretical—data misuse, privacy and compliance violations, hallucination-driven errors, permission overreach, and non-traceable audits can all escalate into reputational or regulatory crises.

BNY’s governance philosophy avoids adding yet another “AI approval layer.” Instead, governance is built into the platform itself:

  • Unified permissions, security protections, and oversight mechanisms;

  • Continuous pre- and post-deployment evaluation of use cases;

  • Governance designed to accelerate action, not suppress innovation. ([bny.com][3])

The lessons for peers are straightforward:

  1. Define accountability boundaries before autonomy: Without accountable autonomy, scalable agents are impossible.

  2. Productize governance, don’t proceduralize it: Governance trapped in documents and meetings cannot scale.

  3. Treat training as infrastructure: The real bottleneck is often the distribution of capability, not model performance.

Overview of AI Application Impact in BNY Scenarios

Application ScenarioAI Capabilities UsedPractical ImpactQuantified ResultsStrategic Significance
Supplier Contract ReviewNLP + Retrieval-Augmented Generation (RAG) + Structured SummarizationFaster legal review and greater consistencyReview time reduced from 4 hours to 1 hour (-75%); 3,000+ contracts/year ([OpenAI][2])Transforms high-risk knowledge work into auditable workflows
HR Policy Q&AEnterprise knowledge Q&A + Permission controlFewer manual requests; unified responsesReduced manual requests and improved consistency (no disclosed figures) ([OpenAI][2])Reduces organizational friction through knowledge reuse
Risk Insight AgentMulti-step reasoning + internal/external data fusionEarly identification of emerging risk signalsNo specific lead time disclosed (described as pre-emptive intervention) ([OpenAI][2])Enhances risk resilience through cognitive front-loading
Enterprise-Scale Platform (Eliza)Agent building/sharing + unified governance + controlled environmentExpands innovation from experts to the entire workforce125+ active use cases; 20,000 employees building agents ([bny.com][3])Turns AI into the organization’s operating system

HaxiTAG-Style Intelligent Leap: Delivering Experience and Value Transformation, Not a Technical Checklist

BNY’s case is representative not because of which model it adopted, but because it designed a replicable diffusion path for generative AI: platform-level boundaries, governance-driven acceleration, culture-shaping training, and trust built on measurable outcomes. ([OpenAI][2])

For HaxiTAG, this is precisely where productization and delivery methodology converge. With YueLi Engine, knowledge, data, models, and workflows are orchestrated into reusable intelligent pipelines; with EiKM, organizational experience is accumulated into searchable, reviewable knowledge assets; and through systems such as ESGtank, intelligence is embedded directly into compliance and governance frameworks. The result is AI that enters daily enterprise operations in a controllable, auditable, and replicable form.

When AI is truly embedded into an organization’s permission structures, audit trails, and accountability mechanisms, it ceases to be a passing efficiency trend—and becomes a compounding engine of long-term competitive advantage.

Related topic:

Tuesday, May 14, 2024

GPT Search: A Revolutionary Gateway to Information, fan's OpenAI and Google's battle on social media

In recent media reports and on social platforms like Twitter, we can observe a trend: an increasing number of people are discussing and anticipating the launch of OpenAI's so-called "GPT Search" product. Despite the enthusiasm and anticipation in these discussions, the fact remains that OpenAI has not declared the launch of a traditional search product. So, why is there so much focus on the direction of search?

Search as a Crucial Means of Input and Information Retrieval

Search engines have become an indispensable part of daily life because they satisfy the need for quick information retrieval. By simply entering keywords, users can obtain a large amount of relevant information in a short time, which is highly efficient and convenient. Search has become a familiar tool for answering questions, finding information, shopping, and planning travel, playing a key role in various aspects.

Broad Usage Scenarios and High Frequency

The attractiveness of search to tech companies and investors lies in its broad usage scenarios and high frequency. From individual users to enterprises, from academic research to everyday life, search engine applications cover almost every aspect of our lives. The high frequency of use means that any company that makes breakthroughs in search technology can quickly acquire a large user base and accumulate extensive data and user feedback in a short time, continuously optimizing the product and increasing user stickiness.

Commercial Value and Potential

The commercial value and potential of search engines are widely recognized. The existing advertising model has made search engine companies among the most profitable tech giants. By providing precise ad placement and personalized recommendations, search engines bring higher returns on investment for advertisers. With the development of big data and AI technologies, the personalization and intelligence of search engines continue to improve, making their commercial value even more significant. The scale and maturity of the search market mean that any new entrant will attract widespread attention and expectation.

Integration of GPT Technology and Search

Although OpenAI has not explicitly stated it will launch a traditional search product, its GPT technology (Generative Pre-trained Transformer) shows strong potential in information retrieval and processing. Through natural language processing (NLP) capabilities, GPT can understand user inputs and generate natural language text, allowing it to not only answer user questions but also engage in more complex conversations, write articles, generate code, and perform various other tasks.

The integration of GPT technology and search can break the limitations of traditional search. For instance, traditional search engines rely on indexing and keyword matching, whereas GPT, by understanding semantics, can better grasp user intent and provide more suitable answers. This means users no longer need to input precise keywords but can interact with the system through natural language, making the information retrieval process more intuitive and smooth.

Potential in Practical Applications

The potential of GPT Search in practical applications is immense. Firstly, in education and academia, GPT can serve as an intelligent assistant, helping users solve complex problems and providing study materials and suggestions. Secondly, in the business sector, GPT can be used for customer service, market analysis, product recommendations, and more, improving work efficiency and user satisfaction.

In social and content creation fields, GPT can also play an important role. By automatically generating high-quality content, GPT can assist creators in completing more creative work, saving time and effort. Additionally, in professional fields such as healthcare and law, GPT can provide expert consultation and advice, becoming a valuable assistant to professionals.

Continuously Developing Business Prospects

For OpenAI, applying GPT technology to the search domain means opening up a new business opportunity. By providing efficient and intelligent information retrieval services, OpenAI can attract a large number of users and corporate clients. This also brings abundant data resources and feedback, helping to continuously optimize and expand product features.

However, the success of GPT Search also faces some challenges. For example, ensuring the accuracy and reliability of answers, protecting user privacy and data security, and addressing potential biases and discrimination issues are all matters that need careful consideration and resolution.

In summary, the anticipation for OpenAI to launch GPT Search stems not only from the importance and broad application of search for information retrieval but also from the immense potential of GPT technology in natural language processing. Although OpenAI has no plans to launch a traditional search product at present, the application of GPT technology in the search field is indeed poised to change how we obtain information, bringing unprecedented intelligent experiences. In the future, as technology continues to develop and mature, we have reason to expect GPT Search to become a crucial gateway connecting us to the world of information.

Related topic:

OpenAI GPT Search
Generative Pre-trained Transformer (GPT)
Natural Language Processing (NLP)
Search Engine Technology
Information Retrieval Systems
Artificial Intelligence in Search
User Experience in Search Engines
Big Data and AI Integration
Commercial Potential of Search Engines
GPT Applications in Various Domains

GPT-4o: The Dawn of a New Era in Human-Computer Interaction

Mira Murati’s speech unveiled the mystery of OpenAI’s latest AI model, GPT-4o. This launch not only marks a significant technical breakthrough but also brings tremendous improvements in usability and user experience in human-computer interactions. Here is an in-depth analysis of the launch and the GPT-4o model.

1. Enhanced User Experience and Seamless Use

First, the launch reaffirmed OpenAI’s mission to make advanced AI tools freely available to everyone. Notably, OpenAI is not only offering ChatGPT for free but also striving to lower the barriers to its use. For example, they recently removed the account registration step, allowing users to access ChatGPT without a cumbersome process. Additionally, the launch announced the release of a desktop version of ChatGPT, further facilitating user access and operation.

Another standout feature of GPT-4o is its significant enhancement of the user experience. The new user interface design is more straightforward and intuitive, aiming to let users focus on interacting with ChatGPT rather than spending too much time on the interface.

2. Comprehensive Upgrades of GPT-4o

GPT-4o integrates the core intelligence of GPT-4 and has significantly improved its speed, text, visual, and audio comprehension abilities. Over the past few years, OpenAI has been committed to enhancing the intelligence level of its models, and GPT-4o represents a qualitative leap. The new model excels in multimodal interactions, processing and generating text, understanding, and responding to audio and visual content. This new capability propels ChatGPT to a new level of application, transforming it from a text-based tool to a truly multifunctional assistant.

3. Breakthrough in Voice Mode

The launch showcased a groundbreaking advancement in voice interaction with GPT-4o. Previous voice modes required integrating multiple models (e.g., transcription, speech synthesis) to provide voice interaction functionality, which increased latency and reduced interaction fluidity. GPT-4o has natively integrated these functions, significantly reducing latency issues.

This innovation allows for more natural and real-time voice conversations. GPT-4o can respond instantly and capture and reflect emotions during the conversation. For example, in a case demonstrated at the launch, ChatGPT could help users perform deep breathing exercises to reduce tension and provide instant feedback based on the user’s speech speed and breathing rate. This capability makes human-computer dialogue more humane and natural, offering users an unprecedented experience.

4. New Experience of Multimodal Interaction

The visual capabilities of GPT-4o were another highlight of the launch. With this feature, users can directly upload screenshots, photos, or files containing text and images and have conversations with ChatGPT. Whether interpreting text in images or helping solve practical problems like solving math equations or analyzing code, GPT-4o performs effortlessly.

In a case demonstrated at the launch, users could take a photo of a linear equation with their phone camera, and ChatGPT could automatically recognize the equation and provide step-by-step problem-solving guidance. Additionally, GPT-4o can analyze scenes in pictures and recognize emotions in the facial expressions of people in photos, providing a more interactive experience. For example, when users show a selfie, GPT-4o can immediately analyze and provide feedback on the user’s emotional state, further bridging the gap between the user and the AI.

5. Stronger Support for Developers

The launch also announced the release of GPT-4o’s API, enabling developers to integrate this advanced model into their applications. This not only greatly expands the application scenarios of GPT-4o but also provides more innovative space for developers. The new model is not only faster than previous versions but also more cost-effective, which is a significant advantage for developers looking to deploy AI tools on a large scale.

6. Security and Ethical Considerations

With the enhancement of GPT-4o’s multimodal capabilities, security issues become increasingly important. Real-time audio and video processing bring new challenges such as privacy breaches and fake information generation. Therefore, OpenAI emphasized that they have been working with multiple stakeholders, including governments, media, entertainment, and various social institutions, to ensure the safe and responsible launch of new technologies.

These efforts include built-in anti-abuse mechanisms and long-term research to effectively address potential risks in various application scenarios. OpenAI showcased their efforts in protecting user privacy, data security, and preventing technology abuse, ensuring that every interaction with GPT-4o occurs in a safe and controlled environment.

7. Impressive Live Demonstrations

In addition to technical enhancements, the launch featured multiple live demonstrations showcasing the new features of GPT-4o. For example, using GPT-4o for real-time language translation not only instantly translated conversation content but also adjusted translation quality and style based on context and semantics. Additionally, GPT-4o demonstrated the ability to judge user emotions through eye contact and provide corresponding feedback, offering a fresh interactive experience.

Through these live demonstrations, the audience could intuitively feel the powerful capabilities and humanized design of GPT-4o. This not only enhanced user trust in new technology but also inspired more people to imagine and expect AI application scenarios.

8. Conclusion and Outlook

In summary, the release of GPT-4o is not only a technological advancement but also a revolution in human-computer interaction experience. From lowering the usage threshold and enhancing interaction naturalness to introducing multimodal capabilities and stronger developer support, GPT-4o truly elevates AI technology to a new height. Meanwhile, OpenAI demonstrates a high level of responsibility and foresight in ensuring the safety and ethical standards of the technology.

Looking forward, as GPT-4o's capabilities gradually roll out, we can expect to see more innovative and practical AI applications in fields such as education, healthcare, entertainment, and enterprise services. This will not only significantly improve work efficiency but also provide users with a richer and more personalized experience.

Finally, the release of GPT-4o not only showcases OpenAI’s leading position in the AI field but also sets a new standard for the entire AI community. Based on this, we have reason to believe that future AI technologies will be more intelligent, more humane, and bring more positive changes and possibilities to human society. Whether you are a regular user, developer, or industry expert, GPT-4o will be a new era tool worth anticipating and exploring.

Related topic:

1. GPT-4o Launch Analysis
2. Enhanced User Experience with ChatGPT Desktop Version
3. Multimodal Interaction Capabilities of GPT-4o
4. Voice Mode Advancements in GPT-4o
5. Visual Comprehension and Image Analysis with GPT-4o
6. Developer Support for GPT-4o API Integration
7. Security and Ethical Considerations in GPT-4o Deployment
8. Live Demonstrations of GPT-4o Features
9. Real-time Language Translation by GPT-4o
10. Emotion Recognition and Feedback in GPT-4o Interactions

Friday, May 8, 2026

LLMs Enter Enterprise Core Systems — The Real Question Is No Longer "Is the Model Strong Enough?"

 In the past two years, enterprise AI infrastructure has undergone a distinct transformation.

Enterprises no longer lack models.

From OpenAI, Anthropic, Google Gemini to DeepSeek, vLLM, SGLang, and Ollama, model capabilities and inference performance are evolving rapidly. Yet, once enterprises enter real production environments, they begin confronting another set of more pragmatic challenges:

  • AI answers "look correct" but cannot prove their basis;
  • Different models exhibit vast capability disparities, making business systems increasingly difficult to maintain;
  • Enterprise knowledge is scattered across documents, databases, emails, and audio-visual content, unable to coalesce into a unified understanding;
  • Inference costs, model routing, data security, and protocol compatibility gradually become new sources of system complexity;
  • Enterprises have already adopted AI, yet still cannot truly "trust AI in production."

This is precisely why Yueli KGM Computing is now open-source.

It is an enterprise production-grade AI application framework.

More accurately, it is:

The "knowledge computation and inference orchestration infrastructure layer" for the enterprise AI application era.


What Is Yueli KGM Computing?

An "Inference Orchestration + Compatible Gateway + Knowledge Computation" Middleware for Enterprise AI

Yueli KGM Computing is an open-source, enterprise-grade knowledge computation engine and inference orchestration middleware.

Its core positioning is unequivocal:

Use the determinism of knowledge graphs to constrain the probabilistic nature of large language models.

It doesn't seek to "make models smarter."

Instead, it addresses:

  • How to make enterprise AI more trustworthy;
  • How to make multi-model systems governable;
  • How to truly embed inference capabilities into enterprise business systems;
  • How to equip AI infrastructure with observability, replaceability, and auditability.

It can serve as:

  • An OpenAI / Anthropic compatible gateway;
  • A multi-model routing and scheduling layer;
  • An enterprise knowledge graph and GraphRAG engine;
  • A privatized AI infrastructure control plane;
  • An enterprise AI middleware embedded into existing systems.

It can also:

  • Connect to local vLLM / Ollama / SGLang;
  • Integrate with OpenAI-compatible cloud services;
  • Orchestrate a hybrid of local inference and cloud MaaS;
  • Deliver model governance and knowledge augmentation under a unified API gateway and scheduling controller.

Why Does Enterprise AI Need a "Knowledge Computation Layer"?

For many enterprise AI projects today, the real problem is not model performance.

It is this:

Enterprise Knowledge Is Not Entering the Inference Pipeline

The problem with traditional RAG is:

  • Retrieval results are merely "similar text";
  • They lack relational structures;
  • They lack domain ontologies;
  • They lack factual boundaries;
  • They lack source verifiability.

The result:

The model generates a wrong answer that "looks exactly like the right answer."

In industries such as finance, healthcare, government, manufacturing, new energy, intellectual property, and compliance, such problems are unacceptable.

Therefore, the core capability of Yueli KGM Computing is not simple vector retrieval.

It is:

KGM (Knowledge Generation Modeling)

That is:

An LLM Inference System Constrained by Knowledge Graphs

It will:

  1. Extract entities and relationships from enterprise documents, databases, audio-visual content, and business systems;
  2. Construct an enterprise private domain ontology;
  3. Organize knowledge into a reasonable graph;
  4. Perform GraphRAG retrieval before inference;
  5. Inject factual nodes as constraint context into the LLM;
  6. Output traceable, verifiable results.

This means:

AI is no longer "freestyling."

Instead:

It performs controlled reasoning within the boundaries of enterprise knowledge.


What Does Yueli KGM Computing Actually Deliver?

A Unified Industrial Protocol AI Gateway Layer

Within the same process, KGM simultaneously provides:

  • OpenAI Compatible API
  • Anthropic Claude Compatible API

Including:

  • /v1/chat/completions
  • /v1/responses
  • /v1/messages

And automatically completes:

  • tool_calls
  • tool_use

Dual-protocol semantic mapping.

This means:

Enterprise applications only need to connect to a single Base URL.

No matter how the underlying models change, business systems remain agnostic.


Dynamic Inference Orchestration and Model Scheduling

KGM supports:

  • Local inference;
  • Cloud MaaS;
  • Multi-model hybrid scheduling;
  • Cost-based scheduling;
  • Performance-based scheduling;
  • Dynamic routing by task type.

For example:

  • Sensitive data → On-premise Ollama;
  • Long text → Gemini;
  • Highly complex reasoning → Claude;
  • High throughput → vLLM;
  • Low cost → DeepSeek.

All of this can be accomplished through declarative configuration.

Rather than rewriting a routing layer for every project.


Knowledge Graph-Driven GraphRAG

This is KGM's most central capability.

Compared to traditional vector RAG:

KGM constructs:

  • Enterprise domain ontology;
  • Relationship graphs;
  • Contextual reasoning paths;
  • Structured factual constraints.

Therefore, it not only knows:

"Which texts are similar."

It also knows:

"What relationships exist among pieces of knowledge."

This is the critical leap for enterprise AI from "chat tool" to "business system."


Enterprise-Grade Control Plane and Observability

After going live, a significant number of AI projects rapidly descend into an "ungovernable state."

Enterprises find themselves unable to answer:

  • Which model is providing the service?
  • Which requests are the most costly?
  • Which inference node is failing?
  • Which API has abnormal latency?
  • Which model has a higher hallucination rate?

KGM provides:

  • Prometheus Metrics;
  • Runtime lifecycle management;
  • Circuit breaker mechanisms;
  • Structured logging;
  • Model asset governance;
  • Runtime control plane;
  • Multi-tenant isolation;
  • Data security policies.

It is not a simple proxy.

It is a genuinely operable AI middleware.


How Do Enterprises Embed Yueli KGM?

Scenario One: Enterprise Knowledge Q&A

The typical path:

Enterprise Documents / Databases / Wikis / Emails
                    ↓
            KGM Semantic Parsing
                    ↓
          GraphRAG Knowledge Graph
                    ↓
            LLM Constrained Inference
                    ↓
        Traceable, Trustworthy Answers

R&D teams no longer depend on:

"Who remembers the solution from back then?"

Instead, they directly ask:

  • In which version did this issue appear?
  • How was it fixed at the time?
  • Which systems were affected?
  • Who was involved in the decision?

KGM will construct a complete knowledge chain from:

  • Git;
  • Confluence;
  • Emails;
  • Meeting records;
  • Technical documentation.

Scenario Two: Finance and Compliance Review

The biggest risk with traditional LLMs:

Citing non-existent regulations.

KGM's approach is:

  • Build a regulatory knowledge graph;
  • Structure regulatory clauses;
  • Restrict reasoning within knowledge boundaries;
  • Directly trigger a "knowledge gap" alert beyond those boundaries.

This means:

AI no longer "guesses."

It reasons within the enterprise's rule system.


Scenario Three: AI-Native Product Embedding

For engineering teams:

KGM can serve as the underlying AI Runtime.

Including:

  • Multi-model scheduling;
  • GraphRAG;
  • Tool Calling;
  • MCP;
  • Memory;
  • Knowledge Runtime;
  • Prompt orchestration;
  • Runtime Observability.

Engineering teams no longer need to rebuild:

  • Gateways;
  • Routing;
  • Metrics;
  • Tool Runtime;
  • Protocol adaptation;
  • Multi-model compatibility layers.

Scenario Four: Audio-Visual Semantic Computing

This is a direction often overlooked by enterprises today but is exceptionally high-value.

KGM supports:

  • Video caption parsing;
  • Semantic label extraction;
  • Meeting content knowledge transformation;
  • Training video knowledge graphs;
  • Audio-visual Q&A.

For example:

An enterprise can directly ask:

"In last quarter's product meetings, what were the disputes regarding pricing strategy?"

The system will automatically locate:

  • The corresponding meeting;
  • The corresponding individuals;
  • The corresponding viewpoints;
  • The corresponding timeline.

What Is Its Relationship to LangChain, LlamaIndex, and vLLM?

This is not a competitive relationship.

Rather, it is:

A Layered Relationship

LayerRepresentative ProjectCore Responsibility
InferencevLLM / SGLangHigh-performance inference
ApplicationLangChain / DifyAgent and Workflow
DataLlamaIndexData connection and retrieval
MiddlewareYueli KGMInference orchestration + Protocol compatibility + Knowledge constraints

Therefore, the most rational enterprise architecture often is:

  • vLLM for inference;
  • LangChain for business agents;
  • Dify or BotFactory for low-code workflows;
  • KGM as the unified AI middleware and knowledge computation layer.

Why MIT Open Source?

The Yueli KGM Computing GitHub Repository and NPM package are open-sourced under the MIT License.

This means:

  • Enterprises can use it freely for commercial purposes;
  • They can modify it for private deployment;
  • They can deeply integrate it;
  • They can build their own industry-specific versions.

The true value of Yueli KGM Computing does not lie in closed-source code.

It lies in:

  • Enterprise AI infrastructure capability;
  • Industry knowledge modeling experience;
  • Private deployment delivery capability;
  • Knowledge engineering systems;
  • Data intelligence and inference architecture practices.

The Next Phase of Enterprise AI Is Shifting from "Model Competition" to "Knowledge Governance"

Over the past two years, the industry has been discussing:

Whose model is stronger.

But in the next five years, the questions enterprises will truly care about will become:

  • Who can make AI more trustworthy?
  • Who can make AI more stable?
  • Who can make AI truly enter business systems?
  • Who can equip AI with enterprise-grade governance capabilities?

The significance of Yueli KGM Computing lies precisely here.

It is a crucial middleware layer for enterprise AI transitioning from the experimental stage to production-grade infrastructure.

Related topic:

Thursday, October 31, 2024

Enhancing Workforce Productivity and Human-AI Collaboration Through Generative AI

Generative AI's Impact on the Workforce

It's interesting to see the growing influence of generative AI on the workforce as suggested by the recent paper. The estimates provided offer a window into the potential impact of AI on labor productivity. Here's a brief summary of the key points:

- The paper estimates that between 0.5% and 3.5% of all work hours in the U.S. are currently being assisted by generative AI.

- This translates to an increase in labor productivity of between 0.125 and 0.875 percentage points.

These figures indicate that generative AI could be contributing significantly to productivity gains in the American workforce. It's important to consider the following implications:

1. Economic Growth: Higher labor productivity could contribute to overall economic growth and competitiveness.

2. Job Transformation: The role of human workers may evolve as AI takes on more tasks. This could lead to the creation of new job categories and the retraining of the workforce.

3. Skill Requirements: There may be a shift in the types of skills that are in demand, with a growing need for workers who can collaborate with AI systems effectively.

4. Ethical and Social Considerations: As AI becomes more integrated into the workforce, there could be ethical questions regarding privacy, job displacement, and the overall impact on society.

Understanding the dynamics of AI's role in the workforce is crucial for policymakers, businesses, and individuals as they navigate the future of work.

Generative AI in Practice

The recent paper's estimate suggests that generative AI is already playing a significant role in the U.S. workforce, potentially impacting up to 3.5 percent of all work hours. This could translate to a notable increase in labor productivity, ranging from 0.125 to 0.875 percentage points.

Sarah Friar, CFO of OpenAI, reinforces this trend, emphasizing that AI is not just an experimental technology but is actively being integrated into various sectors. She points out that OpenAI's major enterprise clients are in education and healthcare, with financial services, including investment banks, also being a significant market.

Friar's comments hint at the potential for artificial general intelligence (AGI) to arrive sooner than anticipated, with tangible value already being realized in current AI products. She shares an anecdote where a lawyer used OpenAI's GPT-3 (o1) to create a legal brief, noting the lawyer's willingness to pay significantly more for paralegal services for the same task. However, the cost savings from using AI in this context are questionable, given the average hourly pay for paralegals.

Despite these advancements, OpenAI's foray into the enterprise sector appears to be facing challenges. Friar notes that 75% of the company's business revenue comes from consumer users, with only a small percentage of the 250 million weekly active users converting to paying customers at a rate of $20+ per month. This suggests that while AI technology is advancing rapidly, the enterprise adoption and monetization may be slower than anticipated.

Related Topic