Contact

Contact HaxiTAG for enterprise services, consulting, and product trials.

Showing posts with label openAI. Show all posts
Showing posts with label openAI. Show all posts

Friday, January 23, 2026

From “Controlled Experiments” to “Replicable Scale”: How BNY’s Eliza Platform Turns Generative AI into a Bank-Grade Operating System

Opening: Context and Inflection Point

The Bank of New York Mellon (BNY) is not an institution that can afford to “experiment at leisure.” It operates at the infrastructural core of the global financial system—asset custody, clearing, and the movement and safeguarding of data and cash. As of the third quarter of 2025, the value of assets under custody and/or administration reached approximately USD 57.8 trillion. Any error, delay, or compliance lapse in its processes is therefore magnified into systemic risk. ([bny.com][1])

When ChatGPT ignited the wave of generative AI at the end of 2022, BNY did not confine its exploration to a small circle of engineers or innovation labs. Instead, it elevated the question to the level of how the enterprise itself should operate. If AI is destined to become the operating system of future technology, then within a systemically important financial institution it cannot exist as a peripheral tool. It must scale within clearly defined boundaries of governance, permissions, auditability, and accountability. ([OpenAI][2])

This marked the inflection point. BNY chose to build a centralized platform—Eliza—integrating model capabilities, governance mechanisms, and workforce enablement into a single, scalable system of work, developed in collaboration with frontier model providers such as OpenAI. ([OpenAI][2])

Problem Recognition and Internal Reflection: The Bottleneck Was Not Models, but Structural Imbalance

In large financial institutions, the main barrier to scaling AI is rarely compute or model availability. More often, it lies in three forms of structural imbalance:

  • Information silos and fragmented permissions: Data and knowledge across legal, compliance, business, and engineering functions fail to flow within a unified boundary, resulting in “usable data that cannot be used” and “available knowledge that cannot be found.”

  • Knowledge discontinuity and poor reuse: Point-solution proofs of concept generate prompts, agents, and best practices that are difficult to replicate across teams. Innovation is repeatedly reinvented rather than compounded.

  • Tension between risk review and experimentation speed: In high-risk industries, governance is often layered into approval stacks, slowing experimentation and deployment until both governance and innovation lose momentum.

BNY reached a clear conclusion: governance should not be the brake on AI at scale—it should be the accelerator. The prerequisite is to design governance into the system itself, rather than applying it as an after-the-fact patch. Both OpenAI’s case narrative and BNY’s official communications emphasize that Eliza’s defining characteristic is governance embedded at the system level. Prompts, agent development, model selection, and sharing all occur within a controlled environment, with use cases continuously reviewed through cross-functional mechanisms. ([OpenAI][2])

Strategic Inflection and the Introduction of an AI Platform: From “Using AI” to “Re-architecting Work”

BNY did not define generative AI as a point-efficiency tool. It positioned it as a system of work and a platform capability. This strategic stance is reflected in three concrete moves:

  1. Centralized AI Hub + Enterprise Platform Eliza
    A single entry point, a unified capability stack, and consistent governance and audit boundaries. ([OpenAI][2])

  2. From Use-Case Driven to Platform-Driven Adoption
    Every department is empowered to build first, with sharing and reuse enabling scale. Eliza now supports 125+ active use cases, with 20,000 employees actively building agents. ([OpenAI][2])

  3. Embedding “Deep Research” into the Decision Chain
    For complex tasks such as legal analysis, risk modeling, and scenario planning, multi-step reasoning is combined with internal and external data as a pre-decision thinking partner, working in tandem with agents to trigger follow-on actions. ([OpenAI][2])

Organizational Intelligence Re-architecture: From Departmental Coordination to Integrated Knowledge, Workflow, and Accountability

Eliza is not “another chat tool.” It represents a reconfiguration of how the organization operates. The transformation can be summarized along three linked pathways:

1. Departmental Coordination → Knowledge-Sharing Mechanisms

Within Eliza, BNY developed a mode of collaboration characterized by joint experimentation, shared prompts, reusable agents, and continuous iteration. Collaboration no longer means more meetings; it means faster collective validation and reuse. ([OpenAI][2])

2. Data Reuse → Formation of Intelligent Workflows

By unifying permissions, controls, and oversight at the platform level, Eliza allows “usable data” and “usable knowledge” to enter controlled workflows. This reduces redundant labor and gray processes while laying the foundation for scalable reuse. ([bny.com][3])

3. Decision Models → Model-Based Consensus

In high-risk environments, model outputs must be tied to accountability. BNY’s approach productizes governance itself: cross-functional review and visible, in-platform controls ensure that use cases evolve from the outset within a consistent risk and oversight framework. ([bny.com][3])

From HaxiTAG’s perspective, the abstraction is clear: the deliverable of AI transformation is not a single model, but a replicable intelligent work system. In product terms, this often corresponds to a composable platform architecture—such as YueLi Engine (knowledge computation and orchestration), EiKM (knowledge accumulation and reuse), and vertical systems like ESGtank—that connects knowledge, tools, workflows, and auditability within a unified boundary.

Performance and Quantified Impact: Proving That Scale Is More Than a Slogan

What makes BNY’s case persuasive is that early use cases were both measurable and repeatable:

  • Contract Review Assistant: For more than 3,000 supplier contracts per year, legal review time was reduced from four hours to one hour, a 75% reduction. ([OpenAI][2])

  • Platform Scale Metrics: With 125+ active use cases and 20,000 employees building agents, capability has expanded from a small group of experts to the organizational mainstream. ([bny.com][3])

  • Cultural and Capability Diffusion: Training programs and community-based initiatives encouraged employees to see themselves as problem solvers and agent builders, reinforced through cross-functional hackathons. ([OpenAI][2])

Together, these indicators point to a deeper outcome: AI’s value lies not merely in time savings, but in upgrading knowledge work from manual handling to controlled, autonomous workflows, thereby increasing organizational resilience and responsiveness.

Governance and Reflection: Balancing Technology and Ethics Through “Endogenous Governance”

In financial services, AI risks are tangible rather than theoretical—data misuse, privacy and compliance violations, hallucination-driven errors, permission overreach, and non-traceable audits can all escalate into reputational or regulatory crises.

BNY’s governance philosophy avoids adding yet another “AI approval layer.” Instead, governance is built into the platform itself:

  • Unified permissions, security protections, and oversight mechanisms;

  • Continuous pre- and post-deployment evaluation of use cases;

  • Governance designed to accelerate action, not suppress innovation. ([bny.com][3])

The lessons for peers are straightforward:

  1. Define accountability boundaries before autonomy: Without accountable autonomy, scalable agents are impossible.

  2. Productize governance, don’t proceduralize it: Governance trapped in documents and meetings cannot scale.

  3. Treat training as infrastructure: The real bottleneck is often the distribution of capability, not model performance.

Overview of AI Application Impact in BNY Scenarios

Application ScenarioAI Capabilities UsedPractical ImpactQuantified ResultsStrategic Significance
Supplier Contract ReviewNLP + Retrieval-Augmented Generation (RAG) + Structured SummarizationFaster legal review and greater consistencyReview time reduced from 4 hours to 1 hour (-75%); 3,000+ contracts/year ([OpenAI][2])Transforms high-risk knowledge work into auditable workflows
HR Policy Q&AEnterprise knowledge Q&A + Permission controlFewer manual requests; unified responsesReduced manual requests and improved consistency (no disclosed figures) ([OpenAI][2])Reduces organizational friction through knowledge reuse
Risk Insight AgentMulti-step reasoning + internal/external data fusionEarly identification of emerging risk signalsNo specific lead time disclosed (described as pre-emptive intervention) ([OpenAI][2])Enhances risk resilience through cognitive front-loading
Enterprise-Scale Platform (Eliza)Agent building/sharing + unified governance + controlled environmentExpands innovation from experts to the entire workforce125+ active use cases; 20,000 employees building agents ([bny.com][3])Turns AI into the organization’s operating system

HaxiTAG-Style Intelligent Leap: Delivering Experience and Value Transformation, Not a Technical Checklist

BNY’s case is representative not because of which model it adopted, but because it designed a replicable diffusion path for generative AI: platform-level boundaries, governance-driven acceleration, culture-shaping training, and trust built on measurable outcomes. ([OpenAI][2])

For HaxiTAG, this is precisely where productization and delivery methodology converge. With YueLi Engine, knowledge, data, models, and workflows are orchestrated into reusable intelligent pipelines; with EiKM, organizational experience is accumulated into searchable, reviewable knowledge assets; and through systems such as ESGtank, intelligence is embedded directly into compliance and governance frameworks. The result is AI that enters daily enterprise operations in a controllable, auditable, and replicable form.

When AI is truly embedded into an organization’s permission structures, audit trails, and accountability mechanisms, it ceases to be a passing efficiency trend—and becomes a compounding engine of long-term competitive advantage.

Related topic:

Tuesday, April 22, 2025

Analysis and Interpretation of OpenAI's Research Report "Identifying and Scaling AI Use Cases"

Since the advent of artificial intelligence (AI) technology in the public sphere, its applications have permeated every aspect of the business world. Research conducted by OpenAI in collaboration with leading industry players shows that AI is reshaping productivity dynamics in the workplace. Based on in-depth analysis of 300 successful case studies, 4,000 adoption surveys, and data from over 2 million business users, this report systematically outlines the key paths and strategies for AI application deployment. The study shows that early adopters have achieved 1.5 times faster revenue growth, 1.6 times higher shareholder returns, and 1.4 times better capital efficiency compared to industry averages. However, it is noteworthy that only 1% of companies believe their AI investments have reached full maturity, highlighting a significant gap between the depth of technological application and the realization of business value.

AI Generative AI Opportunity Identification Framework

Repetitive Low-Value Tasks

The research team found that knowledge workers spend an average of 12.7 hours per week on tasks such as document organization and data entry. For instance, at LaunchDarkly, the Chief Product Officer created an "Anti-To-Do List," delegating 17 routine tasks such as competitor tracking and KPI monitoring to AI, which resulted in a 40% increase in strategic decision-making time. This shift not only improved efficiency but also reshaped the value evaluation system for roles. For example, a financial services company used AI to automate 82% of its invoice verification work, enabling its finance team to focus on optimizing cash flow forecasting models, resulting in a 23% improvement in cash turnover efficiency.

Breaking Through Skill Bottlenecks

AI has demonstrated its unique bridging role in cross-departmental collaboration scenarios. A biotech company’s product team used natural language to generate prototype design documents, reducing the product requirement review cycle from an average of three weeks to five days. More notably, the use of AI tools for coding by non-technical personnel is becoming increasingly common. Surveys indicate that the proportion of marketing department employees using AI to write Python scripts jumped from 12% in 2023 to 47% in 2025, with 38% of automated reporting systems being independently developed by business staff.

Handling Ambiguity in Scenarios

When facing open-ended business challenges, AI's heuristic thinking demonstrates its unique value. A retail brand's marketing team used voice interaction to brainstorm advertising ideas, increasing quarterly marketing plan output by 2.3 times. In the strategic planning field, AI-assisted SWOT analysis tools helped a manufacturing company identify four potential blue ocean markets, two of which saw market share in the top three within six months.

Six Core Application Paradigms

The Content Creation Revolution

AI-generated content has surpassed simple text reproduction. In Promega's case, by uploading five of its best blog posts to train a custom model, the company increased email open rates by 19% and reduced content production cycles by 67%. Another noteworthy innovation is style transfer technology—financial institutions have developed models trained on historical report data that automatically maintain consistency in technical terminology, improving compliance review pass rates by 31%.

Empowering Deep Research

The new agentic research system can autonomously complete multi-step information processing. A consulting company used AI's deep research functionality to analyze trends in the healthcare industry. The system completed the analysis of 3,000 annual reports within 72 hours and generated a cross-verified industry map, achieving 15% greater accuracy than manual analysis. This capability is particularly outstanding in competitive intelligence—one technology company leveraged AI to monitor 23 technical forums in real-time, improving product iteration response times by 40%.

Democratization of Coding Capabilities

Tinder's engineering team revealed how AI reshapes development workflows. In Bash script writing scenarios, AI assistance reduced unconventional syntax errors by 82% and increased code review pass rates by 56%. Non-technical departments are also significantly adopting coding applications—at a retail company, the marketing department independently developed a customer segmentation model that increased promotion conversion rates by 28%, with a development cycle that was only one-fifth of the traditional method.

The Transformation of Data Analysis

Traditional data analysis processes are undergoing fundamental changes. After uploading quarterly sales data, an e-commerce platform's AI not only generated visual charts but also identified three previously unnoticed inventory turnover anomalies, preventing potential losses of $1.2 million after verification. In the finance field, AI-driven data coordination systems shortened the monthly closing cycle from nine days to three days, with an anomaly detection accuracy rate of 99.7%.

Workflow Automation

Intelligent automation has evolved from simple rule execution to a cognitive level. A logistics company integrated AI with IoT devices to create a dynamic route planning system, reducing transportation costs by 18% and increasing on-time delivery rates to 99.4%. In customer service, a bank deployed an intelligent ticketing system that autonomously handled 89% of common issues, routing the remaining cases to the appropriate experts, leading to a 22% increase in customer satisfaction.

Evolution of Strategic Thinking

AI is changing the methodology for strategic formulation. A pharmaceutical company used generative models to simulate clinical trial plans, speeding up R&D pipeline decision-making by 40% and reducing resource misallocation risks by 35%. In merger and acquisition assessments, a private equity firm leveraged AI for in-depth data penetration analysis of target companies, identifying three financial anomalies and avoiding potential investment losses of $450 million.

Implementation Path and Risk Warnings

The research found that successful companies generally adopt a "three-layer advancement" strategy: leadership sets strategic direction, middle management establishes cross-departmental collaboration mechanisms, and grassroots innovation is stimulated through hackathons. A multinational group demonstrated that setting up an "AI Ambassador" system could increase the efficiency of use case discovery by three times. However, caution is needed regarding the "technology romanticism" trap—one retail company overly pursued complex models, leading to 50% of AI projects being discontinued due to insufficient ROI.

HaxiTAG’s team, after reading OpenAI's research report openai-identifying-and-scaling-ai-use-cases.pdf, analyzed its implementation value and conflicts. The report emphasizes the need for leadership-driven initiatives, with generative AI enterprise applications as a future investment. Although 92% of effective use cases come from grassroots practices, balancing top-down design with bottom-up innovation requires more detailed contingency strategies. Additionally, while the research emphasizes data-driven decision-making, the lack of a specific discussion on data governance systems in the case studies may affect the implementation effectiveness. It is recommended that a dynamic evaluation mechanism be established during implementation to match technological maturity with organizational readiness, ensuring a clear and measurable value realization path.

Related Topic

Unlocking the Potential of RAG: A Novel Approach to Enhance Language Model's Output Quality - HaxiTAG
Enterprise-Level LLMs and GenAI Application Development: Fine-Tuning vs. RAG Approach - HaxiTAG
Innovative Application and Performance Analysis of RAG Technology in Addressing Large Model Challenges - HaxiTAG
Revolutionizing AI with RAG and Fine-Tuning: A Comprehensive Analysis - HaxiTAG
The Synergy of RAG and Fine-tuning: A New Paradigm in Large Language Model Applications - HaxiTAG
How to Build a Powerful QA System Using Retrieval-Augmented Generation (RAG) Techniques - HaxiTAG
The Path to Enterprise Application Reform: New Value and Challenges Brought by LLM and GenAI - HaxiTAG
LLM and GenAI: The New Engines for Enterprise Application Software System Innovation - HaxiTAG
Exploring Information Retrieval Systems in the Era of LLMs: Complexity, Innovation, and Opportunities - HaxiTAG
AI Search Engines: A Professional Analysis for RAG Applications and AI Agents - GenAI USECASE