AgenticUniverse - Previously Formi
  1. Our Technical Note
AgenticUniverse - Previously Formi
  • Our Technical Note
    • Why Open AI is not Enough
    • How business Outcomes would Change Radically with AgenticUniverse
    • Our Research
      • STT - Nuances and Insights
      • Solving for STT Constraints
  • Generate Token
    • Login/Generate Token
      POST
  • Agent Configuration
    • Model Configuration
      • Configuration Helpers
        • Supported Providers
        • Supported Models
        • Supported Parameters
      • Get Model Configuration
      • Set Model Configuration
    • State Machine
      • Edge
        • Legacy
          • Create Edge
          • Edge Details
          • Update Edge
          • Delete Edge
        • Update edge properties
        • Get edge details
        • Delete an edge
        • Create an edge (transition) between two states
      • State
        • Create State from Template
        • Get State Information
        • Update State
        • Delete State
      • Get State Machine Structure
    • Prompt Templates
      • Get All Templates
      • Render Template
    • Tools
      • Get Tools List
      • Add Tool
      • Update Tool
      • Delete Tool
    • Get All Agents
      GET
    • Single Agent Details
      GET
    • Create Agent
      POST
    • Update Agent Details
      PUT
    • Enable Dashboard For An Outlet
      POST
    • Disable Dashboard For An Outlet
      POST
    • Get Call queue Sheet ID
      GET
  • Interactions
    • Pre-Interaction Context
      • Schedule an Interaction
      • Update an Interaction Id
      • Delete an Interaction Id
      • Clear all interactions
      • Get Summarized Interaction Info
    • Interaction Modalities
      • Video
        • Generation
          • Generate Welcome Video
      • Text
        • Start Interaction
        • Create Response
        • End Interaction
      • Voice
        • Connection Configuration
          • Quickstart
          • Connecting Twilio
          • Connecting Exotel
          • Formi WebSocket Configuration Guide
          • Create a New Connection Vendor
          • Get All Connection Vendors
          • Update a Connection Vendor
          • Delete a Connection Vendor
          • Get Agent's Connection Config
          • Add or Update Agent's Connection Config
    • Post Interaction Configuration
      • Email Destination Configuration
      • Variables CRUD
        • Get all required variables for the outlet with map
        • Modify variable definition for outlet
        • Add a new variable for the outlet
        • DELETE variable for outlet
        • Connect Variable to a destination
      • Destinations CRUD
        • Get all destinations for the outlet
        • Modify Destination for outlet
        • Add a new Destination for the outlet
        • DELETE Destinations for outlet
    • Get Interaction Summary
      GET
    • Resolve an Escalated Interaction
      POST
    • Get the Interaction list
      GET
    • Get Information regarding Single Interaction
      GET
  • Agent Utilisation
    • Get Credits Available
      GET
    • Interaction Utilisation
      GET
    • Model Utilisation
      GET
  • Webhooks
    • Get webhook URL
    • Update webhook URL
    • Get webhook metadata
    • Modify webhook metadata
    • Get reservation ingestion metadata
  • Untitled Endpoint
    POST
  1. Our Technical Note

Why Open AI is not Enough

Scaling Law#

Scaling Laws for LLMs describe the power-law relationships between model performance and key scaling factors;
We are approaching a critical inflection point where traditional scaling laws are showing signs of stagnation, primarily due to data limitations;
We'll exhaust high-quality human-generated text data between 2025-2030, with median estimates pointing to 2028 if current scaling trends continue. The total effective stock is approximately 4×10¹⁴ tokens
OpenAI's internal struggles with "Orion" (originally intended as GPT-5) exemplify this: the performance gains over GPT-4 were significantly smaller than the GPT-3→GPT-4 jump, with some tasks showing no reliable improvement.

Reference:#

Business Case: https://www.businessinsider.com/openai-orion-model-scaling-law-silicon-valley-chatgpt-2024-11
Research Reference: https://arxiv.org/pdf/2211.04325

Impact:#

Public vs. Private Data Asymmetry
Public data: ~300T tokens of general knowledge
Private business data: Exponentially larger but inaccessible by the model trainers;
Current scaling laws require ~20 tokens per model parameter for optimal training
Required Model Size for Business Reasoning: ~10^15 parameters
Required Training Data: 20 × 10^15 = 2×10^16 tokens
Available High-Quality Data: ~3×10^14 tokens
Data Gap: 67x more data needed than exists

Why can fine-tuning/distillation not solve the problem?#

Information-Theoretic Quality Requirements: Fine-tuning requires exponentially higher data quality than pretraining due to the signal extraction problem
Margin-Based Learning Mathematics: Fine-tuning operates in the low-margin regime where small errors have large impacts
High plasticity: Model can learn new patterns but forgets old ones;
High stability: Model retains old knowledge but can't adapt to new patterns;
In summary, the general capabilities of the pre-trained LLMs is necessary to drive the outcomes in the right direction, while grounding it to the business requirements and rules is also a challenge, solve one does not directly solve the other;
Additionally, cost implications, the amount of technical competency required, etc. are additional factors to consider other than the technical implication;

Why does the above problem exist mathematically and in simple language?#

The "Memory Interference" Problem
Think of Your Brain Learning a New Language
Imagine your brain has 100 "memory slots" and you've used 90 of them to learn English. Now you want to learn Chinese:
1.
Total brain capacity: 100 slots
2.
Used for English: 90 slots
3.
Available for Chinese: 10 slots
Problem: Chinese needs 50 slots to be useful
Available capacity: Only 10 slots
Result: Either bad Chinese OR forget English
In neural networks, this is exactly what happens:
1.
Model has billions of parameters (like memory slots)
2.
Most are used for general knowledge (like English)
3.
Business knowledge needs many parameters (like Chinese)
4.
Mathematical constraint: You can't exceed total capacity
The "Tug of War" Mathematics
When fine-tuning tries to learn business knowledge:
1.
simpleGeneral knowledge wants: Parameter = Value A
2.
Business knowledge wants: Parameter = Value B
During training:
1.
Step 1: Move toward A (general knowledge improves, business degrades)
2.
Step 2: Move toward B (business improves, general knowledge degrades)
3.
Result: Model "bounces" between A and B, never settling
Mathematical reality: A ≠ B, so no single value satisfies both
Modified at 2025-08-09 11:58:10
Previous
Our Technical Note
Next
How business Outcomes would Change Radically with AgenticUniverse
Built with