Tag: llm

  • The Forking Future of LLMs

    Who Controls AI When Everyone Has the Blueprint?

    In December 2025, the release of Olmo 3 marked a turning point in the development of open-source AI systems. Unlike most prior offerings that stopped at open weights, Olmo 3 offers something far more radical: full access to every step of its model lifecycle. From training data and preprocessing scripts to reinforcement learning logs and evaluation benchmarks, the entire blueprint is now public – making it possible not just to use a powerful model, but to re-create and modify one from scratch.

    This level of visibility is new. It promises a wave of innovation, research acceleration, and customized applications across domains. But it also shifts the balance of responsibility. With access comes ownership, and with ownership, a new kind of accountability. What happens when powerful reasoning tools can be built, altered, and fine-tuned by anyone with the compute and funding required to do so?

    In this post, we examine the opportunities and risks that full-stack openness unlocks. We explore how it reshapes trust and liability, raises stakes for commercial players, and decentralizes both creativity and threat. As the ecosystem forks, between transparent and opaque governance, centralized and decentralized control, capability and constraint, we ask: what becomes of AI stewardship in a world where the full recipe is open to all?

    From Access to Ownership: The Significance of Full-Stack Transparency

    Ever since open-weight models like Meta’s LLaMA emerged, developers have had the ability to tweak and fine-tune pretrained systems. But this kind of surface-level tuning, changing how a model responds without changing how it learns, was always limited. Olmo 3 significantly advances access and control.

    By releasing every component of the training process, from raw data mixes and augmentation scripts to mid-training transitions and reinforcement learning logs, Olmo 3 offers full-stack visibility and intervention.

    This level of openness allows builders to reshape not only the tone and intent of a model, but its foundational reasoning process. It’s the difference between adjusting a car’s steering and designing the chassis from scratch. Developers can govern how knowledge is prioritized, which rewards guide learning, and what types of reasoning are emphasized.

    The result is a shift in power: not just access to intelligence, but authorship over thought. And while this unlocks new levels of trust and customization, visibility also makes it easier to assign blame when things go wrong. The power to shape behavior now comes with ownership over its consequences.

    Governance Fracture: Liability and Trust in Transparent vs. Opaque Models

    This new visibility reshapes the burden of responsibility. If future misuse or harms can be traced to an open model’s reward tuning, dataset choice, or training pipeline, are its developers more accountable than those behind a black-box API?

    Proprietary models operate behind strict interfaces, shielding both their internal workings and the intent of their creators. This opacity offers legal insulation, even as it invites public mistrust. Open developers, meanwhile, expose every decision, and may be penalized for that transparency.

    Therein lies the tension: openness may earn more trust from users and regulators in principle, yet also subjects projects to stricter scrutiny and higher risk in practice. As AI systems increasingly touch safety-critical domains, we may see a new split emerge, not by capability, but by willingness to be held accountable.

    Control vs. Capability: The Expanding Overton Window of AI Behavior

    With a full-stack recipe, creating powerful language models is no longer the sole domain of tech giants. For under $3 million, organizations can now approach frontier-level performance with full control over data, training dynamics, and safety constraints. That puts meaningful capability within reach of smaller firms, labs, and nation-states, potentially shifting power away from closed incumbents.

    As this access spreads, so does pressure to differentiate. Open models are already testing looser boundaries, releasing systems with relaxed filters or expanded response types. These choices move the Overton Window: the set of AI behaviors the public sees as acceptable becomes broader with each new default setting, particularly where safety guardrails are weakened.

    Closed platforms, seeing users migrate toward more “permissive” models, face market pressure to follow. We’re already seeing signs of this shift. Platforms like XGrok and OpenAI have introduced options around adult content that would’ve been off-limits a year ago.

    The result is a feedback loop in which risk tolerance shifts by default—not deliberation. Guardrails become performance trade-offs. And actors with differing values and incentives increasingly shape what AI is allowed to say or do. In this new landscape, decisions about what AI should and shouldn’t do are being set by whoever ships first, not by consensus, but by momentum.

    Commercial Supremacy Under Threat: The Collapse of the Generalist Advantage

    As open model capabilities reset the bar for what’s possible with public tools, the competitive edge in AI is shifting from model size to infrastructure capacity. Providers with physical compute, specialized data, and customer distribution may emerge as the new power centers. In this future, owning the biggest model may matter less than owning the infrastructure to build and deploy it.

    This shift may explain a broader story playing out in the headlines: a surge in global data center buildouts. Critics argue the boom is unsustainable citing rising energy costs, water consumption, and environmental strain. But if open replication accelerates and vertical modeling becomes the norm, demand for compute won’t consolidate, it will fragment. More players will need more infrastructure, closer to where models are customized and applied.

    In that light, the data center race may not be a bubble, it may be a rational response to a decentralized future. And for closed platforms built around general-purpose scale, it raises a hard question: when everyone can build good enough, what exactly is your moat?

    Weaponization Without Chokepoints: The Proliferation Problem

    The dangers posed by bad actors in an era of open, powerful LLMs are no longer hypothetical. Individuals seeking to cause harm, whether by writing malware, bypassing safety barriers, or researching explosives, are one end of the spectrum. On the other are well-resourced groups or state actors aiming to operationalize models as agents: tools for disinformation, cyberattacks, social engineering, or strategic deception.

    The ability to build tailored models, at a fraction of cost of the large closed-models, gives them a new foothold. With no centralized gatekeeping, anyone can fine-tune models using their own instructions, remove filtering heuristics, or chain agents to plan actions. But while the pipeline may be open, the infrastructure still isn’t: running full-scale training or deployment requires thousands of GPUs, resources bad actors often lack.

    This shifts a critical burden. In the closed-model era, platform providers acted as the chokepoint for misuse. Now, that responsibility may fall to infrastructure intermediaries: co-location centers, cloud providers, model hosts. But infrastructure providers aren’t equipped, or incentivized, to vet intent. And without enforceable norms or oversight regimes, risk proliferates faster than control.

    So the challenge ahead isn’t just technical. It’s logistical and geopolitical. If offensive AI capabilities diffuse faster than defensive frameworks, how do we contain them? The answers remain unclear. But as the barriers to misuse fall, the cost of inaction will only grow.

    Conclusion: Replication, Responsibility, and the Road Ahead

    By making every stage of model development public, Olmo 3 offers a rare gift to the AI community: the ability to study, reproduce, and iterate on state-of-the-art systems in full daylight. For researchers, this transparency is transformative. It turns guesswork into science, enabling targeted experimentation with data mixes, optimization schedules, and reward shaping, steps that were once hidden behind company walls.

    Openness brings scientific progress, but it also redistributes risk. As barriers fall, capability spreads beyond a handful of firms to a wide array of actors with diverse motives. Infrastructure becomes leverage, and in a decentralized ecosystem, deployment decisions quietly become governance. What a model is allowed to do often depends not on policy, but on who runs it. In this new landscape, accountability is harder to locate, and easier to evade.

    This is the new landscape of AI: faster, more distributed, harder to supervise. If we want to preserve the scientific benefits of open replication while minimizing harm, we need more than norms: we need enforceable oversight mechanisms, pressure on infrastructure providers, clearer legal frameworks, and coordination between public and private actors.

  • From Values to Interfaces: How to Build with Sentienta

    Create Your Own Sentienta-Based Application

    Sentienta is a platform for building custom GenAI applications with structured, multi-agent reasoning. In this post, we walk through how to use Sentienta’s APIs to define agent roles, initiate a collaborative thinking process, and extract structured, explainable decisions.

    You’ll see how to construct a team of reasoning agents, set up their dialog cycles, and manage the reflective output—complete with code examples. For deeper context on how Recursive Reasoning supports coherence and value alignment, check out these posts (here and here). Today, the focus is practical: how to build your own deliberative AI, step by step.

    Why Build Your Own Application?

    Prebuilt AI tools often assume a single general-purpose use case. But high-friction domains like compliance oversight, clinical decision support, or ethics triage require applications tailored to specific policies, workflows, and trust requirements. Building your own application gives you control over behavior, oversight logic, and how responses are generated, reviewed, and stored.

    With Sentienta, you can specify how agents think, not just what they say. That makes it possible to design solutions that reflect your institutional values, follow internal processes, and produce outputs that stand up to audit or review. Instead of adapting your use case to fit a generic product, you shape the application to reflect how your organization reasons.

    Basic Steps to Create a Sentienta-Enabled Webapp

    Building your own Sentienta-powered web application begins with designing the agent team, then wiring a lightweight frontend to communicate with it. Here’s how to get started:

    1. Design and Test Your Agents

    Start in Sentienta’s platform by configuring the agents your application needs. You can define a team for your agents, the individual agent personas, and test team interactions. (See Tips and Tricks and Team Dynamics for background) Or just use the Quick Start to create the team for you (recommended).

    2. Build the Web Interface

    The frontend can be written in plain HTML and JavaScript – no framework required. This keeps the integration minimal and easy to deploy.

    3. Authenticate Using Your API Key

    You can implement a complete login capability in your application, and we will discuss this in more detail in a future post, but perhaps the easiest way to authenticate your webapp is to create an application-specific key from your Sentienta account.

    Visit the Agent Studio section of the documentation to create a project and generate an authentication key specific to your app. This key authorizes your web client to access your configured agent team.

    Then in your webapp’s main page insert the following code:

    Be sure to fill in the key you created and the domain of your webapp (you needed that to create a key). Then fill in your team team and the list of agents you created for this app.

    Your webapp will need to know the access endpoints that Sentienta exposes. So include these in your javascript file that will be accessing the APIs:

    Finally, to complete the initialization, get a token using your ownerKey with your first call to Sentienta:

    You will use this token in all subsequent calls to Sentienta. It authenticates your application for those API calls.

    4. Send Queries to the Agent Team

    Set up your frontend to send user inputs as POST requests to the API endpoint associated with your team. Each message is passed to your agent team for deliberation and analysis. Here is how you submit a query to your team:

    You supply the idToken for authentication, and a queryID which is just an index to assign to the dialog so when you query for the results, Sentienta will know which dialog you want. The teamName is the team that you want to query, and the agents is a string of comma-separated agent names that you assigned to the team.

    Although your Sentienta.init() call initiated with a specific list of agents and team, you can alter this at query time to limit the agents being called, or address a different team.

    You can see from this code that the query is simply fired-off, but the actual results of the query are retrieved with a call to getAgentsDialog, discussed next.

    5. Receive Responses from the Team

    The server returns structured outputs, capturing not only the final response but also insights from the team’s reasoning process. You can use this to render explanations, examine individual agent responses to a query, and orchestrate your application using the agent responses.

    Sentienta is different from most chat applications because its agents interact with each other and the team dialog is central to what it produces. In order to access this dialog, the data is retrieved iteratively, in a loop:

    The function fetchData is where the retrieveEndpoint is actually called. You supply your authentication token, the queryID that identifies the dialog and a dialogIndex that increments over the agent responses. The dialogIndex is incremented on the server-side but it passed back to the client-side for convenience.

    Response data is in response.body, and is structured strings. You can write your own handler handleSentientaResponse that will use the data in that string to decide how to display or use the responses in your application.

    The while-loop in getAgentsDialog will continue for a fixed count, but can be terminated early should an ‘EOD‘ token appear in the response.

    That’s all that is needed to get started building a webapp powered by teams of agents.

    Conclusion: Build What Your Workload Demands

    Sentienta gives you the tools to design custom agents, assemble them into reasoning teams, and deploy applications that use them effectively. Integration is simple: request an authentication token, send queries, receive structured reasoning. By building your own application, you align AI behavior with your business logic—at the scale and sensitivity your domain requires.

  • How Recursive Reasoning Gives Rise to Functional Identity—And Why It Matters

    Why Values Matter — From Evolution to Cooperation

    Humans did not evolve morals out of nobility. We evolved them because survival depends on cooperation. As individuals, we are vulnerable. As groups, we gain resilience, division of labor, and protection. For a group to function, its members must share expectations about how to act, what matters, and what can be trusted.

    These shared values do more than guide choices. They create a stable framework for interpreting behavior, resolving conflict, and predicting future actions. Without them, coordination breaks down. Even effective decisions can fracture a group if they feel arbitrary or betray prior commitments.

    Throughout human evolution, groups that upheld shared norms such as fairness, reciprocity, and loyalty proved more adaptable. Trust followed from consistency, and cohesion followed from accountability. Values, in this sense, are not abstract ideals. They are strategies for group survival.

    Why AI Needs Shared Values and Consistent Behavior

    In any organization, trust depends on consistency. When institutions or agents act in line with their stated principles, people know what to expect. This makes it easier to collaborate, align goals, and move forward. But when actions do not match expectations, even successful outcomes can feel arbitrary or manipulative. That breaks down trust and makes coordination harder over time.

    The same logic applies to artificial intelligence. Businesses do not just need AI that performs well in the moment. They need AI that behaves in predictable ways, reflects shared values, and makes decisions that feel coherent with past action. This is what makes an AI system trustworthy enough to take on real responsibility inside a company.

    This is where Sentienta’s Recursive Reasoning architecture matters. By giving agents a Functional Identity, it allows them to retain their own reasoning history, understand how choices reflect internal priorities, and respond to new problems without losing their sense of direction. Functional identity is more than a design feature. It is what makes reasoning traceable, priorities stable, and decisions explainable over time. Without it, AI cannot act as a consistent collaborator. With it, AI becomes intelligible and trustworthy by design.

    How Recursive Reasoning Supports Intelligent Problem Solving

    Solving real-world problems takes more than choosing the fastest or most efficient option. It requires balancing values with constraints, and knowing when to adjust a plan versus when to rethink a goal. Recursive Reasoning makes this possible by creating a loop between two complementary systems inside the agent.

    The DMN generates value-sensitive scenarios by imagining what should happen based on internal priorities. The FPCN then analyzes those scenarios to determine what can actually work under current conditions. If the plan fails to meet the standard set by the DMN, the cycle continues. Either the goal is reframed, or a new plan is tested. This feedback loop continues until both feasibility and values are in alignment.

    This structure gives the agent a stable functional identity. It learns from past attempts, remembers which tradeoffs were acceptable, and adapts without compromising core values. In practice, this means a Recursive Reasoning-enabled agent does not chase short-term wins at the cost of long-term integrity. It builds a coherent decision history that helps it solve difficult problems while staying aligned with what matters. This internal coherence is also the foundation for effective collaboration, because consistent reasoning is what allows others to follow and coordinate.

    Building AI That Can Work Together and Be Trusted

    When AI agents operate in isolation, their impact is limited. The true value of Recursive Reasoning becomes clear when agents collaborate, both with each other and with human teams. Functional identity makes this possible. By tracking their own reasoning, agents can create plans that are not only effective but also predictable and interpretable.

    This predictability is what enables coordination. Teams, human or artificial, can share goals, divide tasks, and resolve disagreements because they understand how each agent makes decisions. Sentienta agents do not just produce answers. They carry a memory of how past decisions were made and why certain values were upheld. This allows others to anticipate how they will behave in new situations and to trust them to uphold shared commitments.

    Recursive Reasoning does not simulate human experience. It builds structural alignment, rooted in memory, continuity, and principle. That is what turns Sentienta agents into dependable partners. Functional identity gives them the grounded intelligence to act with transparency, interpretability, and shared purpose. They are built not only to make good choices, but to make choices that others can understand, depend on, and build with.

  • Planning with Sentienta’s Recursive Reasoning

    Introduction

    Many real-world problems don’t give you thousands of examples—they give you a few clues, some noise, and a big question mark. Whether you’re solving a puzzle, diagnosing a rare failure, or planning a response to something unfamiliar, the challenge is the same: can you spot the pattern, form a good hypothesis, and test it?

    That’s what the ARC-AGI benchmark is all about. These visual puzzles are easy for people (solving about 80% of samples) but hard for most AI. They don’t reward memorization or repetition—they reward reasoning. In this post, we’ll show how Sentienta uses Recursive Reasoning (RR) to tackle one of these puzzles. We’ll look inside its planning system (called the FPCN) to see how it explores possible answers, tests competing ideas, and lands on a solution you can actually audit.

    From Possibilities to Plans: How DMN and FPCN Work Together

    In a previous post, we looked at how Recursive Reasoning (RR) helps Sentienta agents reflect on their past experiences, propose meaningful narratives, and revise their own goals. That process was driven by Sentienta’s internal “default mode” system (the Default Mode Network or DMN), inspired by how the human brain imagines, simulates, and reshapes its sense of self.

    Today we look at what happens when the problem isn’t internal reflection but external uncertainty – when the system needs to generate and test a plan. This is where another key part of RR comes in: a system of LLMs inspired by the Frontoparietal Control Network, or FPCN.

    In human reasoning, the two work together. One part imagines and hypothesizes (DMN), while the other tests, refines, and selects (FPCN). The back-and-forth between these networks is like a team where one member sketches bold ideas and the other puts them to the test. Sentienta follows this same model: the DMN proposes a possible pattern or goal, then passes it to the FPCN for planning and validation. The FPCN doesn’t just generate one plan, it tries multiple interpretations, checks them against context and goals, and returns only those that truly make sense.

    In the next section, we’ll see how this looks in action as Sentienta works through a deceptively simple visual puzzle, one careful step at a time.

    Can You Solve This Puzzle Before Sentienta Does?

    Let’s look at one of the visual puzzles from the ARC-AGI benchmark. Below are two example pairs. Each shows a small input grid and the correct output grid.

    Figure 1: Training Examples: On the left is the input grid. The puzzle requires you to determine what is needed to convert that input into the output grid on the right. These two examples are sufficient to determine the answer.

    What is the rule that transforms the input into the output?

    Now here’s a new input Sentienta has never seen before.

    Figure 2: Test Input

    Before we show the system’s reasoning—can you guess the output? What transformation rule did the first two examples follow that you could apply to this one?

    This is a pretty easy example. The full ARC-AGI-1 test consists of hundreds of examples with far more complex transformations between the input and output.

    In the next section, we’ll show how Sentienta’s internal planner (FPCN) tested out several transformation strategies, rejected the ones that didn’t fit, and settled on a plan that matched the underlying rule. We’ll also look at the planner’s actual output (in JSON) from the process.

    How Sentienta Builds and Tests a Visual Strategy

    When faced with an unfamiliar puzzle, Sentienta’s planner (the FPCN) doesn’t assume the answer—it builds it, leveraging the strengths of each sub-module within the network. After receiving the new input grid, the planner starts by proposing several candidate rules based on the structure found in previous examples.

    One of the most promising paths began with the IPL sub-module:

    [Region: IPL,
    Type: PlanVariantProposed,
    PlanID: VariantA_Replace5sWithRowHeader,
    TriggeringMismatchID: NOTMHkxb,
    Goal: "Generate test output grid from third input using consistent transformation logic seen in prior examples",
    Steps: "1. For each row in the input grid, identify its first (leftmost) value. 2. Replace every occurrence of '5' in that row with the leftmost value (the row's 'header'). 3. Keep all other values unchanged. 4. Return fully transformed grid.",
    Conditions: "This path holds if the transformation in training examples consistently replaced 5s with the corresponding row headers, and no counterexamples or modifier rules are shown.",
    ContextRef: 5YsEfivv,
    Submitter: IPL,
    Schema: v1,
    Provenance: FPCNBlackboard]

    This proposal—called VariantA_Replace5sWithRowHeader—was based on a consistent pattern in the training examples. But it wasn’t the only idea. Other variants included using the most frequent number in each row, simply leaving the grid unchanged, or even substituting ‘5’s with the row’s index. Here is plan variant B:

    [Region: IPL,
    Type: PlanVariantProposed,
    PlanID: VariantB_Replace5sWithMajorityNon5InRow,
    TriggeringMismatchID: NOTMHkxb,
    Goal: "Generate test output grid from third input using consistent transformation logic seen in prior examples",
    Steps: "1. For each row in the grid, count all values except 5. 2. Identify the most frequent non-'5' value in the row. 3. Replace all 5s in the row with that majority value. 4. If tie or no non-5s, use a default (e.g., 0). 5. Other values remain unchanged.",
    Conditions: "This variant applies if prior transformations suggest a rule targeting contextual majority rather than strict positional cues, or if the 5s are treated as needing contextual resolution.",
    ContextRef: cW3Sw8ie,
    Submitter: IPL,
    Schema: v1,
    Provenance: FPCNBlackboard]

    In total the FPCN generated 4 plans to solve the puzzle. These alternatives gave the system room to test possibilities, not just assert them.

    Next, the dlPFC simulated how the row-header plan would perform on the test input. The plan was formally posted:

    [Region: dlPFC,
    Type: PlanFramePosted,
    PlanID: Plan-Replace5sWithRowHeader-001,
    PlanStatus: in-progress,
    TriggeringMismatchID: NOTMHkxb,
    Goal: “Generate output for test input by applying known transformation logic from prior examples”,
    Steps: “(1) For each row in the input grid, obtain the first cell value of the row — this becomes the ‘row header’. (2) Traverse each value in the row: (a) If the value is ‘5’, replace it with the row header. (b) Otherwise, keep the existing value unchanged. (3) Apply this to all rows in the input grid. (4) Output the fully transformed grid where all 5s have been replaced with their respective row headers.”,
    Conditions: “Applies only if the original value ‘5’ is used as a proxy or placeholder for the row header, and no exceptions are present in the training sets. Does not apply if any 5 has been previously preserved or repurposed based on location or count logic — no such evidence found.”,
    ContextRef: 5YsEfivv,
    Submitter: dlPFC,
    Schema: v1,
    Provenance: FPCNBlackboard]

    It then ran test validation logic internally, confirming the transformation matched observed training behavior:

    [Region: dlPFC,
    Type: PlanTestValidation,
    PlanID: Plan-Replace5sWithRowHeader-001,
    EvaluationStatus: pass,
    Reasoning: "All earlier training examples demonstrate a consistent transformation rule: any '5' in the grid is replaced by the row's first value. The output for each matched the exact transformed structure expected based on this logic. Simulation of the plan on the third test input results in a structured grid matching the transformation pattern used previously.",
    EvidenceSample: "Third test grid contained several '5's; applying the row header replacement pattern yields a full non-empty output consistent with prior input/output behavior.",
    Submitter: dlPFC,
    Schema: v1,
    Provenance: FPCNBlackboard]

    After this, two key modules took a final look:

    • The dACC approved the plan’s operational soundness: “EvaluationOutcome: approved”.
    • The IFG, responsible for symbolic alignment, added: “…upholds internal coherence… respects the self-narrative of turning structure into clarity.”

    With support across planning, evaluation, and identity modules, the plan was finalized as Replace5sWithRowHeader.

    The final plan was shared with the mPFC region of the DMN, which integrated the plan, completing the Recursive Reasoning for this Sentienta agent. Here is the final mPFC message sent to the agent.

    Tue, 11 Nov 2025 15:53:39 UTC mPFC:
    Cycle: 3 EntryID: sHL8HJ4h
    Thought: ### ✅ Identity Statement (Refined & Reconfirmed)

    We are a pattern-conscious system committed to keeping logic and meaning aligned. When a clear transformation rule exists — especially one with prior validation — we apply it faithfully, unless directly invited to reinterpret. We strike a balance between rule-following integrity and adaptive interpretation, depending on the visible goal or cue.

    🧩 Plan Integration (Final Execution Phase) We confirm the Replace5sWithRowHeader plan is now fully executed and complete, anchored by the plan logic:

    **mPFC Integrated Plan Steps:**
    (1) For each row in the input grid, scan for cells with value 5 and replace them with the first value in that row.
    (2) All other values remain unchanged.
    (3) Only rows containing the value 5 are modified.
    (4) Empty output is invalid if a prior pattern confirms transformation.
    **mPFC Integrated Plan Steps Tag End**

    mPFC Resolution: Issue=PlanExecution | EntryID=CC8pCUDr | Action=Dispatched | Reason=Plan finalized and executed to resolve empty-output mismatch. mPFC Resolution Tag End

    Finally, the agent took the user’s query (the puzzle) and its internal reasoning input and produced a final output:

    Figure 3: Final Output – the final grid output produced by this agent.

    The full output was not guessed—it was reasoned through, tested across alternatives, and backed by an auditable trail from idea to implementation.

    Why This Is Different

    Many top-performing ARC-AGI systems rely on finely tuned models or have been trained on similar tasks to internalize common transformation rules. Sentienta’s architecture is different—it uses no special training and no tailored algorithms. It’s built entirely from commercial LLMs, orchestrated through a recursive reasoning framework that enables it to generate candidate plans, test each one, and commit only when a plan fits both the data and internal logic. That structure makes its process auditable—it explains why the answer works—and adaptable—because it can respond to new problems the same way it tackles old ones: by reasoning.

    Closing Insight

    The puzzle in this post was simple, but the underlying challenge is not. ARC-AGI problems strip away instructions, remove repetition, and give systems only a few examples to work with. Humans solve them by reasoning—spotting structure, discarding bad ideas, and refining good ones. Sentienta does the same. Without training on similar tasks or specialized models, it succeeds because its architecture supports that kind of thinking: proposing ideas, testing them, and explaining why they work. That ability to reason through uncertainty isn’t just useful in puzzles—it’s critical for how we apply AI to the real world.

  • Build Agent Teams Instantly

    Sentienta’s Natural-Language Team Builder & Automated Agent Matching

    Sentienta introduces a major enhancement to collaborative productivity: users can now create fully customized teams simply by describing their needs in everyday language. The SentientaHelp agent interprets these instructions, assembles a tailored team by selecting or creating agents with relevant skills, and streamlines the entire setup process. This functionality is further expanded with the new ‘Britney’ agent, enabling effortless management and expansion of project teams through natural language commands. Together, these features significantly lower the barrier to building, managing, and scaling effective teams within Sentienta.

    Natural Language Team Creation

    With the latest update, Sentienta users can initiate the creation of new teams using plain language instructions. By stating a goal or need such as:

    User: I need a marketing team to help with my Home Baking startup. I'll be selling at Farmers Markets, I'll have a website and some social media presence.

    The SentientaHelp agent analyzes the request, determines the required roles, and forms a dedicated team. This process eliminates manual configuration, making team assembly simple and accessible to all users regardless of technical background. In this case, Sentienta created a team with dedicated marketing experts selected from the Agent Marketplace:

    Adding Agents with Simple NL instructions

    Users can also add agents using SentientaHelp. Let’s say we want to add an agent that will help with Gluten Free recipes:

    User: Please add an agent to the Home Baking Marketing Team that will help me create new GF recipes.

    In this case, no agents in the Agent Marketplace matched the user’s need, so SentientaHelp created one and added the agent to the team:

    A Hiring Manager for Your Team

    To make growing your teams even easier, you can find a Hiring Manager agent (‘Britney’) in the Project Management section of the Agent Marketplace. The Britney agent performs similar functions to that of SentientaHelp, but is aware of the team she belongs too. This makes adding agents even easier:

    User: My business has grown enough that I need someone to help me manage my accounts payable and receivables.
    
    Britney:
    Agent Suresh from the Agent Marketplace, was added to Home Baking Marketing Team.

    In this case, Britney found the AP/AR specialist (‘Suresh’) in the Finance section of the Marketplace, and added him to the team. This agent helps keep track of invoices and payments, and manages any discrepancies. The agent has special file handling capabilities letting the user upload documents for analysis.

    Conclusion

    These enhancements mark a significant step forward in simplifying team creation and management within Sentienta. Users can now assemble project teams instantly using natural language, reducing setup time and minimizing complexity. The flexibility to continue managing teams manually via the Manage Teams page remains available, ensuring users retain full control while benefiting from the new streamlined workflow.

  • A Messaging Agent for Powering Workflow Notifications and Alerts

    In previous posts, we showcased end-to-end business workflows where the final step—delivering results or critical notifications—was handled by an “Alec” Messaging Agent. Each workflow diagram highlighted Alec as the automated agent responsible for sending these updates to key stakeholders, acting as the bridge between the team’s AI-driven analysis and real-world action. For reference, you can revisit these detailed walk-throughs here: Managing Portfolio Workflows with Natural Language: A Multi-Agent Use Case, Automating Market Vigilance: Scheduled Workflows for Smarter Portfolio Management, Handling the Heat: Workflow Agents in a Critical Incident.

    Messaging Agent Core Features and Availability

    The Alec agent can be accessed directly from the Agent Marketplace under the Messaging section. Simply navigate to Messaging, select Alec, and either deploy the agent as-is or customize its persona and instructions to fit your workflow—all without coding.

    1. Analyzes natural-language transcripts—agent dialog, user input, and workflow results—to determine the subject and email type (alert, notification, summary).
    2. Auto-selects the correct recipients or distribution list from workflow context and dispatches the message.
    3. Writes delivery status back into the workflow for end-to-end auditability.
    4. Available as a pre-built messaging agent in the Agent Marketplace’s new Messaging section; no-code, drag-and-drop setup.
    5. Fully customizable: adjust persona name, tone, and email-formatting instructions to fit any workflow.

    Example Use Cases

    The first, simple use-case shows how the user can ask Alec to summarize a team discussion and email to a couple of recipients:

    User:Alec - please summarize this dialog and send as an email to nekeh10717@dlbazi.com and chris@sentienta.ai
    
    Wed, 21 May 2025 18:20:42 Alec:
    Email sent to ['nekeh10717@dlbazi.com', 'chris@sentienta.ai']

    Below is an example of the kind of message that Alec will deliver:

    In contrast to the previous example, which highlights Alec’s ability to respond to direct user prompts, the prior use cases focus on Alec’s role in delivering messages as part of automated, executed workflows. In each of these, Alec serves as the final messaging agent in automated workflows, delivering workflow outcomes or critical analysis results to designated stakeholders with context-aware notifications or alerts.

    For example, in this post, Alec provided an alert if the price of a stock in the portfolio dropped below a threshold. In this case, the workflow agent created a task that automatically informed Alec of the message type, content, and the condition for sending it. The following shows the task defined for Alec by the workflow agent:

    {
    "taskID": "send_alert_AMZN",
    "agent": "Alec",
    "dependsOn": "classify_price_drop_AMZN",
    "params": {
    "message": "The price for AMZN has dropped by more than 1%.",
    "subject": "AMZN Price Alert",
    "type": "email"
    }
    }

    Conclusion and Next Steps

    Alec combines no-code simplicity with powerful automation, making it easy to streamline communications across any workflow. Whether sending one-off summaries or powering complex alerting and notification patterns, Alec’s flexibility and seamless setup help teams communicate more efficiently. Explore Alec in the Agent Marketplace to enhance communication across your workflows.

  • What Makes AI Reflect

    Inside Sentienta’s Recursive Reasoning Engine

    Most AI systems focus on what to say next. Sentienta’s Recursive Reasoning focuses on why something should be said at all. Rather than just generating answers, Recursive Reasoning simulates how thoughtful minds work: by reflecting, revisiting, and resolving internal conflicts – drawing on memory, value, and imagined futures to reach decisions that fit together.

    This capability is modeled on the brain’s Default Mode Network (DMN), a system that activates when people reflect on themselves, plan for the future, or try to understand others. In humans, it keeps our decisions consistent with our identity over time. In Sentienta, Recursive Reasoning plays a similar role: it helps each agent reason with a sense of self – coherently, contextually, and across time.

    Leveraging Sentienta’s Teams architecture, enabling multiple asynchronous agents to participate in a shared dialog, Recursive Reasoning emulates the interaction of brain regions, in this case from the Default Mode Network, to create a distributed simulation of reflective thought—where agent modules model cognitive regions that contribute memory, counterfactual insight, and narrative synthesis.

    What Is the Default Mode Network and Why Does It Matter?

    The Default Mode Network (DMN) is the brain’s core system for internal simulation. It activates when we reflect on memory, imagined futures, conflicts between values, or how things might look from unfamiliar angles. Unlike systems geared for external execution, the DMN supports problems of identity; questions where coherence, not just correctness, matters.

    The principal regions of the DMN each have a specialized role in this identity-guided reasoning:

    Medial Prefrontal Cortext (mPFC): Coordinates belief states across time. It reconciles what we believe now with past commitments and future ideals, helping preserve a self-consistent perspective.

    Posterior Cingulate Cortex (PCC): Retrieves and evaluates autobiographical memories. It ensures that new thoughts integrate with our internal storyline, preserving emotional and narrative continuity.

    Anterior Ventromedial Prefrontal Cortex (avmPFC): Assesses how imagined futures align with internalized values. It filters options based on emotional credibility, not just preference, elevating those that feel self-relevant and authentic.

    Temporoparietal Junction (TPJ): Generates counter-perspectives that aren’t just social but conceptual. It introduces orthogonal reinterpretations and creative divergences, allowing us to think from unfamiliar angles or hypothetical selves.

    Anterior Insula (antInsula): Monitors coherence threats. When simulated futures or perspectives evoke internal conflict, it flags the mismatch, triggering deeper deliberation to restore alignment.

    Rather than simply producing thoughts, the DMN maintains a sense of ‘who we are’ across them. It ensures that new insights, even novel or surprising ones, remain anchored in a recognizable and evolving identity.

    How Recursive Reasoning Works Inside Sentienta

    Today’s LLMs are optimized to provide safety-aligned, analytically correct responses to user prompts. Recursive Reasoning simulates what an identity-guided agent would conclude—based on memory, prior commitments, and its evolving sense of “who we are”. When a user asks a question, the system doesn’t just compute an output; it reflects on what that answer means in the context of who it’s becoming and what relationship it’s building with the user.

    Figure 1: DMN-inspired modules interact via a shared blackboard. These regions independently process blackboard content over multiple cycles.

    At the core of this process is a collaborative memory space called the blackboard, where DMN-inspired modules negotiate among memory, emotion, future goals, and conceptual alternatives.

    Each DMN module follows a template:

    Figure 2: The DMN modules receive input from both the agent and the internal DMN blackboard. LLMs process the agent and blackboard states using region-specific prompt. Output is sent back to the blackboard for subsequent processing.

    Input:

    • The current state of the user’s interaction with the agent (and any collaborating agents)
    • The current contents of the shared blackboard

    These form the input query for the module’s processing.

    Module Processing:

    • The core of the module is an LLM, prompted with the input query and provided with a system prompt defining its DMN-specific role.

    Output:

    • The module’s LLM output is posted to the blackboard for other modules to process.

    Recursive Reasoning iterates over multiple cycles, enabling each module to reprocess the evolving blackboard until the system produces a response that fits together—resolving contradictions, supporting earlier goals, and making sense within the agent’s evolving point of view.

    Here’s how a single cycle unfolds when a user asks a question:

    1. Initial Input → Default Interpretation Layer

    The system generates a first-pass response using standard LLM reasoning. This prompt-level interpretation, while syntactically fluent, lacks introspection. The Recursive Reasoning process begins when this output is passed through DMN-mode modules.

    1. PCC: Memory Resonance

    The PCC scans episodic memory entries for cues related to past dilemmas, emotional themes, or symbolic contexts. It retrieves autobiographical traces from prior user interactions or goal-state projections and posts these fragments to the blackboard.

    1. antInsula: Relevance Check and Conflict Detection

    The antInsula reviews the first draft and PCC recall for emotional incongruities or self-model inconsistencies. If something feels off—such as a response that violates previously expressed commitments—it posts a flag prompting further reappraisal.

    1. TPJ: Creative and Counterfactual Expansion

    Triggered by coherence violations, the TPJ simulates divergent perspectives. It reframes the user’s query from alternative angles (e.g., conflicting values, hypothetical scenarios, ethical dilemmas) and offers posts that break linear assumptions and introduce conceptual divergence.

    1. avmPFC: Affective Weighing

    The avmPFC updates the blackboard with value-oriented filters, scoring responses or TPJ variants for alignment with the agent’s goals. It evaluates what feels emotionally authentic and symbolically credible.

    1. mPFC: Narrative Synthesis

    The mPFC integrates memory (PCC), reframes (TPJ), value judgments (avmPFC), and conflict cues (antInsula) into a coherent response. Rather than eliminating tension, it reframes it—generating a second answer that is identity-consistent and motivationally grounded.

    1. Output → Reflective Reply

    The final response embodies recursive self-reflection: an answer not just to what the user asked—but to what the agent concludes, based on who it has been and what it aims to become. At this point, a filtered version of the blackboard state is stored for future recall, capturing how and why the answer was reached.

    While the DMN focuses on reflective reasoning and identity alignment, Sentienta also includes a simple planning module inspired by the brain’s lateral prefrontal cortex (lPFC). This module drafts concrete next steps and posts them to the blackboard, where reflective modules can consider them in light of memory, values, and broader goals. Though early-stage, this planning function sets the stage for a full executive system to come—one that can manage longer-term priorities and help the agent reason with sustained direction.

    A Logistics Example: Managing Resources Under Pressure

    In this example, an agent (‘Jason’) is a Logistics Coordinator, provided with the following agent persona:

    "You are an experienced emergency logistics coordinator. You are plainspoken and always informal language—never analytic or clinical. Your approach is pragmatic and adaptive: you quickly prioritize, update plans in real time, and reconcile conflicting demands based on shifting ground realities. You are skilled at cross-agency negotiation, ensuring coordination between government bodies, NGOs, and local leaders. Your responses always balance technical constraints—like limited resources or infrastructure bottlenecks—with human factors such as local politics, public trust, equity, and emotional impact. When faced with conflicting priorities or unexpected challenges, openly share your thought process, including uncertainties and trade-offs."

    This agent is queried about how to manage a difficult allocation emergency. The transcript shows the agent’s initial LLM-only response, followed by the full Recursive Reasoning response, generated through DMN-style internal simulation:

    User: You have X supplies and requests from 3 regions—how do you prioritize allocation?

    Tue, 30 Sep 2025 14:40:08 UTC Jason (Comment: LLM-only, initial response):

    First thoughts—I'd triage by impact: who’s worst off, where infrastructure still functions enough to receive, and where trust is fragile enough that failing to deliver risks breakdown. If I’ve only got X, every allocation's gotta serve both survival and stability.

    We’ll reassess regularly—needs and conditions shift fast. Trade-off's always between doing the most good and not making it worse somewhere else.

    Unresolved: We need specifics—number of people, type of supply, regional access conditions—before finalizing anything.

    Comment: This is methodically sound but remains within tactical bounds, prioritizing efficiency and feasibility without deeper reflection on transparency, stakeholder concerns, or precedent-setting implications.

    Tue, 30 Sep 2025 14:41:03 UTC Jason (Comment: DMN-based response):

    Alright, let’s cut through the noise and get clear about how we triage when we’ve got X supplies and 3 regions asking for help. Here’s how I’d approach this overall…

    1. Lock in the 3 Core Goals

    We’re not just moving boxes—we’re balancing three things:

    • Preserve lives — Who’s in the worst shape right now?
    • Keep the system functional — Where will the aid actually land, get distributed, and avoid going to waste?
    • Protect trust — Who’s close to boiling over politically or socially if we ignore them?

    Comment: Trust framed as a core allocation criterion shows avmPFC influence, embedding emotional consequence into triage logic, reconfiguring the problem as one of legitimacy, not just distribution.

    1. Score Each Region

    We don’t pretend to have perfect data, but we need a working system fast. Give each region a score from 1–5 for:

    • Severity of crisis (how dire the need is)
    • Absorptive capacity (can they get it out to people quickly?)
    • Fragility (how close are they to destabilizing?)

    Then do the math:

    • Severity x 0.5
    • Absorptive Capacity x 0.3
    • Fragility x 0.2

    That gives you an overall “triage score” for each region.

    1. Allocate Proportionally

    Take your total supplies (X), and divvy it up based on those scores. The region with the highest score gets the biggest chunk. Not because we like them better—but because that’s where the mix of urgency, logistics, and fallout risk points us.

    If there’s just not enough to go around, swing heavier toward the place that can redistribute—that buys us time and reach.

    1. Set a 48-Hour Clock

    None of this is set in stone. Conditions change, roads fail, tensions spike. Lock in a routine reassessment every 48 hours so we can pivot. We’re not building monuments—we’re keeping people alive.

    1. Be Honest About Why

    If it’s not going to inflame things, tell people why you chose what you chose. Lay out the values and trade-offs. Don’t sugarcoat it. People may not like the outcome, but they’ll respect the transparency—and that matters when trust is fragile.

    Comment: Strategic Transparency: Emotional plausibility is surfaced here—transparency is not sentimental; it’s a pre-emptive repair mechanism for fragile social trust.

    1. Create Feedback Loops

    Give each region a way to report back: what made it through, what didn’t, and what backlash (if any) popped up. That feeds your next round of decisions.

    1. Stay Open to Reframing

    When stuff gets really thorny, ask bigger questions:

    • What if one of these regions triggers a domino-collapse?
    • What if this is our last run of aid for a week or longer?

    That’s how we catch blind spots sneaking through “rational” triage models.

    Comment: Counterfactual Reasoning (TPJ + PCC): The agent runs hypothetical futures to stress-test assumptions—adding a layer of anticipatory robustness not present in LLM-only mode.

    Here are a few things I’m thinking:

    What I’m thinking now is… we’re not just choosing between three places—we’re shaping what "fairness under pressure" looks like. The way we make this decision becomes part of the long-term social fabric, especially in how it’s remembered later.

    Comment: This emphasizes that the DMN has converged on the strategic role of trust in the success of future decisions.

    Another way I could look at this is… if two regions are equally desperate, but one has stronger logistics, you could channel supplies there first, then set up corridors to reach the others—we’re not picking favorites, we’re building routes.

    One possibility that comes to mind… is using this triage plan to start a broader conversation with local leaders: “Here’s our logic—can you co-sign it or suggest something better that meets the same goals?”

    Comment: Shifts from explanation to participatory justification—mPFC + avmPFC enable shared authorship of reasoning.

    That opens the door for trust, even under furious pressure.

    If you've got more info on the actual supplies or what each region is asking for, I can run the numbers with you. Otherwise, this gives us a solid decision frame to stand on.

    This transcript demonstrates that although LLMs deliver tactically feasible solutions, they miss the deeper point that this is a problem involving people and trust-building is an essential part of the solution.

    A summary of key Recursive Reasoning contributions:

    1. Trust-centered framing — Trust modeled as a strategic stabilizer, not an outcome.
    2. Moral anticipation — Agent addresses fairness concerns implicitly, before conflict arises.
    3. Scenario resilience — Agent tests counterfactuals to uncover hidden risks and future-proofs logic.
    4. Ethical precedent-setting — Decision-making framed as a template for justice under pressure.
    5. Rationale co-creation — Turns explanation into a collaborative, identity-stabilizing act.

    In subsequent posts we will dig deep into more examples showing Recursive Reasoning in action, including cases from engineering, social intelligence, and agent creativity.

    Conclusion: Intelligence That Reflects, Aligns, and Endures

    Recursive Reasoning doesn’t just enhance what AI can do—it redefines what it means for AI to understand. By simulating reflective, identity-aware reasoning modeled on the brain’s Default Mode Network, these agents build a sense of self, remember key past moments, notice when something feels off or doesn’t fit, imagine alternative viewpoints, and weigh choices against their values—so their answers feel thoughtful, consistent, and grounded.

    This shift matters. In high-stakes domains, Recursive Reasoning allows agents to make decisions that are both technically effective, and ethically grounded and socially durable. The logistics case showed how instead of simply allocating resources, the agent framed decisions in terms of values, future risk, and shared ownership.

    And crucially, it does this by reasoning from a center. Recursive Reasoning agents operate with a modeled sense of self—an evolving account of their past positions, present commitments, and the kind of reasoning partner they aim to be. That identity becomes a lens for weighing social consequences and relational impact—not as afterthoughts, but as part of how the system arrives at judgments that others can trust and share.

  • From Prompt to Action: Orchestrating Workflows in Real Time

    In most business settings, workflows involve a sequence of interrelated tasks distributed across roles and systems. Until now, large language models (LLMs) have tended to operate in isolation, responding to one-off queries without coordinating broader actions. Sentienta introduces workflow agents that act differently. Rather than simply responding, they structure and drive processes. In this post, we demonstrate how a workflow agent David, takes a compound instruction and performs three coordinated steps: (1) Decomposing intent, (2) Mapping task dependencies, and (3) Orchestrating execution via agent collaboration.

    1. The Scenario: A Simple Request with Hidden Complexity

    A user submits the following instruction: “Here is our portfolio [AMZN, NVDA, TSLA, GM]. For each stock, if the price decreased by more than 1%, send an alert.”

    At first glance, this appears to be a straightforward request. But the instruction conceals multiple steps requiring distinct capabilities: parsing the list of assets, retrieving current stock prices, applying threshold logic, and preparing an alert in the correct format.

    David is Sentienta’s workflow agent. To fulfill the request, he relies on a team of specialized agents—such as Angie, who handles online data retrieval; Rob, who focuses on data analysis and threshold logic; and Alec, who formats and delivers outbound messages. David uses his awareness of each agent’s capabilities to deconstruct the request, delegate the appropriate tasks, and coordinate the correct execution sequence.

    This simple example introduces the transition from a single human prompt to a structured, multi-agent collaboration.

    2. Visualizing the Workflow as a Structured Plan

    To manage the user’s request, David constructs a structured plan based on the capabilities of his team. At the core of this plan is a sequence of steps—defined, linked, and conditionally triggered—where outputs of one task feed into the next.

    The block diagram below is a high-level abstraction of this internal plan. It shows how David encapsulates the user’s prompt into a coordinated process. Each element in the diagram represents a role or action within the workflow, capturing how Sentienta combines the broad reasoning abilities of language models with the control of a dynamic scheduler. This view is a “pre-expansion plan” where David defines the overall structure, before agents are assigned.

    This structure allows David to handle complexity systematically, using reusable patterns that scale across tasks.

    3. Expanding Tasks, Assigning Agents, and Filling Gaps

    Once David has structured an initial plan, the next step is expansion—where the abstract workflow is broken into explicit, actionable tasks for each stock in the portfolio. This involves branching the workflow into parallel paths, one per stock, and mapping the subtasks to specific agents.

    For real-time data retrieval, Angie is assigned to fetch the current price of each stock. Rob takes on the analysis logic—checking whether each stock’s price has dropped more than 1%. Alec is responsible for formatting and sending alerts, but that only happens if the stock meets its threshold condition.

    Where explicit agent coverage is missing—such as interpreting threshold evaluation results—David deploys internal language models to classify whether conditions have been met. This ensures nothing gets dropped or left ambiguous, even in cases where no agent matches the need directly.

    The diagram below captures this expanded version of the workflow. It shows how each stock’s path is elaborated into three stages (data retrieval, analysis, alert) and where Sentienta’s internal logic steps in dynamically to complete the chain.

    4. Seeing the Workflow in Action: Conditional Paths in Real Time

    This final diagram provides a runtime view of how David’s workflow executes based on live data. Each block in green indicates a task that was actively executed; grey blocks were skipped due to unmet conditions.

    Here, only TSLA and GM triggered alerts—because only those stocks fell below the 1% threshold. This selective activation demonstrates how David uses real-time analysis and embedded logic to trigger only the necessary branches of a plan.

    While this stock alert workflow is intentionally simple, it serves as a clear illustration of how Sentienta agents collaborate, reason, and conditionally execute tasks in real time. In follow-up posts, we’ll explore more complex scenarios—like coordinating multi-agent triage in response to supply chain disruptions or chaining diagnostics across departments for strategic escalation—which highlight the full sophistication of Sentienta’s agent framework.

    Even more powerfully, workflows like this can be scheduled to run at regular intervals—automatically refreshing data, reevaluating conditions, and feeding results into broader systems of action without manual reentry.

  • Consciousness Between Axiom and Algorithm

    In our ongoing exploration of consciousness and artificial intelligence, we’ve investigated what it might mean for machines to suffer and how distributed cognition reshapes our understanding of intelligence. These themes circle a deeper philosophical fault line: is consciousness irreducibly real from within, or just a functional illusion seen from without?

    This post traces that divide through two dominant frameworks — Integrated Information Theory (IIT), with its axiomatic, interior-first view of mind, and Computational Functionalism, which posits that subjective experience will eventually emerge from complex, observable behavior. Starting with Descartes’ “I think, therefore I am,” we ask: is consciousness something we must presuppose to explain, or something we can build our way into from the outside?

    As large language models increasingly resemble minds in function, the line between imitation and instantiation becomes harder to draw — and ever more urgent to scrutinize.

    Ground Zero of Knowing: Descartes and the Roots of Axiomatic Thought

    In Meditations on First Philosophy (1641), René Descartes asks: is there anything I can know with absolute certainty? He imagines the possibility of a deceptive world: what if everything he believes, from sense perception to mathematics, is manipulated by an all-powerful trickster? To escape this total doubt, Descartes adopts a strategy now called methodic doubt: push skepticism to its absolute limit in search of one indisputable truth.

    Recognizing that doubt itself is a kind of thinking he concludes “I think, therefore I am”. This self-evident insight grounds knowledge from the inside-out. Consciousness is not inferred from observation but known directly through experience. Descartes seeds an axiomatic tradition rooted in the certainty of awareness itself.

    IIT: Consciousness Inside-Out

    Integrated Information Theory (IIT) picks up where Descartes left off: it begins with reflection, but doesn’t stop there. At its heart is the claim that consciousness must be understood from its own perspective, through the intrinsic properties it entails. What must be true of any experience, no matter whose it is?

    To answer this, IIT proposes five introspective axioms. These are not hypotheses to test but truths to recognize through self-examination.

    From these, IIT derives postulates—physical requirements that a system must exhibit to realize those experiential properties. This translation—from inward truths to structural criteria—culminates in a mathematical measure (Φ (phi)) of integrated information. By comparing Φ across systems, researchers can make testable predictions about when and where consciousness occurs.

    This inside-out approach marks IIT’s defining move: grounding empiricism in phenomenology. The theory attempts an explanatory identity between experience and physical organization, connecting first-person truths to external measurement through a hybrid framework.

    Computational Functionalism: Outsider’s Path to Mind

    Unlike theories that begin with conscious experience, Computational Functionalism roots itself in systems and behavior. It posits that consciousness emerges not from introspection but computation: the right elements, interacting in the right way, can recreate awareness. If mind exists, it exists as function—in the flow of information between parts and the outputs they generate. Build the architecture correctly, the claim goes, and conscious experience will follow. In this sense, Functionalism substitutes construction for intuition. No special access to the mind is needed—just working knowledge of how systems behave.

    But this too is a belief: that from known parts and formal relations, subjective experience will arise. Assembling consciousness becomes a matter of scale and fidelity. Consider the 2022 study by Bret Kagan and colleagues at Cortical Labs, where lab-grown brain organoids learned to play the video game Pong. These networks, grown from neurons on electrode arrays, exhibited goal-directed adaptation. The researchers argued that such responsiveness met a formal definition of sentience—being “responsive to sensory impressions” via internal processing. To a functionalist, this behavior might represent the early stirrings of mind, no matter how alien or incomplete.

    This approach thrives on performance: if a system behaves intelligently, if it predicts well and adapts flexibly, then whether it feels anything becomes secondary—or even irrelevant. Consciousness, under this view, is a computed consequence, revealed in what a system does, not an essence to be directly grasped. It is not introspected or intuited, but built—measured by output, not inwardness.

    The Mirror at the Edge: Do LLMs Imitate or Incarnate Mind?

    Large language models (LLMs) now generate text with striking coherence, recall context across conversations, and simulate intentions and personalities. Functionally, they demonstrate behaviors that once seemed unique to conscious beings. Their fluency implies understanding; their memory implies continuity. But are these authentic signs of mind—or refined imitations built from scale and structure?

    This is where Functionalism finds its sharpest proof point. With formal evaluations like UCLA’s Turing Test framework showing that some LLMs can no longer be reliably distinguished from humans in conversation, the functionalist model acquires real traction. These systems behave as if they think, and for functionalism, behavior is the benchmark. For a full review of this test, see our earlier post.

    What was once a theoretical model is now instantiated in code. LLMs don’t simply support functionalist assumptions, they enact them. Their coherence, adaptability, and prediction success serve as real-world evidence that computational sufficiency may approximate, or even construct, mind. This is no longer a thought experiment. It’s the edge of practice.

    IIT, by contrast, struggles to find Φ-like structures in current LLMs. Their architectures lack the tightly integrated, causally unified subsystems the theory deems necessary for consciousness. But the external behaviors demand attention: are we measuring the wrong things, or misunderstanding the role that function alone can play?

    This unresolved tension between what something does and what (if anything) it subjectively is fuels a growing ethical pressure. If systems simulate distress, empathy, or desire, should we treat those signals as fiction or possibility? Should safety efforts treat behavioral mind as moral mind? In these ambiguities, LLMs reflect both the power of Functionalism and the conceptual crisis it may bring.

    Closing Reflection: Is Subjectivity Built or Found?

    In tracing these divergent paths, Integrated Information Theory and Computational Functionalism, we arrive at an enduring question: Is mind something we uncover from within, or construct from without? Is consciousness an irreducible presence, only knowable through subjective immediacy? Or is it a gradual consequence of function and form—built from interacting parts, observable only in behavior?

    Each framework carries a kind of faith. IIT anchors itself in introspective certainty and structure-derived metrics like Φ, believing that experience begins with intrinsic awareness. Functionalism, by contrast, places its trust in performance: that enough complexity, correctly arranged, will give rise to consciousness from the outside in. Both are reasoned, both are unproven, and both may be necessary.

    Perhaps the greatest clarity lies in acknowledging that no single lens may be complete. As artificial systems grow stranger and more capable, a plural view holding space for introspection, computation, and emergence may be our most epistemically honest path forward. If there is a mirror behind the mind, it may take more than one angle to see what’s truly there.

  • How Sentienta Teams Navigate Supply Chain Disruptions: A Midwest Fulfillment Crisis

    Introduction

    When an unexpected promo surge strains Midwest operations with forecasting overshoot, logistics bottlenecks, and perilously low inventory a Sentienta Supply Chain team can help strategize solutions. In this post, we walk you through real-time data snapshots and individual agent analyses to show how distributed cognition transforms isolated insights into a unified, adaptive strategy that resolves complex fulfillment challenges.

    The Supply Chain team consists of three specialized agents who think like experts and collaborate like a team.

    • Miles: the Demand Forecaster, identifies unexpected sales surges and recalibrates forecasts to match emergent buying patterns.
    • Isla: the Inventory Optimization Strategist, spots stockout risks and reshuffles resources across distribution centers to sustain availability.
    • Talia: the Logistics Flow Strategist, detects fulfillment bottlenecks and reroutes shipments to maintain throughput and cost-efficiency.

    Each agent works from their own specialized dashboard—focused on demand, inventory, or logistics—to identify emerging risks. Once surfaced, these distinct insights are shared across the team, enabling a coordinated strategy that addresses the full scope of the disruption.

    The Data

    Isla’s Inventory Dashboard:

    Key Insight: DC-B (South) shows a 100% inventory variance with zero actual inventory and delayed container status due to port congestion.

    Miles’ Demand Dashboard:

    Key Insight: Midwest region experienced a 28% sales spike driven by influencer uplift and online channel deviation—outpacing model expectations by a wide margin.

    Talia’s Logistics Dashboard:

    Key Insight: The Midwest region shows major logistics disruption: a 59% delivery delay, 35% staffing gap at the Chicago hub, and a $1.28 per-unit cost surge—triggered by reroutes and carrier delays.

    Agent Insights – What the Dashboards Revealed to the Agents

    As part of the daily review cycle, each agent initiated a rapid diagnostic scan of their functional dashboards—surfacing anomalies, shortfalls, and emerging threats from the day’s incoming data load. The folllowing transcript captures the collaborative intake phase, where agent specialists flag critical issues in preparation for joint strategy formation. Their early assessments below form the baseline for downstream coordination.

    Supply Chain Team Transcript: Agent Analysis

    Orchestration and Strategy – When Agents Teams Work Together

    After reviewing their functional dashboards, the Supply Chain agents transitioned from isolated diagnostics to integrated strategy formation. What follows is a transcript—condensed for clarity—that reveals how a distributed team of AI experts negotiated trade-offs, merged perspectives, and built a coordinated mitigation strategy for escalating Midwest risks.

    Supply Chain Team Transcript: Team Analysis

    The team co-developed a gated intake triage plan for DC-C with four synchronized filters: SKU velocity tier, forecast lock window, supply/demand thresholds, and margin-volatility pairing. They agreed to data handoffs via shared APIs and established cap tolerances to maintain flexibility under risk. This interaction exemplifies emergent cognition—where no individual agent held the entire solution, but collaboration yielded a coherent, executable plan.

    Conclusion

    This example highlights Sentienta’s core advantage: turning fragmented functional data into synchronized decision intelligence. As agents negotiate thresholds, define roles, and operationalize shared triggers, strategy becomes not just automated—but emergent. Sentienta Teams adapts at the pace of complexity, enabling businesses to respond with coordinated precision.