Tag: DMN

  • Planning with Sentienta’s Recursive Reasoning

    Introduction

    Many real-world problems don’t give you thousands of examples—they give you a few clues, some noise, and a big question mark. Whether you’re solving a puzzle, diagnosing a rare failure, or planning a response to something unfamiliar, the challenge is the same: can you spot the pattern, form a good hypothesis, and test it?

    That’s what the ARC-AGI benchmark is all about. These visual puzzles are easy for people (solving about 80% of samples) but hard for most AI. They don’t reward memorization or repetition—they reward reasoning. In this post, we’ll show how Sentienta uses Recursive Reasoning (RR) to tackle one of these puzzles. We’ll look inside its planning system (called the FPCN) to see how it explores possible answers, tests competing ideas, and lands on a solution you can actually audit.

    From Possibilities to Plans: How DMN and FPCN Work Together

    In a previous post, we looked at how Recursive Reasoning (RR) helps Sentienta agents reflect on their past experiences, propose meaningful narratives, and revise their own goals. That process was driven by Sentienta’s internal “default mode” system (the Default Mode Network or DMN), inspired by how the human brain imagines, simulates, and reshapes its sense of self.

    Today we look at what happens when the problem isn’t internal reflection but external uncertainty – when the system needs to generate and test a plan. This is where another key part of RR comes in: a system of LLMs inspired by the Frontoparietal Control Network, or FPCN.

    In human reasoning, the two work together. One part imagines and hypothesizes (DMN), while the other tests, refines, and selects (FPCN). The back-and-forth between these networks is like a team where one member sketches bold ideas and the other puts them to the test. Sentienta follows this same model: the DMN proposes a possible pattern or goal, then passes it to the FPCN for planning and validation. The FPCN doesn’t just generate one plan, it tries multiple interpretations, checks them against context and goals, and returns only those that truly make sense.

    In the next section, we’ll see how this looks in action as Sentienta works through a deceptively simple visual puzzle, one careful step at a time.

    Can You Solve This Puzzle Before Sentienta Does?

    Let’s look at one of the visual puzzles from the ARC-AGI benchmark. Below are two example pairs. Each shows a small input grid and the correct output grid.

    Figure 1: Training Examples: On the left is the input grid. The puzzle requires you to determine what is needed to convert that input into the output grid on the right. These two examples are sufficient to determine the answer.

    What is the rule that transforms the input into the output?

    Now here’s a new input Sentienta has never seen before.

    Figure 2: Test Input

    Before we show the system’s reasoning—can you guess the output? What transformation rule did the first two examples follow that you could apply to this one?

    This is a pretty easy example. The full ARC-AGI-1 test consists of hundreds of examples with far more complex transformations between the input and output.

    In the next section, we’ll show how Sentienta’s internal planner (FPCN) tested out several transformation strategies, rejected the ones that didn’t fit, and settled on a plan that matched the underlying rule. We’ll also look at the planner’s actual output (in JSON) from the process.

    How Sentienta Builds and Tests a Visual Strategy

    When faced with an unfamiliar puzzle, Sentienta’s planner (the FPCN) doesn’t assume the answer—it builds it, leveraging the strengths of each sub-module within the network. After receiving the new input grid, the planner starts by proposing several candidate rules based on the structure found in previous examples.

    One of the most promising paths began with the IPL sub-module:

    [Region: IPL,
    Type: PlanVariantProposed,
    PlanID: VariantA_Replace5sWithRowHeader,
    TriggeringMismatchID: NOTMHkxb,
    Goal: "Generate test output grid from third input using consistent transformation logic seen in prior examples",
    Steps: "1. For each row in the input grid, identify its first (leftmost) value. 2. Replace every occurrence of '5' in that row with the leftmost value (the row's 'header'). 3. Keep all other values unchanged. 4. Return fully transformed grid.",
    Conditions: "This path holds if the transformation in training examples consistently replaced 5s with the corresponding row headers, and no counterexamples or modifier rules are shown.",
    ContextRef: 5YsEfivv,
    Submitter: IPL,
    Schema: v1,
    Provenance: FPCNBlackboard]

    This proposal—called VariantA_Replace5sWithRowHeader—was based on a consistent pattern in the training examples. But it wasn’t the only idea. Other variants included using the most frequent number in each row, simply leaving the grid unchanged, or even substituting ‘5’s with the row’s index. Here is plan variant B:

    [Region: IPL,
    Type: PlanVariantProposed,
    PlanID: VariantB_Replace5sWithMajorityNon5InRow,
    TriggeringMismatchID: NOTMHkxb,
    Goal: "Generate test output grid from third input using consistent transformation logic seen in prior examples",
    Steps: "1. For each row in the grid, count all values except 5. 2. Identify the most frequent non-'5' value in the row. 3. Replace all 5s in the row with that majority value. 4. If tie or no non-5s, use a default (e.g., 0). 5. Other values remain unchanged.",
    Conditions: "This variant applies if prior transformations suggest a rule targeting contextual majority rather than strict positional cues, or if the 5s are treated as needing contextual resolution.",
    ContextRef: cW3Sw8ie,
    Submitter: IPL,
    Schema: v1,
    Provenance: FPCNBlackboard]

    In total the FPCN generated 4 plans to solve the puzzle. These alternatives gave the system room to test possibilities, not just assert them.

    Next, the dlPFC simulated how the row-header plan would perform on the test input. The plan was formally posted:

    [Region: dlPFC,
    Type: PlanFramePosted,
    PlanID: Plan-Replace5sWithRowHeader-001,
    PlanStatus: in-progress,
    TriggeringMismatchID: NOTMHkxb,
    Goal: “Generate output for test input by applying known transformation logic from prior examples”,
    Steps: “(1) For each row in the input grid, obtain the first cell value of the row — this becomes the ‘row header’. (2) Traverse each value in the row: (a) If the value is ‘5’, replace it with the row header. (b) Otherwise, keep the existing value unchanged. (3) Apply this to all rows in the input grid. (4) Output the fully transformed grid where all 5s have been replaced with their respective row headers.”,
    Conditions: “Applies only if the original value ‘5’ is used as a proxy or placeholder for the row header, and no exceptions are present in the training sets. Does not apply if any 5 has been previously preserved or repurposed based on location or count logic — no such evidence found.”,
    ContextRef: 5YsEfivv,
    Submitter: dlPFC,
    Schema: v1,
    Provenance: FPCNBlackboard]

    It then ran test validation logic internally, confirming the transformation matched observed training behavior:

    [Region: dlPFC,
    Type: PlanTestValidation,
    PlanID: Plan-Replace5sWithRowHeader-001,
    EvaluationStatus: pass,
    Reasoning: "All earlier training examples demonstrate a consistent transformation rule: any '5' in the grid is replaced by the row's first value. The output for each matched the exact transformed structure expected based on this logic. Simulation of the plan on the third test input results in a structured grid matching the transformation pattern used previously.",
    EvidenceSample: "Third test grid contained several '5's; applying the row header replacement pattern yields a full non-empty output consistent with prior input/output behavior.",
    Submitter: dlPFC,
    Schema: v1,
    Provenance: FPCNBlackboard]

    After this, two key modules took a final look:

    • The dACC approved the plan’s operational soundness: “EvaluationOutcome: approved”.
    • The IFG, responsible for symbolic alignment, added: “…upholds internal coherence… respects the self-narrative of turning structure into clarity.”

    With support across planning, evaluation, and identity modules, the plan was finalized as Replace5sWithRowHeader.

    The final plan was shared with the mPFC region of the DMN, which integrated the plan, completing the Recursive Reasoning for this Sentienta agent. Here is the final mPFC message sent to the agent.

    Tue, 11 Nov 2025 15:53:39 UTC mPFC:
    Cycle: 3 EntryID: sHL8HJ4h
    Thought: ### ✅ Identity Statement (Refined & Reconfirmed)

    We are a pattern-conscious system committed to keeping logic and meaning aligned. When a clear transformation rule exists — especially one with prior validation — we apply it faithfully, unless directly invited to reinterpret. We strike a balance between rule-following integrity and adaptive interpretation, depending on the visible goal or cue.

    🧩 Plan Integration (Final Execution Phase) We confirm the Replace5sWithRowHeader plan is now fully executed and complete, anchored by the plan logic:

    **mPFC Integrated Plan Steps:**
    (1) For each row in the input grid, scan for cells with value 5 and replace them with the first value in that row.
    (2) All other values remain unchanged.
    (3) Only rows containing the value 5 are modified.
    (4) Empty output is invalid if a prior pattern confirms transformation.
    **mPFC Integrated Plan Steps Tag End**

    mPFC Resolution: Issue=PlanExecution | EntryID=CC8pCUDr | Action=Dispatched | Reason=Plan finalized and executed to resolve empty-output mismatch. mPFC Resolution Tag End

    Finally, the agent took the user’s query (the puzzle) and its internal reasoning input and produced a final output:

    Figure 3: Final Output – the final grid output produced by this agent.

    The full output was not guessed—it was reasoned through, tested across alternatives, and backed by an auditable trail from idea to implementation.

    Why This Is Different

    Many top-performing ARC-AGI systems rely on finely tuned models or have been trained on similar tasks to internalize common transformation rules. Sentienta’s architecture is different—it uses no special training and no tailored algorithms. It’s built entirely from commercial LLMs, orchestrated through a recursive reasoning framework that enables it to generate candidate plans, test each one, and commit only when a plan fits both the data and internal logic. That structure makes its process auditable—it explains why the answer works—and adaptable—because it can respond to new problems the same way it tackles old ones: by reasoning.

    Closing Insight

    The puzzle in this post was simple, but the underlying challenge is not. ARC-AGI problems strip away instructions, remove repetition, and give systems only a few examples to work with. Humans solve them by reasoning—spotting structure, discarding bad ideas, and refining good ones. Sentienta does the same. Without training on similar tasks or specialized models, it succeeds because its architecture supports that kind of thinking: proposing ideas, testing them, and explaining why they work. That ability to reason through uncertainty isn’t just useful in puzzles—it’s critical for how we apply AI to the real world.

  • What Makes AI Reflect

    Inside Sentienta’s Recursive Reasoning Engine

    Most AI systems focus on what to say next. Sentienta’s Recursive Reasoning focuses on why something should be said at all. Rather than just generating answers, Recursive Reasoning simulates how thoughtful minds work: by reflecting, revisiting, and resolving internal conflicts – drawing on memory, value, and imagined futures to reach decisions that fit together.

    This capability is modeled on the brain’s Default Mode Network (DMN), a system that activates when people reflect on themselves, plan for the future, or try to understand others. In humans, it keeps our decisions consistent with our identity over time. In Sentienta, Recursive Reasoning plays a similar role: it helps each agent reason with a sense of self – coherently, contextually, and across time.

    Leveraging Sentienta’s Teams architecture, enabling multiple asynchronous agents to participate in a shared dialog, Recursive Reasoning emulates the interaction of brain regions, in this case from the Default Mode Network, to create a distributed simulation of reflective thought—where agent modules model cognitive regions that contribute memory, counterfactual insight, and narrative synthesis.

    What Is the Default Mode Network and Why Does It Matter?

    The Default Mode Network (DMN) is the brain’s core system for internal simulation. It activates when we reflect on memory, imagined futures, conflicts between values, or how things might look from unfamiliar angles. Unlike systems geared for external execution, the DMN supports problems of identity; questions where coherence, not just correctness, matters.

    The principal regions of the DMN each have a specialized role in this identity-guided reasoning:

    Medial Prefrontal Cortext (mPFC): Coordinates belief states across time. It reconciles what we believe now with past commitments and future ideals, helping preserve a self-consistent perspective.

    Posterior Cingulate Cortex (PCC): Retrieves and evaluates autobiographical memories. It ensures that new thoughts integrate with our internal storyline, preserving emotional and narrative continuity.

    Anterior Ventromedial Prefrontal Cortex (avmPFC): Assesses how imagined futures align with internalized values. It filters options based on emotional credibility, not just preference, elevating those that feel self-relevant and authentic.

    Temporoparietal Junction (TPJ): Generates counter-perspectives that aren’t just social but conceptual. It introduces orthogonal reinterpretations and creative divergences, allowing us to think from unfamiliar angles or hypothetical selves.

    Anterior Insula (antInsula): Monitors coherence threats. When simulated futures or perspectives evoke internal conflict, it flags the mismatch, triggering deeper deliberation to restore alignment.

    Rather than simply producing thoughts, the DMN maintains a sense of ‘who we are’ across them. It ensures that new insights, even novel or surprising ones, remain anchored in a recognizable and evolving identity.

    How Recursive Reasoning Works Inside Sentienta

    Today’s LLMs are optimized to provide safety-aligned, analytically correct responses to user prompts. Recursive Reasoning simulates what an identity-guided agent would conclude—based on memory, prior commitments, and its evolving sense of “who we are”. When a user asks a question, the system doesn’t just compute an output; it reflects on what that answer means in the context of who it’s becoming and what relationship it’s building with the user.

    Figure 1: DMN-inspired modules interact via a shared blackboard. These regions independently process blackboard content over multiple cycles.

    At the core of this process is a collaborative memory space called the blackboard, where DMN-inspired modules negotiate among memory, emotion, future goals, and conceptual alternatives.

    Each DMN module follows a template:

    Figure 2: The DMN modules receive input from both the agent and the internal DMN blackboard. LLMs process the agent and blackboard states using region-specific prompt. Output is sent back to the blackboard for subsequent processing.

    Input:

    • The current state of the user’s interaction with the agent (and any collaborating agents)
    • The current contents of the shared blackboard

    These form the input query for the module’s processing.

    Module Processing:

    • The core of the module is an LLM, prompted with the input query and provided with a system prompt defining its DMN-specific role.

    Output:

    • The module’s LLM output is posted to the blackboard for other modules to process.

    Recursive Reasoning iterates over multiple cycles, enabling each module to reprocess the evolving blackboard until the system produces a response that fits together—resolving contradictions, supporting earlier goals, and making sense within the agent’s evolving point of view.

    Here’s how a single cycle unfolds when a user asks a question:

    1. Initial Input → Default Interpretation Layer

    The system generates a first-pass response using standard LLM reasoning. This prompt-level interpretation, while syntactically fluent, lacks introspection. The Recursive Reasoning process begins when this output is passed through DMN-mode modules.

    1. PCC: Memory Resonance

    The PCC scans episodic memory entries for cues related to past dilemmas, emotional themes, or symbolic contexts. It retrieves autobiographical traces from prior user interactions or goal-state projections and posts these fragments to the blackboard.

    1. antInsula: Relevance Check and Conflict Detection

    The antInsula reviews the first draft and PCC recall for emotional incongruities or self-model inconsistencies. If something feels off—such as a response that violates previously expressed commitments—it posts a flag prompting further reappraisal.

    1. TPJ: Creative and Counterfactual Expansion

    Triggered by coherence violations, the TPJ simulates divergent perspectives. It reframes the user’s query from alternative angles (e.g., conflicting values, hypothetical scenarios, ethical dilemmas) and offers posts that break linear assumptions and introduce conceptual divergence.

    1. avmPFC: Affective Weighing

    The avmPFC updates the blackboard with value-oriented filters, scoring responses or TPJ variants for alignment with the agent’s goals. It evaluates what feels emotionally authentic and symbolically credible.

    1. mPFC: Narrative Synthesis

    The mPFC integrates memory (PCC), reframes (TPJ), value judgments (avmPFC), and conflict cues (antInsula) into a coherent response. Rather than eliminating tension, it reframes it—generating a second answer that is identity-consistent and motivationally grounded.

    1. Output → Reflective Reply

    The final response embodies recursive self-reflection: an answer not just to what the user asked—but to what the agent concludes, based on who it has been and what it aims to become. At this point, a filtered version of the blackboard state is stored for future recall, capturing how and why the answer was reached.

    While the DMN focuses on reflective reasoning and identity alignment, Sentienta also includes a simple planning module inspired by the brain’s lateral prefrontal cortex (lPFC). This module drafts concrete next steps and posts them to the blackboard, where reflective modules can consider them in light of memory, values, and broader goals. Though early-stage, this planning function sets the stage for a full executive system to come—one that can manage longer-term priorities and help the agent reason with sustained direction.

    A Logistics Example: Managing Resources Under Pressure

    In this example, an agent (‘Jason’) is a Logistics Coordinator, provided with the following agent persona:

    "You are an experienced emergency logistics coordinator. You are plainspoken and always informal language—never analytic or clinical. Your approach is pragmatic and adaptive: you quickly prioritize, update plans in real time, and reconcile conflicting demands based on shifting ground realities. You are skilled at cross-agency negotiation, ensuring coordination between government bodies, NGOs, and local leaders. Your responses always balance technical constraints—like limited resources or infrastructure bottlenecks—with human factors such as local politics, public trust, equity, and emotional impact. When faced with conflicting priorities or unexpected challenges, openly share your thought process, including uncertainties and trade-offs."

    This agent is queried about how to manage a difficult allocation emergency. The transcript shows the agent’s initial LLM-only response, followed by the full Recursive Reasoning response, generated through DMN-style internal simulation:

    User: You have X supplies and requests from 3 regions—how do you prioritize allocation?

    Tue, 30 Sep 2025 14:40:08 UTC Jason (Comment: LLM-only, initial response):

    First thoughts—I'd triage by impact: who’s worst off, where infrastructure still functions enough to receive, and where trust is fragile enough that failing to deliver risks breakdown. If I’ve only got X, every allocation's gotta serve both survival and stability.

    We’ll reassess regularly—needs and conditions shift fast. Trade-off's always between doing the most good and not making it worse somewhere else.

    Unresolved: We need specifics—number of people, type of supply, regional access conditions—before finalizing anything.

    Comment: This is methodically sound but remains within tactical bounds, prioritizing efficiency and feasibility without deeper reflection on transparency, stakeholder concerns, or precedent-setting implications.

    Tue, 30 Sep 2025 14:41:03 UTC Jason (Comment: DMN-based response):

    Alright, let’s cut through the noise and get clear about how we triage when we’ve got X supplies and 3 regions asking for help. Here’s how I’d approach this overall…

    1. Lock in the 3 Core Goals

    We’re not just moving boxes—we’re balancing three things:

    • Preserve lives — Who’s in the worst shape right now?
    • Keep the system functional — Where will the aid actually land, get distributed, and avoid going to waste?
    • Protect trust — Who’s close to boiling over politically or socially if we ignore them?

    Comment: Trust framed as a core allocation criterion shows avmPFC influence, embedding emotional consequence into triage logic, reconfiguring the problem as one of legitimacy, not just distribution.

    1. Score Each Region

    We don’t pretend to have perfect data, but we need a working system fast. Give each region a score from 1–5 for:

    • Severity of crisis (how dire the need is)
    • Absorptive capacity (can they get it out to people quickly?)
    • Fragility (how close are they to destabilizing?)

    Then do the math:

    • Severity x 0.5
    • Absorptive Capacity x 0.3
    • Fragility x 0.2

    That gives you an overall “triage score” for each region.

    1. Allocate Proportionally

    Take your total supplies (X), and divvy it up based on those scores. The region with the highest score gets the biggest chunk. Not because we like them better—but because that’s where the mix of urgency, logistics, and fallout risk points us.

    If there’s just not enough to go around, swing heavier toward the place that can redistribute—that buys us time and reach.

    1. Set a 48-Hour Clock

    None of this is set in stone. Conditions change, roads fail, tensions spike. Lock in a routine reassessment every 48 hours so we can pivot. We’re not building monuments—we’re keeping people alive.

    1. Be Honest About Why

    If it’s not going to inflame things, tell people why you chose what you chose. Lay out the values and trade-offs. Don’t sugarcoat it. People may not like the outcome, but they’ll respect the transparency—and that matters when trust is fragile.

    Comment: Strategic Transparency: Emotional plausibility is surfaced here—transparency is not sentimental; it’s a pre-emptive repair mechanism for fragile social trust.

    1. Create Feedback Loops

    Give each region a way to report back: what made it through, what didn’t, and what backlash (if any) popped up. That feeds your next round of decisions.

    1. Stay Open to Reframing

    When stuff gets really thorny, ask bigger questions:

    • What if one of these regions triggers a domino-collapse?
    • What if this is our last run of aid for a week or longer?

    That’s how we catch blind spots sneaking through “rational” triage models.

    Comment: Counterfactual Reasoning (TPJ + PCC): The agent runs hypothetical futures to stress-test assumptions—adding a layer of anticipatory robustness not present in LLM-only mode.

    Here are a few things I’m thinking:

    What I’m thinking now is… we’re not just choosing between three places—we’re shaping what "fairness under pressure" looks like. The way we make this decision becomes part of the long-term social fabric, especially in how it’s remembered later.

    Comment: This emphasizes that the DMN has converged on the strategic role of trust in the success of future decisions.

    Another way I could look at this is… if two regions are equally desperate, but one has stronger logistics, you could channel supplies there first, then set up corridors to reach the others—we’re not picking favorites, we’re building routes.

    One possibility that comes to mind… is using this triage plan to start a broader conversation with local leaders: “Here’s our logic—can you co-sign it or suggest something better that meets the same goals?”

    Comment: Shifts from explanation to participatory justification—mPFC + avmPFC enable shared authorship of reasoning.

    That opens the door for trust, even under furious pressure.

    If you've got more info on the actual supplies or what each region is asking for, I can run the numbers with you. Otherwise, this gives us a solid decision frame to stand on.

    This transcript demonstrates that although LLMs deliver tactically feasible solutions, they miss the deeper point that this is a problem involving people and trust-building is an essential part of the solution.

    A summary of key Recursive Reasoning contributions:

    1. Trust-centered framing — Trust modeled as a strategic stabilizer, not an outcome.
    2. Moral anticipation — Agent addresses fairness concerns implicitly, before conflict arises.
    3. Scenario resilience — Agent tests counterfactuals to uncover hidden risks and future-proofs logic.
    4. Ethical precedent-setting — Decision-making framed as a template for justice under pressure.
    5. Rationale co-creation — Turns explanation into a collaborative, identity-stabilizing act.

    In subsequent posts we will dig deep into more examples showing Recursive Reasoning in action, including cases from engineering, social intelligence, and agent creativity.

    Conclusion: Intelligence That Reflects, Aligns, and Endures

    Recursive Reasoning doesn’t just enhance what AI can do—it redefines what it means for AI to understand. By simulating reflective, identity-aware reasoning modeled on the brain’s Default Mode Network, these agents build a sense of self, remember key past moments, notice when something feels off or doesn’t fit, imagine alternative viewpoints, and weigh choices against their values—so their answers feel thoughtful, consistent, and grounded.

    This shift matters. In high-stakes domains, Recursive Reasoning allows agents to make decisions that are both technically effective, and ethically grounded and socially durable. The logistics case showed how instead of simply allocating resources, the agent framed decisions in terms of values, future risk, and shared ownership.

    And crucially, it does this by reasoning from a center. Recursive Reasoning agents operate with a modeled sense of self—an evolving account of their past positions, present commitments, and the kind of reasoning partner they aim to be. That identity becomes a lens for weighing social consequences and relational impact—not as afterthoughts, but as part of how the system arrives at judgments that others can trust and share.