Files
FusionAGI/fusionagi/core/super_big_brain.py
defiQUG c052b07662
Some checks failed
Tests / test (3.10) (push) Has been cancelled
Tests / test (3.11) (push) Has been cancelled
Tests / test (3.12) (push) Has been cancelled
Tests / lint (push) Has been cancelled
Tests / docker (push) Has been cancelled
Initial commit: add .gitignore and README
2026-02-09 21:51:42 -08:00

137 lines
5.3 KiB
Python

"""Super Big Brain orchestrator: tokenless, recursive, graph-backed reasoning."""
from __future__ import annotations
from dataclasses import dataclass, field
from typing import Any
from fusionagi.schemas.atomic import AtomicSemanticUnit, DecompositionResult
from fusionagi.schemas.head import HeadId, HeadOutput, HeadClaim, HeadRisk
from fusionagi.schemas.grounding import Citation
from fusionagi.reasoning.decomposition import decompose_recursive
from fusionagi.reasoning.context_loader import load_context_for_reasoning, build_compact_prompt
from fusionagi.reasoning.tot import ThoughtNode, expand_node, prune_subtree, merge_subtrees
from fusionagi.reasoning.multi_path import generate_and_score_parallel
from fusionagi.reasoning.recomposition import recompose, RecomposedResponse
from fusionagi.reasoning.meta_reasoning import challenge_assumptions, detect_contradictions
from fusionagi.memory.semantic_graph import SemanticGraphMemory
from fusionagi.memory.sharding import shard_context
from fusionagi.memory.scratchpad import LatentScratchpad
from fusionagi.memory.thought_versioning import ThoughtVersioning
from fusionagi._logger import logger
@dataclass
class SuperBigBrainConfig:
"""Configuration for Super Big Brain pipeline."""
max_decomposition_depth: int = 3
min_depth_before_conclusion: int = 1
parallel_hypotheses: int = 3
prune_threshold: float = 0.3
max_context_chars: int = 4000
def run_super_big_brain(
prompt: str,
semantic_graph: SemanticGraphMemory,
config: SuperBigBrainConfig | None = None,
adapter: Any | None = None,
) -> RecomposedResponse:
"""
End-to-end Super Big Brain pipeline:
1. Decompose prompt -> atomic units
2. Shard and load context
3. Run hierarchical ToT with multi-path inference
4. Recompose with traceability
5. Persist units/relations to semantic graph
"""
cfg = config or SuperBigBrainConfig()
decomp = decompose_recursive(prompt, max_depth=cfg.max_decomposition_depth)
if not decomp.units:
return RecomposedResponse(summary="No content to reason over.", confidence=0.0)
semantic_graph.ingest_decomposition(decomp.units, decomp.relations)
ctx = load_context_for_reasoning(decomp.units, semantic_graph=semantic_graph, sharder=shard_context)
compact = build_compact_prompt(decomp.units, max_chars=cfg.max_context_chars)
hypotheses = [u.content for u in decomp.units[:cfg.parallel_hypotheses] if u.content]
if not hypotheses:
hypotheses = [compact[:500]]
scored = generate_and_score_parallel(hypotheses, decomp.units)
nodes = [n for n, _ in sorted(scored, key=lambda x: x[1], reverse=True)]
best = nodes[0] if nodes else ThoughtNode(thought=compact[:300], unit_refs=[u.unit_id for u in decomp.units[:5]])
if cfg.min_depth_before_conclusion > 0 and best.depth < cfg.min_depth_before_conclusion:
child = expand_node(best, compact[:200], unit_refs=best.unit_refs)
child.score = best.score
best = child
prune_subtree(best, cfg.prune_threshold)
assumptions = challenge_assumptions(decomp.units, best.thought)
contradictions = detect_contradictions(decomp.units)
recomp = recompose([best], decomp.units)
recomp.metadata["assumptions_flagged"] = len(assumptions)
recomp.metadata["contradictions"] = len(contradictions)
recomp.metadata["depth"] = best.depth
logger.info(
"Super Big Brain complete",
extra={"units": len(decomp.units), "confidence": recomp.confidence},
)
return recomp
def _recomposed_to_head_output(
recomp: RecomposedResponse,
head_id: HeadId,
) -> HeadOutput:
"""Convert RecomposedResponse to HeadOutput for Dvādaśa integration."""
claims = [
HeadClaim(
claim_text=c,
confidence=recomp.confidence,
evidence=[Citation(source_id=uid, excerpt="", confidence=recomp.confidence) for uid in recomp.unit_refs[:3]],
assumptions=[],
)
for c in recomp.key_claims[:5]
]
if not claims:
claims = [
HeadClaim(claim_text=recomp.summary, confidence=recomp.confidence, evidence=[], assumptions=[]),
]
risks = []
if recomp.metadata.get("assumptions_flagged", 0) > 0:
risks.append(HeadRisk(description="Assumptions flagged; verify before acting", severity="medium"))
if recomp.metadata.get("contradictions", 0) > 0:
risks.append(HeadRisk(description="Contradictions detected in context", severity="high"))
return HeadOutput(
head_id=head_id,
summary=recomp.summary,
claims=claims,
risks=risks,
questions=[],
recommended_actions=["Consider flagged assumptions", "Resolve contradictions if any"],
tone_guidance="",
)
class SuperBigBrainReasoningProvider:
"""ReasoningProvider for HeadAgent: uses Super Big Brain pipeline."""
def __init__(
self,
semantic_graph: SemanticGraphMemory | None = None,
config: SuperBigBrainConfig | None = None,
) -> None:
self._graph = semantic_graph or SemanticGraphMemory()
self._config = config or SuperBigBrainConfig()
def produce_head_output(self, head_id: HeadId, prompt: str) -> HeadOutput:
"""Produce HeadOutput using Super Big Brain pipeline."""
recomp = run_super_big_brain(prompt, self._graph, self._config)
return _recomposed_to_head_output(recomp, head_id)