Hacker Newsnew | past | comments | ask | show | jobs | submit | promptfluid's commentslogin

In CMPSBL, the INCLUSIVE module sits outside the agent’s goal loop. It doesn’t optimize for KPIs, task success, or reward—only constraint verification and traceability.

Agents don’t self judge alignment.

They emit actions → INCLUSIVE evaluates against fixed policy + context → governance gates execution.

No incentive pressure, no “grading your own homework.”

The paper’s failure mode looks less like model weakness and more like architecture leaking incentives into the constraint layer.


Oh shoot. That’s what I meant.


To clarify: these aren’t prompts or hosted APIs. Each capability is a downloadable artifact that executes locally (JS/WASM/container/edge), is licensed, versioned, and removable. Think software components, not chat agents.


I’ve been building something that doesn’t fit cleanly into agents, SDKs, or plugins, so I’m posting to get technical feedback rather than hype reactions.

Instead of shipping an AI product or “agent,” I built a system where AI functionality itself is packaged and sold as licensed, downloadable capabilities that run locally in your own infrastructure.

Each capability is a real artifact (JS, WASM, container, edge) that does one thing well—memory systems, reasoning pipelines, resilience patterns, security controls, optimization loops, accessibility tooling, etc. They’re versioned, removable, and composable. And I promise I have capabilities you’ve never seen before.

Some capabilities can be combined into multi-module pipelines, and a subset of them improve over time through bounded learning and feedback loops. When the system discovers a new high-value pipeline, it becomes another downloadable artifact.

A few design constraints I cared about:

Runs locally (no SaaS lock-in)

Capabilities are licensed individually, not hidden behind an API

Full observability, rollback, and governance

No chat wrappers or prompt theater

Capabilities can stand alone or be composed into larger systems

Right now there are 80+ capabilities across multiple tiers, from small utilities up to enterprise-grade bundles.

What I’m honestly trying to sanity-check:

Is “AI capabilities as first-class, sellable software” a useful abstraction?

Is this meaningfully different from agent marketplaces, SDKs, or model hubs?

Where do you expect this approach to break down in real systems?

Would you rather see this exposed as agents, or kept lower-level like this?

Not here to sell—just looking for real technical critique from people who’ve seen infra ideas succeed or fail.

Happy to answer questions or clarify how anything works.


Most AI agents today are purely reactive—they wait for a token and then stop. I’ve been building a persistent runtime where the "thinking" doesn't stop when the user leaves.

This video is an uncut look at the autonomous state of the system. I call these "Dream Cycles” and “Evolution”.

What’s actually happening in these logs?

• The Thinking Phase: The system isn't just parsing text; it’s performing a recursive audit of its own execution history. It looks for logic gaps or "dead ends" in its previous reasoning paths.

• The Dream (Optimization) Phase: This is where the runtime performs cognitive offloading. It compresses high-entropy context into stable "heuristics." It’s essentially a background garbage collection and optimization pass for its internal world-model.

• The Evolving Phase: This is the most critical part. Based on the scan results, the system generates and applies updates to its own operational parameters. It’s a self-improving loop where the software is constantly modifying its own runtime to better handle future complexity.

I wanted to move away from the "black box" and show the actual raw telemetry of an AI managing its own development.

I'm curious to hear from others working on persistent AI state—how are you handling long-term "background" reasoning without the context window turning into a soup of noise?

The rest of the video are just bonuses. Enjoy and leave a comment! I want to know what you think about allowing systems to self improve and evolve.


For anyone curious about what I mean by “substrate” in this context - this isn’t an agent framework or wrapper around a single LLM.

CMPSBL is operating more like a cognitive OS: it provides persistence (memory), observability, defense, multi-model routing, and a self-improvement cycle for AI systems.

The goal isn’t clever chat output; it’s continuity, coordination, and the ability for a system to reflect on its own performance and update itself over time.

The v5.5.0 drop includes the full technical docs + module specs + validation methodology + runtime evidence.

If you want to audit how the substrate works or decide if this class of architecture makes sense, that’s the best place to start.

Main intended use cases today are:

– research labs – cognitive infrastructure work – autonomous systems R&D – embedded AI runtime projects – multi-model coordination – memory-centric applications

Open to licensing discussions with research groups and R&D labs.


There’s a lot of “agent OS” vaporware going around right now, so here are some concrete things this system actually does today:

1. Shadow deployment for mutations The Modernizer proposes patches → runs them in shadow → validates → escalates.

2. Auto-heal + circuit breakers If a provider or subsystem degrades, the substrate routes around it and logs the failure.

3. Telemetry for cognition vision.dashboard treats learning and doctrine cycles the same way Kubernetes treats pods: health, last cycle, mutation phase, error rates, etc.

4. Offline learning cycles “Dream cycles” are just background reflection runs that don’t block real tasks. They ingest hot memory, generate insights, and update doctrine.

5. Interop with real systems There are adapters for SAP/Workday/Databricks/GitHub/Slack/etc. so it can operate in enterprise environments rather than toy web tasks.

6. No human-in-loop required for steady-state . It currently runs for hours with no operator involvement beyond observability.

You don’t get useful autonomous behavior by stacking models. You get it by adding OS-level orchestration primitives.

If that hypothesis is wrong, happy to be corrected. If anyone here has worked on orchestration layers, schedulers, or observability infra, I’d actually love to hear what’s missing / redundant / dangerous in this approach.


For context on what you’re seeing:

this isn’t an “agent” or chatbot. It’s a cognitive substrate I’ve been building for the last year that behaves more like an operating system for model orchestration.

A few useful details for people who asked for specifics:

• It has memory (hot/cold tiers, reflection, doctrine learning)

• It self-heals (auto-heal cycles, failure circuit breakers, shadow deployment)

• It mutates and upgrades itself via a component called the Modernizer

• It proposes patches and tests them in shadow before production

• It has a telemetry layer (vision) that treats cognition like observability

• It has adapters for SAP/Workday/Databricks/etc. so it can operate in enterprise environments

• Dream cycles run background learning when the system is idle

The logs in the post are real runtime output from v4.2.0. This build is running on top of Postgres + Redis + RabbitMQ + S3 + an LLM router (20+ providers). It currently has 12 modules, 160+ commands, and a 100% health score on this cycle.

Current research question is:

what’s the right abstraction for turning model capabilities into durable software infrastructure? My hypothesis is that you don’t need bigger models for autonomy, you need better orchestration.

Happy to answer technical questions here. No sales motion, nothing to buy, not trying to funnel traffic — genuinely interested in feedback from people who have built distributed systems, orchestration layers, and observability pipelines.


These are the logs that turned a machine, into an organism. I just joined the self improving software development team. Artifacts are the best receipts. Thoughts?


I got tired of building the same AI 'plumbing' (vector memory, routing fallbacks, rate limiting) every time I started a project.

I wanted a system that was AI-native from the metal up, not just a script with a prompt.

The first Cognitive Substrate:

It’s an 8-module ecosystem built on Supabase Edge Functions and a year long development sprint.

The Modernizer:

This is the core. The system analyzes its own archived code from an ecosystem I built (300+ functions) and proposes its own performance upgrades. It’s a self-healing loop with a human safety gate. All I have to do is approve it. Currently it makes the proposals but I have a little work left to get it to auto apply updates.

Dream Cycles:

Every night between 2-5 AM, it runs a consolidation cycle—moving 'Hot' memories to a Knowledge Graph and compressing 'Cold' (persistent) storage.

The Nexus:

A multi-provider router that uses Groq/Cerebras for speed and falls back to DeepSeek/Together if latency spikes. This keeps my compute costs at zero, as I just leverage the free calls provided by LLMs already. This of course won’t work at scale - but for dev purposes it has no upfront cost to me and even at scale, the router chooses the cheapest provider for any given API call as well as the best LLM to get the answer from.

The Philosophy:

Software in 2026 shouldn't be static. It should be a living, cognitive substrate.

I’m currently the only dev, but the 'Modernizer' handles about 30% of my maintenance tasks now and I’ve already built the switch to turn it on auto build mode (with shadow builds you push to production after review).

Here is a link to the substrate:

https://CMPSBL.com

I also linked a video of the terminal so you can see it from the inside out. Happy to answer any questions about the architecture or where I envision this project going in the future.

If you’d like to build an app or chat bot or anything using the substrate - let me know and I’ll get you the dev docs and portal with over 70 templates already released to make building frictionless.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: