
Agent with memory + reflection logic. Reads context, reacts to emotional cues.
Why I Built It
Most AI systems today are black boxes. Pretrained, stateless, non-reflective.
That’s a dead end for autonomy.
I needed a way to give agents memory, context, and the ability to reflect on previous actions — not just respond.
So I wrote my own language.
It’s called Sentience. It’s a REPL where everything an agent does is contextual, stored, and queryable. A full loop:
Input → Evaluation → Memory → Reflection → Adjustment
What It Does
- Custom REPL (written in Rust)
- Memory access:
mem.short,mem.long,mem.latent - Reflective blocks:
reflect { ... } - Conditional logic based on context (
if context includes "X") - Embed, train, recall — all scoped to the agent’s active memory state
- Supports structured agent declarations
What's Unique
- Not Turing-complete on purpose
- Designed for introspection, not general-purpose coding
- Agents can read what they’ve done, compare, and adjust
- REPL is deterministic, readable, hackable
Current Status
REPL parser, lexer, AST
Memory evaluator
Reflection blocks (reflect, embed)
WIP Training mechanism stubbed
WIP Agent context loop (Inception) in progress
Everything runs in isolated Rust runtime (no external LLMs)
Links
- Code: github.com/nbursa/sentience
- Architecture WIP: github.com/nbursa/SynthaMind
What's Next
- Complete the training block (
train {}) - Support agent-defined behavior loops
- Integrate latent similarity (
similar_to "X") for semantic recall - Finish reflection runtime (Inception project)
Final Thought
I didn’t build Sentience to compete with anything.
I built it because I needed a tool to think with.
If it helps someone else build something weird — even better.
— Nenad