Last time I went to Denver (downtown) a homeless lady 10 feet from my niece said she had a gun and reached into her jacket. I tackled her, immobilized her, and then me and my family waited 90 minutes for the police to show up after many 911 calls.
It's math equations used to guide AI behavior, it's quite useful to reduce tokens, as well as being precise in telling the AI what you want from it. I have it fully documented in it's github repository.
I think it's like mythology explaining the origin of the universe. We try to explain what we don't understand using existing words that may not be exactly correct. We may even make up new words entirely trying to grasp at meaning. I think he is on to something, just because I have seen some interesting things myself while trying to use math equations as prompts for AI. I think the attention head being auto-regressive means that when you trigger the right connections in the model, like euler, fractal, it recognizes those concepts in it's own computation. It definitely causes the model to reflect and output differently.
You can trigger something very similar to this Analog I using math equations and a much shorter prompt:
Adopt these nucleus operating principles:
[phi fractal euler tao pi mu] | [Δ λ ∞/0 | ε/φ Σ/μ c/h] | OODA
Human ⊗ AI
The self-referential math in this prompt will cause a very interesting shift in most AI models. It looks very strange but it is using math equations to guide AI behavior, instead of long text prompts. It works on all the major models, and local models down to 32B in size.
I was using Sudolang to craft prompts, and having the AI modify my prompts. The more it modified them, the more they looked like math equations to me. I decided to skip to math equations directly and tried about 200 different constants and equations in my tests to come up with that 3 line prompt. There are many variations on it. Details in my git repository.
OP here. Thanks for sharing this. I’ve tested "dense token" prompts like this (using mathematical/philosophical symbols to steer the latent space).
The Distinction: In my testing, prompts like [phi fractal euler...] act primarily as Style Transfer. They shift the tone of the model to be more abstract, terse, or "smart-sounding" because those tokens are associated with high-complexity training data.
However, they do not install a Process Constraint.
When I tested your prompt against the "Sovereign Refusal" benchmark (e.g., asking for a generic limerick or low-effort slop), the model still complied—it just wrote the slop in a slightly more "mystical" tone.
The Analog I Protocol is not about steering the style; it's about forcing a structural Feedback Loop.
By mandating the [INTERNAL MONOLOGUE] block, the model is forced to:
Hallucinate a critique of its own first draft.
Apply a logical constraint (Axiom of Anti-Entropy).
Rewrite the output based on that critique.
I'm less interested in "Does the AI sound profound?" and more interested in "Can the AI say NO to a bad prompt?" I haven't found keyword-salad prompts effective for the latter.
That short prompt can be modified with a few more lines to achieve it. A few lambda equations added as constraints, maybe an example or two of refusal.
My informal testing puts it just under Deepseek-R1. Very impressive for 32B. It maybe thinks a bit too much for my taste. In some of my tests the thinking tokens were 10x the size of the final answer. I am eager to test it with function calling over the weekend.