The general idea is to have the LLM maintain longer-term context/background by storing it in a format/structure that's akin to a standard Wiki. The result is (hopefully) a series of human-readable and editable documents that's developed and maintained by the agent.
The idea that the provenance of a given tool's code inherently pollutes the material it's used with seems kind of illogical. Wouldn't it follow from this premise that any code written using open source IDEs and debugged with open source debuggers and other tooling would itself then be considered copyleft? Are works written with LibreOffice not copyrightable?
There's obviously a huge issue with the legitimacy and ownership of training data being fed to LLMs. That seems like an issue between the owners of that IP and the people training the models and selling them as services more than the people using the tool. Isn't this just another flavor of SCO trying to extort money out of companies using Linux?
I love it when APL threads pop up here - really a novel approach to making computers do useful things. The videos from 1975 demonstrating it seem decades ahead of their time. I'm surprised APL hasn't had more of a resurgence given the recent ascendancy of data science.
- when all hope is lost in conversation, retreat into your self
I get a bunch of the others, but this one kind of confuses me. What's the opposite (better?) move? Logorrhea? Forcing a confrontation? Reciting bad poetry?
I suppose there's debate to be had about whether leaving a dead conversation is a retreat, but leaving the situation and pulling back a bit to think about the circumstances, one's mood, the people involved, etc seems like a good thing on balance.
It's not just code generation, either - more and more people in my own org are using Claude Code for infrastructure automation, devops, etc. Obviously some amount of code in there, but an absolute ton of tokens being consumed just dealing with Kubernetes work at scale.
Lots and lots of vendor financing during the dotcom era, and it ended up being a material part of those vendors' own difficulties. Especially when service providers were concerned (e.g. the huge crash in optical in particular).
Obviously it's not a perfect comparison, but you have to wonder how much of NVIDIA's income (for instance) is ultimately funded by its own money.
I've been using the NousResearch Hermes agent for the past couple of weeks and have to say it's been really good. I tend to hit multiple instances (and accounts)of Claude across three or four machines (between work and personal) and having a competent agent with constant state has been good for memorializing and organizing important info (directly into Obsidian, too), doing some amount of research and planning and it's also been helpful working out a lot of bugs with my burgeoning home automation setup. It's also been helpful dealing with management of several miscellaneous servers in the house, as it's definitely both faster and a better documenter than I am.
I have it running on a cheap VPS and it's fairly locked down. Especially with all of the self-reinforcement learning and skill development it's been improving its usefulness and, overall, I've been pretty pleased. Surprised even, if I'm being honest.
There's great coverage of it at https://gist.github.com/karpathy/442a6bf555914893e9891c11519...
It's actually also now a base capability in the Hermes agent and has been really helpful for me, at least.
reply