If you’re downvoting, could you say why? I’m new to this site but there are very few posts on ai like my own. There’s a few on different ai tools, but what I have isn’t a tool, the output generated by it is for itself.
ollie_
It started out with the idea to make a system that can improve itself, without me having to prompt it to improve itself. The “question” is the prompt. And crafting the prompt is very difficult. So it’s an experiment in constructing a question that a system can use to improve itself without having to explicitly say “improve your codebase” or “give yourself more actions or freedoms” or others like that. I want the llm to conjure up ideas.
So as the prompt increases in complexity, with different sections that could be categorised as kinds of human thought, will it get better? Can we find parallels between how the human minds works that can be translated into a prompt. A key area that developed was using the future as well as the past in the prompt. Having memories is an obvious inclusion, as is including what happened in the past runs of the program, but what wasn’t obvious was including predictions for the future. Including the ability to make long and short term predictions, and have it be able to change these, and record the outcomes, saw big improvements in the directions it would take over time. It also seemed to ‘ground it’. Without the predictions space, it became overly concerned with its performance metrics as a proxy for improvement and it began over-optimising.
Defining ‘better’ or ‘improving’ is very difficult. Right now, i’m using token input size growth whilst maintaining clarity of thought as the rough measure.
I believe that opensource advancements like R1 will drive wider adoption of ai systems.
I think that the pricing models will change soon. Everyone talks about cost per million tokens to contact a hosted service, but I think it’ll switch to be cloud costs to provide infrastructure that can run models. Virtual machines running something like ollama.
This solves another huge problem, privacy and how prompt data is handled. If you’re using an api to a hosted service you need to have a very good understanding of how your submitted prompt data is handled. This is key for organisations. I feel like the lack of understanding here is preventing widespread adoption, especially for communication tools that handle sensitive data.
For example, you could run Deepseek R1 using ollama on an Azure virtual machine (nc series) that you pay per hour for, and then your cost isn’t based on usage of your ai. Right now it’s expensive to provision the infra to support decent models, but these costs fall continuously.
I can imagine a world where organisations provision cloud infrastructure in their environments running open source models.
https://learn.microsoft.com/en-us/azure/virtual-machines/sizes/gpu-accelerated/nc-series?tabs=sizebasic
https://huggingface.co/deepseek-ai/DeepSeek-R1