- everlierIf you're following local/open-weight LLMs, this might be an interesting retrospective of 2025.
- Just a fun little workflow I made. You can find the code here: https://github.com/av/harbor/blob/main/boost/src/modules/dnd...
- I am also confused, I can't see any recent news/updates for this project either.
- Hi,
I've been rocking my LLM setup for more than two years now and it grew quite extensive enough to warrant an "awesome" list.
You may read in detail about my experience with all the services here on r/LocalLLaMa: https://www.reddit.com/r/LocalLLaMA/comments/1oclug7/getting...
- I'm glad to hear I'm not alone. Due to the nature of what I do, I'm often accumulating ~800-900GB of Docker images and volumes on my machine, sometimes running 20-30 containers at once starting/stopping them concurrently. Somehow, very rarely, but still quite often (once every couple of weeks) - it leads to a complete deadlock somewhere inside of the kernel due to some crazy race condition that I'm absolutely in no way able to reliably reproduce.
- There are too many LLM-related projects. Setting up multiple runtimes, Python, Node, Go, Rust and then some environments, different CUDA versions, dependencies is tedious. Managing updates later is even worse.
So, I'm building a toolkit that allows to keep things simple for the end user. Run Ollama and Open WebUI configured to work together: `harbor up ollama webui`. Don't like Ollama? Then `harbor up llamacpp webui`. There are 17 backends, 14 frontends and 50+ different satellite projects, config profiles that can be imported from a URL, tunnels, and a helper desktop app.
https://github.com/av/harbor?tab=readme-ov-file#what-can-har...
- Most notable part is seeing it learning more and more with every iteration.
You can watch a live recording here: https://www.youtube.com/watch?v=Of3Nx9b71Mg
- Made a fun little experiment this evening. You can watch a tiny transformer (Karpathy's NanoGPT) model learning features and structure of English language in just over 12 minutes in real time from the tiny_shakespeare dataset.
You can see how it slowly picks up more and more coherence over 5000 training steps.
- So many comments about reasoning here, yet none about the very obvious one, it's not stability of the infrastructure, it's future direction of a product like Claude Code. They need to know how to continue their optimisation machine to fit developers needs the best way possible (for good or for worse).
I guess we should wait for some opt-out telemetry some time soon. It'll be nothing too crazy at first, but we'll see how hungry they are for the data.
- I was, for a long time, scared of my future due to the low/no-code, automation, LLMs, outsourcing, etc. Until, at some point, I realised something simple - the risk factor for my job is not determined by how good new tools are, but only by how lazy people are about learning and adopting them. And here history gives another lesson - we never learn, eternal cycle of mistakes will continue.
- I feel like there's a story behind multimodal architectures not being integrated into the main package by default
- Nice use of Flutter!
- I'm getting disillusioned with the big clouds, all of the reports about account hacking, 20m to find how much you have to pay for a thing in a thing you didn't know that needs to be paid for...
Co-location or a VPS seems like a best choice in all cases except when you're an enterprise.