- taormina parentI can’t go a full a full conversation without obviously false claims. They will insist you are correct and that your correction is completely correct despite that also being wrong.
- For a bit of shameless plug, we're actively building a mobile and desktop game that avoids these patterns.
You can learn more about Danger World at https://danger.world
- Taxi-by-app existed pre-Uber, the innovation was making the taxi actual show up. Austin had the apps, and I would order one, and the taxis would get distracted on the way to my house and pick up other fares, so I couldn't go where I wanted. They had every chance to not be outcompeted by Uber, but they couldn't stop being taxis. And, here we are.
- I'm still working on Danger World (https://danger.world), my casual 2D narrative adventure with turn-based RPG elements. Built in Flame, on top of Flutter for iOS, Android, Windows and MacOS.
We're getting close! It's just a matter of polishing and polishing and polishing, but I'm really excited about how close we are to launch.
- The idea has been tried before and it failed because people don’t actually want this product at the scale the inventors thought. Amazon has never stopped doing this. Adding an element of indeterminism to the mix doesn’t make this a better product. Imagine what the LLM is going to hallucinate with your credit card attached.
- I mean, this really isn't a large codebase, this is a small-medium sized codebase as judged by prior jobs/projects. 9000 lines of code?
When I give them the same task I tried to give them the day before, and the output gets noticeably worse than their last model version, is that better? When the day by day performance feels like it's degrading?
They are definitely not as good as I would like them to be but that's to be expected of professionals who beg for money hyping them up.
- Grok hasn't gotten better. OpenAI hasn't gotten better. Claude Code with Opus and Sonnet I swear are getting actively worse. Maybe you only use them for toy projects, but attempting to get them to do real work in my real codebase is an exercise in frustration. Yes, I've done meaningful prompting work, and I've set up all the CLAUDE.md files, and then it proceeds to completely ignores everything I said, all of the context I gave, and just craps out something completely useless. It has accomplished a small amount of meaningful work, exactly enough that I think I'm neutral instead of in the negative in terms of work:time if I have just done it all myself.
I get to tell myself that it's worth it because at least I'm "keeping up with the industry" but I honestly just don't get the hype train one bit. Maybe I'm too senior? Maybe the frameworks I use, despite being completely open source and available as training data for every model on the planet are too esoteric?
And then the top post today on the front page is telling me that my problem is that I'm bothering to supervise and that I should be writing an agent framework so that it can spew out the crap in record time..... But I need to know what is absolute garbage and what needs to be reverted. I will admit that my usual pattern has been to try and prompt it into better test coverage/specific feature additions/etc on the nights and weekends, and then I focus my daytime working hours on reviewing what was produced. About half the time I review it and have to heavily clean it up to make it usable, but more often than not, I revert the whole thing and just start on it myself from scratch. I don't see how this counts as "better".
- There aren’t enough GPUs for average gamers to buy anything vaguely recent and they would love to be able to. Making the best GPUs on the planet is still huge and the market is quite large. Scalping might finally die at this rate, but NVDA wasn’t making any of the scalping money anyway so who cares? Data centers and gamers still need every GPU NVDA can make.
- > LLMs are eliminating the need to have a vast array of positions on payrolls. From copywriters to customer support, and even creative activities such as illustration and even authoring books, today's LLMs are already more than good enough to justify replacing people with the output of any commercial chatbot service.
I'd love a source to these claims. Many companies are claiming that they are able to layoff folks because of AI, but in fact, AI is just a scapegoat to counteract the reckless overhiring due to free money in the market over the last 5-10 years and investors are demanding to see a real business plan and ROI. "We can eliminate this headcount due to the efficiency of our AI" is just a fancy way to make the stock price go up while cleaning up the useless folks.
People have ideas. There are substantially more ideas than people who can implement ideas. As with most technology, the reasonable expectation is to assume that people are just going to want more done by the now tool powered humans, not less things.
- I've used a wide variety of the "best" models, and I've mostly settled on Opus 4 and Sonnet 4 with Claude Code, but they don't ever actually get better. Grok 3-4 and GPT4 were worse, but like, at a certain point you don't get brownie points for not tripping over how low the bar is set.
- Github Pages STILL don't have any sort of built-in analytics available. I shouldn't need GA or something else to track the basic website metrics when you absolutely know that MS and GH have been tracking these things the whole time. People have had issues up asking for this for literal years.
- Alright, well, Opus 4.1 seems exactly as useless as Opus 4 was, but it's probably eating my tokens faster. Wish they let you tell somehow.
At least Sonnet 4 is still usable, but I'll be honest, it's been producing worse and worse slob all day.
I've basically wasted the morning on Claude Code when I should've just been doing it all myself.
- Gas costs money. The car costs money. You can only do the same hike that's an hour away so many times, before you're traveling to go to new places, and hotels cost money at that point. Pickleball courts cost money. The pickleball equipment costs money. People do go to the library, and then they go home and don't interact with other people.