Preferences

samwho
Joined 1,290 karma
I write interactive visualisations of programming topics at https://samwho.dev

  1. I love this, the end result looks so good.

    Something you don’t really mention in the post is why do this? Do you have an end goal or utility in mind for the book shelf? Is it literally just to track ownership? What do you do with that information?

  2. Thank you <3
  3. Thank you! <3

    These are all built with React and CSS animations (or the Web Animations API where I needed it). I’m not very good at React so the code is a real mess. 2 of the components also use threejs for the 3D bits.

    For the stuff on my personal site, which simonw graciously linked to in another reply, you can see all the code behind my work at https://github.com/samwho/visualisations

  4. Simon, you’re too kind. Thank you. <3
  5. When I was writing this, GPT 5.1 was the latest and it got it right away. It’s the sequence of prime numbers fwiw :)
  6. How would information leak, though? There’s no difference in the probability distribution the model outputs when caching vs not caching.
  7. The only thing that comes to mind is some kind of timing attack. Send loads of requests specific to a company you’re trying to spy on and if it comes back cached you know someone has sent that prompt recently. Expensive attack, though, with a large search space.
  8. I was wondering about this when I was reading around the topic. I can’t personally think of a reason you would need to segregate, though it wouldn’t surprise me if they do for some sort of compliance reasons. I’m not sure though, would love to hear something first-party.
  9. With KV caching as it’s described there it has to be a prefix match. OpenAI state in their docs they don’t cache anything below 1024 tokens long, and I’m sure I read somewhere that they only cache in 1024 token blocks (so 1024, 2048, 3072, etc) but I can’t find it now.

    There’s been some research into how to cache chunks in the middle, but I don’t think any of the providers are doing it yet because it needs the prompt to be structured in a very specific way.

  10. Could you tell me what browser/OS/device you’re using? A few people have said this and I haven’t been able to reproduce it.
  11. Another person had this problem as well and we couldn’t figure out what causes it. We suspect something to do with WebGL support. What browser/device are you using? Does it still break if you disable all extensions? I’d love to fix this.
  12. It’s funny, I didn’t set out for that to be the case. When I pitched the idea internally, I wanted to scratch my own itch (what on earth is a cached token?) and produce a good post. But then I realised I had to go deeper and deeper to get to my answer and accidentally made a very long explainer.
  13. Yay, glad I could help! The sampling process is so interesting on its own that I really want to do a piece on it as well.
  14. Yeah! It’s planned for sure. It won’t be the direct next one, though. I’m taking a detour into another aspect of LLMs first.

    I’m really glad you liked it, and seriously the resources I link at the end are fantastic.

  15. Huh, when I was writing the article it was GPT-5.1 and I remember it got it no problem.
  16. Thank you so much <3
  17. It is the same ngrok!

    The product has grown a lot since the mid 2010s. Still got free localhost tunnelling, but we also have a whole bunch of production-grade API gateway tooling and, as of recently, AI gateway stuff too.

  18. Thank you so much <3

    Yes, I recently wrote https://github.com/samwho/llmwalk and had a similar experience with cache vs no cache. It’s so impactful.

  19. The one where he does a landscape is one of my favourite videos of all time.
  20. Someone gave me the advice to use animals, ideally animals of very different sizes or colours. People instantly picture them and remember them.
  21. This is what I use and I’ve been very happy with it for many years. It hasn’t caused me any trouble and as far as I can tell it hasn’t changed in the whole time I’ve used it.
  22. Hehe, yes, Ned sent me a lovely email the other day about it. Happy to be doing my part.
  23. Can’t imagine what it might be.
  24. I’m not sure that’s quite my position. Happy to cede that I lack fluency, and I appreciate your time and the time others have given to help me understand.
  25. I did push up a rewording that hopefully lands better. Fingers crossed.

This user hasn’t submitted anything.

Keyboard Shortcuts

Story Lists

j
Next story
k
Previous story
Shift+j
Last story
Shift+k
First story
o Enter
Go to story URL
c
Go to comments
u
Go to author

Navigation

Shift+t
Go to top stories
Shift+n
Go to new stories
Shift+b
Go to best stories
Shift+a
Go to Ask HN
Shift+s
Go to Show HN

Miscellaneous

?
Show this modal