Emiledel
Joined 6 karma
- Emiledel parenthttps://github.com/its-emile/memory-safe-agent
- I've shared a repo here with deterministic, policy driven routing of user inputs so as to operate with it without influencing agent decisions (though it's up to tool calls to take precautions with what they return) https://github.com/its-emile/memory-safe-agent The teams at owasp are great, join us !
- This is a valuable thread to pull thank you: how about differentiating targeted, sticky advertising vs context based advertising. On my local paper website, I value being shown local stuff (to your point on small business). On security websites, show me security products. On HN, Show me tech and science.
The ADD incidence rate being 10x for adults since 2005 (not to even mention kids), we'd all appreciate relevance to what we're exploring/thinking about/learning, rather than the genuine nuisance of nagging for something out of context because we're tracked all around the web.
- I think your idea has already worked for some companies to filter out AI applications, why not try? Especially in a font color identical to the background. You can also scaffold your way to generate questions that get the worst LLM performance, while still being very clear to understand, one side validating the clarity and theoretical tractability for the age, and one side actually solving it. Actor and two critics maybe. I have a container somewhere to create and use this kind of chain visually, could put it on GitHub but I'm sure there are dozens already
- I'm hiring, and discussions of how we want to respond to engineer candidates who get stuck are interesting. I'm personally more interested in their collaboration (wildcard) than their chat-fu (assumed at this point). So my advice to people reading this with interviews in the next year (or next week) is to consider getting off the screen and solving something with a person. We will all get plenty of self-solving time, but it helps if you can show that you can explain yourself during rapid fire situations involving others, or to bring them along with your plan, or building an unfamiliar plan B with others when two AZ are down in us-east-1 and noone planned for XYZ to be unavailable (eg something that the LLM site depended on) Not that I'm certain it'll happen, but I think calculators (to go back to this story) were more reliable than anything we've typed into the past month, and for me that includes their batteries.
- functions work fairly well for that https://platform.openai.com/docs/guides/function-calling
- I'm excited to share that this is already supported, and I highly recommend leveraging it for safer application deployments. https://platform.openai.com/docs/guides/function-calling
- What FLI are asking for seems critical - noone should take satisfaction with the disclaimers ARC had to give relative to emergent behavior just with gpt4. Kudos for openai getting their eyes on it, but that work seems unfinished in the paper, or at least the runtime powers that are now released around have to get more complete red team efforts documented if we are to trust its isolation/containment and alignment.