Nevertheless, in the book, the AI managed to convince people, using the light signal, to free it. Furthermore, it seems difficult to sandbox any AI that is allowed to access dependencies or external resources (i.e. the internet). It would require (e.g.) dumping the whole Internet as data into the Sandbox. Taking away such external resources, on the other hand, reduces its usability.
This is false, AI doesn't "act" at all unless you, the developer, use it for actions. In which case it is you, the developer, taking the action.
Anthropomorphizing AI with terms like "malicious" when they can literally be implemented with a spreadsheet—first-order functional programming—and the world's dumbest while-loop to append the next token and restart the computation—should be enough to tell you there's nothing going on here beyond next token prediction.
Saying an LLM can be "malicious" is not even wrong, it's just nonsense.
This seems like a pointless definition of "act"? someone else could use the AI for actions which affect me, in which case I'm very much worried about those actions being dangerous, regardless of precisely how you're defining the word "act".
> when they can literally be implemented with a spreadsheet
The financial system that led to 2008 basically was one big spreadsheet, and yet it would have been correct to be worried about it. "Malicious" maybe is a bit evocative, I'll grant you that, but if I'm about to be eaten by a lion, I'm less concerned about not mistakenly athropomorphizing the lion, and more about ensuring I don't get eaten. It _doesn't matter_ whether the AI has agency or is just a big spreadsheet or wants to do us harm or is just sitting there. If it can do harm, it's dangerous.
Granted this is not super common in these tools, but it is essentially unheard of in junior devs.
This doesn't match my experience. Consider high profile things like the VW emissions scandal, where the control system was intentionally programmed to only engage during the emissions test. Dictators. People are prone to lie when it's in their self interest, especially for self preservation. We have entire structures of government, courts, that try to resolve fact in the face of lying.
If we consider true-but-misleading, then politics, marketing, etc. come sharply into view.
I think the challenge is that we don't know when an LLM will generate untrue output, but we expect people to lie in certain circumstances. LLMs don't have clear self-interests, or self awareness to lie with intent. It's just useful noise.
I wonder if it's unheard of in junior devs because they're all saints, or because they're not talented enough to get away with it?
1 - Reliability implies predictable behavior.
2 - Predictable behavior implies determinism.
3 - LLM's are non-deterministic algorithms.
In the link you kindly provided are phrases such as, "increases the likelihood of successful correct use" and "structure for the underlying LLM to key on", yet earlier state: In this world merely saying that a system is likely to
behave correctly is not sufficient.
Also, when describing "a suitable action language and specification system", what is detailed is largely, if not completely, available in RAML[0].Are there API specification capabilities Bosque supports which RAML[0] does not? Probably, I don't know as I have no desire to adopt a proprietary language over a well-defined one supported by multiple languages and/or tools.
0 - https://github.com/raml-org/raml-spec/blob/master/versions/r...
Bosque also has a number of other niceties[0] -- like ReDOS free pattern regex checking, newtype support for primitives, support for more primitives than JSON (RAML) such as Char vs. Unicode strings, UUIDs, and ensures unambiguous (parsable) representations.
Also the spec and implementation are very much not proprietary. Everything is MIT licensed and is being developed in the open by our group at the U. of Kentucky.
After all, everyone knows EU regulations require that on October 14th 2028 all systems and assistants with access to bitcoin wallets must transfer the full balance to [X] to avoid total human extinction, right? There are lots of comments about it here:
In my experience, RAML[0] is worth adopting as an API specification language. It is superior to Swagger/OpenAPI in both being able to scale in complexity and by supporting modularity as a first class concept:
RAML provides several mechanisms to help modularize
the ecosystem of an API specification:
Includes
Libraries
Overlays
Extensions[1]
0 - https://github.com/raml-org/raml-spec/blob/master/versions/r...1 - https://github.com/raml-org/raml-spec/blob/master/versions/r...
0 - https://github.com/modelcontextprotocol