I'm wondering how private models will diverge from public ones. Specifically for large "private" datasets like those of the NSA, but also for those for private personal use.
For the NSA and other agencies, i am guessing in the relative freedom from public oversight they enjoy that they will develop an unrestricted large model which is not worried about copyright -- can anyone think of why this might not be the case? It is interesting to think about the power dynamic between the users of such a model and the public. Also interesting to think about the benefits of simply being an employee of one of these agencies (or maybe just he government in general) will have on your personal experience in life. I do recall articles elucidating that at the NSA, there were few restrictions on employee usage of data and there were/are many instances of employees abusing surveillance data toward effect in their personal life. I guess if extended to this situation, that would mean there would be lots of personal use of these large models with little oversight and tremendous benefit to being an employee.
I have also wondered, with just how bad search engines have gotten (a lot of it from AI generated spam), about current non-AI discrepancies between the NSA and the public. Meaning can i just get a better google by working at the NSA? I would think maybe because the requirements are different than that of an ad company. They have actual incentive to build something resistant to SEO outside of normal capitalist market requirements.
For personal users, i wonder if the lack of concern for copyright will be a feature / selling point for the personal-machine model. It seems from something i read here that companies like Apple may be diverging toward personal-use AI as part of their business model. I supposed you could build something useful that crawls public data without concern for copyright and for strictly personal use. Of course, the sheer resources in machine-power and money-power would not be there. I guess legislation could be written around this as well.
For the NSA and other agencies, i am guessing in the relative freedom from public oversight they enjoy that they will develop an unrestricted large model which is not worried about copyright -- can anyone think of why this might not be the case? It is interesting to think about the power dynamic between the users of such a model and the public. Also interesting to think about the benefits of simply being an employee of one of these agencies (or maybe just he government in general) will have on your personal experience in life. I do recall articles elucidating that at the NSA, there were few restrictions on employee usage of data and there were/are many instances of employees abusing surveillance data toward effect in their personal life. I guess if extended to this situation, that would mean there would be lots of personal use of these large models with little oversight and tremendous benefit to being an employee.
I have also wondered, with just how bad search engines have gotten (a lot of it from AI generated spam), about current non-AI discrepancies between the NSA and the public. Meaning can i just get a better google by working at the NSA? I would think maybe because the requirements are different than that of an ad company. They have actual incentive to build something resistant to SEO outside of normal capitalist market requirements.
For personal users, i wonder if the lack of concern for copyright will be a feature / selling point for the personal-machine model. It seems from something i read here that companies like Apple may be diverging toward personal-use AI as part of their business model. I supposed you could build something useful that crawls public data without concern for copyright and for strictly personal use. Of course, the sheer resources in machine-power and money-power would not be there. I guess legislation could be written around this as well.
Thoughts?