Proportion of American workers who say that their company or workplace has an employee policy in place govern AI usage.https://www.conference-board.org/press/us-workers-and-generative-ai
In 2023, generative AI went mainstream and for the first time ever, employees at companies around the world were able to put a general-purpose AI platform to work performing real tasks, productively.
This all came with a catch, however—like a human assistant, ChatGPT can only perform a task when it has all of the information needed to do so. And this has meant that as employees ask ChatGPT to write letters, analyze reports, run numbers, summarize filings, and perform other office tasks, it is being provided with data that would normally never be shared with someone that isn't a company employee.
The risks of concern run the gamut from the obvious concern that ChatGPT might "learn" from sensitive company data and recall it later to another user entirely, to the fact that providing much of this data to a third party (which is what ChatGPT is at the end of the day) runs afoul of compliance requirements.
A number of solutions have emerged in the time since ChatGPT's explosion into the mainstream, many of them ad-hoc. At the simplest level, some organizations have simply banned ChatGPT use and blocked ChatGPT on their networks. Others have managed to shoehorn security controls for ChatGPT into existing DLP or network security tools.
More elegant solutions, like Plurilock AI PromptGuard, intelligently analyze the interaction between an employee and ChatGPT and then "clean up" the communication, hiding sensitive data from ChatGPT before ChatGPT is able to see it. As a solution, this is generally superior to network blocks because it enables continued AI productivity and avoids creating incentives for users to try to work around controls.