AI in the Workplace: Balancing Potential, Risks, and Regulations for Employers

Nutter McClennen & Fish LLP
Contact

Since ChatGPT was released in late 2022, and the subsequent release of Claude, Bing AI, Llama, and Bard, the use of generative artificial intelligence (“AI”) has dramatically increased. Unsurprisingly, employees now use AI at work to complete work assignments that range from the routine (organizing slide decks) to the complex (preparing memoranda and client deliverables). This is no surprise given AI can increase productivity and cut costs.

Despite its enormous potential, the use of AI at work is not without its risks for employers. First, all AI tools have the possibility to generate false answers, otherwise referred to as “hallucinations.” Hallucinations occur because AI models are trained on data available on the internet; however, if the data is incomplete, biased, or false, the AI model may learn an incorrect pattern and “predict” the incorrect answer. Accordingly, there is no means to definitively guarantee the quality and accuracy of work generated by AI. Second, the legal landscape for the regulation of AI is incredibly fragmented. Over a third of states have enacted comprehensive data privacy laws which include a provision regulating the use of AI in profiling, data privacy, and accountability of the AI models. The state regulation available today is not harmonized and is often at odds with regulation set forth in the European Union, Israel, Brazil, China, and elsewhere.

Given that employees are understandably attracted to the use of AI for efficiency at work, it is almost certain that employees will use AI regardless of whether employers are directing their employees to use such products. Therefore, employers should implement a comprehensive AI policy given that the failure to do so could lead to (i) incorrect work product being distributed to clients and/or (ii) violations of applicable laws and regulations.

An AI policy should address at a minimum whether your company seeks to encourage or limit the use of AI at work. Companies in highly regulated industries (for example, financial services or healthcare) should consider restricting employees’ use of AI to certain limited functions. Similarly, companies that maintain and store client sensitive data (law firms, accounting firms) should be mindful of what data is potentially uploaded to open system AI. If client sensitive data is uploaded to an open system, it is likely that a company and/or its employees have inadvertently breached confidentiality obligations and, further, violated applicable laws and regulations.

Companies with lower-risk profiles (marketing/branding/fashion) may wish to encourage the use of AI to improve efficiency with the strong caveat that such companies should educate all employees on how to effectively use AI and the applicable legal, ethical, and moral considerations.

Given the above, all companies, regardless of industry, should address the following key criteria in a company AI policy:

  1. Use of AI is subject to each employee undergoing AI training. AI training should cover how AI is trained and that hallucinations exist.
  2. All AI-generated work should be treated as a first draft that should always be reviewed by a human before circulated outside of the company.
  3. Employees should be aware of regulations applicable to their industry and understand that their usage must be limited to permitted work.
  4. Employees should be trained on company confidentiality obligations inclusive of obligations relating to client data and company trade secrets. Given that the most popular AI platforms are open system platforms, employees must understand that uploading any company data onto such a platform may put such data into the realm of public consumption. Accordingly, companies should expressly prohibit employees from inputting confidential or sensitive information onto an AI platform. Furthermore, companies should ensure that all employees are using AI with the appropriate settings to mitigate the risk that their interactions are used to improperly train the AI (a capability available with the enterprise version of AI or additional changes to baseline settings).
  5. Companies who need to track the source of data utilized in AI platforms should regulate reliance on open AI given that current AI platforms utilize large data sets with complex algorithms which may not be able to provide concrete and accurate answers as to how certain data is sourced. Therefore, usage of untraceable data may violate applicable data privacy risks and rules.

If a company chooses to allow AI in the workplace, employers would be wise to develop a comprehensive AI data policy with the advice and counsel of legal counsel familiar with such company’s particular regulatory concerns and jurisdictional reach. Nutter attorneys are able to assist employers in navigating these new developments.

DISCLAIMER: Because of the generality of this update, the information provided herein may not be applicable in all situations and should not be acted upon without specific legal advice based on particular situations.

© Nutter McClennen & Fish LLP | Attorney Advertising

Written by:

Nutter McClennen & Fish LLP
Contact
more
less

PUBLISH YOUR CONTENT ON JD SUPRA NOW

  • Increased visibility
  • Actionable analytics
  • Ongoing guidance

Nutter McClennen & Fish LLP on:

Reporters on Deadline

"My best business intelligence, in one easy email…"

Your first step to building a free, personalized, morning email brief covering pertinent authors and topics on JD Supra:
*By using the service, you signify your acceptance of JD Supra's Privacy Policy.
Custom Email Digest
- hide
- hide