The Top Five Questions Legal Should Ask IT During Copilot for Microsoft 365 Adoption

Epiq
Contact

Epiq

It’s no longer hype. Organizations are broadly looking to adopt Gen AI to drive productivity across their workforce.  

There’s also end-user demand to satisfy. People use Gen AI in their personal lives and want to use it at work. Workers are evaluating employment options based on the prospective company's available Gen AI tools. From the top down, organizations must ensure they’re keeping up with the competition for both talent and productivity.  

And it’s proving to be effective. A recent Forrester study projected increased productivity, cost reductions, reduced onboarding time, and an impressive ROI estimated at up to 450%. 

All of this is driving mass adoption and in ways that have never been seen before.
 
Large, conservative enterprise organizations, who have historically been slow to adopt technology, are now purchasing tens of thousands of M365 Copilot licenses and pushing adoption at a rapid pace.   

Amidst all these exciting developments, the key question on the mind of every General Counsel or Chief Legal Officer is: How do we implement Copilot for Microsoft 365 responsibly?  

Here are five questions legal can ask IT to help ensure Copilot for Microsoft 365 adoption addresses legal considerations. The good news is addressing these considerations will also help drive greater Copilot for Microsoft 365 efficiency and better results! 

Question #1: How can we ensure Copilot for Microsoft 365 doesn’t expose sensitive information to employees who shouldn’t have access? 

An end-user will typically have access to Outlook, OneDrive, Teams, and other applications, but they likely have access to information they aren’t aware of, such as files on certain SharePoint or OneDrive sites. Suppose an employee has shared a file using the selection ‘share with anyone in my organization.’ This was not previously a significant risk when ‘anyone in my organization’ would also need to know the link to access the file. However, when an end-user sends a prompt to Copilot for Microsoft 365, it’s grounded in utilizing all files the end-user can access. Therefore, Copilot’s response could contain information the end-user wasn’t intended to have access to or they don’t know they have access to.  

To address this concern, organizations can leverage M365 Purview’s data security and compliance capabilities to identify and label sensitive content, including data subject to HIPAA, PCI, Privacy regulations, and sensitive business materials defined by the organization, e.g., a confidential M&A target, intellectual property, etc.  To speed up Copilot implementation, organizations can protect data overexposed to Copilot in stages based on risk categories. For example, first identifying and remediating overexposed data that contains sensitive content allows end-users to get access to Copilot quickly. Later, the team can address permission hygiene on overexposed data that does not have sensitive content.   

Question #2: How can we avoid having Copilot for Microsoft 365 use outdated or inaccurate data?  

As Copilot for Microsoft 365 grounds on accessible data, that data could be outdated, or it could contain factually incorrect information. Many organizations retain broadly due to a “keep everything” data culture or have not implemented automated disposition practices aligned with corporate retention policies.  

Copilot for Microsoft 365 will be a better tool for its end-users if it uses accurate and precise data. When an organization is rolling out Copilot for Microsoft 365, it is a great time to revisit data retention and disposition policies. Good data retention and disposition practices reduce data volumes based on timeframe and content duplicity, helping to avoid poor Copilot for Microsoft 365 responses. These same retention policies will also reduce data subject to future legal holds, lowering litigation costs and making it a win-win for both legal and IT.   

Question #3: How can we prevent data loss or leakage? 

By taking the steps above to define sensitive content and implement automatic data classification, organizations can take advantage of their good data hygiene and apply proactive controls to reduce the risk of data loss or leakage. With Microsoft Purview Data Loss Prevent (DLP), organizations can block users from sharing sensitive information to a personal email account or other unauthorized destinations. Proactive controls can warn end-users, block the activity, log the activity, and direct a centralized person to act on it. It can also identify any responses from Copilot that contain information from an already classified file, and the new file will inherit the classification and associated controls. 

Question #4: How do you monitor and control the prompts and responses in Copilot for Microsoft 365? 

Out of the box, Gen AI technology is smart. It won’t let end-users ask foolish questions.  

In addition, proper implementation of a classification schema and DLP policy controls will protect prompts and responses at an individual level.
 
If concerns remain that end-users will ask for sensitive or confidential information, then additional layers of protection can be applied. This includes limiting what end-users can ask in a prompt or preventing them from asking sensitive questions to an organization.  Responses will also be evaluated for compliance with organizational policies to prevent sensitive information from being surfaced to an end-user who shouldn’t have
access to that content. 

Question #5: If Copilot creates new documents, how do we align our data retention policies with our eDiscovery practices? 

Today, the retention of Copilot prompts is tied to Microsoft Teams. If an end-user asks Copilot for Microsoft 365 to summarize a Word document, the end-user’s mailbox retains the prompt and response.  

There is industry debate around whether this should be considered another data source, like a chat, but technically it isn’t. It’s transient content or can be thought of as a convenience copy, and it doesn’t need to be kept for records retention purposes. Still, an organization may want to retain Copilot for Microsoft 365-generated data for employment litigation scenarios, forensic investigations, etc. Organizations should consider and put policies in place regarding this type of data.  

Bonus question that everyone should be asking: What is a realistic timeframe for responsibly adopting Gen AI?  

One of the most important and time-consuming aspects of preparing for Gen AI adoption is ensuring data is controlled and secure. On average, an enterprise-level organization’s full deployment will take six months. One way to speed up time to value is to utilize an agile approach and roll out Gen AI in stages. In this way, Gen AI tools  can get into in the hands of the most critical end-users, and learnings can be applied  from these early adopters when rolling out the program to other groups within an organization.  

This approach can accelerate adoption and is more effective than trying to fix problems and implement controls after your broader teams start adopting. While an agile approach is not new, layering analysis of sensitive information and access controls allows an organization to reduce risk associated with the deployment of Gen AI. 

[View source.]

Written by:

Epiq
Contact
more
less

PUBLISH YOUR CONTENT ON JD SUPRA NOW

  • Increased visibility
  • Actionable analytics
  • Ongoing guidance

Epiq on:

Reporters on Deadline

"My best business intelligence, in one easy email…"

Your first step to building a free, personalized, morning email brief covering pertinent authors and topics on JD Supra:
*By using the service, you signify your acceptance of JD Supra's Privacy Policy.
Custom Email Digest
- hide
- hide