Generative AI Security Key points
What is Generative AI?
Generative AI, or generative artificial intelligence, is a form of machine learning that is able to produce text, video, images, and other types of content. ChatGPT, DALL-E, and Bard are examples of generative AI applications that produce text or images based on user-given prompts or dialogue.
Security with Generative AI-
Below are some points to be noted for reducing risk associated with security while using generative AI-
1. Data-related risk involves threat actors stealing the data set used to train a generative AI model. Without sufficient encryption and controls around data access, any sensitive information contained in a model’s training data could become visible to attackers who obtain the data set.
2. Security Vulnerabilities in AI tools- Like any other software, generative AI tools themselves can contain vulnerabilities that expose companies to cyberthreats. for example, OpenAI took ChatGPT offline to fix a bug in the chatbot’s open source library that had enabled some users to see chat titles from another active user’s chat history. It was also possible to see the first message of a newly created conversation in someone else’s chat history if both users were active around the same time.
3. Data poisoning and theft — Generative AI tools must be fed with massive amounts of data to work properly. This training data comes from various sources, many of which are publicly available on the internet — and, in some cases, could include an enterprise’s previous interactions with clients.
In a data poisoning attack, threat actors could manipulate the pre-training phase of the AI model’s development. By injecting malicious information into the training data set, adversaries could influence the model’s prediction behavior down the line, leading to false or otherwise harmful responses.
Another data-related risk involves threat actors stealing the data set used to train a generative AI model. Without sufficient encryption and controls around data access, any sensitive information contained in a model’s training data could become visible to attackers who obtain the data set.
4. Breaching compliance obligations — When using AI-powered chatbots in enterprise environments, IT leaders should evaluate the following risks related to violating relevant regulations:
· Incorrect responses. AI-powered tools sometimes give false or superficial answers. Exposing customers to misleading information could give rise to legal liability in addition to negatively affecting the enterprise’s reputation.
· Data leakage. Employees could share sensitive work information, including customers’ PII or protected health information (PHI), during conversations with an AI chatbot. This, in turn, could violate regulatory standards such as GDPR, PCI DSS and HIPAA, risking fines and legal action.
· Bias. AI models’ responses sometimes demonstrate bias on the basis of race, gender or other protected characteristics, which could violate anti-discrimination laws.
- Breaching intellectual property and copyright laws. AI-powered tools are trained on massive amounts of data and are typically unable to accurately provide specific sources for their responses. Some of that training data might include copyrighted materials, such as books, magazines and academic journals. Using AI output based on copyrighted works without citation could subject enterprises to legal fines.
- Laws concerning chatbot use. Many enterprises have begun integrating ChatGPT and other generative AI tools into their applications, with some using AI-powered chatbots to answer their customers’ inquiries immediately. But doing so without informing customers in advance risks penalties under statutes such as California’s bot disclosure law.
- Data privacy. Some enterprises might want to develop their own generative AI models, a process likely to involve collecting large amounts of training data. If threat actors successfully breach enterprise, IT infrastructure and gain unauthorized access to training data, the resulting exposure of sensitive information contained in compromised data sets could violate data privacy laws.
5. Malicious use of deepfakes — Voice and facial recognition are getting used more as an access control security measure. AI is an opportunity for bad actors to create deepfakes that get around that security, as has been reported.
Best practices for security when using generative AI tools in the enterprise-
To address the numerous security risks associated with generative AI, enterprises should keep the following strategies in mind when implementing generative AI tools.
1. Classify, anonymize, and encrypt data before building or integrating generative AI
Enterprises should classify their data before feeding it to chatbots or using it to train generative AI models. Determine which data is acceptable for those use cases, and do not share any other information with AI systems.
Likewise, anonymize sensitive data in training data sets to avoid revealing sensitive information. Encrypt data sets for AI models and all connections to them and protect the organization’s most sensitive data with robust security policies and controls.
2. Train employees on generative AI security risks and create internal usage policies
Employee training is the most critical protective measure to mitigate the risk of generative AI-related cyber-attacks. To implement generative AI responsibly, organizations must educate employees about the risks associated with using this technology.
Organizations can set guidelines for generative AI use at work by developing a security and acceptable use policy. Although specifics will vary from organization to organization, a general best practice is to require human oversight. Don’t automatically trust content generated by AI; humans should review and edit everything AI tools create.
AI use and security policies should also specifically mention what data can be included in queries to chatbots and what is not permitted. For example, developers should never feed intellectual property, copyrighted materials, PII or PHI into AI tools.
3. Vet generative AI tools for security
Conduct security audits and regular penetration testing exercises against generative AI tools to identify security vulnerabilities before deploying them into production.
Security teams can also train AI tools to recognize and withstand attack attempts by feeding them with examples of cyber-attacks. This reduces the likelihood that a hacker will successfully exploit the organization’s AI systems.
4. Govern employees’ access to sensitive work data
Apply the principle of least privilege within enterprise environments, enabling only authorized personnel to access AI training data sets and the underlying IT infrastructure.
Using an identity and access management tool can help centralize and control employees’ access credentials and rights. Likewise, implementing multifactor authentication can help safeguard AI systems and data access.
5. Ensure underlying networks and infrastructure are secure
Deploy AI systems on a dedicated network segment. Using a separate network segment with restricted access to host AI tools enhances both security and availability.
For organizations hosting AI tools in the cloud, select a reputable cloud provider that implements strict security controls and has valid compliance certifications. Ensure all connections to and from cloud infrastructure are encrypted.
6. Keep an eye on compliance requirements, including regularly auditing vendors
Compliance regulations are constantly evolving, and with the uptick in enterprise AI adoption, organizations will likely see more compliance requirements related to generative AI technology.
Enterprises should closely monitor compliance regulations affecting their industry for any changes related to the use of AI systems. As part of this process, when using AI tools from a third-party vendor, regularly review the vendor’s security controls and vulnerability assessments. This helps ensure any security weaknesses in the vendor’s systems do not traverse into the enterprise’s IT environment.