Engaging With Artificial Intelligence (AI)
Engaging With Artificial Intelligence (AI)
Intelligence (AI)
Engaging with Artificial Intelligence (AI) 2
Contents
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
What is AI?. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Further Reading. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Another type of input manipulation attack is known as ‘adversarial examples’. In the context
of AI, adversarial examples are the crafting of specialised inputs that, when given to an AI,
intentionally cause it to produce incorrect outputs, such as misclassifications. Inputs can be
crafted to pass a confidence test, return an incorrect result or bypass a detection mechanism.
Note that, in adversarial examples, inputs to the AI are manipulated while it is in use, rather
than when it is being trained. For example, consider a music sharing service that requires
user submitted music to pass an AI powered copyright check before it is published. In an
adversarial example attack, a user might slightly speed up a copyright protected song so
that it passes the AI powered copyright check while remaining recognisable to listeners.
3. Generative AI hallucinations
Outputs generated by an AI system may not always be accurate or factually correct.
Generative AI systems are known to hallucinate information that is not factually correct.
Organisational functions that rely on the accuracy of generative AI outputs could be
negatively impacted by hallucinations, unless appropriate mitigations are implemented.
¹ Nasr, M., Carlini, N., Hayase, J., Jagielski, M., Cooper, A.F., Ippolito, D., Choquette-
Choo, C.A., Wallace, E., Tramèr, F. and Lee, K., 2023. Scalable extraction of training
data from (production) language models. arXiv preprint arXiv:2311.17035.
• AI systems are often hosted in the cloud and may send data between different regions. Ensure that
any AI system your organisation uses can meet your data residency or sovereignty obligations.
• If using a third-party AI system, ensure your organisation is aware of how your data will
be managed in the event your commercial agreement with the third-party ends. This
information is typically outlined in the vendor’s privacy policy or terms of service.
The supply chains of AI systems can be complex and, as a result, are likely to carry inherent
risks. Conducting a supply chain evaluation can help you to identify and manage these risks.
If your organisation is involved in training the AI system it uses, consider the supply
chain of foundational training data and fine-tuning data as well, to aid in preventing
data poisoning. The security of the data and model parameters is critical.
Does your organisation understand the limits and constraints of the AI system?
AI systems can be incredibly complex. While it is often not practical, or possible, to understand
the intricacies of how AI systems work, it is still helpful to understand their general limits and
constraints. For example, is the AI system prone to hallucinations? If the system is involved in
data classification, what is its rate of false positives and false negatives? Understanding the
system’s limits and constraints will assist your organisation to account for them in its processes.
Does your organisation have suitably qualified staff to ensure the AI system is set-up,
maintained and used securely?
Ensure that your organisation is adequately resourced to securely set-up, maintain and use
the AI system.
Consider which staff would be interacting with the AI system, what these staff would be required
to know to interact with the system securely and how this knowledge can be developed.
Staff that use the system should be trained on what data can and cannot be input to the system,
for example, personally identifiable information or the organisation’s intellectual property. Staff
should also be trained on the extent to which the system’s outputs can be relied upon and any
organisational processes for output validation.
• Log and monitor outputs from the AI system to detect any change in behaviour
or performance that may indicate a compromise or data drift.
• Log and monitor inputs to the AI system to ensure compliance obligations are met
and to aid investigation and remediation efforts in the event of an incident.
• Log and monitor the network and endpoints that host your AI system
to detect attempts to access, modify or copy system data.
• Log and monitor logins to repositories that hold training data, the AI system’s
development and production environments and backups. Consider how any logging
and monitoring tools your organisation employs may integrate with your AI system.
• Log and monitor for high frequency, repetitive prompts. These can
be a sign of automated prompt injection attacks.
What will your organisation do if something goes wrong with the AI system?
Consider how your organisation may be impacted if an incident or error affects the AI
system so that you can implement proportionate mitigations and contingencies.
If you are using a third-party AI system, familiarise yourself with any up-time or availability
commitments the vendor has made. Ensure that vendor and customer responsibilities
regarding incident management are clearly defined in the service contract.
Ensure that your organisation’s incident response plan accounts for issues arising from,
or to, its AI systems and consider how business continuity can be achieved in the event
of a serious incident. Your incident response plan should also clearly define the roles and
responsibilities that are critical to addressing any incident that affects the AI system.
Copyright.
© Commonwealth of Australia 2023.
With the exception of the Coat of Arms and where otherwise stated, all material
presented in this publication is provided under a Creative Commons Attribution
4.0 International licence (www.creativecommons.org/licenses).
For the avoidance of doubt, this means this licence only applies to material
as set out in this document.
The details of the relevant licence conditions are available on the Creative
Commons website as is the full legal code for the CC BY 4.0 licence
(www.creativecommons.org/licenses).