The Office of the Australian Information Commissioner (OAIC) has issued two new guides to clarify the application of Australia’s privacy obligations for users of commercially available artificial intelligence (AI) products and developers of generative AI (GenAI) models.
This development signals the importance of complying with the Privacy Act 1988 (Cth) (Privacy Act) when using personal information in AI systems and products, especially given the prevalence of AI, including GenAI. It also reflects the Federal Government’s continued focus on promoting and facilitating compliance with Australia’s privacy laws, especially in relation to AI, as new AI systems and products continue to be used by more and more Australians.
General guidance on AI and the Privacy Act
The Privacy Act applies to all uses of AI involving personal information, including where information is used to develop, test, train or use an AI system. The guidance notes that, where there is doubt about the application of the Privacy Act to AI activities, developers and users should err on the side of caution and assume it applies.
Further, the OAIC recommends that both GenAI developers and users of commercially available AI systems take a ‘privacy by design’ approach. This includes the conduct of privacy impact assessments to understand the impact of use of a particular AI product on the privacy of individuals and the identification of ways to manage, minimise or eliminate those impacts.
Guidance on privacy and the use of commercially available AI products
The guidance for users of AI products is specifically targeted at organisations that deploy AI systems that are built with, collect, store, use or disclose personal information. This is irrespective of whether the product is deployed for internal purposes (e.g. use within an organisation) or external purposes (e.g. use affecting customers).
- Privacy obligations apply to AI inputs and outputs. Privacy obligations apply to personal information input into an AI system and AI-generated data that contains personal information. Personal information includes inferred, false or artificially generated information that may identify an individual. Organisations should be aware that inputting personal information into an AI system may raise privacy risks, especially as the information may be difficult to track, control or remove once it enters the system. Further, if output data (generated by AI) includes personal information, that data is subject to privacy obligations (e.g. Australian Privacy Principle (APP) 10 would apply and require that the information be accurate, up-to-date and complete). See our article on a recent decision of the Victorian Privacy Commissioner on this very issue.
- Privacy policies and notifications should be transparent about the use of AI. Privacy policies and notifications should contain clear and transparent information about an organisation’s use of AI. This includes ensuring that privacy notifications specify any AI-related purposes for which personal information is being collected, and that public facing AI tools (such as chatbots) are clearly identified to users. This point of guidance follows the proposed amendments to the Privacy Act recently introduced into Federal Parliament, which included the recommendation that the use of personal information in an automated decision-making system will require disclosure under an organisation’s privacy policy.
- AI-generated or inferred personal information must comply with collection requirements. Where AI systems are used to generate or infer personal information (including incorrect information such as hallucinations and deepfakes), this will be taken as a collection of personal information, and therefore the requirements under APP 3 will apply. This means organisations must ensure that the collection or generation of personal information by AI is reasonably necessary for its functions or activities, and only done by lawful and fair means.
- Personal information input into an AI system can only be used for the primary purpose for which it was collected. If personal information is input into an AI system, the information can only be used or disclosed for the primary purpose for which it was collected (in accordance with APP 6), unless consent is obtained or if a secondary use that would be reasonably expected by the individual can be established. The guidance clarifies that a secondary use may be within an individual’s reasonable expectations if it was expressly outlined in a notice at the time of collection and in the organisation’s privacy policy.
- Personal information should not be entered into publicly available GenAI tools. Personal information, particularly sensitive information, should not be entered into publicly available GenAI tools, as there are significant and complex privacy risks that arise from doing so.
Key takeaways and next steps
- Check that the personal information can be used in the AI model by reference to the purpose for which the personal information was collected.
- When using sensitive information in AI, clear consent of the individual must be obtained (unless an exception applies).
- Ensure that privacy policies and notices are amended to reflect the use of AI, as appropriate.
- Organisations must take reasonable steps to ensure the accuracy of the personal information it collects, generates, uses and discloses when using an AI product. This includes taking steps to ensure that the data being used to train or develop the AI product is accurate, which may involve an organisation’s ongoing review of its data holdings and data quality.
- Confirm that the use of personal information is reasonably necessary for the functions or activities of the business.
For further details and practical guidance, refer to the OAIC’s checklist of privacy considerations when using commercially available AI products.
Guidance on privacy and developing and training generative AI models
The guidance for developers of generative AI models is targeted at organisations that design, build, train, adapt or combine AI models and applications. Although the guidance references GenAI models, it is useful in the context of any kind of AI model that uses personal information.
- Developers must take reasonable steps to ensure accuracy in GenAI models. Reasonable steps should be commensurate with the level of risk of the model and can include undertaking testing, using high quality data sets, and clearly disclaiming the limitations of the AI model.
- Consider whether publicly available data can legally be used to train GenAI models. A developer should not assume that publicly available data can be used lawfully to train or fine-tune a model and should consider whether any data used or collected for use in an AI model contains personal information. This is a critical issue and has been the subject of numerous OAIC determinations.
- The collection of sensitive information generally requires consent. Particular care should be taken when collecting photographs or recordings of individuals (including artificially generated ones), as they may contain sensitive information. If information constitutes sensitive information, it cannot be scraped from the web or collected from a third-party dataset without obtaining prior consent.
- Consider whether previously collected personal information can be lawfully used in training an AI model. A developer should consider whether it can lawfully use personal information to train an AI model if that is not the primary purpose for which the information was originally collected. It may be lawful to do so if it can be established that training an AI model is a secondary use that would be reasonably expected by the individual. Where such secondary use cannot be established, consent should be sought and / or a meaningful and informed ability to opt-out should be offered.
- Seek explicit consent prior to using personal information to train AI. Notwithstanding evolving community expectations, the OAIC recommends seeking explicit consent, and offering individuals a meaningful and informed ability to opt out, before allowing AI to use personal or sensitive information to train itself.
Key takeaways and next steps
- Developers must consider whether the data they are collecting or intending to use contains personal information. This includes publicly available information, and/or when a developer is using a third-party dataset. Additional steps, such as deleting information, may be necessary to comply with privacy obligations.
- Organisations providing information to developers, and developers reusing personal information they already hold to train generative AI, should consider data minimisation techniques, including using collection criteria and data sanitisation methods.
- When using sensitive information to train AI models, clear consent of the individual must be obtained (unless an exception applies).
- Developers and organisations providing information to developers should ensure their privacy policy and any APP 5 collection notifications are up-to-date and explain the use of data for AI training purposes.
For further details and practical guidance, refer to the OAIC’s checklist of privacy considerations when developing or training generative AI models.
This publication is introductory in nature. Its content is current at the date of publication. It does not constitute legal advice and should not be relied upon as such. You should always obtain legal advice based on your specific circumstances before taking any action relating to matters covered by this publication. Some information may have been obtained from external sources, and we cannot guarantee the accuracy or currency of any such information.