New report to help businesses implement responsible AI
Practices such as impact assessments, data curation, fairness measures, pilot studies and organizational training are some of the simple but effective approaches described in a new report that can help Australian businesses develop robust and responsible artificial intelligence (AI).
Released today, the new AI report for Australia's National AI Center, coordinated by CSIRO, explores practical steps to implement the Australian Government's eight AI ethics principles.
The report, "Implementing Australia's AI Ethics Principles: A selection of Responsible AI practices and resources," was developed by Gradient Institute.
It comes as the recent Australian Responsible AI Index found that despite 82% of businesses believing they were practicing AI responsibly, less than 24% had actual measures in place to ensure they were aligned with responsible AI practices. National AI Center Director, Stela Solar said that although businesses acknowledge the commercial opportunities of AI, many don't know how to responsibly navigate the fast-paced environment and meet customer expectations.
"This report is designed to help businesses build responsible AI approaches in their organization," said Solar.
"We hear from businesses, that their ability to innovate with AI is directly correlated with their ability to earn trust from the communities they serve. This is also reflected in Australia's AI Ecosystem Momentum report which found that trust, privacy, security, data quality and skills were the top blockers to AI projects.
"AI systems that are developed without appropriate checks and balances can have unintended consequences that can significantly damage company reputation and customer loyalty."
Bill Simpson-Young, CEO of Gradient Institute said he hoped the report would encourage more businesses to start the journey towards responsible AI practices.
"Even though Responsible AI practices, resources and standards will keep evolving at a fast pace, this should not distract organizations from implementing practices that are known to be effective today," Mr. Simpson-Young said.
"For example, when an AI system is engaging with people, informing users of an AI's operation builds trust and empowers them to make informed decisions. Transparency for impacted individuals could be as simple as informing the user when they are interacting with an AI system.
"While it is broadly accepted that fairness is important, what constitutes fair outcomes or fair treatment is open to interpretation and highly contextual. What constitutes a fair outcome can depend on the harms and benefits of the system and how impactful they are.
"It is the role of the system owner to consult relevant affected parties, domain and legal experts and system stakeholders to determine how to contextualize fairness to their specific AI use‑case. The report helps organizations address these challenges," he said.
The report is the first major publication developed through National AI Center's recently announced Responsible AI Network—a pioneering cross-ecosystem collaboration aimed at uplifting responsible AI practice and sovereign capability across Australia.
It is being released under a creative commons license by the National AI Center and Gradient Institute to invite organizations to actively share their experiences and develop practices of implementing AI responsibly.
More information: Implementing Australia's AI Ethics Principles report: www.csiro.au/en/work-with-us/i … cs-Principles-report