Australia Evaluates Mandatory AI Rules in High-Risk Areas
The Australian Government is actively considering the introduction of mandatory regulations for high-risk AI development. This move follows growing public concerns over the safety and ethical implications of rapidly advancing AI technologies, and publishers' demands for fair compensation for premium content used in AI training.
The government's approach is shaped by five principles: using a risk-based framework, avoiding undue burdens, open engagement, consistency with the Bletchley Declaration, and prioritizing people and communities in regulatory development. Concerns addressed include inaccuracies in AI model inputs and outputs, biased training data, lack of transparency, and the potential for discriminatory outputs.
Current initiatives to address AI-related risks include the AI in Government Taskforce, reforms to privacy laws, cyber security considerations, and the development of a regulatory framework for automated vehicles. These efforts are in line with the Australian Government's commitment to safe and responsible AI deployment.
The focus is on high-risk AI applications, such as those in healthcare, employment, and law enforcement. The government proposes a mix of mandatory and voluntary measures to mitigate these risks. Transparency in AI, including labelling AI-generated content, is also a key consideration.
Internationally, Australia's regulatory stance on AI is more aligned with the US and UK, which favor a softer approach, unlike the EU's more stringent AI Act. This balanced approach allows the government to address both known and potential risks of AI technologies, ensuring safety and ethical use.
The Australian Government will continue to work with states and territories to strengthen regulatory frameworks. Possible steps include introducing mandatory safeguards for high-risk AI settings, considering legislative vehicles for these guardrails, and specific obligations for the development of frontier AI models. An interim expert advisory group is also being established to guide the development of AI guardrails.
In summary, while embracing the potential of AI to improve quality of life and economic growth, the Australian Government is taking careful steps to ensure the safe, responsible, and ethical development and deployment of AI technologies, particularly in high-risk scenarios.
Image source: Shutterstock
Read More
US States Challenge CBDC as Legal Tender
Jan 18, 2024 0 Min Read
Ethereum's Layer 2 Debate: Buterin Aligns with Daniel Wang on Validium Classification
Jan 18, 2024 0 Min Read
Chinese Premier Li Qiang Advocates Ethical AI Boundaries at Davos
Jan 18, 2024 0 Min Read
EBA Issues Updated Guidelines for Crypto Asset Service Providers
Jan 18, 2024 0 Min Read
IRS Delays Crypto Reporting Requirements for US Businesses
Jan 18, 2024 0 Min Read