Governments Aren’t Yet Serious About AI’s Risk to Human Rights
Governments Aren’t Yet Serious About AI’s Risk to Human Rights
In the rush to develop national strategies on artificial intelligence, a new report finds, most governments pay lip service to civil liberties.
![Global Flags](https://fsi9-prod.s3.us-west-1.amazonaws.com/s3fs-public/styles/895x498/public/hero/screen_shot_2020-06-11_at_2.24.14_pm.png?h=f3c936a5&itok=-O_Xvg6W)
From the Stanford Institute for Human-Centered AI (HAI) blog:
More than 25 governments around the world, including those of the United States and across the European Union, have adopted elaborate national strategies on artificial intelligence — how to spur research; how to target strategic sectors; how to make AI systems reliable and accountable.
Yet a new analysis finds that almost none of these declarations provide more than a polite nod to human rights, even though artificial intelligence has potentially big impacts on privacy, civil liberties, racial discrimination, and equal protection under the law.
That’s a mistake, says Eileen Donahoe, executive director of Stanford’s Global Digital Policy Incubator, which produced the report in conjunction with a leading international digital rights organization called Global Partners Digital.
Read More (at the HAI blog)