INSUBCONTINENT EXCLUSIVE:
After months of work, a set of guidelines designed to protect humanity from a range of threats posed by artificial intelligence have been
proposed.
Now, a privacy group wants the U.S
government to adopt them too.
The set of 12 universal guidelines revealed at a meeting in Brussels last week are designed to &inform and
improve the design and use of AI& by maximizing the benefits while reducing the risks
AI has been for years a blanket term for machine-based decision making, but as the technology gets better and is more widely adopted, the
results of AI-based outcomes are having a greater effect on human lives — from gaining credit, employment, and even tocriminal
sentencing.
But often those decisions are made with proprietary and closed-off algorithms, making it near-impossible to know if the
decisions are fair or justified.
These guidelines, according to the Electronic Privacy Information Center (EPIC), are designed to be baked
in to AI to ensure the protection of human rights
That includes a right to know the factors, logic and the techniques used to the outcome of a decision; a fairness obligation that removes
discriminatory decision making; and an obligation to secure systems against cybersecurity threats
The principles also include a prohibition on unitary scoring — to prevent governments from using AI to score their citizens and residents
— a subtle jab at China controversial social credit system.
Now, EPIC wants to bring those principles stateside, where many of the
next-generation AI technologies are under development.
In a letter to the National Science Foundation, EPIC called on the little-known
government agency to adopt the universal guidelines, months after it opened its doors to proposals on a national AI policy.
&By investing in
AI systems that strive to meet the [universal] principles, NSF can promote the development of systems that are accurate, transparent, and
accountable from the outset,& wrote Marc Rotenberg, EPIC president and executive director
&Ethically developed, implemented, and maintained AI systems can and should cost more than systems that are not, and therefore merit
investment and research.&
EPIC said that the 12 principles fit neatly within the seven strategies already set out by the U.S
so far — making the case for their adoption easier.
More than 200 experts and 50 organizations have signed on to the guidelines —
including the Federal of American Scientists and the Government Accountability Project.
With the government request for information now
closed, it likely to be many more weeks — if not months — before the government decides what its next steps will be — if any
It not so much up to the NSF to decide, but likely the White House Office of Science and Technology Policy.
A White House spokesperson did
not respond to a request for comment.
You can read the full set of guidelines below:
Right to Transparency
All individuals have the right to know the basis of an AI decision that concerns them
This includes access to the factors, the logic, and techniques that produced the outcome.
Right to Human Determination
All individuals have the right to a final determination made by a person.
Identification Obligation
The institution responsible for an AI system must be made known to the public.
Fairness Obligation
Institutions must ensure that AI systems do not reflect unfair bias or make impermissible discriminatory decisions.
Assessment and
Accountability Obligation
An AI system should be deployed only after an adequate evaluation of its purpose and objectives, its benefits, as well as its risks
Institutions must be responsible for decisions made by an AI system.
Accuracy, Reliability, and Validity Obligations
Institutions must ensure the accuracy, reliability, and validity of decisions.
Data Quality Obligation
Institutions must establish data provenance, and assure quality and relevance for the data input into algorithms.
Public Safety Obligation
Institutions must assess the public safety risks that arise from the deployment of AI systems that direct or control physical devices, and
implement safety controls.
Cybersecurity Obligation
Institutions must secure AI systems against cybersecurity threats.
Prohibition on Secret Profiling
No institution shall establish or maintain a secret profiling system.
Prohibition on Unitary Scoring
No national government shall establish or maintain a general-purpose score on its citizens or residents.
Termination Obligation
An institution that has established an AI system has an affirmative obligation to terminate the system if human control of the system is no
longer possible.
The future of AI relies on a code of ethics