Transparency Principle

The elements of the Transparency Principle can be found in several modern privacy laws, including the US Privacy Act, the EU Data Protection Directive, the GDPR, and the Council of Europe Convention 108. The aim of this principle is to enable independent accountability for automated decisions, with a primary emphasis on the right of the individual to know the basis of an adverse determination. In practical terms, it may not be possible for an individual to interpret the basis of a particular decision, but this does not obviate the need to ensure that such an explanation is possible.
Principle: Universal Guidelines for AI, Oct, 2018

Published by Center for AI and Digital Policy

Related Principles

3.1 Explainability and verifiability

One of the basic characteristics of human consciousness is that it perceives the environment, seeks answers to questions. i.e. explanations of why and how something is or is not. That trait influenced the evolution of man and the development of science, and therefore artificial intelligence. Man's need to understand and make things clear to him found its foothold in this principle. Clarity in the context of these Guidelines means that all processes: development, testing, commissioning, system monitoring and shutdown must be transparent. The purpose and capabilities of the artificial intelligence system itself must be explainable, especially the decisions (recommendations) which it brings (to the extent that it is expedient) to all who are affected by the System (directly or indirectly). lf certain results of the System's work cannot be explained, it is necessary to mark them as a system with a "black box" model. Verifiability is a complementary element of this principle, which ensures that the System can check in all processes, ie. during the entire life cycle. Verifiability includes the actions and procedures of checking artificial intelligence systems during testing and implementation, as well as checking the short term and long term impact that such a system has on humans.

Published by Republic of Serbia in ETHICAL GUIDELINES FOR THE DEVELOPMENT, APPLICATION AND USE OF RELIABLE AND RESPONSIBLE ARTIFICIAL INTELLIGENCE, Febrary, 2023

3.2 Dignity

It is the duty of all members of society to mutually respect and protect this right as one of the basic and inviolable rights of every human being. Every individual has the right to protect own dignity: violation or non respect of this right is sanctioned by law. Human dignity (further: dignity) should be understood as a starting principle (principle) that focuses on the preservation of human integrity. Based on that premise, the persons to whom these Guidelines refershould at all times, regardless of the stage in which the concrete artificial intelligence solution isIdevelopment, application or use), keep in mind the person and his integrity as a central concept. In thisregard, it is necessary to develop systems that, at every stage, make it imperative to respect theperson's personality, his freedom and autonomy. Respecting human personality means creating a system that will respect the cognitive, social andcultural characteristics of each individual. The artificial inteligence systems that are being developed must be in accordance with the above, therefore it is necessary to take care that they cannot in any waylead to the subordination of man to the functions of the system, as well as endangering his dignity andintegrity. In order to ensure respect for the principle of dignity, artificial intelligence systems must not be such that in the processes of work and application they grossly ignore the autonomy of human choice. The Constitution of the Republic of Serbia emphasizes that dignity "is inviolable and everyone is obliged to respectand protect it., Evervone has the right to free personal development, if it does not violate the rights of others guaranteed by the Constitution. The Convention on Human Rights states the following: "Human dignity (dignity) is not only a basic human right but also the foundation of human rights." Human dignity is inherent in every human being. In the Republic of Serbia, this term is regulated in the following ways: "The dignity of the person (honor, reputation, or piety) of the person to whom the information refers.'it is legally protected.” "Whoever abuses another or treats him in a way that offends a human being. dignity, shall be punished by imprisonment for up to one year. "Work in the public interest is any socially useful work that does not offend human dignity and is not done for the purpose of making a profit." This principle emphasizes that the integrity and dignity of all who may be affected by the Artificial inteligence Systemmust be taken care of at all times, As it is a general concept, to which life, in addition to the law, gives different sides although the essence is the same, it is appropriate to attach to the concept itself: honor, reputation, that is, piety.

Published by Republic of Serbia in ETHICAL GUIDELINES FOR THE DEVELOPMENT, APPLICATION AND USE OF RELIABLE AND RESPONSIBLE ARTIFICIAL INTELLIGENCE, Febrary, 2023

1. Right to Transparency.

All individuals have the right to know the basis of an AI decision that concerns them. This includes access to the factors, the logic, and techniques that produced the outcome. [Explanatory Memorandum] The elements of the Transparency Principle can be found in several modern privacy laws, including the US Privacy Act, the EU Data Protection Directive, the GDPR, and the Council of Europe Convention 108. The aim of this principle is to enable independent accountability for automated decisions, with a primary emphasis on the right of the individual to know the basis of an adverse determination. In practical terms, it may not be possible for an individual to interpret the basis of a particular decision, but this does not obviate the need to ensure that such an explanation is possible.

Published by The Public Voice coalition, established by Electronic Privacy Information Center (EPIC) in Universal Guidelines for Artificial Intelligence, Oct 23, 2018

· Transparency and explainability

37. The transparency and explainability of AI systems are often essential preconditions to ensure the respect, protection and promotion of human rights, fundamental freedoms and ethical principles. Transparency is necessary for relevant national and international liability regimes to work effectively. A lack of transparency could also undermine the possibility of effectively challenging decisions based on outcomes produced by AI systems and may thereby infringe the right to a fair trial and effective remedy, and limits the areas in which these systems can be legally used. 38. While efforts need to be made to increase transparency and explainability of AI systems, including those with extra territorial impact, throughout their life cycle to support democratic governance, the level of transparency and explainability should always be appropriate to the context and impact, as there may be a need to balance between transparency and explainability and other principles such as privacy, safety and security. People should be fully informed when a decision is informed by or is made on the basis of AI algorithms, including when it affects their safety or human rights, and in those circumstances should have the opportunity to request explanatory information from the relevant AI actor or public sector institutions. In addition, individuals should be able to access the reasons for a decision affecting their rights and freedoms, and have the option of making submissions to a designated staff member of the private sector company or public sector institution able to review and correct the decision. AI actors should inform users when a product or service is provided directly or with the assistance of AI systems in a proper and timely manner. 39. From a socio technical lens, greater transparency contributes to more peaceful, just, democratic and inclusive societies. It allows for public scrutiny that can decrease corruption and discrimination, and can also help detect and prevent negative impacts on human rights. Transparency aims at providing appropriate information to the respective addressees to enable their understanding and foster trust. Specific to the AI system, transparency can enable people to understand how each stage of an AI system is put in place, appropriate to the context and sensitivity of the AI system. It may also include insight into factors that affect a specific prediction or decision, and whether or not appropriate assurances (such as safety or fairness measures) are in place. In cases of serious threats of adverse human rights impacts, transparency may also require the sharing of code or datasets. 40. Explainability refers to making intelligible and providing insight into the outcome of AI systems. The explainability of AI systems also refers to the understandability of the input, output and the functioning of each algorithmic building block and how it contributes to the outcome of the systems. Thus, explainability is closely related to transparency, as outcomes and ub processes leading to outcomes should aim to be understandable and traceable, appropriate to the context. AI actors should commit to ensuring that the algorithms developed are explainable. In the case of AI applications that impact the end user in a way that is not temporary, easily reversible or otherwise low risk, it should be ensured that the meaningful explanation is provided with any decision that resulted in the action taken in order for the outcome to be considered transparent. 41. Transparency and explainability relate closely to adequate responsibility and accountability measures, as well as to the trustworthiness of AI systems.

Published by The United Nations Educational, Scientific and Cultural Organization (UNESCO) in The Recommendation on the Ethics of Artificial Intelligence, Nov 24, 2021

1. Demand That AI Systems Are Transparent

A transparent artificial intelligence system is one in which it is possible to discover how, and why, the system made a decision, or in the case of a robot, acted the way it did. In particular: A. We stress that open source code is neither necessary nor sufficient for transparency – clarity cannot be obfuscated by complexity. B. For users, transparency is important because it builds trust in, and understanding of, the system, by providing a simple way for the user to understand what the system is doing and why. C. For validation and certification of an AI system, transparency is important because it exposes the system’s processes for scrutiny. D. If accidents occur, the AI will need to be transparent and accountable to an accident investigator, so the internal process that led to the accident can be understood. E. Workers must have the right to demand transparency in the decisions and outcomes of AI systems as well as the underlying algorithms (see principle 4 below). This includes the right to appeal decisions made by AI algorithms, and having it reviewed by a human being. F. Workers must be consulted on AI systems’ implementation, development and deployment. G. Following an accident, judges, juries, lawyers, and expert witnesses involved in the trial process require transparency and accountability to inform evidence and decision making. The principle of transparency is a prerequisite for ascertaining that the remaining principles are observed. See Principle 2 below for operational solution.

Published by UNI Global Union in Top 10 Principles For Ethical Artificial Intelligence, Dec 11, 2017