Aisha Naseer is director of research at Huawei Technologies (UK), providing guidance around the standardization of technology development and research within the realms of AI ethics, trustworthiness, and data governance. She also contributes to strategic planning and direction for Huawei’s AI corporate strategy in the U.K. and the European Union. Previously, Naseer led Fujitsu Research of Europe’s AI ethics research program. She is a founding editorial board member of the journal AI and Ethics and was recognized as one of the top 100 Brilliant Women in AI Ethics in 2022.
|RAI constrains AI-related innovation. Disagree||
“With responsible AI, the potential of innovative (autonomous) AI systems gets raised rather than being restricted or constrained. As a phenomenon, most innovations are realized for the social good, which by its intended purpose and nature is not negligent toward the societal obligations it is designed to accomplish. Correspondingly, the AI-related innovations hold the potential to sustain social responsibilities by embedding the “trustworthiness characteristics” within these AI systems.
An interesting perspective is around various trade-offs under the context of responsible AI, such as the trade-off between fairness and fidelity, or between accuracy and interpretability. These trade-offs may incur arguments that favor encumbrance of RAI on AI-related innovations; for example, the more accurate an AI model, the less interpretable it might be from a human understanding point of view. However, this does not hamper the ways in which innovative AI models are diversified and trained, encompassing the core principle of ethical and trustworthy AI.”
|Organizations should tie their responsible AI efforts to their corporate social responsibility efforts. Strongly agree||“From an organizational perspective, the practical implementation of the responsible AI initiative needs to be closely aligned to corporate social responsibility efforts. It involves having the sense of responsibility embedded at the core of organizational values and at the heart of business intonations. Moreover, responsible AI needs to be reflected in the organization’s social and business practices that define the set of rules, norms, and its responsibilities accordingly. In doing so, it is imperative to ensure that suitable governance mechanisms are installed within the organization’s operational supply chains, turning them into value chains that steer responsible business outcomes.”|
|Responsible AI should be a part of the top management agenda. Strongly agree||
“Considering the ramifications of AI and its applications that are being proliferated into the public or general communities, it becomes peremptory for organizations to ensure that adequate considerations are made around responsible AI within the realms of their corporate functions. This requires the topic of responsible AI as part of an organizational agenda, not only at the top management level but also at other levels — i.e., encapsulating operational to strategic.
Although the scale and depth to be covered around responsible AI greatly depend on the nature of the organization and/or business, it needs to be on their mandate and adopted like a culture. Such a paradigm shift (or cultural change) could potentially make significant contributions to perseverance and responsible business.”