Trustworthy AI:

A Computational Perspective

Introduction

The past few decades have witnessed the rise of artificial intelligence (AI) technology. However, recent studies show evidence that AI algorithms may not be trustworthy. For example, they could be vulnerable to slight perturbations of input data; they could undermine fairness by showing bias and stereotypes towards certain groups of people; and their decisions could be hard to explain due to their opaque model architectures. With the widespread use of AI applications in our daily life, whether an AI algorithm is trustworthy or not has become a problem of great concern to researchers, developers and users. Recently, a great amount of research on trustworthy AI has emerged. In this tutorial, we aim to provide a comprehensive overview of the cutting-edge research progress on trustworthy AI from a computational perspective. Specifically, we focus on the six most important dimensions in realizing trustworthy AI: (i) Privacy, (ii) Safety & Robustness, (iii) Explainability (iv) Non-discrimination & Fairness, (v) Environmental Well-Being and (vi) Accountability & Auditability. We will introduce the latest technologies and real-world applications in each dimension according to a taxonomy, and discuss the accordant and conflicting interactions among various dimensions. Besides, we will discuss potential future research directions in this field. We expect that researchers and practitioners can gain a broad overview and a deep insight of trustworthy AI from this tutorial, so as to advance the progress of this field.

Schedule

Time: 14:00-17:15 CET, Monday, April 25th

  • Introduction and Concepts

  • Dimension I: Privacy

  • Dimension II: Safety & Robustness

  • Dimension III: Explainability

  • Dimension IV: Non-discrimination & Fairness

  • Dimension V: Environmental Well- being

  • Dimension VI: Accountability & Auditability

  • Dimension Interactions and Future Directions

Outline

  • Introduction and Concepts

In this part, we will first briefly introduce the background knowledge of the AI technology, including its achievements and current development status. Then we introduce the concept of trustworthy AI, and clarify the motivation for studying trustworthy AI by showing real-world examples of the untrustworthy sides of AI systems. Next, we generally describe the six important and concerning dimensions for trustworthy AI with specific examples. Finally, we further articulate the definition of trustworthy AI by comparing it with similar concepts such as Ethical AI, etc.


  • Privacy

The success of modern AI systems is built upon data which contains a large amount of private and sensitive information, such as credit card data and medical records. To establish trustworthy AI systems, we must guarantee the safety of private and sensitive information carried by the data and models which could be potentially exposed throughout the AI systems. In this part, we will introduce mainstream privacy-preserving AI technologies such as confidential computing, federated learning and differential privacy, followed by discussions on applications in real systems.


  • Safety & Robustness

Safety & Robustness describes the ability of an AI system to be robust to noisy perturbations of the inputs and to be able to make secure decisions. Recent studies show that AI systems can be very sensitive to perturbations of the inputs, which makes it dangerous to apply them in safety-critical scenarios. For example, autonomous vehicles can be fooled by altered road signs. In this part, we will introduce recent research on the safety and robustness of AI, including the advanced model attacking and defense strategies and real-world applications.


  • Explainability

A trustworthy AI system requires explainability, which means that the decision process of an AI system should be able to be explained to stakeholders. For example, in a life-critical scenario such as AI-based disease diagnosis, a black-box decision is not acceptable. The inference mechanism should be transparent to doctors and patients to ensure that the diagnosis is accurate. In this part, we will introduce recent studies on the explainability of AI with real-world applications.


  • Non-discrimination & Fairness

It has been shown that AI algorithms can make biased decisions that are skewed towards a particular individual or a group. For example, some face recognition algorithms have difficulties in detecting faces of African Americans; dialogue models can be biased towards females and African Americans by generating more offensive and negative responses for these groups. In this part, we will introduce recent studies on the fairness of AI, including a taxonomy of bias and fairness in AI, advanced debiasing methods and their real-world applications.


  • Environmental Well-Being

A trustworthy AI system should be sustainable and environmentally friendly. However, recent studies show that training a large-scale AI model can cost a great amount of a carbon emission, which poses a threat to the ecological environment. In this part, we will introduce recent research on the environmental well-being of AI with real-world applications.


  • Accountability & Auditability

Auditability & Accountability requires an AI system to be assessed by a third party, and hold someone responsible for an AI failure, especially in critical applications. In this part, we will introduce the latest auditability & accountability strategies with real-world applications.


  • Dimension Interactions & Future Directions

It has been shown that there are accordance and conflict among different dimensions of trustworthy AI. For example, the robustness and interpretability of deep neural networks are tightly connected and robust models tend to be more interpretable. In this part, we will introduce the recent studies on the interactions among various dimensions of trustworthy AI, and conclude the tutorial with potential future research directions.

Survey

A literature review paper of the same name is recently released. Please check it out via the following link:

https://arxiv.org/abs/2107.06641

Presenters

  • Haochen Liu

Haochen Liu is a Ph.D. student in the Department of Computer Science and Engineering at Michigan State University (MSU). His research interests include fairness in machine learning, natural language processing, AI in education. He has published his research in various highly ranked conference proceedings, such as ACL, EMNLP, COLING, WWW, SIGIR, IJCAI, KDD, etc. More detailed information can be found at http://www.cse.msu.edu/~liuhaoc1/.

  • Yiqi Wang

Yiqi Wang is a Ph.D. student in the Department of Computer Science and Engineering at MSU. Her research interests include graph neural networks and their applications. She has published research papers in top-tier conferences, such as AAAI, ICML, WSDM, WWW, etc. More detailed information can be found at https://www.cse.msu.edu/~wangy206/.

  • Wenqi Fan

Dr. Wenqi Fan is a research assistant professor of the Department of Computing at The Hong Kong Polytechnic University (PolyU). He received his Ph.D. degree from City University of Hong Kong (CityU) in 2020. From 2018 to 2020, he was a visiting research scholar at Michigan State University, under the supervision of Dr. Jiliang Tang. His research interests are in the broad areas of machine learning and data mining, with a particular focus on Explainability in Graph Neural Networks and Adversarial Attacks. He has published innovative papers in top-tier journals and conferences such as IEEE TKDE, KDD, WWW, ICDE, IJCAI, AAAI, RecSys, WSDM, and SDM. More detailed information can be found at https://wenqifan03.github.io.

  • Xiaorui Liu

Xiaorui Liu is a Ph.D. student in the Department of Computer Science and Engineering at MSU. His research mainly focuses on distributed optimization, robust machine learning, and machine learning on graphs. He has published high-quality papers in top-tier conferences such as ICLR, ICML, KDD, AISTATS, SDM, WSDM and ICHI. More detailed information can be found at http://cse.msu.edu/~xiaorui/.

  • Jamell Dacon

Jamell Dacon is a Ph.D. student in the Computer Science and Engineering (CSE) department at Michigan State University (MSU), advised by Dr. Jiliang Tang. His research interests include fairness in machine learning, natural language processing, and text mining. He currently focuses on conceptualizing social harms arising from current corpus linguistic research and the advancements of language technologies highlighting both positive and negative social impacts. He has published his research in highly ranked journals and conference proceedings on social network analysis, computational linguistics and machine learning such as WWW, COLING, IC2S2, SNAM, etc. He has received several fellowships and scholarships from MSU, Google Inc, NASA, and the National Science Foundation (NSF). More information about him can be found at https://cse.msu.edu/~daconjam/.

  • Lingjuan Lyu

Dr. Lingjuan Lyu is currently a senior research scientist and team leader in Sony AI. She was an expert researcher in Ant Group, research fellow at National University of Singapore, and research fellow (Level B3, same level as lecturer/assistant professor) at Australian National University. She received her Ph.D. degree from the University of Melbourne in 2018. She was a winner of the IBM Ph.D. Fellowship Worldwide and contributed to various professional activities. Her current research interests span distributed/federated learning, privacy, robustness, fairness, and edge intelligence. She has published high-quality papers in top-tier conferences and journals such as NeurIPS, ICLR, IJCAI, SIGIR, WWW, CIKM, EMNLP, NAACL, JSAC, JIOT, TII, TPDS, TDSC, etc.

  • Jiliang Tang

Dr. Jiliang Tang is an assistant professor in the Department of Computer Science and Engineering at MSU since Fall 2016. His research interests include social computing, data mining and machine learning, their applications in social media and education, and AI ethics. He has published his research in highly ranked journals and top conference proceedings, which received more than 15,000 citations with h-index 61 and extensive media coverage. He serves as top data science conference organizers (e.g., KDD, WSDM, and SDM) and journal editors (e.g., TKDD). He also serves as the tutor of tutorials in top-tier conferences (e.g., KDD 2021/2020/2017/2014, AAAI 2021/2020, WWW 2014, Recsys 2014). He was the recipient of the 2020 SIGKDD Rising Star Award, 2020 Distinguished Withrow Research Award, Aminer Influential Scholars in AI (2019 and 2020), 2019 NSF Career Award, 2019 IJCAI Early Career Talk Award, the 2015 KDD Best Dissertation runner up and 6 best paper awards (or runner-ups) including WSDM2018 and KDD2016. More detailed information can be found at http://www.cse.msu.edu/~tangjili/.

Slides

WWW tutorial -- Trustworthy AI.pptx