Ethical AI
It goes without saying that developing moral machines is a pressing task for AI research. With the explosion of possible contexts of robot-human interactions, we have every reason to make sure machines shall at least work to minimize moral risks to humans, if not think morally. Failure to produce sufficiently moral behaviour in machines can be morally disastrous for humanity because, unlike moral failures of individual human beings, machine failures are likely to be more systematic (when machines with faulty programming are deployed widely) or far-reaching (when machines are tasked with important decisions, e.g. in autonomous weaponry, or social decision-making). Yet, despite the pressing need for such a task, the challenges are equally arduous. Many authors have suggested the imperfectability of the project of moral AIs. This project aims at developing a better moral system for general AIs, although it cannot perfectly solve all the challenges, but is better than the current proposals.
Trustworthy AI
In the literature on trustworthy AI, there are two main positions. The first view is in order to illuminating trustworthy AI, we need a general account of trust and trustworthiness, which is less anthropocentric. The second prevailing position contends that AI can only be considered reliable, as it lacks autonomy, agency, and the capacity for goodwill. The primary distinction between the two main positions in the existing literature lies in their stance on whether the anthropocentric aspect, such as goodwill, should be compromised in trust relationships when applied to AI, or if we should maintain interpersonal trust while acknowledging that current AI lacks trustworthiness. Each position carries its own limitations. In this project, I develop and advocate for an alternative approach called the network account of trustworthy AI (hereafter NTAI), which aims to overcome the drawbacks inherent in both positions. According to NTAI, a trustworthy AI requires both reliable AI technologies that consistently fulfil commitments, trustworthy human agents, including AI designers, and AI companies, who consistently fulfil commitments out of goodwill; and well-established social systems regarding the AI usage.
Publication
2022 A pluralist hybrid model AIs, AI & Society: Knowledge, Culture and Communication
2023. “Network of AI and trustworthy : response to Simion and Kelps account of trustworthy AI”, Asian Journal of Philosophy, 2(58).
Accepted
"Towards a skilful-expert model for virtuous machine”, American Philosophical Quarterly