When Robots Speak Dialect

On January 29, 2024, the article „‚Ick bin een Berlina‘: dialect proficiency impacts a robot’s trustworthiness and competence evaluation“ was published in Frontiers in Robotics and AI. Authors are Katharina Kühne, Erika Herbold, Oliver Bendel, Yuefang Zhou, and Martin H. Fischer. With the exception of Oliver Bendel – who is a professor at the School of FHNW and an associated researcher in the PECoG group – all of them are members of the University of Potsdam. The paper says about the background: „Robots are increasingly used as interaction partners with humans. Social robots are designed to follow expected behavioral norms when engaging with humans and are available with different voices and even accents. Some studies suggest that people prefer robots to speak in the user’s dialect, while others indicate a preference for different dialects.“ The following results are mentioned: „We found a positive relationship between participants’ self-reported Berlin dialect proficiency and trustworthiness in the dialect-speaking robot. Only when controlled for demographic factors, there was a positive association between participants’ dialect proficiency, dialect performance and their assessment of robot’s competence for the standard German-speaking robot. Participants‘ age, gender, length of residency in Berlin, and device used to respond also influenced assessments. Finally, the robot’s competence positively predicted its trustworthiness.“ The article can be accessed at www.frontiersin.org/articles/10.3389/frobt.2023.1241519/full.

Fig.: A Berlin robot (Image: Ideogram)

Paper on Robotic Hugs

From March 27-29, 2023, the AAAI 2023 Spring Symposia featured the symposium „Socially Responsible AI for Well-being“ by Takashi Kido (Teikyo University, Japan) and Keiki Takadama (The University of Electro-Communications, Japan). This time the venue was exceptionally not Stanford University, but the Hyatt Regency SFO Airport. On March 28, Prof. Dr. Oliver Bendel presented the paper „Increasing Well-being through Robotic Hugs“, written by himself, Andrea Puljic, Robin Heiz, Furkan Tömen, and Ivan De Paola. It has now been published and can be downloaded via ceur-ws.org/Vol-3527/. From the abstract: „This paper addresses the question of how to increase the acceptability of a robot hug and whether such a hug contributes to well-being. It combines the lead author’s own research with pioneering research by Alexis E. Block and Katherine J. Kuchenbecker. First, the basics of this area are laid out with particular attention to the work of the two scientists. The authors then present HUGGIE Project I, which largely consisted of an online survey with nearly 300 participants, followed by HUGGIE Project II, which involved building a hugging robot and testing it on 136 people. At the end, the results are linked to current research by Block and Kuchenbecker, who have equipped their hugging robot with artificial intelligence to better respond to the needs of subjects.“ More information on the conference via aaai.org/conference/spring-symposia/sss23/.

Fig.: Hugging a robot (Image: DALL-E 3)

Artificial Intelligence & Animals

The online event „Artificial Intelligence & Animals“ took take place on 16 September 2023. Speakers were Prof. Dr. Oliver Bendel (FHNW University of Applied Sciences and Arts Northwestern Switzerland), Yip Fai Tse (University Center for Human Values, Center for Information Technology Policy, Princeton University), and Sam Tucker (CEO VegCatalyst, AI-Powered Marketing, Melbourne). Panelists were Ian McDougall (Executive Vice President and General Counsel, LexisNexis London), Jamie McLaughlin (Animal Law Commission Vice President, UIA), and Joan Schaffner (Associate Professor of Law, George Washington University). Oliver Bendel „has been thinking on animal ethics since the 1980s and on information and machine ethics since the 1990s“. „Since 2012, he has been systematically researching machine ethics, combining it with animal ethics and animal welfare. With his changing teams, he develops animal-friendly robots and AI systems.“ (Website Eventbrite) Yip Fai Tse co-wrote the article „AI ethics: the case for including animals“ with Peter Singer. Sam Tucker is an animal rights activist.

Fig.: One topic was facial recognition for bears

Ein neuartiger Chatbot, der auf Constitutional AI beruht

Die Maschinenethik widmet sich maschineller oder künstlicher Moral bzw. moralischen Maschinen. Üblicherweise werden moralische Regeln in Maschinen gepflanzt, an die sich diese strikt halten. Seit einiger Zeit verwendet man auch Machine Learning, etwa im Kontext der Pflege. Die Maschine lernt aus ihren Erfahrungen bzw. aus den Daten und passt ihr Verhalten an. Unter dem Begriff der Constitutional AI erhält diese Richtung nun Auftrieb. Im Vordergrund stehen Sprachmodelle wie GPT-3 und GPT-4, die mit Leitlinien, Richtlinien und Gesetzen trainiert werden. An der Hochschule für Wirtschaft FHNW beginnt am 19. September 2023 das Projekt „CAIBOT: Der Chatbot, der auf Constitutional AI beruht“. Initiiert hat es Prof. Dr. Oliver Bendel, der seit vielen Jahren im Bereich der Maschinenethik forscht und mit seinen wechselnden Teams zahlreiche Prototypen gebaut hat. Für das Projekt konnte Cédric Wespi gewonnen werden. Er wird im Rahmen seiner Abschlussarbeit Grundlagen zu Constitutional AI erarbeiten und den CAIBOT prototypisch implementieren. Die Ergebnisse werden im Januar 2024 präsentiert.

Abb.: Der CAIBOT benutzt Sprachmodelle wie GPT

Call for Proposals for the AAAI Spring Symposium Series

In late August 2023, AAAI announced the continuation of the AAAI Spring Symposium Series, to be held at Stanford University from 25-27 March 2024. Due to staff shortages, the prestigious conference had to be held at the Hyatt Regency SFO Airport in San Francisco in March 2023 – and will now return to its traditional venue. The call for proposals is available on the AAAI Spring Symposium Series page. Proposals are due by 6 October 2023. They should be submitted to the symposium co-chairs, Christopher Geib (SIFT, USA) and Ron Petrick (Heriot-Watt University, UK), via the online submission page. Over the past ten years, the AAAI Spring Symposia have been relevant not only to classical AI, but also to roboethics and machine ethics. Groundbreaking symposia were, for example, „Ethical and Moral Considerations in Non-Human Agents“ in 2016, „AI for Social Good“ in 2017, or „AI and Society: Ethics, Safety and Trustworthiness in Intelligent Agents“ in 2018. More information is available at aaai.org/conference/spring-symposia/sss24/.

Fig.: Hoover Tower on the campus of Stanford University

AI or what the ChatGPT

Am 20. Juni 2023 hält Prof. Dr. Oliver Bendel einen Onlinevortrag für LeLa, das Lernlabor Hochschuldidaktik für Digital Skills, ein Kooperationsprojekt der fünf Zürcher Hochschulen ETH Zürich, PHZH, UZH, ZHAW und ZHdK. „Ethische Implikationen generativer KI“ sind das Thema. Zunächst klärt der Technikphilosoph den Begriff der generativen KI (engl. „generative AI“). Dann stellt er aus Sicht von Informations- und Maschinenethik verschiedene Überlegungen zu diesem Bereich an, wobei er vor allem auf ChatGPT fokussiert. So ist die Frage, woher die Daten für das zugrunde liegende Sprachmodell kommen und unter welchen Bedingungen das Reinforcement Learning from Human Feedback abläuft. Zudem dürfte relevant sein, welche Daten man beim Prompt preisgibt und welche Prompts auf welche Art und Weise zurückgewiesen werden. Ein weiteres Problemfeld ist das Halluzinieren der Sprachmodelle bzw. der darauf basierenden Anwendungen. Diese verdrehen Fakten und erfinden Referenzen. Mit Visual ChatGPT soll man über Texteingaben Bilder generieren und dann einzelne Elemente editieren können. Solche und andere Bildgeneratoren wie DALL-E, Stable Diffusion und Midjourney werfen wiederum zahlreiche ethische Fragen auf, auch mit Blick auf den Kunstbereich. GPT-3 und GPT-4 bzw. ChatGPT sind nicht nur für bzw. als Chatbots und Contentgeneratoren relevant, sondern auch für Industrie- und Serviceroboter. In diesem Bereich hat indes vor allem das Sprachmodell PaLM-E Aufmerksamkeit erregt. Indem Bilddaten und Daten zu Zuständen und Ereignissen integriert werden, werden Virtualität und Realität verbunden. Konkret kann der Benutzer mit Hilfe eines Prompts einem Roboter eine Anweisung geben, die dieser dann in seiner Umgebung ausführt, die er vorher beobachtet hat und weiter beobachtet. Dabei sind wiederum Herausforderungen vorhanden, etwa mit Blick auf Verlässlichkeit und Haftung. Weitere Informationen zur Reihe „AI or what the ChatGPT“ über dizh.ch/event/was-sind-die-ethische-implikationen-generativer-ki/.

Abb.: Ein mit einem Bildgenerator erstelltes Bild

Generative KI im Wirtschaftslexikon

„Generative KI (‚KI‘ steht für ‚künstliche Intelligenz‘) ist ein Sammelbegriff für KI-basierte Systeme, mit denen auf scheinbar professionelle und kreative Weise alle möglichen Ergebnisse produziert werden können, etwa Bilder, Video, Audio, Text, Code, 3D-Modelle und Simulationen. Menschliche Fertigkeiten sollen erreicht oder übertroffen werden. Generative KI kann Schüler, Studenten, Lehrkräfte, Büromitarbeiter, Politiker, Künstler und Wissenschaftler unterstützen und Bestandteil von komplexeren Systemen sein. Man spricht auch, dem englischen Wort folgend, von Generative AI, wobei ‚AI‘ die Abkürzung für ‚Artificial Intelligence‘ ist.“ Mit diesen Worten beginnt ein Beitrag, der am 22. Februar 2023 im Gabler Wirtschaftslexikon erschienen ist. Es werden Informationen zu Entwicklung und Hintergrund sowie einige Beispiele für Anwendungen gegeben. Am Ende wird die Perspektive der Ethik eingenommen, unter besonderer Berücksichtigung der Informationsethik. Der Beitrag von Oliver Bendel kann über wirtschaftslexikon.gabler.de/definition/generative-ki-124952 abgerufen werden.

Abb.: Von Midjourney generiertes Porträt

AAAI Symposium on AI for Well-being

As part of the AAAI 2023 Spring Symposia in San Francisco, the symposium „Socially Responsible AI for Well-being“ is organized by Takashi Kido (Teikyo University, Japan) and Keiki Takadama (The University of Electro-Communications, Japan). The AAAI website states: „For our happiness, AI is not enough to be productive in exponential growth or economic/financial supremacies but should be socially responsible from the viewpoint of fairness, transparency, accountability, reliability, safety, privacy, and security. For example, AI diagnosis system should provide responsible results (e.g., a high-accuracy of diagnostics result with an understandable explanation) but the results should be socially accepted (e.g., data for AI (machine learning) should not be biased (i.e., the amount of data for learning should be equal among races and/or locations). Like this example, a decision of AI affects our well-being, which suggests the importance of discussing ‚What is socially responsible?‘ in several potential situations of well-being in the coming AI age.“ (Website AAAI) According to the organizers, the first perspective is „(Individually) Responsible AI“, which aims to clarify what kinds of mechanisms or issues should be taken into consideration to design Responsible AI for well-being. The second perspective is „Socially Responsible AI“, which aims to clarify what kinds of mechanisms or issues should be taken into consideration to implement social aspects in Responsible AI for well-being. More information via www.aaai.org/Symposia/Spring/sss23.php#ss09.

Fig.: Golden Gate Bridge

AAAI Spring Symposia at Hyatt Regency, SFO Airport

The Association for the Advancement of Artificial Intelligence (AAAI) is pleased to present the AAAI 2023 Spring Symposia, to be held at the Hyatt Regency, San Francisco Airport, California, March 27-29. According to the organizers, Stanford University cannot act as host this time because of insufficient staff. Symposia of particular interest from a philosophical point of view are „AI Climate Tipping-Point Discovery“, „AI Trustworthiness Assessment“, „Computational Approaches to Scientific Discovery“, „Evaluation and Design of Generalist Systems (EDGeS): Challenges and methods for assessing the new generation of AI“, and „Socially Responsible AI for Well-being“. According to AAAI, symposia generally range from 40–75 participants each. „Participation will be open to active participants as well as other interested individuals on a first-come, first-served basis.“ (Website AAAI) Over the past decade, the conference has become one of the most important venues in the world for discussions on robot ethics, machine ethics, and AI ethics. It will be held again at History Corner from 2024. Further information via www.aaai.org/Symposia/Spring/sss23.php.

Fig.: The conference will be held in California

Proceedings von „How Fair is Fair?“ erschienen

Am 17. November 2022 sind die Proceedings von „How Fair is Fair? Achieving Wellbeing AI“ (Organisatoren: Takashi Kido und Keiki Takadama) auf CEUR-WS erschienen. Das AAAI 2022 Spring Symposium fand vom 21. bis 23. März 2022 an der Stanford University statt. Im elektronischen Band finden sich sieben Full Papers mit 6 – 8 Seiten: „Should Social Robots in Retail Manipulate Customers?“ von Oliver Bendel und Liliana Margarida Dos Santos Alves (3. Platz der Best Presentation Awards), „The SPACE THEA Project“ von Martin Spathelf und Oliver Bendel (2. Platz der Best Presentation Awards), „Monitoring and Maintaining Student Online Classroom Participation Using Cobots, Edge Intelligence, Virtual Reality, and Artificial Ethnographies“ von Ana Djuric, Meina Zhu, Weisong Shi, Thomas Palazzolo und Robert G. Reynolds, „AI Agents for Facilitating Social Interactions and Wellbeing“ von Hiro Taiyo Hamada und Ryota Kanai (1. Platz der Best Presentation Awards) , „Sense and Sensitivity: Knowledge Graphs as Training Data for Processing Cognitive Bias, Context and Information Not Uttered in Spoken Interaction“ von Christina Alexandris, „Fairness-aware Naive Bayes Classifier for Data with Multiple Sensitive Features“ von Stelios Boulitsakis-Logothetis und „A Thermal Environment that Promotes Efficient Napping“ von Miki Nakai, Tomoyoshi Ashikaga, Takahiro Ohga und Keiki Takadama. Zudem gibt es mehrere Short Papers und Extended Abstracts. Die Proceedings können über ceur-ws.org/Vol-3276/ abgerufen werden.

Abb.: Der Stanford Bunny

Responsible AI

„HASLER RESPONSIBLE AI“ is a research program of the Hasler Foundation open to research institutions within the higher education sector or non-commercial research institutions outside the higher education sector. The foundation explains the goals of the program in a call for project proposals: „The HASLER RESPONSIBLE AI program will support research projects that investigate machine-learning algorithms and artificial intelligence systems whose results meet requirements on responsibility and trustworthiness. Projects are expected to seriously engage in the application of the new models and methods in scenarios that are relevant to society. In addition, projects should respect the interdisciplinary character of research in the area of RESPONSIBLE AI by involving the necessary expertise.“ (CfPP by Hasler Foundation) Deadline for submission of short proposals is 24 January 2021. More information at haslerstiftung.ch.

Fig.: Responsible AI

Care Robots with Sexual Assistance Functions?

The paper „Care Robots with Sexual Assistance Functions“ by Oliver Bendel, accepted at the AAAI 2020 Spring Symposium „Applied AI in Healthcare: Safety, Community, and the Environment“, can be accessed via arxiv.org/abs/2004.04428. From the abstract: „Residents in retirement and nursing homes have sexual needs just like other people. However, the semi-public situation makes it difficult for them to satisfy these existential concerns. In addition, they may not be able to meet a suitable partner or find it difficult to have a relationship for mental or physical reasons. People who live or are cared for at home can also be affected by this problem. Perhaps they can host someone more easily and discreetly than the residents of a health facility, but some elderly and disabled people may be restricted in some ways. This article examines the opportunities and risks that arise with regard to care robots with sexual assistance functions. First of all, it deals with sexual well-being. Then it presents robotic systems ranging from sex robots to care robots. Finally, the focus is on care robots, with the author exploring technical and design issues. A brief ethical discussion completes the article. The result is that care robots with sexual assistance functions could be an enrichment of the everyday life of people in need of care, but that we also have to consider some technical, design and moral aspects.“ Due to the outbreak of the COVID-19 pandemic, the physical meeting to be held at Stanford University was postponed. It will take place in November 2020 in Washington (AAAI 2020 Fall Symposium Series).

Fig.: Can robots complement or replace sexual relationships?

AI, ME und MC

Immer wieder hört man, oft von Theologen, manchmal von Philosophen, dass Maschinen nicht autonom seien, nicht intelligent, nicht moralisch etc. Sie übertragen den Begriff, den sie aus ihrem Bereich kennen, auf technische Wissenschaften wie Informatik, Künstliche Intelligenz (KI) und Maschinenethik (die technisch geprägt ist und eng mit KI und Robotik zusammenarbeitet). Sie anerkennen nicht, dass jede Disziplin ihre eigenen Begriffe haben kann (und in der Regel hat). Bei einer Tagung im Jahre 2015 beschimpfte Bundestagspräsident Prof. Dr. Norbert Lammert, ein zutiefst religiöser Mann, die Referenten mit den Worten, Maschinen seien nicht autonom, sie hätten sich nämlich nicht selbst ein Gesetz gegeben. Nun sprechen Informatik und Robotik aber nun einmal von autonomen Systemen und Maschinen, und selbstverständlich dürfen sie das, wenn sie darlegen, wie sie das meinen. Eine solche Begriffsklärung und -aneignung steht sogar am Anfang jeder wissenschaftlichen Betätigung, und dass die Begriffe gleich lauten wie die anderer Bereiche, heißt keineswegs, dass sie dasselbe bedeuten und bedeuten müssen. Eine neue Grafik von Prof. Dr. Oliver Bendel, die auf früheren Entwürfen aufbaut, stellt dar, was der Gegenstandsbereich der Disziplinen oder Arbeitsbereiche der KI, der Maschinenethik und des Maschinellen Bewusstseins ist, und macht für sie terminologische Vorschläge. Im Kern geht es diesen darum, etwas in bestimmten Aspekten ab- oder nachzubilden bzw. zu simulieren. So schafft es die Künstliche Intelligenz eben, künstliche Intelligenz hervorzubringen, etwa Dialogsysteme oder Maschinen, die bestimmte Probleme lösen. Ob diese „wirklich“ intelligent sind oder nicht, ist keine sinnvolle Frage, und der Terminus technicus benutzt nicht umsonst das Adjektiv „künstlich“ – hier wäre noch einfacher als im Falle von „autonom“ zu verstehen, dass es sich um eine „neue“ (immerhin seit über 50 Jahren erklärte) Bedeutung handelt.

Abb.: AI, ME und MC

AI’s Next Ethical Challenge

Parallel to his work in machine ethics, Oliver Bendel is trying to establish animal-machine interaction (AMI) as a discipline. He was very impressed by Clara Mancini’s paper „Animal-Computer Interaction (ACI): A Manifesto“ on animal-computer interaction. In his AMI research, he mainly investigates robots, gadgets, and devices and their behavior towards animals. There are not only moral questions, but also questions concerning the design of outer appearance and the ability to speak. The general background for his considerations is that more and more machines and animals meet in closed, half-open and open worlds. He believes that semi-autonomous and autonomous systems should have rules so that they treat animals well. They should not disturb, frighten, injure or kill them. Examples are toy robots, domestic robots, service robots in shopping malls and agricultural robots. Jackie Snow, who writes for New York Times, National Geographic, and Wall Street Journal, has talked to several experts about the topic. In an article for Fast Company, she quotes the ethicists Oliver Bendel and Peter Singer. Clara Mancini is also expressing her point of view. The article with the title „AI’s next ethical challenge: how to treat animals“ can be downloaded here.

Fig.: A couple of elephants

An Award for AI Devoted to the Social Good

The Association for the Advancement of Artificial Intelligence (AAAI) and Squirrel AI Learning announced the establishment of a new one million dollars annual award for societal benefits of AI. According to a press release of the AAAI, the award will be sponsored by Squirrel AI Learning as part of its mission to promote the use of artificial intelligence with lasting positive effects for society. „This new international award will recognize significant contributions in the field of artificial intelligence with profound societal impact that have generated otherwise unattainable value for humanity. The award nomination and selection process will be designed by a committee led by AAAI that will include representatives from international organizations with relevant expertise that will be designated by Squirrel AI Learning.“ (AAAI Press Release, 28 May 2019) The AAAI Spring Symposia have repeatedly devoted themselves to social good, also from the perspective of machine ethics. Further information via aaai.org/Pressroom/Releases//release-19-0528.php.

Fig.: An award for AI

AI Love You

Using an interdisciplinary approach, the book „AI love you“ explores the emerging topics and rapid technological developments of robotics and artificial intelligence through the lens of the evolving role of sex robots, and how they should best be designed to serve human needs. „An international panel of authors provides the most up-to-date, evidence-based empirical research on the potential sexual applications of artificial intelligence. Early chapters discuss the objections to sexual activity with robots while also providing a counterargument to each objection. Subsequent chapters present the implications of robot sex as well as the security and data privacy issues associated with sexual interactions with artificial intelligence.“ (Information by Springer) Topics featured in this book include: the Sexual Interaction Illusion Model, the personal companion system, Harmony, designed by Realbotix, an exposition of the challenges of personal data control and protection when dealing with artificial intelligence, and the current and future technological possibilities of projecting three-dimensional holograms. Oliver Bendel is the author of the contribution to the latter topic, entitled „Hologram Girl“. The book is edited by Yuefang Zhou and Martin H. Fischer and will be published in summer 2019. More information via www.springer.com/gp/book/9783030197339.

Fig.: AI love you