Library
  • XR Guild Library 2.0
  • AI Ethics
    • Artificial Intelligence and the Future of Work: Mapping the Ethical Issues
    • Why the Godfather of AI Now Fears His Creation
    • Is 'Ethical AI' a Fantasy? - TRIP Annual Symposium 2024
    • Publics’ views on ethical challenges of artificial intelligence: a scoping review
    • Ethics in AI Seminar: Responsible Research and Publication in AI
    • Trustworthy AI and Ethics with IBM Consulting's Phaedra Boinodiris
    • AI is an energy hog. This is what it means for climate change.
    • Non-Consensual Synthetic Intimate Imagery: Prevalence, Attitudes, and Knowledge in 10 Countries
    • Unraveling the Ethical Conundrum of Artificial Intelligence: A Synthesis of Literature and Case Stud
    • Ethics of AI: Challenges and Governance
    • AI hiring tools may be filtering out the best job applicants
    • AI Ethics: A Bibliometric Analysis, Critical Issues, and Key Gaps
    • Tackling problems, harvesting benefits -- A systematic review of the regulatory debate around AI
    • Global Dialogues on AI
    • UN Global Digital Compact
  • Current Peer-Reviewed Immersive Technologies
    • Towards Understanding Diminished Reality
    • BlendMR: A Computational Method To Create Ambient Mixed Reality Interfaces
    • RealityReplay: Detecting and Replaying Temporal Changes In Situ using Mixed Reality
    • A Survey on Remote Assistance and Training in Mixed Reality Environments
    • SemanticAdapt: Optimization-based Adaptation of Mixed Reality Layouts Leveraging Virtual-Physical Se
    • Auptimize: Optimal Placement of Spatial Audio Cues for Extended Reality
    • TurnAware: Motion-aware Augmented Reality Information Delivery While Walking
    • SecureMR - Security and Privacy for Camera Access in XR Applications
    • The Poetics of Augmented Space
    • Integrating Immersive Technologies with STEM Education
    • Educational Administration: Theory and Practice
    • Immersive technologies as an innovative tool to increase academic success and motivation
    • Metaverse as a Cutting-edge Platform for Attaining Sustainable Development Goals (SDGs)
    • Defining immersion and immersive technologies
    • A bibliometric analysis of immersive technology in museum exhibitions: exploring user experience
    • The Combination of Artificial Intelligence and Extended Reality: A Systematic Review
    • Technology Integration Methods for Bi-directional Brain-computer Interfaces and XR-based Interventio
    • Towards an Eye-Brain-Computer Interface: Combining Gaze with the Stimulus-Preceding Negativity
    • Immersive interfaces for clinical applications: current status and future perspective
    • Hypersphere - XR Design for Metaverse by Synthesizing Neuro Reality and Virtual Reality
    • CAVE: An Emerging Immersive Technology - A Review
    • Experiences with a Virtual Reality System for Immersive Decision Making and Learning
    • Virtual Reality: How Much Immersion Is Enough?
  • Ethical Responsibility
    • Cybersecurity and Privacy Challenges in Extended Reality: Threats, Solutions, and Risk Mitigation...
    • The Poetics of Augmented Space
    • Ethical Considerations When Designing and Implementing Immersive Realities in Nursing Education
    • Ethical considerations in designing virtual and augmented reality products - virtual and augmented r
    • XRAI-Ethics: Towards a Robust Ethical Analysis Framework for Extended Artificial Intelligence
    • An XR Ethics Framework
    • Extended Reality (XR) and the Erosion of Anonymity and Privacy
    • Who Owns Our Second Lives: Virtual Clones and the Right to Your Identity
    • The IEEE Global Initiative on Ethics of Extended Reality (XR) Report - Metaverse and Its Governance
    • The Ethics of Realism in Virtual and Augmented Reality
    • Smart Workplaces for Older Adults: Coping 'Ethically' with Technology Pervasiveness - Ageing@Work an
    • 2089-2021 - IEEE Standard for an Age Appropriate Digital Services Framework Based on the 5Rights Pri
    • State of the Art in Ethics for AR
    • Mapping of the ethical Issues in XR-overview of Ethical Frameworks: A Scoping Review
    • Quality Considerations for Ethical Design of Virtual and Augmented Reality.
    • Beyond Speculation About the Ethics of Virtual Reality: The Need for Empirical Results
    • Getting rescued by RoboCop? Legal and ethical challenges of the use of extended reality in Frontex’s
    • Who will govern the metaverse? Examining governance initiatives for extended reality (XR) technologi
    • A Survey of Extended Reality (XR) Standards
  • XR Industry news/events/Sources
    • VRARA Events
    • XRSI News
    • IEEE News
    • XRA News
    • XR TODAY Events
    • VR Marketing News
    • MSF Newsroom
    • AWE Blog
    • OpenXR (Khronos Group)
    • CES
    • Augmented Reality for Enterprise Alliance
    • XR Fair Tokyo
    • Mobile World Congress
    • Stereopsia
    • XR Expo
    • Laval Virtual
    • XR & Metaverse Standards Register
    • XRWomen Research Library
  • Neurotechnology, Brain Tech, and Ethical Challenges
    • Brain augmentation and neuroscience technologies
    • U.S. Public Perceptions of the Sensitivity of Brain Data
    • Protecting Neural Data Privacy—First, Do No Harm
    • Neurotechnology: Current Developments and Ethical Issues
    • Karen Rommelfanger - Neuroethics and the Future of Neurotechnology
    • My Brain Made Me Buy It: The Neuroethics of Advertising - Exploring Ethics
    • IEEE Brain Episode 17: Q&A with Dr. Cristin Welle, Associate Professor, University of Colorado
    • Mindscape 229: Nita Farahany on Ethics, Law, and Neurotechnology
    • Ethics of Neurotechnology
    • Understanding the Ethical Issues of Brain-Computer Interfaces (BCIs): A Blessing or the Beginning of
    • Ethics and Governance of Neurotechnology in Africa: Lessons From AI
    • TED Radio Hour: Your brain is the next tech frontier
    • Use of Invasive Brain-Computer Interfaces in Pediatric Neurosurgery: Technical and Ethical Considera
    • FUTURES: Designing Brain-Computer Interfaces" with Conor Russomanno
    • Neurotechnology and the Battle for Your Brain
    • Ethical Frontiers: Navigating the Intersection of Neurotechnology and Cybersecurity
    • Neuroethics and AI Ethics: A Proposal for Collaboration
    • An Integrated Embodiment Concept Combines Neuroethics and AI Ethics
    • Neuroethics Today
  • Mind Control and Privacy at Work
    • Brain Recording, Mind-Reading, and Neurotechnology: Ethical Issues from Consumer Devices
    • The Convergence of Virtual Reality and Social Networks: Threats to Privacy and Autonomy
    • Ethics Emerging: the Story of Privacy and Security Perceptions in Virtual Reality
    • Neuroinsights in Immersive Worlds: Safeguarding Cognitive Freedom in Virtual Realms
    • Digital body, identity and privacy in social virtual reality: A systematic review
    • The concept “freedom” in a virtual reality of the information society
    • Freedom of Thinking in the Terms
    • Mind control: The metaverse may be the ultimate tool of persuasion
    • Metaverse in Mental Health: The Beginning of a Long History
    • “Playing God”: How the metaverse will challenge our very notion of free will
    • The Metaverse: the Ultimate Tool of Persuasion
    • Searching for the Metaverse: Neuroscience of Physical and Digital Communities
    • The neurosociological paradigm of the metaverse
    • Metaverse and Privacy
    • The enterprise metaverse, cognitive science, and interpersonal communication
    • The Metaverse: from Marketing to Mind Control
  • Physiological and Psychological effects of XR
    • Empathy and Perspective Taking
    • Building long-term empathy: A large-scale comparison of traditional and virtual reality...
    • A VIRTUAL SAFE SPACE? AN APPROACH OF INTERSECTIONALITY AND SOCIAL IDENTITY TO BEHAVIOR IN VIRTUAL...
    • Who Do You Think You Are? What Does Your Avatar Say About You?
    • LET'S CONNECT IN METAVERSE. BRAND'S NEW DESTINATION TO INCREASE CONSUMERS' AFFECTIVE BRAND...
    • The psychological impact of the Metaverse
    • The effect of virtual reality forest and urban environments on physiological and psychological...
    • The Effectiveness of Virtual Reality Exercise on Individual’s Physiological, Psychological and...
    • A perspective on potential psychological risks and solutions of using virtual reality in...
    • The impact of immersive virtual reality meditation for depression and anxiety among inpatients...
    • Digital body, identity and privacy in social virtual reality: A systematic review
    • Virtual Reality and Emotion: A 5-Year Systematic Review of Empirical Research
    • Virtual Reality Games and the Role of Body Involvement in Enhancing Positive Emotions and...
    • Empowering Social Growth Through Virtual Reality–Based Intervention for Children With Attention...
    • Influence of Avatar Identification on the Attraction of Virtual Reality Games: Survey Study
    • How Avatar Identification Affects Enjoyment in the Metaverse: The Roles of Avatar Customization...
    • Influence of Avatar Identification on the Attraction of Virtual Reality Games: Survey Study
    • Beyond the pixelated mirror: Understanding avatar identity and its impact on in-game advertising...
    • Exploring the user-avatar relationship in videogames: A systematic review of the Proteus effect
    • VR: A Path to Mental Wellness
  • Privacy & Policy
    • Voices Of VR #1091: IEEE XR Ethics: The Erosion of Privacy & Anonymity
    • Privacy and Ethical Considerations for Extended Reality Settings - On Tech Ethics
    • Rethinking Privacy in the AI Era
    • AI Data, Governance and Privacy: SYNERGIES AND AREAS OF INTERNATIONAL CO-OPERATION
    • Safety and Privacy in Immersive Extended Reality: An Analysis and Policy Recommendations
    • Cybersecurity and Privacy Challenges in Extended Reality: Threats, Solutions, and Risk Mitigation...
    • Extended Reality (XR) Ethics
    • Voices Of VR #997: Debating XR Privacy Tech Policy with Ellysse Dick
    • XR In the Classroom: Student Privacy and Safety
    • Augmenting Security and Privacy in the Virtual. Realm: An Analysis of Extended Reality Devices
    • Privacy Preservation in Artificial Intelligence and Extended Reality (AI-XR) Metaverses: A Survey
    • On Tech Ethics Podcast – Privacy and Ethical Considerations for Extended Reality Settings
    • OVRseen: Auditing Network Traffic and Privacy Policies in Oculus VR
    • Reality Check: Why the U.S. Government Should Nurture XR Development
    • The Tech Law Hangout
  • Security & Safety
    • Safety and Privacy in Immersive Extended Reality: An Analysis and Policy Recommendations
    • Implications of XR on Privacy, Security and Behaviour: Insights
    • XR Trends 2022 (11/12): XR Safety and Security
    • Futuristic Metaverse: Security and Counter Measures
    • Navigating the XR Educational Landscape: Privacy, Safety, and Ethical Guidelines
    • VR headsets may lead users to the ER
    • Concerns with Privacy in Virtual Reality
    • What are the Security and Privacy Risks of VR and AR
    • VR Risks For Kids And Teens
    • We Share Pro Tips To Implement VR Industrial Training Successfully
    • Virtual Reality Headsets
    • Voices Of VR #786: XR Safety Initiative: Security, Privacy, & Ethics in XR
    • XR and AI for Security and Privacy with Luis Bravo Martins
    • AMXRA Guidelines on Extended Reality and Children
  • Glossary
  • Archive
    • Videos
      • XR Guild
      • Rolando Masís-Obando (XRMasiso)
      • Avi Bar-Zeev
      • Louis Rosenberg
      • Sarah A. Barker
        • VR| A Path to Mental Wellness: Meet VR Experience Designer Owen Harris
        • VR: A Path to Mental Wellness: Meet VR Experience Designer Owen Harris
        • VR: A Path to Mental Wellness: Meet Brennan Spiegel
    • Extended Principles
    • Newsletters
      • Metaversethics.org
    • Podcasts
      • Voices of VR Podcast
      • VR: A Path to Mental Wellness
      • The Virtual World Society Podcast
      • Experiment: Google NotebookLM version of existing works
    • Written Works
      • ACM Whitepapers on Ethics
      • IEEE Whitepapers on Ethics
      • Louis Rosenberg
      • Brittan Heller
      • Avi Bar-Zeev
      • Nita Farahany
      • W3C Papers and Principles
      • Additional Papers
    • XR Guild Library Database
      • Issues around Safety
      • Physiological and Psychological effects of XR
      • Industry news/implications
      • Data Privacy & Policy
      • Current Peer-Reviewed Immersive Technologies, Human Psychology and Ethics
    • Science and Ethics Council Topics
      • Cybersickness and Nauseation
      • Accessibility
      • Brain Computer Interfaces
      • The Magic (and Limits) of Human Visual Perception
      • Conversational Agents
Powered by GitBook
On this page
  • Conversational Agents
  • Conversational Agents and the AI Manipulation Problem:
  • Bibliography
  1. Archive
  2. Science and Ethics Council Topics

Conversational Agents

And the Ethical Risks of Human Manipulation

Last updated 7 months ago

By Louis B. Rosenberg

Conversational Agents

Conversational agents are AI-powered systems that engage users in natural interactive dialog. When the interactions are text-based, these systems are generally referred to as chatbots. When enabled through natural spoken language (via real-time voice processing and voice generation) they are more commonly called voice bots and can be deployed for a wide range of uses, including personal assistants and customer service representatives. When deployed with a simulated appearance through an avatar (either on flat screens or on immersive displays) they become embodied entities that are sometimes referred to as virtual spokespersons (VSPs) or virtual assistants [1]. When implemented with sufficient visual and audio fidelity, these interactive avatars can be indistinguishable from authentic human representatives [2].

Whether conversational agents communicate through text, audio, or a visual avatar, all conversational dialog is generally converted to text before processing by a Large Language Model (LLM). In addition, Multimodal Large Language Models (MLLMs) can now process audio and video signal streams (in near real-time) to supplement language processing. This enables rapid analysis of the emotional affect of human users from audio cues (including vocal inflections) and visual cues (including facial expressions, body posture, and gestures.) In this way, conversational agents are poised to engage humans naturally through convincing avatars that can express empathetic facial expressions and vocal inflections and can react to the facial expressions, vocal inflections, body posture, and other cues of the engaged human.

The Benefits: Conversational Agents with this level of sophistication have significant potential to make computing more natural and intuitive, enabling us to interact with software-based systems using the most human of all skills – language. This, more than any advancement over the last 20 years, will enable mainstream computing to finally move away from traditional interface models based on flat screens and keyboards/touchscreens. In addition, this will likely bolster the adoption of XR interfaces (AR, VR, MR) by enabling hands-free interactions with embodied agents through natural conversational interactions. It will also transform traditional flat-screen computing. In the near future, it will become commonplace to embed conversational agents on website to act as salespeople, customer service representatives, and as the “friendly face” of critical software utilities such as search engines. In most cases, interactions will likely be perceived as highly convenient and intuitive.

The Risks: While offering value, Conversational Agents also pose a unique threat to human agency as they represent an interactive and adaptive form of media that can be used to impart targeted influence. While many forms of media can be used as tools of influence, Conversational Agents are unique in that they can easily target users through personalized content in real-time. This creates new risks that extend beyond the dangers of traditional forms of media, including social media. As described below, it is helpful to formalize these risks in the context of Control Theory, as this can help stakeholders appreciate that interactive agents represent a very threat which requires new safeguards, guidelines, and ethical use policies. Without enhanced protections, conversational agents could be deeply misused, easily crossing the line from deploying targeted influence, to driving targeted manipulation.

Conversational Agents and the AI Manipulation Problem:

All forms of media can be used to impart influence on individuals. Therefore, technology developers and business leaders need to be mindful of the consequences of misuse or abuse. Conversational agents are a unique form of media in that they can easily be designed to adapt their influence tactics during real-time conversations, optimizing their impact on individual users [3]. The AI Manipulation Problem formalizes this risk by identifying the four basic steps by which a conversational agent could be used to excessively influence a user. These steps create a “feedback loop” around the user as follows:

  1. Impart real-time targeted influence on an individual user through AI-crafted dialog.

  2. Sense the user’s real-time reaction to the imparted conversational influence.

  3. Adjust conversational tactics to increase the persuasive impact on the user.

  4. Repeat steps i, ii, iii to maximize influence effectiveness.

Any system that follows these four steps could create a “feedback loop” around the user in which the individual is repeatedly targeted, assessed, and re-targeted to progressively optimize the influence. In the field of engineering, this is typically called a “Feedback Control System.” As shown in Figure 1 below, the basic pieces of any control system include a System you aim to control, Sensors for detecting the system’s behavior, and a Controller that adjusts the influence on the system to optimize the desired impact. A classic example is a thermostat for setting a temperature goal in a house. If the house (i.e. the System) falls below the goal, the thermostat turns the heat on. If the temperature rises above the goal, the thermostat turns the heat off. This simple feedback loop keeps the temperature close to the specified goal (i.e. imparts optimized influence).

The AI Manipulation Problem considers the case where the target system is a human user, the controller is a conversational agent, and the sensors are microphones and cameras that monitor user’s response via language processing, supplemented with analysis of vocal inflections, facial expressions and other physical cues. As shown below in Figure 2 below, this “human control system” is essentially the same as a simple thermostat, but the Input is not a temperature goal but an Influence Objective to be imparted on the user. The AI Agent will engage the user in interactive dialog, gradually adapting its conversational tactics based on the measured behavior of the system (i.e., the verbal responses of the user, potentially supplemented with the emotional content in the user’s vocal inflections, factual expressions, and body posture.

Consider a conversational agent deployed to convey misinformation about a political candidate. That AI agent will engage a target user in interactive dialog, likely adopting a conversational style that is custom selected for the specific user based on stored personal data about that user. For example, the user might be targeted with an AI Agent that speaks in a casual style and makes emotional arguments if that user’s personal data suggests this approach will be effective. Conversely, a different user might be targeted with more formal language and logical arguments, if that user’s personal data suggests it will be effective.

With the style selected, the conversational content can then be custom crafted to optimize impact on the target user, for example by referencing that user’s personal interests, profession, or political values. The user will react through a conversational response. The controller will assess the user’s reaction in real-time, for example by determining if their resistance to the influence is based on factual and/or emotional barriers. The controller will then adapt its tactics to overcome the resistance, offering custom tailored counterpoints. This process is repeated as conversation continues, the controller working to iteratively overcome objections and efficiently guide the user toward accepting the influence [2,3]. Such a controller could be deployed to persuade an individual user into beliefs or behaviors or that he or she may not normally adopt through traditional media.

To make these dangers clear for XR environments, the 2023 short film Privacy Lost gives a quick fictionalized example of conversational manipulation through AI-powered XR glasses:

Some argue that conversational manipulation is not a new threat, as human salespeople already use interactive dialogue to influence customers by overcoming objections. While true, AI-powered systems are likely to be significantly more effective agents of manipulation. For example, a 2024 study performed at Swiss Federal Institute of Technology found that when human users debated an AI-agent powered by GPT-4, they had 81.7% (p < 0.01) higher odds of increased agreement with their opponent as compared to participants who debated humans [4]. In addition, the AI agents could be trained on aggressive sales tactics, methods of persuasion, human psychology, and cognitive biases.

In addition, AI agents are likely to have super-human abilities to sense human emotions during interactive conversations. Already, AI system can detect micro-expressions on human faces that are far too subtle for human observers [5]. Similarly, AI-systems can read faint changes in human complexion known as facial blood flow patterns and subtle changes in pupil dilation to assess emotions in real-time. In addition, these platforms are likely to store data during conversational interactions over time, tracking and analyzing which types of arguments and approaches are most effective on each user personally.

For example, the system could be designed to learn whether a target user is more easily swayed by factual data, emotional appeals, or by playing on their insecurities or fear of missing out. In other words, these systems will not only adapt to real-time emotions, but they could get progressively better at influencing the target user over time, potentially learning how to draw them into conversations, guide them into accept new ideas, or convince them to buy things they don’t need, believe things that are untrue, or even support extreme policies or politicians that they’d otherwise reject.

In these ways, the AI Manipulation Problem cautions technology developers, business leaders, and policymakers that human users will be increasingly vulnerable to targeted manipulation unless ethical guidelines and/or regulations are put in place to restrict or ban the use of real-time feedback loops in which AI agents can adapt and optimize their tactics during interactive conversations to maximize impact. Protections could also include banning the storage of data that characterizes the responsiveness of individual human users to different conversational styles and/or persuasive tactics, thereby preventing systems from learning over time how to best target individual users with conversational influence [6,7].

Bibliography

[3] Rosenberg, L (2023). The Manipulation Problem: Conversational AI as a Threat to Epistemic Agency. 2023 CHI Workshop on Generative AI and HCI (GenAICHI 2023). Association for Computing Machinery, Hamburg Germany (April 28, 2023)

[4] Salvi, F., et. al, (2024). “On the Conversational Persuasiveness of Large Language Models: A Randomized Controlled Trial.” ArXiv abs/2403.14380

[5] Li, X. et al. (2018). "Towards Reading Hidden Emotions: A Comparative Study of Spontaneous Micro-Expression Spotting and Recognition Methods," in IEEE Transactions on Affective Computing, vol. 9, no. 4, pp. 563-577, 1 Oct.-Dec. 2018, doi: 10.1109/TAFFC.2017.2667642.

[7] Graylin, A., Rosenberg, L.“Our Next Reality,” Nicholas Brealey Publishing (June 4, 2024)

[1] Rosenberg, L. (2022) Regulation of the Metaverse: A Roadmap: The risks and regulatory solutions for largescale consumer platforms. In Proceedings of the 6th International Conference on Virtual and Augmented Reality Simulations (ICVARS '22). Association for Computing Machinery, New York, NY, USA, 21–26.

[2] Rosenberg, L. (2023). The Metaverse as an AI-mediated Tool of Targeted Persuasion. In: Matteo Zallio (eds) Human-Centered Metaverse and Digital Environments AHFE (2023) International Conference. AHFE Open Access, vol 99. AHFE International, USA.

[6] Wallace C., Rosenberg L., Pearlman K., et al. (2023) Whitepaper the Metaverse and Standards. Sydney, Australia: Standards Australia.

https://doi.org/10.1145/3546607.3546611
http://doi.org/10.54941/ahfe1003938
https://www.standards.org.au/documents/h2-3061-metaverse-report
Figure 1: Standard Diagram of Feedback Control System.
Figure 2: AI Manipulation Problem as a Control System that optimizes Conversational Influence.
Privacy Lost ()
link to YouTube video