Call For Papers


ABUSE AND MISUSE OF SOCIAL AGENTS
Special Issue of Interacting with Computers

For a txt copy of this CFP click here


(Note: For Misuse and Abuse of Interactive Technologies CFP for Special Issue of Interaction Studies Go to this webpage for details)

:: Introduction ::
For decades, science fiction writers have envisioned a world in which robots and computers act like human assistants, virtual companions, and artificial helpmates. Nowadays, for better or for worse, that vision is becoming reality. A number of human-like interfaces and machines are under development in research centers around the world, and several prototypes have already been deployed on the Internet and in businesses. Even in our homes, service robots, such as vacuum cleaners and lawn movers, are becoming increasingly common. These creatures are the first-generation social agents: machines designed to build relationships with users while performing tasks with some degree of autonomy. Social agents display a range of human-like behaviors: they communicate using natural language, gesture, display and recognize emotions, and are even designed to mirror our facial expressions and show empathy.

Until recently, scientific investigations into the psychological aspects of human relationships with social agents have mainly addressed the positive effects of this relationship, such as an increase in trust and task facilitation (e.g., Bickmore & Picard, 2005). Nevertheless, as the interaction bandwidth evolves to encompass a broader range of social and emotional expressiveness, there is the possibility of the user and social agent displaying anti-social, hostile, and disinhibited behaviors. Workshops held at Interact2005 and CHI2006 (De Angeli, Brahnam, & Wallis, 2005; De Angeli, Brahnam, Wallis, & Dix, 2006) have suggested that anthropomorphic metaphors can inadvertently rouse the user to display dissatisfaction through angry interactions, sexual harassment, and volleys of verbal abuse.

At first glance, verbally or even physically abusing social agents and service robots may not appear to pose much of a problem—nothing that could be accurately labelled abuse since computers and machines are not people and thus not capable of being harmed. Nevertheless, the fact that abuse, or the threat of it, is part of the interaction opens important moral, ethical, and design issues. As machines begin to look and behave more like people, it is important to ask how they should behave when threatened and verbally and physically attacked. Another concern is the potential that socially intelligent agents have of taking advantage of users, especially children, who are prone to attribute to these characters more warmth and human qualities then they actually posses. Many parents, for instance, are disturbed by the amount of information social agents are able to obtain in their interactions with children. It is feared that these relationship-building agents could be used as potent means of marketing and advertising.

For this special issue of Interacting with Computers, we are soliciting papers from a range of disciplines (psychology, HCI, robotics, and cultural studies) that address the negative side of human-computer interaction.

:: Topics ::
Papers on all aspects of the topic are welcome, but we are particularly interested in papers that address the following questions:
* How does the misuse and abuse of social agents affect the user’s computing experience?
* How does disinhibition with social agents differ from Internet disinhibition?
* What are the psychological, sociological, and technological factors involved in negative interactions between users and social agents?
* What design factors (e.g., embodiment, communication styles, functionality) trigger or restrain disinhibited behaviours?
* What are the social and psychological consequences of negative human-computer interactions?
* How can we develop machines that learn to avoid user abuses?
* How can agent technology be exploited to take advantage of users and what ethical and design mechanisms can counteract this possibility?
* Is it appropriate for machines to ignore aggression? If conversational agents do not acknowledge verbal abuse will this only serve to aggravate the situation?
* If potential clients are abusing virtual business representatives, then to what extent are they abusing the businesses or the social groups the human-like interfaces represent?

:: Guest Editors ::
Antonella De Angeli (University of Manchester), UK
Sheryl Brahnam (Missouri State University), US

:: Submissions ::
Contributions should not exceed 10,000 words and should follow the IwC Guideliness for Authors.

Authors intending to submit a paper should send an Abstract to the contact person below).

All papers submitted will be double blind, peer reviewed.

:: Important Dates ::
March 20th: Abstract Submission
April 5, 2007: Paper Submission
May 25: Notification of acceptance
June 15: Camera ready
Contact: Sheryl Brahnam sbrahnam: (*) facescience (*) org (replace * with appropriate email symbols).