For Full-Text PDF, please login, if you are a member of IEICE,|
or go to Pay Per View on menu list, if you are a nonmember of IEICE.
Mood-Learning Public Display: Adapting Content Design Evolutionarily through Viewers' Involuntary Gestures and Movements
Ken NAGAO Issei FUJISHIRO
IEICE TRANSACTIONS on Information and Systems
Publication Date: 2014/08/01
Online ISSN: 1745-1361
Type of Manuscript: Special Section PAPER (Special Section on Cyberworlds)
public display, image processing, human behavior recognition, genetic algorithm,
Full Text: PDF>>
Due to the recent development of underlying hardware technology and improvement in installing environments, public display has been becoming more common and attracting more attention as a new type of signage. Any signage is required to make its content more attractive to its viewers by evaluating the current attractiveness on the fly, in order to deliver the message from the sender more effectively. However, most previous methods for public display require time to reflect the viewers' evaluations. In this paper, we present a novel system, called Mood-Learning Public Display, which automatically adapts its content design. This system utilizes viewers' involuntary behaviors as a sign of evaluation to make the content design more adapted to local viewers' tastes evolutionarily on site. The system removes the current gap between viewers' expectations and the content actually displayed on the display, and makes efficient mutual transmission of information between the cyberworld and the reality.