Adrian David Cheok is featured in the Channel News Asia documentary series “Becoming Human”, which explores love and artificial intelligence. It will be televised on 10th March 2019, Sunday, 9PM (GMT +8, Singapore, Jakarta and Delhi).
Following its success in London, the International Congress on Love and Sex with Robots will hold its fourth conference at the University of Montana in the United States on 13 – 14 December 2018.
Matt McMullen, CEO, artist and design director of Realbotix, will be the keynote speaker at the conference. He will be delivering a speech on the topic of “Artificial Intelligence, Robotics, and Intimacy: A new Alternative form of relationship”
For the past 20 years, McMullen has been creating “The World’s Finest Love Dolls”, as well as undertaking multiple custom projects along the way. He started Abyss Creations, the manufacturer of Realdoll, out of his garage in 1997. McMullen’s dolls have popped up on more than 20 television shows, and co-starred in 10 films. Now, he focuses on integrating cutting-edge technology with silicone doll artistry and believes that AI driven robots can become companions and partners to human beings, and can connect with us in ways that are often overlooked when considering these technologies.
The Love and Sex with Robots congress is the only annual academic event on the topic of human-robot intimate relationships. The congress was founded and chaired by David Levy, author of the eponymous New York Times bestseller “Love and Sex with Robots”, and Adrian David Cheok, Professor of Pervasive Computing at City, University of London and Director of Imagineering Institute, Malaysia.
ACE 2018 is calling for submissions to its 15th conference to be held at the University of Montana, USA, on 10-12 December 2018.
We seek all submissions on Affective Computing, Augmented, Mixed and Virtual Reality, Educational and Fitness Games, Game Design, Interactive Storytelling, Mobile and Ubiquitous Entertainment, Sound and Music Design, Tangible Interfaces etc. A full list of topics is available on our website: http://ace2018.info/topics/
Papers may range from 2 to 20 pages long, in the Springer CS Proceedings format. Deadline for submission is 15th July, 2018 (GMT). Detailed instructions for submission format is found on http://ace2018.info/call-for-submissions/.
This year, our keynote speaker is Professor Peter Gray, research professor of psychology at Boston College who has conducted and published research in neuroendocrinology, developmental psychology, anthropology, and education. He is author of an internationally acclaimed introductory psychology textbook (Psychology, Worth Publishers, now in its 8th edition), which views all of psychology from an evolutionary perspective. His recent research focuses on the role of play in human evolution and how children educate themselves, through play and exploration, when they are free to do so. He has expanded on these ideas in his book, Free to Learn: Why Unleashing the Instinct to Play Will Make Our Children Happier, More Self-Reliant, and Better Students for Life (Basic Books). He also authors a regular blog called Freedom to Learn, for Psychology Today magazine. He earned his undergraduate degree at Columbia College and Ph.D. in biological sciences, many years ago, at the Rockefeller University. His own current play includes kayaking, long distance bicycling, backwoods skiing, and vegetable gardening. http://ace2018.info/keynote/
Director of the Imagineering Institute Malaysia, Chair Professor of Pervasive Computing at City University London, Founder and Director of the Mixed Reality Lab Singapore
Virtual senses for the internet
Nowadays, we mostly perceive our digital world by viewing text and images or hearing audio. But what about the rest of our senses?
Adrian Cheok, Director of the Imagineering Institute Malaysia, founded the Mixed Reality Lab in Singapore and tries to integrate the rest of our senses, like smelling or tasting, into our digital experience. „What do we need digital smells and tastes for anyway?“, one might ask.
Those senses are of big importance for influencing emotion. Every day we can experience how smells can change our mood, e.g. when eating a delicious meal.
The possibilities for using digital senses are widely spread: games, films, messaging and communication over social networks or telephone, but also commercial usage like advertisement.
Cheok’s Mixed Reality Lab develops devices that stimulate our senses with electrical signals, „because we can’t send the chemicals over the internet“, that are normally needed for these body-reactions, explains Prof. Cheok. As he puts it, we already live in our own „analog virtual reality“ with our brain as a device to perceive the world surrounding us. That’s why he believes that the step to a digital virtual reality will not be a very big one.
Might there be a danger in replacing our „real life“ with digital virtual reality?
According to Cheok, differences will become smaller, „but society will adapt“. Kissing or marrying a robot in the future may be as normal as human marriage today.
Nonetheless, the focus of his research is not in replacing our life with digital experiences, but in expanding the analog life by adding digital impressions.
Professor Adrian David Cheok will give a keynote speech at IT Innovation Day on 28 September 2017, in Amersfoort, Netherlands.
Title: Love and Sex with Robots Time: 14:33, 28 September 2017 Location: Prodentfabriek, Amersfoort, Netherlands
In his speech, Professor Cheok will look at the tangible aspect (touch) of technology, and the ways in which this will contribute to an overall experience for people, including sexual behaviour. Adrian Cheok will outline a more controversial view of the future, along with how tangible technology will enhance experiences at all levels of human behaviour. When is good, good enough, real enough, and how can quality be improved? Adrian will also share his latest tech inventions with the public.
Digital Food is an exhibition that focuses on the futuristic idea of how taste or flavour have been evolving, i.e. from natural, to artificial or synthetic flavours. The exhibition also explores how senses can be manipulated using digital technology in future. It challenges us to think about how digital food is going to enhance living quality and improve our health. This exhibition is only available for a limited period so do catch it while it last! It is jointly developed by Science Centre Singapore and Imagineering Institute.
20 Sep 2017 – 20 Nov 2017
Hall A, Science Centre
Typical time required:
Digital Candy Shop
The Digital Candy Shop is the main highlight of this exhibition with two interactive stations, i.e. the Digital Cream Pot and the Digital Lollipop, which allow you to “taste” food using technology.
Can you smell and taste colours?
This is part of the story of building up an artificial food experience. Visual cues are part of our perception of flavour and food. From the moment we see the food, our brains begin to build expectations using memories of previous experiences linked with the food’s colour, smell or appearance. Come have a “taste” of the smell.
Checking your sensitivity in smell
This exhibit challenges visitor if they have a super nose to distinguish different smells and identify the common ones.
Professor Adrian Cheok was invited to give a keynote speech at the International Conference on Foundations of Digital Games 2017, at Cape Cod, USA.
FDG 2017 is a major international event in-cooperation with ACM SIGAI, SIGCHI, and SIGGRAPH. It seeks to promote the exchange of information concerning the foundations of digital games, technology used to develop digital games, and the study of digital games and their design, broadly construed. The goal of the conference is the advancement of the study of digital games, including but not limited to new game technologies, critical analysis, innovative designs, theories on play, empirical studies, and data analysis.
Professor Cheok’s keynote speech will be covering the trending topic of “Love and Sex with Robots”.
Time: 15 Aug 2017, 9am
Venue: The Resort and Conference Center at Hyannis, Cape Cod, MA, USA
Title: Love and Sex with Robots
Abstract: “Love and Sex with Robots” has recently become a serious academic topic within the fields of Human Machine Interaction and Human Robot Interaction. This topic has also witnessed a strong upsurge of interest amongst the general public, print media, TV documentaries and feature films. This talk covers the personal aspects of human relationships and interactions with robots and artificial partners. New technologies and research prototypes have been developed to allow more intimate interactions with robot companions like sex robots, emotional robots, humanoid robots, and artificial intelligent systems that can simulate human emotions. Such technologies and systems also engage the users with all their senses, including touch, taste and smell, creating multisensory and immersive interactive experiences. In this talk, we will conclude that humans will marry robots by 2050.
Date: August 7, 2017
Adrian David Cheok, Kasun Karunanayaka, Surina Hariri, Hanis Camelia, and Sharon Kalu Ufere Imagineering Institute, Iskandar Puteri, Malaysia & City, University of London,UK.
Phone: +607 509 6568
Fax: +607 509 6713
Here we are excited to introduce the world’s ﬁrst computer controlled digital device developed to stimulate olfactory receptor neurons with the aim of producing smell sensations purely using electrical pulses. Using this device, now we can easily stimulate the various areas of nasal cavity with different kinds of electric pulses. During the initial user experiments, some participants experienced smell sensations including ﬂoral, fruity, chemical, and woody. In addition, we have observed a dif- ference in the ability of smelling odorants before and after the electrical stimulation. These results suggest that this technology could be enhanced to artiﬁcially create and modify smell sensations. By conducting more experiments with human subjects, we are expecting to uncover the patterns of electrical stimulations, that can effectively generate, modify, and recall smell sensations. This invention can lead to internet and virtual reality digital smell.
To date, almost all smell regeneration methods used in both academia and industry are based on chemicals. These methods have several limitations such as being expensive for long term use, complex, need of routine maintenance, require reﬁlling, less controllability, and non-uniform distribution in the air. More importantly, these chemical based smells cannot be transmitted over the digital networks and regenerate remotely, as we do for the visual and auditory data. Therefore, discovering a method to produce smell sensations without us- ing chemical odorants is a necessity for digitizing the sense of smell. Our concept is illustrated in the Figure 1, which is electrically stimulating the olfactory receptor neurons (ORN) and study whether this approach can produce or modify smell sensations. During a medical experiment in 1973, electrical stimulation of olfactory receptors reported some smell sensations including almond, bitter almond, and vanilla . However, three other similar experiments that used electrical stimulation failed to reproduce any smell sensations [2, 3, 4]. Therefore, ﬁnding a proper method to electrically reproduce smell sensations was still undiscovered.
Our approach is different from the previous research mentioned above. Our main objective is to develop a controllable and repeatable digital technology, a device that connects to computers and be easily able to programmed and controlled. Also this device needs to generate electric pulses of different frequencies, cur- rents, pulse widths and stimulation times. To provide more stimulation possibilities, we wanted this device to be capable of stimulating diverse sites at the ventral surface of the inferior, middle, and superior nasal concha. Fig. 2 shows the computer controlled digital device we have developed to stimulate olfactory receptors. The amount of current output by the circuit can be controlled using one of the ﬁve push buttons shown in Figure 2 and the respective LED near the push button will lights up after the selection. The frequency of the stimulation pulses and stimulation time is controlled by the microcontroller program. It is possible to vary the stimulation frequency from 0Hz to 33kHz and pulse width using the programming. The pair of silver electrodes combined with the endoscopic camera was used to stimulate olfactory receptor neurons, and during the stimulation, one electrode is conﬁgured as the positive and the other electrode as the ground. Fig 3 and Fig 4 shows testing our device with human subjects.
During our ﬁrst user study, we have stimulated the 30 subjects using 1mA to 5mA range with frequencies 2Hz, 10Hz, 70Hz, and 180Hz. 1mA at 10Hz and 1mA at 70Hz were the stimulation parameters which gave most prominent results for the smell related responses. Electrical stimulation with 1mA and 70Hz induced the highest odor perceptions. 27% of the participants reported the perceived fragrant and chemical sensa- tions. Other smell sensations that are reported for include, 20% fruity, 20% sweet, 17% tosted and nutty, 10% minty, and 13% woody. Stimulation parameters 1mA/10Hz reported 17% fragrant, 27% sweet 27%, chemical 10%, woody 10%. Meanwhile, results for the 4mA/70Hz reported 82% for pain and 64% reported pressure sensations. We have also probed the effect of electrical stimulation on the nose after stimulation. Therefore, we asked participants to repeat the snifﬁng of known odorants immediately after stimulation and rate the intensity. Most of the participants reported higher intensity after stimulation. This showed that the electrical stimulation increased the intensity of the odorants in the nose.
We are planning to extend this user experiment with more number of participants. The effects of the differ- ent electrical stimulation parameters such as frequency, current, and stimulation period will be more closely studied in future. By analyzing the results, we plan to identify various stimulation patterns that can produce different smell sensations. If the electrical stimulation of olfactory receptors effectively produce smell sen- sations, it will revolutionize the ﬁeld of communication. Multisensory communication is currently limited to text, audio and video contents. Digitizing touch sense are already been achieved experimentally in the research level and will be embedded to daily communication near future. If the digitization of smell be- comes possible it will paved the way for sensing, communicating and reproducing ﬂavor sensations over the internet. This will create more applications in the ﬁelds such as human computer interaction, virtual reality, telepresence, and internet shopping.
1.Uziel, A.: Stimulation of human olfactory neuro-epithelium by long-term continuous electrical currents. Journal de physiologie 66(4) (1973) 409422
2.Weiss, T., Shushan, S., Ravia, A., Hahamy, A., Secundo, L., Weissbrod, A., Ben-Yakov, A., Holtzman, Y., Cohen- Atsmoni, S., Roth, Y., et al.: From nose to brain: Un-sensed electrical currents applied in the nose alter activity in deep brain structures. Cerebral Cortex (2016)
3.Straschill, M., Stahl, H., Gorkisch, K.: Effects of electrical stimulation of the human olfactory mucosa.Stereotactic and Functional Neurosurgery 46(5-6) (1984) 286289
4.Ishimaru, T., Shimada, T., Sakumoto, M., Miwa, T., Kimura, Y., Furukawa, M.: Olfactory evoked potential produced by electrical stimulation of the human olfactory mucosa. Chemical senses 22(1) (1997) 7781
Professor Adrian David Cheok will give a keynote speech at Visual SG in Singapore Science Centre on 28 July 2017.
Topic: Everysense Everywhere Human Communication
Time: 11:10am, 28 July 2017
Location: Singapore Science Centre
Visual SG is South East Asia’s signature Visualisation Festival. The Festival celebrates beauty through its bold emphasis on the visual aesthetics, insights and narratives that reside in data and scientific visualisation. Envisaged as both a serious study and playful showcase, VisualSG presents a full on visual spectacle of data through the lens of artistic and creative expression. Through its line-up of interactive displays, forums and workshops, Visual SG not only raises awareness of the burgeoning field of big data, it also aims to provoke conversations on the significant role of data analytics in today’s business and societal context.
The theme for Visual SG this year is “Make Visual!”. This theme takes us back on a journey to discover our intrinsic roots; that of the intrepid explorer, creator and inventor. It encourages all of us to take that first step to discover the magic that is all around us through unbridled curiosity.
Visual SG’s 2017 line up is an eclectic collection of artists and scientists who are all pushing the boundaries to tell their stories of science through visually stunning and engaging media.