We got two full papers accepted at the next ACM International Conference on Multimodal Interactions:
Both articles revolve around mental health issues and, in particular, around the possibility to infer the mental conditions of people from their observable behaviour. In the case of depression, the approach takes into account what people say and how they say it. The focus is on the possibility to detect depression using only a few seconds of speech, something that it is important because depressed people find it difficult to speak for long time. In the case of attachment, the proposed approach analyses the way children use an interactive system and the results show that secure children tend to better interact with the design of the system, in line with the expectations of Attachment Theory.
It is a great pleasure to say that the publications have been obtained with the help of one of my most senior PhD students (Nujud Aloshban) and an experienced postdoctoral researchers with whom I have been collaborating for 5 years.
A paper submitted in collaboration with Fuxiang Tao and Anna Exposition has been accepted for presentation at Interspeech 2020 (the conference is rated A by CORE):
F.Tao, A.Esposito and A.Vinciarelli, “Spotting the traces of depression in read speech: An Approach Based on Computational Paralinguistics and Social Signal Processing“, Proceedings of Interspeech, 2020.
The most interesting aspect of the article is that it uses a classifier to show that three particular behavioural cues (reading speed, number of pauses and average length of the pauses) can improve by close to 20 points the accuracy of classifier that discriminates between depressed and non-depressed readers. In other words, the paper proposes the use of a classifier as a means to test the association between depression and behavioural cues, in alternative to the common approaches used in psychology for the same purpose (e.g., correlational analysis). In addition, the paper shows that neuroscience findings, in particular the interplay between depression and neural mechanisms underlying language processing, can help to improve technology.
My recent work on Visual Social Distancing:
M.Cristani, A.Del Bue, V.Murino, F.Setti, A.Vinciarelli
The Visual SocialDistancing Problem
IEEE Access, Vol. 8, pp. 126876-126886, 2020
has been featured in an article by ANSA, the most important press agency in Italy:
The article proposes an approach for the analysis of the way people gather in public spaces that takes into account not only physical, but also social aspects of interpersonal distance. The work is a nice example of how multiple disciplines (Computer Vision and Social Signal Processing in this case) can collaborate to address a technological problem.
I had the great pleasure to participate as a speaker in the Boundless Podcast, an initiative led by Richard Foster Fletcher to spread awareness about Artificial Intelligence and its latest developments. The discussion revolved around socially intelligent agents and, in particular, around their potential as a means to improve the life of people in the world.
The full recording of my interview is available at the following link:
Alessandro Vinciarelli at the Boundless Podcast
while the many other interesting episodes of the series can be accessed here.
I have been mentioned in a recent article posted in the blog of the “Scientific American”, where I have been asked to give my opinion about the role of context in the recognition of the emotions and, more in general, in the understanding of human behaviour:
The article provides a very extensive account of the available opinions about the matter and it confirms how elusive the concept of context is. Any persona that has been working in technology for at least a decade has certainly passed through successive waves of “context-aware” or “context-sensitive” approaches, but the meaning of the word context remains overall uncertain. The blog post of the Scientific American tries to shed some light by asking a large number of experts what they think about and the result is highly instructive. From my point of view, the most important aspects are the interplay between the analysis of context and the application of multimodal technologies on one side and, on the other side, the increasingly growing amount of information that can be gathered through the pervasive network of sensors and technologies that constellates out life.
On September 3rd, I have successfully co-chaired with Olga Perepelkina the first International Workshop on Social and Emotion AI for Industry, in collaboration with Neurodata Lab:
The workshop was organised in Cambridge, in conjunction with the IEEE International Conference on Affective Computing and Intelligent Interaction and, according to the organisers of the conference, it has attracted the largest number of participants among the different workshops organised in the same day.
The overall goal of the workshop was to gather academic researchers and industry practitioners active in social and emotion AI, the domains aimed at developing technologies that deal with affective and interactional aspects of their users. The program included refereed papers (50% acceptance rate) as well as keynote presentations given by speakers active in both academia and industry.
We learned about the use AI driven technologies that help people affected by different neurological diseases thanks to Emteq and its CEO Charles Nduka, about the effectiveness of AI technologies for the affective analysis of audio thanks to Audeering and its CEO Dagmar Schuller, and about the activities of Microsoft about empathic computing thanks to Daniel McDuff.
From an academic point of view, the workshop has benefited from the talks by Catherine Pelachaud (an introduction to impression management fore virtual agents), Antonio Camurri (an introduction to his new project on next generation motion capture systems) and Andrew McStay (an ethical perspective on AI).
Since I study nonverbal communication, I have always been fascinated by our innate tendency to leak information about ourselves through our behaviour. The new article “What and ‘Ehm’ leaks about you: Mapping Fillers into Personality Traits with Quantum Evolutionary Feature Selection Algorithms“, just accepted for publication by the IEEE Transactions on Affective Computing, is further step in such a direction. The work shows that one of our most minor behavioural cues, the short vocalisations like “ehm” and “uhm” that we utter when we do not know what to say next, conveys information about our personality traits. The results have been obtained by analysing around 3,000 fillers uttered by 120 persons.
The work has been done in collaboration with Mohammad Tayarani, former postdoc of mine that recently obtained a fellowship at the University of Hertfordshire, and Anna Esposito, one of my very first mentors and longtime friend and colleague. Mohammad has developed a new feature selection approach that allows one to spot the features (physical measurements we extract from fillers) most likely to account for personality. Anna has contributed with her deep knowledge of speech processing and her major experience in interdisciplinary work between Computing Science and Psychology. The IEEE Transactions on Affective Computing are in the top 5% of the Scimago Ranking and have an impact factor of 6.28.
I have published a new article that analyses the preferences of roughly 5,000 prospective students attending one of the Open Days organised by the University of Glasgow:
The full citation of the paper is as follows: A.Vinciarelli, W.Riviera, F.Dalmasso, S.Raue, C.Abeyratna, “What Do Prospective Students Want? An Observational Study of Preferences About Subject of Study in Higher Education“, in “Innovations in Big Data Mining and Embedded Knowledge”, A.Esposito, A.M.Esposito, L.C.Jain (eds.), pp 83-97.
The article has been written in collaboration with Bizvento, a start-up founded by a few students of the School of Computing Science of our University, audit was supported by The Data Lab. It is the very first time I write a paper with a sociological slant (it was a chance for me to read some sociological literature about the relationship between family condition and education level that people attain). In parallel, it was an interesting exercise on how much information can be obtained by crossing multiple publicly available repositories and data.
I am having the great pleasure of attending WIRN (Italian Neural Network Workshops) where I have been given the chance to give a keynote talk. The workshop gathers experts on Artificial Intelligence from all over Italy (and the rest of the world) to discuss about the latest advances in machine learning and similar technologies. For me it is a chance to have interesting discussions with the members of a vibrant community and, most important, to meet a lot of talented young researchers that do their first steps in the friendly environment that WIRN provides. This year, my talk has been on attachment and its relationship with Artificial Intelligence (the problem is to automatically diagnose the attachment condition of a child) and Human Computer Interaction (the problem is to find traces of the attachment condition in the way children interact with software).
According to Arnet Miner, the search engine supported by the Chinese Government, I am one of the top 100 researchers in Multimedia in the decade 2007-2017. According to the message they sent me (see below), this results from the analysis of 230 millions of documents collected over 368,402 venues.
Dear Alessandro Vinciarelli,
We are pleased to inform you that you have been recognized as a Most Influential Scholar for your outstanding and vibrant contributions to the field of Multimedia. Congratulations!
In 2018, the AMiner Most Influential Scholar List names the world’s top-cited research scholars from the fields of Artificial Intelligence (AI). The list is conferred in recognition of outstanding technical achievements with lasting contribution and impact to the research community. The 2018 winners are among the most-cited scholars from the top venues of their respective subject fields in recent ten years (between 2007 and 2017). Recipients are automatically determined by a computer algorithm deployed in the AMiner system that tracks and ranks scholars based on citation counts collected by top-venue publications.
AMiner (https://aminer.org) is a free online service for academic social network analysis and mining. As of 2018, the system has collected information on over 136 million researchers, 230 million publication papers, and 368,402 venues. The system has been in operation on the Internet since 2006 and has been visited by nearly 8.32 million independent IP accesses. It provides various search/mining services for publishers, NSFC, and research venues such as ACM/IEEE Transactions, ACM SIGKDD, ACM WSDM, and IEEE ICDM. Further details can be found online at the AMiner Wikipedia page: https://en.wikipedia.org/wiki/Arnetminer.
As part of the recognition, your research profile extracted from the AMiner database is being featured this month (March, 2019) on AMiner homepage. The full list of the most influential scholars can be found here: https://www.aminer.cn/ai10/multimedia. For your information, you can sign up for an AMiner account and keep your personal profile and publications updated (https://aminer.org/profile/5405e3cbdabfae450f3de5dd).