GESPIN 2023 Nijmegen
GESPIN 2023 Nijmegen | |
---|---|
Type | Conference |
Categories (tags) | Uncategorized |
Dates | 2023/09/13 - 2023/09/15 |
Link | http://bit.ly/3WgsMfO |
Address | |
Geolocation | 51° 50' 42", 5° 50' 34" |
Abstract due | |
Submission deadline | 2023/03/15 |
Final version due | |
Notification date | 2023/05/31 |
Tweet | Submissions for Gesture and Speech in Interaction 2023 Conference are open. Submissions close 15 Mar 2023. More information and how to submit here: http://bit.ly/3WgsMfO |
Export for iCalendar |
Gesture and Speech in Interaction 2023 Nijmegen:
Details:
GESPIN 2023
"Broadening perspectives, integrating views"
Location: Nijmegen
Date: Wed 13- Fri 15th of September
Paper submission opens: January 10th, 2023
Paper submission deadline: March 15th, 2023
Notification of acceptance/rejection: end of May, 2023
Registration open: TBA
Registration open: TBA
GeSpIn is an interdisciplinary event for researchers working on the interaction between speech and visual communicative signals, such as articulatory, manual, and bodily gestures co-occurring with speech. At GeSpIn 2023 we hope to bring together researchers working on visual signals together with vocalization or speech, from multidisciplinary perspectives in order to exchange ideas and present the cutting edge of their field. This 8th edition of GeSpIn will be held in Nijmegen, the Netherlands and will focus on the theme of “Broadening Perspectives, Integrating Views: Towards General Principles of Multimodal Signaling Systems”.
As such, we encourage researchers working on (multimodal) prosody, social anthropology, philosophy, (psycho)linguistics, psychology, cognitive science, neuroscience, human movement science, computer science (e.g., human-computer interaction), comparative biology, and more to submit their research to address topics such as:
- Do principles of speech-gesture interaction generalize to, or interact with, other multimodal interactions and forms of audiovisual integration (e.g., speech interacting with head gestures or facial signals)?
- What methods in computer science can be used to characterize and synthesize the (temporal) interactions between speech and gesture, within and between agents?
- How is speech-gesture coupling influenced by the immediate dialogic context (e.g., behavior of the interlocutor, or speech act being performed)
- Can multimodal signaling as studied in non-human animals teach us something fundamental about multimodal communication systems that also applies to humans?
- What can cross-linguistic comparisons of speech-gesture interaction teach us about the underlying principles of multimodal coordination?
- Development of gesture-speech coordination: Can general principles of development be identified? Are there sensitive periods and developmental stages?
- What is the role of basic biomechanical or neural processes in visual and auditory signaling and the perception of said multimodal signals?
Please note that all researchers and theoreticians/philosophers working on the interaction between gesture/visual and sound-producing cues (e.g., in terms of pragmatics, prosody, semantics) should feel invited, also if their particular study does not fit these topics exactly.
Organizers
Wim Pouw & James Trujillo (main contacts: wim.pouw@donders.ru.nl/james.trujillo@donders.ru.nl)
Hans Rutger Bosker
Linda Drijvers
Marieke Hoetjes
Judith Holler
Lieke van Maastricht
Asli Ozyurek