A picture containing indoor, table, dining table, restaurant

Description automatically generated

AAIS - Advanced Adaptive Intelligent Systems
Funding: KTH Digital Futures
Duration: 2019 - 2024

The aim is to develop adaptive social robots that can understand humans’ communicative behaviour and task-related physical actions, and adapt their interaction to suit. We aim to investigate and demonstrate fluid and seamless adaptation of intelligent systems to users’ context, needs or preferences. [more]

 

Map

Description automatically generated

Adapt - A multi-modal dialogue system for finding apartments in Stockholm
Funding: CTT
Duration: 1998-01-01 - 2004-12-31A

The AdApt project had as its goal to be the foundation for the development and evaluation of advanced multimodala spoken dialogue systems. [more]

 

A close-up of a logo

Description automatically generated with medium confidence

ASPI - Audiovisual to Articulatory Speech Inversion
Funding: EU
Duration: 2005 - 2008

[more]

 

A picture containing text

Description automatically generated

August -
Funding: CTT
Duration: 1998-01-01 - 2000-12-01

The August system was a conversational spoken dialogue system featuring an animated agent called August, whose persona was inspired by August Strindberg, the famous Swedish 19th century author. [more]

 

BabyRobot - Child-Robot Communication and Collaboration
Funding: EU
Duration: 2015-12-01 - 2018-12-31

The main goal of the project is to create robots that analyze and track human behavior over time in the context of their surroundings using audio-visual monitoring in order to establish common ground and intention-reading capabilities. [more]

 

A close up of a spider

Description automatically generated with low confidence

BioASU - Biologically inspired statistical methods for flexible automatic speech understanding
Funding: VR
Duration: 2010 - 2014

The project will develop machine learning methods for speech understanding that more closely resemble the biological approach to learning. [more]

 

A picture containing indoor, cup, tableware, dishware

Description automatically generated

BLARK/SNK - An Infrastructure for Swedish language technology
Funding: VR
Duration: 2007-01-01 - 2008-12-31

Research and development on language technology needs an infrastructure of publicly available and standardized basic resources. [more]

 

Bostran - The bowed string - bow motion analysis and influence of bow properties
Funding: VR
Duration: 2002-01-01 - 2006-12-31

The basic physical process for the sound generation in the bowed instruments is a stick-slip motion triggered by the wave motion on the string. The project comprises three sub-studies related to the modelling of the bow-string interaction and its control mechanisms:

more]

 

BUG - Children's voice development and gender differences
Funding: VR
Duration: 2003 - 2007

The project investigates some developmental aspects of children's voices. [more]

 

A picture containing text, underpants

Description automatically generated

CALATEA - Computer-Animated LAnguage TEAchers
Funding: VR
Duration: 2009-01 - 2011-12

The key contribution of CALATEA is to use computer-animated virtual teachers and speech technology in combination toimprove self-studies in second language learning. [more]

 

Graphical user interface, application, website

Description automatically generated

CALST - Computer-Assisted Listening and Speaking Tutor
Funding: NTNU + Norgesuniversitetet
Period: 2009-08-01 - 2012

The project aims at developing a computer program that will be used to train Norwegian as a second language. [more]

 

A picture containing text, clipart

Description automatically generated

CAVE - Caller Verification
Funding: EU
Duration: 1996 - 1998

The CAVE project focussed on the development and testing of speaker verification systems for use in telephone applications like calling-card or financial services. [more]

 

A stack of books

Description automatically generated with medium confidence

Centlex - Centlex
Funding: CTT
Duration: 2000 - 2007

CentLex is the central lexicon resource produced and maintained by the Centre for Speech Technology. [more]

 

CHIL - Computers in the Human Interaction Loop
Funding: EU
Duration: 2004-01-01 - 2007-08-31

The goal of the CHIL project is to make everybody’s daily life easier. [more]

 

A group of people in a room

Description automatically generated with low confidence

CityCrowd - Personalized spatially-aware dialogue systems
Funding: VR
Duration: 2014 - 2019

A project exploring the intersection between spoken dialogue systems, geographic databases and crowd-sourcing. [more]

 

CLARIN - Common Language Resources and Technology Infrastructure
Funding: EU
Duration: 2008-01-01 - 2010-12-01

The CLARIN project is a large-scale pan-European collaborative effort to create, coordinate and make language resources and technology available and readily useable. [more]

 

COIN - Co-adaptive human-robot interactive systems
Funding: SSF (Stiftelsen för Strategisk Forskning)
Duration: 2016-05-01 - 2020-12-31

The main goal is to develop a systematic, bi-directional short- and long-term adaptive framework that yields safe, effective, efficient, and socially acceptable robot behaviors and human-robot interactions. [more]

 

Connected - Context-aware speech synthesis for conversational AI
Funding: VR
Duration: 2019 - 2024

Speakers use a variety of strategies to continuously adjust their speech delivery in response to the situational context of the conversation. In this project the aim is to develop a context-aware conversational speech synthesizer that allows for implicit control over the manner of speaking to cater for the communicative needs... [more]

 

A picture containing text, person, indoor

Description automatically generated

CORDIAL - Coordination of Attention and Turn-taking in Situated Interaction
Funding: VR
Duration: 2014-01-01 - 2018-12-31

[more]

 

A picture containing text, blackboard, sign

Description automatically generated

COST 278 - Spoken language interaction in telecommunication
Funding: EU
Duration: 2001 - 2005

[more]

 

COST2103 - Advanced Voice Function Assessment
Funding: EU, KTH
Duration: 2006-09-21 - 2010-12-18

The main objective of the Action is to combine previously unexploited techniques with new theoretical developments to improve the assessment of voice for as many European languages as possible, while acquiring in parallel data with a view to elaborating better voice production models.

Progress in the clinical assessment and enhancement of... [more]

 

Text

Description automatically generated with low confidence

DEAL - Role-playing and Dialogue System for Second Language Learners
Funding: CTT
Duration: 2006-12-12 - 2010-12-12

DEAL is a multidisciplinary research platform where challenges and potential benefits of combining elements from computer games, dialogue systems and language learning can be explored. [more]

 

A picture containing text

Description automatically generated

DUMAS - Dynamic Universal Mobility for Adaptive Speech Interfaces
Funding: EU Duration: 2001 - 2004

The DUMAS (Dynamic Universal Mobility for Adaptive Speech Interfaces) project develops multi-lingual speech-based applications, focusing on intelligent and ambient interaction management with a special emphasis on naturalness that takes into account the user's personalised needs. Future communication with mobile electronic systems requires dynamic and adaptive capabilities: systems that can learn... [more]

 

A close-up of a machine

Description automatically generated with low confidence

EACare - Embodied Agent to support elderly mental wellbeing
Funding: SSF (Stiftelsen för Strategisk Forskning)
Duration: 2016-04-12 - 2020-12-31

The main goal of the proposed multidisciplinary research is to develop a robot head with communicative skills capable of interacting with elderly people at their convenience. [more]

 

A yellow bicycle with a handlebar

Description automatically generated with low confidence

ENABL - ENAbler for computer-Based vocational tasks with Language and speech
Funding: EU
Duration: 1997 - 2000

One main objective of the ENABL project was to provide access by voice, via speech recognition, to an engineering design system, ICAD. Since persons with manual disabilities also often have a vocal disability, a part of the ENABL project was dedicated to evaluating the performance of dysarthric speech on a speech recognition engine.... [more]

 

Error - Error and Miscommunication in Human-Computer Dialogue Systems
Funding: VR
Duration: 2002-01-01 - 2004-12-31

The aims of the current project are to analyze miscommunication and to investigate means of improving error handling in spoken and multimodal dialogue systems. Generally, the goal is to attain a greater understanding of the mechanisms and structures underlying problematic human-interaction. By studying users’ linguistic behavior as they interact with... [more]

 

Logo, company name

Description automatically generated

EUNISON - Extensive UNIfied-domain SimulatiON of the human voice
Funding: EU, KTH
Duration: 2013-03-01 - 2016-02-28

This project seeks to build physics-based simulations of the human voice that are more detailed and more complete than before, using numerical models that have been validated against mechanical replicas. The simulations will be accessible for control in the mechanical, neuromotor and phonetic domains. The project engages seven research groups... [more]

 

A picture containing automaton

Description automatically generated

FACT - Factories of the Future: Human Robot Cooperative Systems
Funding: SSF (Stiftelsen för Strategisk Forskning)
Duration: 2015-12-01 - 2020-12-31

The focus of FACT is on providing safe and flexible feedback in unforeseen situations, enhancement of human-robot cooperation and learning from experience. [more]

 

Chart

Description automatically generated

FonaDyn - Phonatory Dynamics and States
Funding: VR
Duration: 2011-01-01 - 2018-12-31

The voice has several non-linear and context-dependent mechanisms that can give rise to distinct phonatory states. We submit that much of the observed variability in objective voice metrics results from the influence of such states, and will attempt to account for some of them, using a state-based analysis paradigm. [more]

 

GenDial - Modelling utterance generation in conversational dialogue systems
Funding: VR
Duration: 2008 - 2010

The GenDial project spans research in several disciplines including computer science, human-human and human-machine spoken interaction and language technology. Our long term research goal is to develop a new class of conversational spoken dialog systems which to a large extent follows the principles of human-human interaction. A natural part of human conversation... [more]

 

GetHomeSafe - Extended Multimodal Search and Communication Systems for Safe In-Car Application
Funding: EU
Duration: 2011-12 - 2014-12

The aim of the proposed project is to develop a system for safe information access and communication while driving. [more]

 

GROG - Breaks and grouping - structuring of speech in different communicative environments
Funding: VR
Duration: 2002 - 2004

The project is a cooperation between Umeå University and KTH. The project is headed by Eva Strangert Umeå. The purpose of this project is to model the structuring of speech in terms of prosodic boundaries and groupings in various communicative situations. The modeling aims at a structured and optimized description of... [more]

 

A picture containing graphical user interface

Description automatically generated

GULAN - A System for Teaching Spoken Dialogue Systems Technology
Funding: CTT
Duration: 1996 - 1998

The aim of this work has been to put a fully functioning spoken dialogue system into the hands of the students as an instructional aid. [more]

 

HaH - Hearing at home
Funding: EU
Duration: 2006-12-01 - 2009-05-30

The Hearing at Home (HaH) project researches and develops next generation assistive devices that allow the growing number of hearing impaired – which predominantly includes the elderly – equal participation in communication. The project aims to empower the hearing disabled to play their own role in the information society. [more]

 

Hi-Fi Voice - Hi-Fi Voice: the fine structure of the voice signal at high frequencies
Funding: VR
Duration: 2008-01-01 - 2010-12-31

The human voice has weak but perceptually important frequency components at frequencies above 4-5 kHz. This frequency range is studied from production and perception perspectives. [more]

 

A picture containing scissors

Description automatically generated

Higgins - Higgins Funding: CTT
Duration: 2003 - 2010-12

Higgins is a framework project developing modules and techniques for flexible, robust and human-like dialogue system design. [more]

 

IGLU - Interactive Grounded Language Understanding
Funding: EU CHIST-ERA Duration: 2017-01-01 - 2021-12-31

Language is an ability that develops in young children through joint interaction with their caretakers and their physical environment. At this level, human language understanding could be referred as interpreting and expressing semantic concepts (e.g. objects, actions and relations) through what can be perceived (or inferred) from current context in the environment. Previous work... [more]

 

A picture containing text, transport, gear

Description automatically generated

InkSynt - Incremental Text-To-Speech Conversion
Funding: VR
Duration: 2014-01-01 - 2018-01-01

We will develop an incremental text-to-speech converter (TTS), which can be used in dynamically changing situations. In the project we will collect speech databases of how people read incrementally displayed text aloud, which will serve as the basis for the development of methods for incremental TTS with the correct prosody. We will... [more]

 

IURO - Interactive Urban Robot
Funding: EU
Duration: 2010-02-01 - 2013-01-31

The goal of IURO project is to develop a robot that can engage in information-gathering face-to-face interactions in multi-user settings. [more]

 

Icon

Description automatically generated

Jindigo - Java-based Incremental Dialog Framework
Funding: CSC, Young Faculty
Duration: 2009-11 - 2010-12

Jindigo is a framework for developing and experimenting with incremental spoken dialog systems, developed at the Department of Speech Music and Hearing, KTH. [more]

 

Kobra - Knowledge-rich speaker adaptation for speech recognition
Funding: VR
Duration: 2007-01-01 - 2009-12-31

[more]

 

A person wearing headphones

Description automatically generated with medium confidence

Lipread - Lipreading to support Foreign and Second Language Learning by Hearing Impaired and Normally Hearing Persons
Funding: EU
Duration: 2012-01-01 - 2013-12-31

The objective of the project is to develop a syllabus and an e-learning environment to teach lipreading of foreign and second languages. Our target groups: deaf and hard of hearing adults, their therapists and teachers, and normally hearing language learners. [more]

 

A picture containing sport, player, male

Description automatically generated

MASSIVE - Large-scale massively multimodal modelling of non-verbal behaviour in spontaneous dialogue
Funding: VR
Duration: 2011-01-01 - 2013-12-31

The aim is to provide a large-scale kinematic database based on motion capture of human conversational behaviour, as well as to build statistical models of multimodal non-verbal behaviour in dialogue. [more]

 

A picture containing automaton, engine, miller

Description automatically generated

MirrorBot - Data-driven Modelling of Interaction Skills for Social Robots
Funding: SRA/KTH
Duration: 2016 - 2021

A project aiming to use robot-mediated human interaction as a means of collecting data for modelling social signals in human robot interaction [more]

 

MonAMI - Mainstreaming on ambient intelligence
Funding: EU
Duration: 2006-09 - 2010-08

The overall objective of MonAMI is to mainstream accessibility in consumer goods and services, including public services, through applied research and development, using advanced technologies to ensure equal access, independent living and participation for all in the Information Society. [more]

 

A picture containing plant

Description automatically generated

MULTISENSE - Multimodal and multisensory interfaces for interaction with muscolo-skeletal models
Funding: EU
Duration: 2002-07-01 - 2005-06-30

The project concerned the visualization of, and interaction with, data related to musculo-skeletal structures by means of multi-modal and multi-sensorial interfaces. [more]

 

A picture containing text, container

Description automatically generated

MUSCLE - Multimedia Understanding through Semantics, Computation and Learning
Funding: EU
Duration: 2004 - 2007

MUSCLE is an EC-sponsored Network of Excellence that aims at establishing and fostering closer collaboration between research groups in multimedia datamining and machine learning. The Network integrates the expertise of over forty research groups working on image and video processing, speech and text analysis, statistics and machine learning. The goal... [more]

 

A picture containing doll, toy, clipart

Description automatically generated

OLGA -
Funding: NUTEK/VINNOVA
Duration: 1996 - 1997

The ultimate aim of the Olga project is to build an easy to use dialogue interface to a database for the general public. The user interacts with a talking animated agent Olga, visible on a display, either with text, speech or a pointing device. The demonstrator application is a database... [more]

 

A picture containing icon

Description automatically generated

OLP - Ortho-Logo-Paedia
Funding: EU
Duration: 2002 - 2004

The OLP (Ortho-Logo-Paedia) project aims at improving the quality of life of persons with articulatory impairments by applying a novel technological aid to speech therapy, by integrating this training with speech recognition technology and by making these facilities available over the Internet. [more]

 

Diagram

Description automatically generated with medium confidence

PAMIR - Computational Modelling of Perceptual Music Features
Funding: VR
Duration: 2013 - 2017

Today, computers and the Internet are commonly used for all aspects of music culture from production to listening. When a typical online music database contains 5 million songs, new computer tools are needed for characterizing, and indexing music audio. This is the focus of the new research field Music... [more]

 

ParSys - Data-driven Parametric Speech Synthesis
Funding: CTT
Duration: 1997 - 2007

The ParSyn (Parametric Synthesis) project is focused on two challenges: 1) speech synthesis based on both datadriven and rule-based parametric modeling; 2) speech synthesis of spontaneous speech including disfluencies. [more]

 

PF-Star - Preparing future multisensorial interaction research
Funding: EU, KTH
Duration: 2002 - 2004

PF-STAR has contributed to establish future activities in the field of Multisensorial and Multilingual communication (Interface Technologies) on firmer bases by providing technological baselines, comparative evaluations, and assessment of prospects of core technologies, which future research and development efforts can build from. The project has addressed three crucial areas: technologies for... [more]

 

PICASSO - Pioneering Caller Authentication for Secure Service Operation
Funding: EU
Duration: 1998 - 2000

The PICASSO project developed and tested secure telematics transaction services using caller authentication by voice. Specch recognition and speaker verification/identification technologies were integrated to provide interfaces that are at once intuitive, easy to use and secure against intruders. PICASSO built on the results of the CAVE project. New telephone services... [more]

 

RealsimPLE - Combining physical reality with simulations in pedagogical laboratory experiments
Funding: WGLN (Wallenberg Global Learning Network)
Duration: 2005-09-15 - 2007-12-15

[more]

 

A picture containing text, person, indoor, television

Description automatically generated

RoboLearn - Online learning of turn-taking behaviour in spoken human-robot interaction
Funding: VR
Duration: 2016 - 2018

In this project, we will investigate how a robot’s turn-taking behaviour can be learned from experience by interacting with people. [more]

 

Icon

Description automatically generated with low confidence

SAMINK - Incremental processing in multimodal conversational systems
Funding: VR
Duration: 2012-01-01 - 2015-12-31

The aim of this project is to understand how conversational systems can engage in spoken face-to-face interaction in real-time. [more]

 

Icon

Description automatically generated

SAMPAUS - Classifying and deploying pauses for flow control in conversational systems
Funding: VR
Duration: 2012-01 - 2016-12

The project investigates how dialog system can employ pauses and fillers to help users better understand system utterance structure. [more]

 

SAMPROS - Prosody in conversation
Funding: RJ (Bank of Sweden Tercentenary Foundation)

The project investigates how people talking to each other jointly decide who should speak when, and the role of prosody in making these joint decisions. [more]

 

A picture containing text, crossword puzzle

Description automatically generated

SAMRYTM - The rhythm of conversation
Funding: VR
Duration: 2010 - 2012

The project Rhythm of conversation investigates how a set of rhythmic prosodic features contributes to the joint interaction control in conversations. [more]

 

Circle

Description automatically generated with medium confidence

SAMSYNT - Introducing interactional phenomena in speech synthesis
Funding: VR
Duration: 2010-07-01 - 2013-12-31

The project will develop and verify ways of including interactional phenomena in speech synthesis, resulting in well-described and tested methods for synthesizing these phenomena in such a way that they can be employed to recreate human interactional behaviour. [more]

 

A picture containing text, mammal

Description automatically generated

SAVIR - Situated Audio Visual Interaction with Robots
Funding: SRA/KTH
Duration: 2010-01 - 2013-12

The projects investigate how a robot can improve its visual scene understanding by engaging in spoken dialogue with a human. [more]

 

A close-up of a violin

Description automatically generated with medium confidence

Secrets of the violin 
Duration: 2001 - 2010-12

Experimental models and electro-dynamical measurements on real violins. [more]

 

SID - Sonic Interaction Design
Funding: ESF (European Science Foundation)
Duration: 2007-01-01 - 2009-12-31

Sonic Interaction Design is the exploitation of sound as one of the principal channels conveying information, meaning, and aesthetic/emotional qualities in interactive contexts. [more]

 

SIFT - Separating Intonation from Tone
Funding: Bank of Sweden Tercentenary Foundation

[more]

 

SIMULEKT - Simulating Intonational Varieties of Swedish
Funding: VR
Duration: 2007 - 2010

This project represents collaboration between Lund University (Gösta Bruce and Susanne Schötz) and KTH (Björn Granström, Jonas Beskow, Joakim Gustafson and Laura Enflo) The primary goal of the research project is to produce more precise and thorough knowledge about phrase and utterance prosody in some major regional varieties of Swedish.... [more]

 

SingHab - Advanced objective measures of voice in professional female singers
Funding: Fondation Alma et Baxter Ricard, Canada
Duration: 2005 - 2009

The long-term goal of this research project is to improve the evaluation of the singing voice in clinical settings. A tailored assessment of the singing voice is crucial to the effectiveness of rehabilitation work. The result of this research will be to provide useful and clinically accessible evaluation possibilities for... [more]

 

A yellow star with a white background

Description automatically generated with low confidence

SkAT-VG - Sketching Audio Technologies using Vocalizations and Gestures
Funding: EU
Duration: 2014-01-01 - 2016-12-31

The SkAT-VG project will develop a system that enables sound designers to sketch sounds using their own voice and gestures as input. The project partners are IUAV in Venice (coordinator Davide Rocchesso), IRCAM in Paris, KTH in Stockholm, and the company GENESIS S.A. in Aix-en-Provence. [more]

 

Graphical user interface, text

Description automatically generated

Song - Sundberg's Voice Science
Funding: KTH CSC - TMH Sundberg
Duration: 1987 - 2014-12

Kulning - Hard rock - Twang - Belting - Chest/Falsetto - Whisper - High pitch singing - Text intelligibility - MRI [more]

 

SpeakingUp - Making spoken cultural heritage accessible for research
Funding: RJ (Bank of Sweden Tercentenary Foundation)
Duration: 2017-01 - 2021-12

The overall aim of the project is to make Sweden's archival treasure of recorded speech accessible for HS research. SpeakingUp is conducted by the Institute for Language and Folklore (ISOF), KTH and Digisam. [more]

 

Icon

Description automatically generated

SPEDIAL - Spoken Dialogue Analytics
Funding: EU
Duration: 2014-01-01 - 2015-12-31

[more]

 

SpeechDat - Databases for the Creation of Voice Driven Teleservices
Funding: EU
Duration: 1996-03-01 - 1998-02-28

Swedish telephone speech database containing 5000 speakers over the landline telephone network and 1000 speakers recorded over mobile telephone networks [more]

 

A picture containing text

Description automatically generated

SPEECON - Speech Driven Interfaces for Consumer Applications
Funding: EU
Duration: 2001-01-01 - 2003-12-31

Recording of 550 adult speakers and 50 child speakers were recorded with 4 different microphones in different environments, such as office, public place, home, outdoor and car. [more]

 

Diagram

Description automatically generated

Spontal - Multimodal database of spontaneous speech in dialog
Funding: VR, KFI
Duration: 2007-07-01 - 2010-12-31

Massively multimodal (HD video, hifi sound, and motion capture) database of spontaneous speech in dialog. [more]

 

SVP Voice - Detailed multiphysics simulation of human voice production with neural control - a feasibility study
Funding: KTH CSC
Duration: 2011-09-01 - 2012-06-30

This project is a feasibility study in which we examine whether it is possible to make a unified-domain numerical simulation of human voice production that covers the mechanical, fluid and acoustic phenomena involved; and also attempts to control the simulation using representations of (simulated) muscle activation. [more]

 

SYMIR - Musical understanding for music information retrieval
Funding: VR
Duration: 2006-01-01 - 2009-12-31

[more]

 

Graphical user interface, application

Description automatically generated

SynFace -
Funding: EU
Duration: 2001 - 2005

SynFace is a technology that allows an animated talking face to be controlled by a speech signal in real time. [more]

 

The science and technology of music -
Funding: STINT (Swedish Foundation for International Cooperation in Research and Higher Education)
Duration: 2002-09-01 - 2007-08-31

[more]

 

TIG - Timing of intonation and gestures in spoken communication
Funding: RJ (Bank of Sweden Tercentenary Foundation)
Duration: 2012-08 - 2017-01

The goal of the project is to understand timing relationships between intonation and gesture in spontaneous speech. This will be investigated through semi-automatic extraction of co-speech gestures from a large and varied dataset (audio, video, motion-capture), and analysis of function and synchronization of speech and gestures. [more]

 

TIVOLI - Sign learning via game-based interaction
Funding: PTS - Post och Telestyrelsen
Duration: 2011-09 - 2013-09

TIVOLI aims to create learning application for sign language signs, in the form of a computer game featuring sign recogntion via webcam and a signing avatar. The target group is children with communication disorders. [more]

 

A picture containing clipart

Description automatically generated

VariQ - Intonational variation in questions in Swedish
Funding: VR
Period: 2010-01-01 - 2012-12-31

This project investigates questions in dialogue. What is a question, and what makes it into one? [more]

 

VATS - What turns speech into conversation?
Funding: VR
Duration: 2007-01-01 - 2009-12-31

The project What turns speech into conversation? (Vad gör tal till samtal?) investigates features that are specific to conversations among humans - the very features that turn speech into conversation - such as how the speakers know when to speak and when not to. [more]

 

VEMUS - Virtual European music school
Funding: EU
Duration: 2005-01-01 - 2008-09-30

VEMUS (Virtual European Music School) aims at developing an open, highly interactive, and networked multilingual music tuition framework for popular instruments such as the trumpet, the flute, and the clarinet. [more]

 

Ville - The Virtual Language Teacher
Funding: GSLT Graduate School of Language Technology
Duration: 2004-01-01 - 2010-12

The goal of this project is to develop and test a new type of appliction for computer assisted language learning (CALL) and computer assisted pronunciation training (CAPT). [more]

 

A person wearing a virtual reality headset

Description automatically generated with medium confidence

VirtualRobot - Exploring situated interaction with social robots using augmented reality
Funding: SRA/KTH
Duration: 2017 - 2019

In this project, we aim aim to explore the use of Augmented Reality (AR) to investigate the impact of multimodal behaviour (speech, facial expression, full-body motions, conversational formations) and embodiment on turn-taking and joint attention in human-robot interaction. [more]

 

VoiceMech - Numerical modelling of the non-vibratory biomechanics of human voice production
Funding: VR
Duration: 2014-01-01 - 2016-12-31

In the VoiceMech project, a 3D biomechanical model of the non-acoustic parts of the human voice is in development, mainly to serve as a controlling interface for the aerodynamic/acoustic models that are under construction in the Eunison project. VoiceMech is conducted in collaboration with several others, notably the ArtiSynth project... [more]

 

A pair of headphones

Description automatically generated with low confidence

VoxLog - VoxLog: portable voice analyzer
Funding: NUTEK/VINNOVA
Duration: 2010-05-01 - 2013-12-30

A new wearable voice+noise dosimeter has been developed. The project has assessed the validity, usability and commercial potential of the new device. [more]

 

A screenshot of a video game

Description automatically generated with medium confidence

Waxholm - The Waxholm dialog project
Funding: NUTEK/VINNOVA
Duration: 1991 - 1995

Waxholm was a spoken dialogue system for retrieving information about the ferryboat services in the Stockholm archipelago. [more]

 

Logo, company name

Description automatically generated

WikiSpeech -
Funding: PTS - Post och Telestyrelsen
Duration: 2016 - 2018

An open source project that will draw on crowdsourced contributions to make Wikipedia more accessible by adding text-to-speech synthesis that will enable users of the online encyclopedia to have portions of the text read out to them. [more]

 

Logo, company name

Description automatically generated

Wikispeech2 - A speech corpus collector for a more accessible Wikipedia through Wikispeech
Funding: PTS - Post och Telestyrelsen
Duration: 2019 - 2022

The speech corpus collector is an initiative of Wikimedia Sweden, KTH and STTS AB to build a valuable resource for speech technology in Swedish and to make Wikipedia more accessible. the purpose is to develop a set of tools to help using Wikipedia volunteers collect large quantities freely licensed recordings... [more]

 

Workplace noise - The effects of workplace noise on speaking voice
Funding: FAS (Swedish Council for Working Life and Social Research)
Duration: 2003-01-01 - 2005-12-31

[more]

 

A picture containing text, clipart

Description automatically generated

WWAAC - World Wide Augmentative and Alternative Communication
Duration: 2001 - 2004

The project will make several Internet services accessible for persons with language and/or cognitive impairments, including symbol users and a part of the elderly population. An universal coding system will be developed for message encoding which can be used for eMail, eChat or enhancing web-page information. This will be promoted... [more]