the speech recognition virtual kitchen: an initial...

2
The Speech Recognition Virtual Kitchen: An Initial Prototype Florian Metze 1 and Eric Fosler-Lussier 2 1 Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA; U.S.A. 2 Dept. of Computer Science and Engineering, The Ohio State University, Columbus, OH; U.S.A. [email protected], [email protected] Abstract This paper describes a “kitchen” environment to promote com- munity sharing of research techniques, foster innovative exper- imentation, and provide solid reference systems as a tool for education, research, and evaluation with a focus on, but not re- stricted to, speech and language research. The core of the re- search infrastructure is the use of virtual machines (VMs) that provide a consistent environment for experimentation. We liken the virtual machines to a “kitchen” because they provide the in- frastructure into which one can install “appliances” (e.g., speech recognition toolkits), “recipes” (scripts for creating state-of-the art systems), and “ingredients” (language data). In this demo, we show an initial proposal for a community framework based on VM technology used for teaching classes at CMU and OSU. Index Terms: speech recognition, virtualization, educational tools, research infrastructure 1. Introduction Developing and educating a next-generation workforce is be- coming an increasingly challenging task, as building and main- taining a state-of-the-art speech recognition system has moved beyond the ability of a single developer; it is difficult for all but the largest of university laboratories to maintain an end-to- end system. Speech recognizers incorporate knowledge from linguistics, phonetics, acoustics, signal processing, statistical modeling, graph theory, and artificial intelligence. Expecting students to become experts in all of these areas, before at- tempting to work on speech recognition systems, is unrealistic. This poses a bar for developing new research groups, and also makes it difficult for academic institutions without active ASR researchers to integrate ASR projects into the educational cur- ricula, or research projects where ASR is one component of a larger system. What has been missing is a way for academic institutions (and industry) to leverage community resources in maintaining the state-of-the-art, and in ensuring that evaluations and com- parisons are performed on meaningful and relevant tasks, rather than on a plethora of different datasets, tasks, and conditions. While open and closed-source ASR toolkits have become avail- able to create quasi-replicable systems, treating a speech recog- nition system as a “black box” usually results in poor perfor- mance, as some aspect of the system configuration has not been adapted to the requirements of a specific experiment — consti- tuting a significant barrier for use by non-experts, and making it harder for experts to replicate each others’ work. This effort attempts to extend the model of lab-internal knowledge transfer to a community-wide effort through the use of an innovative research infrastructure. We present an infrastructure to facilitate cross-site knowledge sharing: it is es- sentially a way to share “recipes”, demonstrating how to build different kinds of systems, ready-to-run and distributed together with data, log-files, results, etc – everything that is expected from a baseline, and a working environment. Researchers can then modify recipes to run on a different language, a different type of channel, or simply to perform a given test, in order to create a speech recognition system that is flexible and has good performance, and can be integrated in other, bigger projects. The demonstration system we describe here serves two pur- poses: first, it demonstrates how virtual machine technology can be effectively used in the classroom by allowing rapid deploy- ment of resources and consistent environments for classwork. The second purpose of the demo is to engage the community in thinking about a wider-scale community effort to build a com- mon research infrastructure based on VM technology. 2. A Virtual Kitchen with “Recipes” When aspiring chefs enter their training, it is clear that they are not developing their skills in a vacuum. Culinary academies provide a certain amount of infrastructure – that is, a kitchen – in which learning and experimentation can thrive. One can choose to place different kinds of appliances within the kitchen, and can cook with different kinds of ingredients. Recipes, and the skills to use them, are imparted by instructors to initiates. As chefs learn and grow, they take the basic techniques they have learned, and (hopefully) can create new and inspiring dishes by combining ingredients in different ways. In the same way, training in corpus-based language technol- ogy depends on a set of appliances (in the case of ASR, speech recognition toolsets) and ingredients (speech data, transcripts, dictionaries, etc.), and cookbooks containing the recipes. How- ever, established research labs will tend to provide training on one or two appliances, and smaller research labs may try to use an appliance but lack the cookbooks. What is needed in the field is a virtual kitchen that can provide the infrastructure needed for students to try out different techniques. This can lower the bar for entry by smaller research groups, and can enable non-speech labs to use speech technology in their own research beyond a “black-box” approach. The “Speech Recognition Virtual Kitchen” serves as a repository for virtual machines (based on redistributable oper- ating systems such as the Ubuntu Linux derivative) to facilitate the use of toolkits and data. Each virtual machine (VM) will be set up with instructions (typically a collection of scripts) for cre- ating a speech recognition system that is toolkit and data source specific, and will contain the resulting system for reference. The end user is responsible for receiving a distribution of the appro- priate ASR toolkit for the VM from the toolkit provider (by downloading or licensing the software), as well as the speech data. If permitted by the license, data and software can be pre- installed on the end users’ virtual machine fully automatically

Upload: vuthien

Post on 29-Apr-2018

223 views

Category:

Documents


3 download

TRANSCRIPT

Page 1: The Speech Recognition Virtual Kitchen: An Initial …fmetze/interACT/Publications_files/publications/... · The Speech Recognition Virtual Kitchen: An Initial Prototype ... one or

The Speech Recognition Virtual Kitchen: An Initial Prototype

Florian Metze1 and Eric Fosler-Lussier2

1Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA; U.S.A.2Dept. of Computer Science and Engineering, The Ohio State University, Columbus, OH; U.S.A.

[email protected], [email protected]

AbstractThis paper describes a “kitchen” environment to promote com-munity sharing of research techniques, foster innovative exper-imentation, and provide solid reference systems as a tool foreducation, research, and evaluation with a focus on, but not re-stricted to, speech and language research. The core of the re-search infrastructure is the use of virtual machines (VMs) thatprovide a consistent environment for experimentation. We likenthe virtual machines to a “kitchen” because they provide the in-frastructure into which one can install “appliances” (e.g., speechrecognition toolkits), “recipes” (scripts for creating state-of-theart systems), and “ingredients” (language data). In this demo,we show an initial proposal for a community framework basedon VM technology used for teaching classes at CMU and OSU.Index Terms: speech recognition, virtualization, educationaltools, research infrastructure

1. IntroductionDeveloping and educating a next-generation workforce is be-coming an increasingly challenging task, as building and main-taining a state-of-the-art speech recognition system has movedbeyond the ability of a single developer; it is difficult for allbut the largest of university laboratories to maintain an end-to-end system. Speech recognizers incorporate knowledge fromlinguistics, phonetics, acoustics, signal processing, statisticalmodeling, graph theory, and artificial intelligence. Expectingstudents to become experts in all of these areas, before at-tempting to work on speech recognition systems, is unrealistic.This poses a bar for developing new research groups, and alsomakes it difficult for academic institutions without active ASRresearchers to integrate ASR projects into the educational cur-ricula, or research projects where ASR is one component of alarger system.

What has been missing is a way for academic institutions(and industry) to leverage community resources in maintainingthe state-of-the-art, and in ensuring that evaluations and com-parisons are performed on meaningful and relevant tasks, ratherthan on a plethora of different datasets, tasks, and conditions.While open and closed-source ASR toolkits have become avail-able to create quasi-replicable systems, treating a speech recog-nition system as a “black box” usually results in poor perfor-mance, as some aspect of the system configuration has not beenadapted to the requirements of a specific experiment — consti-tuting a significant barrier for use by non-experts, and makingit harder for experts to replicate each others’ work.

This effort attempts to extend the model of lab-internalknowledge transfer to a community-wide effort through theuse of an innovative research infrastructure. We present aninfrastructure to facilitate cross-site knowledge sharing: it is es-sentially a way to share “recipes”, demonstrating how to build

different kinds of systems, ready-to-run and distributed togetherwith data, log-files, results, etc – everything that is expectedfrom a baseline, and a working environment. Researchers canthen modify recipes to run on a different language, a differenttype of channel, or simply to perform a given test, in order tocreate a speech recognition system that is flexible and has goodperformance, and can be integrated in other, bigger projects.

The demonstration system we describe here serves two pur-poses: first, it demonstrates how virtual machine technology canbe effectively used in the classroom by allowing rapid deploy-ment of resources and consistent environments for classwork.The second purpose of the demo is to engage the community inthinking about a wider-scale community effort to build a com-mon research infrastructure based on VM technology.

2. A Virtual Kitchen with “Recipes”When aspiring chefs enter their training, it is clear that they arenot developing their skills in a vacuum. Culinary academiesprovide a certain amount of infrastructure – that is, a kitchen– in which learning and experimentation can thrive. One canchoose to place different kinds of appliances within the kitchen,and can cook with different kinds of ingredients. Recipes, andthe skills to use them, are imparted by instructors to initiates. Aschefs learn and grow, they take the basic techniques they havelearned, and (hopefully) can create new and inspiring dishes bycombining ingredients in different ways.

In the same way, training in corpus-based language technol-ogy depends on a set of appliances (in the case of ASR, speechrecognition toolsets) and ingredients (speech data, transcripts,dictionaries, etc.), and cookbooks containing the recipes. How-ever, established research labs will tend to provide training onone or two appliances, and smaller research labs may try to usean appliance but lack the cookbooks. What is needed in the fieldis a virtual kitchen that can provide the infrastructure needed forstudents to try out different techniques. This can lower the barfor entry by smaller research groups, and can enable non-speechlabs to use speech technology in their own research beyond a“black-box” approach.

The “Speech Recognition Virtual Kitchen” serves as arepository for virtual machines (based on redistributable oper-ating systems such as the Ubuntu Linux derivative) to facilitatethe use of toolkits and data. Each virtual machine (VM) will beset up with instructions (typically a collection of scripts) for cre-ating a speech recognition system that is toolkit and data sourcespecific, and will contain the resulting system for reference. Theend user is responsible for receiving a distribution of the appro-priate ASR toolkit for the VM from the toolkit provider (bydownloading or licensing the software), as well as the speechdata. If permitted by the license, data and software can be pre-installed on the end users’ virtual machine fully automatically

Page 2: The Speech Recognition Virtual Kitchen: An Initial …fmetze/interACT/Publications_files/publications/... · The Speech Recognition Virtual Kitchen: An Initial Prototype ... one or

Figure 1: Screen shot of interactive system developed as part of a class project at CMU: the user controls a character in a virtual worldusing a standard client running on the host computer (bottom-right window), while an automated character is controlled by a systemrunning entirely on the Linux-based VM (top-left window).

(by download from the “kitchen” server, or third-party servers),so that no extra steps are necessary before experimentation.

Under this paradigm, researchers and students can down-load virtual machines, easily reproduce an experiment, changea component (for example, the audio training and testing data)and re-run the experiment, observing the changes.

3. Show & Tell at INTERSPEECH 2012At INTERSPEECH 2012, we will demonstrate a number ofVMs that we have developed in the last year, which showcasethe idea. Figure 1 shows a screenshot of a VM, which has beendeveloped by students in several, subsequent classes and labsat Carnegie Mellon University: in the first semester, studentsimplemented a basic speech recognition system, and an inter-face to Second Life. In the second semester, students improvedthe dialog capabilities of the system, and added a face detectorusing the computer’s web-cam to avoid the need for “push-to-talk”. In the third semester, students added emotion recognitionand performed experiments on a POMDP – all this without theneed for a dedicated machine to host the environment, simplyby passing the VM on from student to student.

A similar approach has been followed at Ohio State, whereexercises in building speech recognition systems were devel-oped around the OpenFST framework [1] as well as the HTKtoolkit [2]. We will demonstrate how VM technology enableseasy distribution of educational materials, including the systemsdiscussed above and other VMs that contain other speech recog-nition experiments and toolkits, such as the open-source Kaldi[3]; we will also discuss our experience in developing and us-ing them. We also plan to engage researchers and educatorsin higher education that typically do not participate in ASR re-search in a workshop at INTERSPEECH 2012 on this topic.

4. OutlookThe “kitchen” has several broader impacts, which we plan todevelop further over the following months. We seek communityinput to tackle some of the issues that have previously hamperedefforts in cross-community sharing: intellectual property issues

often stand in the way of widespread sharing, since differenttoolkits and data may need to be licensed and thus are difficult toaggregate into one standalone package. We will define methodsof distribution that preserve intellectual property rights, while aconsistent environment will allow end users to install open orclosed-sourced systems and data with consistent results.

We will also collect ideas for other scenarios, in which thisinfrastructure will be useful, beyond the ones discussed here.We believe that this infrastructure may be useable by fields otherthan core ASR that are data intensive (synthesis, dialog systems,NLP, computer vision, data mining), and that this may serve asan excellent example for future innovations. Incubating ASR inother fields by providing an easy-to-use, non-trivial research en-vironment will boost the relevance of speech and language tech-nologies in many other fields. The idea also has the potentialto change the way that comparative results are reported: ratherthan just reporting a single number (word error rate) in compar-ing systems, researchers will be able to more easily compare theproducts of two recognition systems.

5. AcknowledgementsThis material is based upon work supported by the National Sci-ence Foundation under a CRI-P grant.

6. References[1] C. Allauzen, M. Riley, J. Schalkwyk, W. Skut, and M. Mohri,

“Openfst: A general and efficient weighted finite-state transducerlibrary,” Proceedings of the Twelfth International Conference onImplementation and Application of Automata, (CIAA 2007), Lec-ture Notes in Computer Science, vol. 4783, pp. 11–23, 2007.

[2] S. Young, G. Evermann, T. Hain, D. Kershaw, G. Moore,J. Odell, D. Ollason, D. Povey, V. Valtchev, and P. Woodland, TheHTK Book. Cambridge University Publishing Department, 2002.[Online]. Available: http://htk.eng.cam.ac.uk

[3] D. Povey, A. Ghoshal, G. Boulianne, L. Burget, O. Glembek,N. Goel, M. Hannemann, P. Motlıcek, Y. Qian, P. Schwarz,J. Silovsky, G. Stemmer, and K. Vesely, “The kaldi speech recog-nition toolkit,” in Proc. ASRU. Big Island, HI; USA: IEEE, 122011.