Editor: Chris Wellekens
Dear Members,
Our Distinguished Lecturers program is quite successful and you will find hereunder the trip reports of prof.Chin-Hui Lee and Mark Swerts.
The validity of job offers published in ISCApad are checked every month and withdrawn as soon as the positions are filled.
Please do not hesitate to send me all important informations concerning the life of our speech community you want to advertise in ISCApad.
The board wishes you a wonderful end of year!
Professor em. Chris Wellekens
Institut Eurecom France
Back to Top
ISCA News
-
Report from our Distinguished Lecturer Professor Chin-Hui Lee
ISCA DL 2007 Trip Report: prepared by Chin-Hui Lee (Nov. 19, 2007)
- Leaving ATL at dawn on Nov. 7 and returning to ATL at night on Nov. 18
- First Talk in Singapore:
- Location: Nanyang Technological University
- Time: 9:30-11:30am on Nov. 9, 2007
- Attendance: 50 people from academia, industry and government
- Topic: Fundamentals and Technical Challenges in Speech Recognition
- Impact: In addition to the increasing speech research effort led by Dr. Haizhou Li at Institute for Infocomm Research, the lecture also raised interest in government agencies to invest more funding in speech R&D.
- Second Talk in Manila:
- Location: De La Salle University
- Time: 9:10-11:00am on Nov. 12, 2007
- Event: "Cutting Edge Researches in Digital Signal Processing Lectures"
- Attendance: about 200 undergraduate students, and 50 DSP researchers
- Topic: Statistical Natural Language Processing: Lessons from Automatic Speech Recognition
- Impact: The study of Tagalog, the key local language, has been gathering momentum. The lecture has opened up dialogues between researchers in the Philippines with the international speech community. We expect a local speech community to be formed soon for initial database collection and evaluation methodologies. There is a recent increase in funding for speech research. Prof. Rachel Edita O. Roxas will take the lead.
- Third Talk in Bangkok:
- Location: Chulalongkorn University
- Time: 1:15-3:00pm on Nov. 15, 2007
- Event: Second Speech Community Forum
- Attendance: 60 people from academia, industry and government
- Topic: Fundamentals and Technical Challenges in Speech Recognition
- Impact: Dr. Chai Wutiwiwatchai of the National Electronics and Computer Technology Center hosted the visit as a part of Thailand's 2nd Speech Forum (the first was held in Feb. 2007). The lecture again opens up dialogues between Thai speech researchers with international speech communities. Dr. Wutiwiwatchai has set a long term goal to bid for an Interspeech conference in Bangkok in the future.
- Summary: I think ISCA is serving developing countries well with the planned DL tours. The Lecturer also acts as an ambassador. I found people are very engaging in such comfortable settings when the DL fits into part of an event, such as in both Manila and Bangkok. The DL is also a contact for follow-up discussions.
-
Report of Marc Swerts on visit to Brazil ( June 2 0 0 7 )
Report Marc Swerts on visit to Brazil ( June 2007 ) as an ISCA distinguished lecturer
1 . Introduction
From June 11, 2007 t ill June 17, I t ravelled to Brazil as an ISCA dist inguished lecturer tovisit two universit ies in the Sao Paolo area. First , I went to the I nst ituto de Estudos da Linguagem of the Linguist ics department at the Unicamp university in Campinas. This is a public universit y which is among the highest - rated universit ies in Brazil, which is also strongly research-oriented. Unfortunately, there was a strike in Campinas against some governmental decisions regarding organisational changes in academia, but that didn't really affect my visit . In the second half of the week I moved to Sao Paolo to visit the linguist ics department of the Catholic University of São Paulo ( PUC-SP) . This university is very large, and is more teaching- rather than research-or iented, compared to Campinas.
The flight and the costs for local t ransportat ion were covered by ISCA; in addit ion, Ireceived a daily allowance from the host universit ies to pay food and lodging. Thesponsoring in Brazil was provided by the FAPEPS project (conducted by the Speech Prosody Studies group in Campinas) and another by the LIACC projecy at PUC-SP.
Contact persons in these respect ive universit ies were Plinio Almeida Barbosa from the Universit y of Campinas and Sandra Madureira from PUC-SP. Below, I will briefly reportabout the lectures I have been giving in these 2 locat ions, and then say something about the outcome of var ious contacts I had with local researchers in Campinas and Sao Paolo.
2 . Lectures
Both in Campinas and Sao Paolo I have given lectures on research I did with colleaguesfrom Tilburg University (mainly Emiel Krahmer and Pashiera Barkhuysen) on aspects of audiovisual prosody. I n my first lecture, I discussed work about the use of facial and auditory cues to communicat ion problems, both in human-human and human-machine interactions. I n my second lecture, I presented the result s of work on audiovisualmarkers of focus and accent , and of ut terance boundaries.
My lectures were well at tended, and led to some useful and inspir ing discussions, both in Campinas and Sao Paolo. I n Campinas, I mainly talked to an audience of MA students,PhD students and senior staff members from the phonet ics and computer sciencedepartment and research groups specialized in language educat ion. There were also people working on distance e- learning who were especially interested in the added valueof mult imodal communicat ion, and feedback and error-handling. I n Sao Paolo, my talkwas scheduled as one of a series of talks in a local workshop on the expressive use of prosody. The other talks, one of which by my local host Sandra Madureira, focused on the use of intonat ion for rendering expressiveness to poet ry and music, which matched nicely with my own studies on audiovisual correlates of posit ive and negat ive emot ions.
The workshop at t racted quite a large audience, about 100 to 150 listeners, most coming from research groups working in the area of speech pathology.
3 . Local contacts
My local host in Campinas was Plinio Barbosa, who is an associate professor in the Inst ituto de Estudos da Linguagem of the Linguist ics department . His main researchspeciality is the analysis and modelling of speech rhythm, part ly for the improvement of synthetic speech. He already has a good internat ional reputat ion. He has done his PhDwork in France, in the ICP lab at Grenoble, where he collaborated with Gerard Bailly, andhas good research contacts with various researchers in the United States (Byrd, Fujimoro, Goldstein, Saltzman) , Japan (Nick Campbell) and especially in Portugal. He is the ISCA liason representat ive of the ISCA sig on I ber ian languages (ht tp: / /www.ilsig.org/ ) , and is guest edit ing a special issue on I ber ian languages, together with I sabel Trancoso, Nestor Becer ro-Yoma, and Kuldip Paliwal (ht tp: / /www.iscaspeech.org/ SpeCom_I L.pdf) . In addit ion, he is editor ial board member of various other|( internat ional and local) journals, including the Internat ional Journal of speech technology. Important to note is that Plinio, together with some other researchers in the Sao Paolo area ( including Sandra Madureira) , will be host ing the next Speech Prosody conference in Spr ing 2008 (his site was selected based on votes from members of the Speech Prosody organisat ion) , with already an impressive list of keynote speakers (http: / /www.sp2008.org/ ) .
During my visit in Campinas, I have also been informed about ongoing research projects through individual presentat ions by local researchers. Plinio Barbosa himself presented results of his work on the development of a speech rhythm model (about which he recent ly got a paper accepted for Speech Communicat ion) , and of his research on the prosody-syntax interface, p-centers, and more general intonat ional modelling.
Pablio Arantes tackles more or less the same issues as Barbosa, but has also performed some addit ional speech percept ion tests, using a click detect ion paradigm; in his view, speech product ion and percept ion are related processes, in line with current views on linguistic entrainment in dialogue modelling. Luciana Lucente works on the development of a Brazilian version of ToBI . The system is highly comparable to exist ing schemes developed for other languages, but differs in that her version aims to result in a more narrow intonat ional t ranscript ion and does not contain a break index t ier . Her Brazilian Portuguese ToB framework also reserves a separate t ier for pragmat ic informat ion. I n addit ion, she works on correlates of narrow focus in Brazilian Portuguese. Sandra Merlo talked about her work of modelling of pauses and disfluencies in different types of spontaneous elicited speech (5 adults producing 20 texts) ; cycles of pause and disfluency structures were semi-automatically determined by means of four er t ransfor mat ion, which produces hierarchical pause st ructures that can easily be explained psycholinguist ically. Aveliny M. Lima talked about her work on P centers, where she analysed CV clusters produced with increasing and decreasing tempos.
José Mario de Martino and Fabio Violina from the department of computer engineering and telecommunicat ion presented work on context -dependent visemes for the development of visual synthesis system. They conducted evaluat ion studies of their synthesic talking head with nonsense clusters, but they are t rying to find new object ive evaluat ion methods. My visit to Sao Paolo was almost ent irely devoted to the local workshop on expressive use of prosody. I n addit ion, I had informal conversat ions with local researchers on var ious issues, including new processing measures to determine degrees of speech pathology and cultural differences in dramat ic expressions of emot ions. With Sandra Madureira I discussed work on finding audiovisual cues to important stretches of informat ion in newsreading. I nterest ingly, both her students and students in my own department are exploring this by annotat ing facial expressions of professional newsreaders from Brazilian and Dutch broadcasts.
4 . Conclusion
Let me end this report by st ressing that my visit to Brazil has been fruit ful for me in morethan one respect . I got some very useful and posit ive feedback on my own lectures,which gave me new ideas for future research. I plan to keep in touch with my Brazilian colleagues, as it would be nice to set up some form of collaborat ion, for instance on aspects of audiovisual speech processing. I n addit ion, it was nice to learn more about local research, which I had known only part ially, but which appears to be very promising indeed.
-
GOOGLE SCHOLAR AND ISCA ARCHIVE
Back to Top
Google Scholar and the ISCA Archive
The indexing of the ISCA Archive (http://www.isca-speech.org/archive/) by the Google Scholar search engine (http://scholar.google.com/) is now thorough enough to be quite useful, so this seems like a good time to give an overview of the service. Google Scholar is a research literature search engine that provides full-text search for ISCA papers whose full text cannot be searched with other search engines. Google Scholar's citation tracking shows what papers have cited a particular paper, which can be very useful for finding follow-up work, related work and corrections. More details about these and other features are given below.
The titles, author lists, and abstracts of ISCA Archive papers are all on the public web, so they can be searched by a general-purpose search engine such as Google. However, the full texts of most ISCA papers are password protected and thus cannot be searched with a general-purpose search engine. Google Scholar, through an arrangement with ISCA, has access to the full text of ISCA papers. Google Scholar has similar arrangements with many other publishers. (On the other hand, general-purpose search engines index all sorts of web pages and other documents accessible through the public web, many of which will not be in the Google Scholar index. So it's often useful to perform the same search using both Google Scholar and a general-purpose search engine.)
Google Scholar automatically extracts citations from the full text of papers. It uses this information to provide a "Cited by" list for each paper in the Google Scholar index. This is a list of papers that have cited that paper. Google Scholar also provides an automatically generated "Related Articles" list for each paper. The "Cited by" and "Related Articles" lists are powerful tools for discovering relevant papers. Furthermore, the length of a paper's "Cited by" list can be used as a convenient (although imperfect) measure of the paper's impact. Discussions about the subtleties of using Google Scholar to measure impact can be found at http://www.harzing.com/resources.htm#/pop_gs.htm and http://blogs.nature.com/nautilus/2007/07/google_scholar_as_a_measure_of.html.
It's possible to restrict Google Scholar searches to papers published by ISCA by using Google Scholar's Advanced Search feature and entering "ISCA" in the "Return articles published in" field. If "ISCA" is entered in that field, and nothing is entered in the main search field, then the search results will show what ISCA papers are the most highly cited.
It should be noted that that there are many papers on ISCA-related topics which are not in the Google Scholar index. For example, it seems many ICPhS papers are missing. And old papers which have been scanned in from paper copies will either not have their full contents indexed, or will be indexed using imperfect OCR technology. Furthermore, as of November 2007 the indexing of the ISCA Archive by Google Scholar is still not 100% complete. There are a few different areas which are not perfectly indexed, but the biggest planned improvement is to start using OCR for the ISCA papers which have been scanned in from paper copies.
There may be a time lag between when a new event is added to the ISCA Archive in the future and when it appears in the Google Scholar index. This time lag may be longer than the usual lag of general-purpose search engines such as Google, because ISCA must create Google Scholar catalog data for every new event and because the Google Scholar index seems to update considerably more slowly than the Google index.
Acknowledgements: ISCA's arrangement with Google Scholar is a project of students Rahul Chitturi, Tiago Falk, David Gelbart, Agustin Gravano, and Francis Tyers, ISCA webmaster Matt Bridger, and ISCA Archive coordinator Wolfgang Hess. Our thanks to Google's Christian DiCarlo and Darcy Dapra, and the rest of the Google Scholar team.
SIG's activities
-
A list of Speech Interest Groups can be found on our web.
Back to Top
Courses, Internships
-
Motorola Labs - Center for Human Interaction Research (CHIR) l
Motorola Labs - Center for Human Interaction Research (CHIR)
located in Schaumburg Illinois, USA,
is offering summer intern positions in 2008 (12 weeks each).
CHIR's mission
Our research lab develops technologies that provide access to rich communication, media and
information services effortless, based on natural, intelligent interaction. Our research
aims on systems that adapt automatically and proactively to changing environments, device
capabilities and to continually evolving knowledge about the user.
Intern profiles
1) Acoustic environment/event detection and classification.
Successful candidate will be a PhD student near the end of his/her PhD study and is skilled
in signal processing and/or pattern recognition; he/she knows Linux and C/C++ programming.
Candidates with knowledge of acoustic environment/event classification are preferred.
2) Speaker adaptation for applications on speech recognition and spoken document retrieval
The successful candidate must currently be pursuing a Ph.D. degree in EE or CS with complete
understanding and hand-on experience on automatic speech recognition related research. Proficiency
in Linux/Unix working environment and C/C++ programming. Strong GPA. A strong background in speaker
adaptation is highly preferred.
3) Development of voice search-based web applications on a smartphone
We are looking for an intern candidate to help create an "experience" prototype based on our
voice search technology. The app will be deployed on a smartphone and demonstrate intuitive and
rich interaction with web resources. This intern project is oriented more towards software engineering
than research. We target an intern with a master's degree and strong software engineering background.
Mastery of C++ and experience with web programming (AJAX and web services) is required.
Development experience on Windows CE/Mobile desired.
4) Integrated Voice Search Technology For Mobile Devices
Candidate should be proficient in information retrieval, pattern recognition and speech recognition.
Candidate should program in C++ and script languages such as Python or Perl in Linux environment.
Also, he/she should have knowledge on information retrieval or search engines.
We offer competitive compensation, fun-to-work environment and Chicago-style pizza.
If you are interested, please send your resume to:
Dusan Macho, CHIR-Motorola Labs
Email: dusan [dot] macho [at] motorola [dot] com
Tel: +1-847-576-6762
Back to Top
Books, Databases, Softwares
-
Books
La production de la parole
Author: Alain Marchal, Universite d'Aix en Provence, France
Publisher: Hermes Lavoisier
Year: 2007Speech enhancement-Theory and Practice
Author: Philipos C. Loizou, University of Texas, Dallas, USA
Publisher: CRC Press
Year:2007Speech and Language Engineering
Editor: Martin Rajman
Publisher: EPFL Press, distributed by CRC Press
Year: 2007Human Communication Disorders/ Speech therapy
This interesting series can be listed on Wiley websiteIncurses em torno do ritmo da fala
Author: Plinio A. Barbosa
Publisher: Pontes Editores (city: Campinas)
Year: 2006 (released 11/24/2006)
(In Portuguese, abstract attached.) WebsiteSpeech Quality of VoIP: Assessment and Prediction
Author: Alexander Raake
Publisher: John Wiley & Sons, UK-Chichester, September 2006
WebsiteSelf-Organization in the Evolution of Speech, Studies in the Evolution of Language
Author: Pierre-Yves Oudeyer
Publisher:Oxford University Press
WebsiteSpeech Recognition Over Digital Channels
Authors: Antonio M. Peinado and Jose C. Segura
Publisher: Wiley, July 2006
WebsiteMultilingual Speech Processing
Editors: Tanja Schultz and Katrin Kirchhoff ,
Elsevier Academic Press, April 2006
WebsiteReconnaissance automatique de la parole: Du signal a l'interpretation
Back to Top
Authors: Jean-Paul Haton
Christophe Cerisara
Dominique Fohr
Yves Laprie
Kamel Smaili
392 Pages
Publisher: Dunod -
ELRA - Language Resources Catalogue - Update
ELRA is happy to announce that 10 new Speech Resources from both LC-STAR
and TC-STAR projects are now available in its catalogue.
*ELRA-S0245 LC-STAR German Phonetic lexicon
*The LC-STAR German Phonetic lexicon comprises 102,169 entries,
including a set of 55,507 common words, a set of 46,662 proper names
(including person names, family names, cities, streets, companies and
brand names) and a list of 6,763 special application words. The lexicon
is provided in XML format and includes phonetic transcriptions in SAMPA. ).
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1028
*ELRA-S0246 LC-STAR German Phonetic lexicon in the Touristic Domain
*The LC-STAR German Phonetic lexicon in the Touristic Domain comprises
8,782 entries from the following categories: nouns, adjectives and
verbs. For each entry the following information is provided:
orthographic form, part-of-speech (POS), phonemic transcription. The
lexicon is provided in XML format and includes phonetic transcriptions
in SAMPA.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1029
*ELRA-S0247 LC-STAR Standard Arabic Phonetic lexicon
*The LC-STAR Standard Arabic Phonetic lexicon comprises 110,271 entries,
including a set of 52,981 common words, a set of 50,135 proper names
(including person names, family names, cities, streets, companies and
brand names) and a list of 7,155 special application words. The lexicon
is provided in XML format and includes phonetic transcriptions in SAMPA.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1030
*ELRA-S0248 LC-STAR English-German Bilingual Aligned Phrasal lexicon
*The LC-STAR English-German Bilingual Aligned Phrasal lexicon comprises
10,733 phrases from the tourist domain. It is based on a list of short
sentences obtained by translation from US-English 10,518 phrasal corpus.
The lexicon is provided in XML format.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1031
*ELRA-S0249 TC-STAR English Training Corpora for ASR: Transcriptions of
EPPS Speech
*This corpus consists of transcriptions from 92 hours of EPPS (European
Parliament Plenary Sessions) speeches held or interpreted in European
English (a mixture of native and non-native English). The transcription
files are stored in Transcriber XML file format.
For corresponding recordings, see ELRA-S0251
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1032
*ELRA-S0250 TC-STAR English-Spanish Training Corpora for Machine
Translation: Aligned Final Text Editions of EPPS
*This corpus consists of respectively 34 million (English) and 38
million (Spanish) running words of bilingual sentence segmented and
aligned texts in English and Spanish obtained from the Final Text
Editions provided by the European Parliament (from April 1996 to Sept.
2004, Dec. 2004 to May 2005, and Dec. 2005 to May 2006. The data is
accompanied by tools for further preprocessing.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1033
*ELRA-S0251 TC-STAR English Training Corpora for ASR: Recordings of EPPS
Speech
*This corpus consists of the recordings of around 290 hours form EPPS
(European Parliament Plenary Sessions) speeches held or interpreted in
European English, 92 hours of which were annotated (transcribed) (the
transcriptions are not provided in the present package). Each file
contains a single channel with 16-bit resolution at a sample rate of 16kHz.
For corresponding transcriptions, see ELRA-S0249.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1034
*ELRA-S0252 TC-STAR Spanish Training Corpora for ASR: Recordings of EPPS
Speech
*This corpus consists of the recordings of around 283 hours from EPPS
(European Parliament Plenary Sessions) speeches held or interpreted in
European Spanish (a mixture of native and non-native Spanish). Each file
contains a single channel with 16-bit resolution at a sample rate of 16kHz.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1036
*ELRA-S0253* *TC-STAR English Test Corpora for ASR
*This corpus consists of 70 hours of recordings of EPPS (European
Parliament Plenary Sessions) speeches held or interpreted in European
English and other European languages. From this corpus, 16 hours of
English speeches (native or non native) were annotated (transcribed).
Each speech file contains a single channel with 16-bit resolution at a
sample rate of 16kHz. The transcription files are stored in Transcriber
XML file format.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1037
*ELRA-S0254* *TC-STAR Spanish Test Corpora for ASR
*This corpus consists of 174 hours of recordings of EPPS (European
Parliament Plenary Sessions) speeches held or interpreted in European
Spanish and other European languages. From this corpus, 16 hours of
Spanish speeches were annotated (transcribed). Each audio file contains
a single channel with 16-bit resolution at a sample rate of 16kHz. The
transcription files are stored in Transcriber XML file format.
For more information, see:
http://catalog.elra.info/product_info.php?products_id=1038
For more information on the catalogue, please contact Valérie Mapelli
mailto:mapelli@elda.org
Visit our on-line catalogue: http://catalog.elra.info
Job openings
-
We invite all laboratories and industrial companies which have job offers to send them to the ISCApad editor: they will appear in the newsletter and on our website for free. (also have a look at http://www.isca-speech.org/jobs.html as well as http://www.elsnet.org/ Jobs)
Back to Top -
Speech Engineer/Senior Speech Engineer at Microsoft, Mountain View, CA,USA
Job Type: Full-Time
Back to Top
Send resume to Bruce Buntschuh
Responsibilities:
Tellme, now a subsidiary of Microsoft, is a company that is focused on delivering the highest quality voice recognition based applications while providing the highest possible automation to its clients. Central to this focus is the speech recognition accuracy and performance that is used by the applications. The candidate will be responsible for the development, performance analysis, and optimization of grammars, as well as overall speech recognition accuracy, in a wide variety of real world applications in all major market segments. This is a unique opportunity to apply and extend state of the art speech recognition technologies to emerging spaces such as information search on mobile devices.
Requirements:
· Strong background in engineering, linguistics, mathematics, machine learning, and or computer science.
· In depth knowledge and expertise in the field of speech recognition.
· Strong analytical skills with a determination to fully understand and solve complex problems.
· Excellent spoken and written communication skills.
· Fluency in English (Spanish a plus).
· Programming capability with scripting tools such as Perl.
Education:
MS, PhD, or equivalent technical experience in an area such as engineering, linguistics, mathematics, or computer science. -
Software Development Engineer, Automatic Speech Recognition at Microsoft Redmond WA, USA
Send resume to Yifan Gong
Back to Top
Speech Technologies and Modeling- Speech Component Group
Microsoft Corporation
Redmond WA, USA
Background:
Microsoft's Speech Component Group has been working on automatic speech recognition (SR) in real environments. We develop SR products for multiple languages for mobile devices, desktop computers, and communication servers. The group now has an open position for software development engineers to work on our acoustic and language modeling technologies. The position offers great opportunities for innovation and technology and product development.
Responsibilities:
Design and implement speech/language modeling and recognition algorithms to improve recognition accuracy.
Create, optimize and deliver quality speech recognition models and other components tailored to our customers' needs.
Identify, investigate and solve challenging problems in the areas of recognition accuracy from speech recognition system deployments.
Improve speech recognition language expansion engineering process that ensures product quality and scalability.
Required competencies and skills:
Passion about speech technology and quality software, demonstrated ability relative to the design and implementation of speech recognition algorithms.
Strong desire for achieving excellent results, strong problem solving skills, ability to multi-task, handle ambiguities, and identify issues in complex SR systems.
Good software development skills, including strong aptitude for software design and coding. 3+ years of experience in C/C++. Programming with scripting languages highly desirable.
MS or PhD degree in Computer Science, Electrical Engineering, Mathematics, or related disciplines, with strong background in speech recognition technology, statistical modeling, or signal processing.
Track record of developing SR algorithms, or experience in linguistic/phonetics, is a plus. -
PhD Research Studentship in Spoken Dialogue Systems- Cambridge UK
Applications are invited for an EPSRC sponsored studentship in Spoken Dialogue Systems leading to the PhD degree. The student will join a team lead by Professor Steve Young working on statistical approaches to building Spoken Dialogue Systems. The overall goal of the team is to develop complete working end-to-end systems which can be trained from real data and which can be continually adapted on-line. The PhD work will focus specifically on the use of Partially Observable Markov Decision Processes for dialogue modelling and techniques for learning and adaptation within that framework. The work will involve statistical modelling, algorithm design and user evaluation. The successful candidate will have a good first degree in a relevant area. Good programming skills in C/C++ are essential and familiarity with Matlab would be useful.
Back to Top
The studentship will be for 3 years starting in October 2007 or January 2008. The studentship covers University and College fees at the Home/EU rate and a maintenance allowance of 13000 pounds per annum. Potential applicants should email Steve Young with a brief CV and statement of interest in the proposed work area -
Elektrobit seeks SW-Engineers (m/f) for multimodal HMI Solutions (Speech Dialog)
Elektrobit Automotive Software is located in Erlangen, Germany and delivers ready-to-mass product implementations of a variety of software standards of the automotive industry and services to implement large software projects. The spectrum is enhanced with tools for HMI and control device development and in-house development, such as a navigation solution. We are developing solutions for multimodal HMIs in automotive infotainment/navigation systems. One focus are speech dialog systems. The challenge lies in realizing natural speech dialogue systems for different applications (navigation, mp3 player etc.) in an embedded environment. You will be designing and developing such speech solutions.
Back to Top
You have know how in one or more of the following areas:
Experience project co-ordination
Programming C/C++, perl, (Java) for windows and/or Linux
Speech recognition
Multimodal dialog systems
Speech synthesis /TTS
SW- Processes and SW-Tests
Experience in Object oriented Programming
Experience with Embedded Operating Systems
Your job description
Project coordination
Coordination of supplier and requirements from different applications
Development and specification of concepts for speech-related SW modules for different applications in embedded environments
Implementation of multimodal HMIs
Integration of speech modules in HMIs
Testing
We expect from you:
A degree in IT, electrical/electronic engineering, computational linguistics or similar
Good working knowledge of German and English
Innovative streak
Willingness to take responsibility in international Teams
We offer you:
A motivating working environment
Challenging work
Support in advancement
Please apply at www.elektrobit.com -> Automotive Software -> jobs
If you have any further questions Mr. Schrör (Tel.-Nr. +49 (9131) 7701-516) or Mr. Huck (-217) will gladly answer them. -
Sound to Sense: 18 Fellowships in speech research
Sound to Sense (S2S) is a Marie Curie Research Training Network involving collaborative speech research amongst 13 universities in 10 countries. 18 Training Fellowships are available, of which 12 are predoctoral and 6 postdoctoral (or equivalent experience). Most but not all are planned to start in September or October 2007.
Back to Top
A research training network's primary aim is to support and train young researchers in professional and inter-disciplinary scientific skills that will equip them for careers in research. S2S's scientific focus is on cross-disciplinary methods for modelling speech recognition by humans and machines. Distinctive aspects of our approach include emphasis on richly-informed phonetic models that emphasize communicative function of utterances, multilingual databases, multiple time domain analyses, hybrid episodic-abstract computational models, and applications and testing in adverse listening conditions and foreign language learning.
Eleven projects are planned. Each can be flexibly tailored to match the Fellows' backgrounds, research interests, and professional development needs, and will fall into one of four broad themes.
1: Multilinguistic and comparative research on Fine Phonetic Detail (4 projects)
2: Imperfect knowledge/imperfect signal (2 projects)
3: Beyond short units of speech (2 projects)
4: Exemplars and abstraction (3 projects)
The institutions and senior scientists involved with S2S are as follows:
* University of Cambridge, UK (S. Hawkins (Coordinator), M. Ford, M. Miozzo, D. Norris. B. Post)
* Katholieke Universiteit, Leuven, Belgium (D. Van Compernolle, H. Van Hamme, K. Demuynck)
* Charles University, Prague, Czech Republic (Z. Palková, T. Dub?da, J. Volín)
* University of Provence, Aix-en-Provence, France (N. Nguyen, M. d'Imperio, C. Meunier)
* University Federico II, Naples, Italy (F. Cutugno, A. Corazza)
* Radboud University, Nijmegen, The Netherlands (L. ten Bosch, H. Baayen, M. Ernestus, C. Gussenhoven, H. Strik)
* Norwegian University of Science and Technology (NTNU), Trondheim, Norway (W. van Dommelen, M. Johnsen, J. Koreman, T. Svendsen)
* Technical University of Cluj-Napoca, Romania (M. Giurgiu)
* University of the Basque Country, Vitoria, Spain (M-L. Garcia Lecumberri, J. Cenoz)
* University of Geneva, Switzerland (U. Frauenfelder)
* University of Bristol, UK (S. Mattys, J. Bowers)
* University of Sheffield, UK (M. Cooke, J. Barker, G. Brown, S. Howard, R. Moore, B. Wells)
* University of York, UK. (R. Ogden, G. Gaskell, J. Local)
Successful applicants will normally have a degree in psychology, computer science, engineering, linguistics, phonetics, or related disciplines, and want to acquire expertise in one or more of the others.
Positions are open until filled, although applications before 1 May 2007 are recommended for starting in October 2007.
Further details are available from the web about:
+ the research network (92kB) and how to apply, + the research projects(328 kB). -
AT&T - Labs Research: Research Staff Positions - Florham Park, NJ
AT&T - Labs Research is seeking exceptional candidates for Research Staff positions. AT&T is the premiere broadband, IP, entertainment, and wireless communications company in the U.S. and one of the largest in the world. Our researchers are dedicated to solving real problems in speech and language processing, and are involved in inventing, creating and deploying innovative services. We also explore fundamental research problems in these areas. Outstanding Ph.D.-level candidates at all levels of experience are encouraged to apply. Candidates must demonstrate excellence in research, a collaborative spirit and strong communication and software skills. Areas of particular interest are
- Large-vocabulary automatic speech recognition
- Acoustic and language modeling
- Robust speech recognition
- Signal processing
- Speaker recognition
- Speech data mining
- Natural language understanding and dialog
- Text and web mining
- Voice and multimodal search
AT&T Companies are Equal Opportunity Employers. All qualified candidates will receive full and fair consideration for employment. More information and application instructions are available on our website at http://www.research.att.com/. Click on "Join us". For more information, contact Mazin Gilbert (mazin at research dot att dot com).
Back to Top -
Research Position in Speech Processing at UGent, Belgium
Background
Since March 2005, the universities of Leuven, Gent, Antwerp and Brussels have joined forces in a big research project, called SPACE (SPeech Algorithms for Clinical and Educational applications). The project aims at contributing to the broader application of speech technology in educational and therapeutic software tools. More specifically, it pursues the automatic detection and classification of reading errors in the context of an automatic reading tutor, and the objective assessment of disordered speech (e.g. speech of the deaf, dysarthric speech, ...) in the context of computer assisted speech therapy assessment. Specific for the target applications is that the speech is either grammatically and lexically incorrect or a-typically pronounced. Therefore, standard technology cannot be applied as such in these applications.
Job description
The person we are looking for will be in charge of the data-driven development of word mispronunciation models that can predict expected reading errors in the context of a reading tutor. These models must be integrated in the linguistic model of the prompted utterance, and achieve that the speech recognizer becomes more specific in its detection and classification of presumed errors than a recognizer which is using a more traditional linguistic model with context-independent garbage and deletion arcs. A challenge is also to make the mispronunciation model adaptive to the progress made by the user.
Profile
We are looking for a person from the EU with a creative mind, and with an interest in speech & language processing and machine learning. The work will require an ability to program algorithms in C and Python. Having experience with Python is not a prerequisite (someone with some software experience is expected to learn this in a short time span). Demonstrated experience with speech & language processing and/or machine learning techniques will give you an advantage over other candidates.
The job is open to a pre-doctoral as well as a post-doctoral researcher who can start in November or December. The job runs until February 28, 2009, but a pre-doctoral candidate aiming for a doctoral degree will get opportunities to do follow-up research in related projects.
Interested persons should send their CV to Jean-Pierre Martens (martens@elis.ugent.be). There is no real deadline, but as soon as a suitable person is found, he/she will get the job.
Back to Top -
Summer Inter positions at Motorola Schaumburg Illinois USA
Motorola Labs - Center for Human Interaction Research (CHIR) located in Schaumburg Illinois, USA, is offering summer intern positions in 2008 (12 weeks each).
CHIR's mission:
Our research lab develops technologies that provide access to rich communication, media and information services effortless, based on natural, intelligent interaction. Our research aims on systems that adapt automatically and proactively to changing environments, device capabilities and to continually evolving knowledge about the user.
Intern profiles:
1) Acoustic environment/event detection and classification.
Successful candidate will be a PhD student near the end of his/her PhD study and is skilled in signal processing and/or pattern recognition; he/she knows Linux and C/C++ programming. Candidates with knowledge of acoustic environment/event classification are preferred.
2) Speaker adaptation for applications on speech recognition and spoken document retrieval.
The successful candidate must currently be pursuing a Ph.D. degree in EE or CS with complete understanding and hand-on experience on automatic speech recognition related research. Proficiency in Linux/Unix working environment and C/C++ programming. Strong GPA. A strong background in speaker adaptation is highly preferred.
3) Development of voice search-based web applications on a smartphone
We are looking for an intern candidate to help create an "experience" prototype based on our voice search technology. The app will be deployed on a smartphone and demonstrate intuitive and rich interaction with web resources. This intern project is oriented more towards software engineering than research. We target an intern with a master's degree and strong software engineering background. Mastery of C++ and experience with web programming (AJAX and web services) is required. Development experience on Windows CE/Mobile desired.
4) Integrated Voice Search Technology For Mobile Devices.
Candidate should be proficient in information retrieval, pattern recognition and speech recognition. Candidate should program in C++ and script languages such as Python or Perl in Linux environment. Also, he/she should have knowledge on information retrieval or search engines.
We offer competitive compensation, fun-to-work environment and Chicago-style pizza.
If you are interested, please send your resume to:
Dusan Macho, CHIR-Motorola Labs
Email: dusan.macho@motorola.com
Tel: +1-847-576-6762
Back to Top -
Post-doc at France Telcom R&D Lannion Brittany France
Post-Doc à France Télécom R&d, Lannion : acquisition de contexte à partir de prise de son ambiante.
DeadLine: 31/12/2007
claude.marro@orange-ftgroup.com
Description du contexte
Des données physiques de toute nature, provenant de l'environnement de l'utilisateur, peuvent être utilisées dans la communication ambiante comme informations de contexte pour offrir des fonctionnalités de service ou d'interface nouvelles, en particulier au niveau de l'adaptation du service à la situation et l'activité des utilisateurs. Ces données sont acquises par divers capteurs répartis dans l'environnement. Les données de scène audio issues de microphones sont parmi les plus riches que l'on puisse exploiter parmi toutes ces données de capteurs, et elles présentent surtout la particularité, dans les applications de communication ambiante, d'être utilisables à la fois comme inputs fonctionnels (communication audio interpersonnelle) et comme inputs de contexte, pour lequel elles peuvent être combinées avec des données issues d'autres types de capteurs. L'objectif ici est de développer des dispositifs permettant d'aboutir à cette double utilisation des données audio. Le système de prise de son doit disparaître de l'attention des utilisateurs et cette dématérialisation en fait la principale difficulté. En effet, l'éloignement de la prise de son entraîne une dégradation de la parole utile et nécessite une localisation de locuteur et une "focalisation" dans sa direction.
Acquisition et restitution audio fonctionnelle
Idéalement, l'objectif très ambitieux pourrait être d'obtenir une acquisition et restitution de son qui puisse être efficace quel que soit l'endroit de la pièce où se trouve une personne, en utilisant des microphones répartis dans l'environnement. En raison des difficultés évoquées ci-dessus, remporter ce challenge est hors de portée de cette étude, on proposera comme alternative un dispositif permettant une prise et restitution du son en un nombre limité de points précisés à l'avance.
Deux approches multi-capteurs sont envisagées : l'antenne acoustique à directivité contrôlée et le microphone ambisonique. L'intérêt d'aborder ces deux techniques réside dans leur complémentarité. En particulier, la première permet un design souple du diagramme de directivité (en fonction de la géométrie, de la fréquence, etc..) et est performante en moyenne et basse fréquence (au détriment de l'encombrement). Quant au microphone ambisonique, il a une taille réduite (au détriment des performances en moyenne et basse fréquence) et permet de reproduire à l'identique un champ acoustique à distance. La première phase de l'étude permettra de déterminer laquelle des approches est la plus adaptée.
Acquisition de contexte sonore
La première fonctionnalité à étudier est la localisation de sources sonores, fonction incontournable pour identifier la source vers lequel le système doit pointer. La détection de présence et la position du locuteur sont les informations contextuelles de base à extraire.
Si l'on considère que la localisation couplée à la prise de son multi-capteur constitue un outil d'analyse du champ sonore, il sera possible d'apporter d'autres informations de contexte. En effet, le développement de traitements spécifiques permettra par exemple de donner le nombre de locuteurs et leurs positions dans la pièce, leur pourcentage de locution, le niveau de bruit de la pièce, etc.
Une analyse plus fine du contexte sonore est à envisager comme une perspective de ce travail et ne sera abordé que si le temps le permet. Ceci concerne les informations à extraire qui nécessiteraient l'usage de technologies telles que la reconnaissance vocale, la classification et l'indexation audio.
Notons que le système de prise de son et les traitements développés dans ce projet constitueront des pré-requis nécessaires pour la continuité des travaux sur l'analyse fine du contexte sonore.
Profil
Aspects pratiques
-
Aucune condition de nationalité n'est requise
-
Le chercheur bénéficiera d'un contrat à durée déterminée de France Télécom, pour une durée de 12 à 18 mois, non renouvelable ni prolongeable.
-
Le chercheur sera intégré à la division R&D de France Télécom sur son site de Lannion, CRD Technologies, Laboratoire « Speech and Sound technologies and Processing».
Compétences Techniques
-
Traitement numérique du signal (analyse spectrale, filtrage, etc.)
-
Traitements multi-microphones
-
Si possible bases en traitement de la parole et en acoustique
-
Goût pour les travaux de recherche applicative (analyse, mise au point et adaptation)
-
Langages Matlab et C
Aptitudes
- Goût pour le travail en équipe.
- Bon niveau en anglais.
Niveau poste : ingénieur en CDD de type post-doctorat - Durée : 12 à 18 mois.
Back to Top -
-
Ph.D. Program CMU-PORTUGAL
Ph.D. Program CMU-PORTUGAL in the area of Language and Information Technologies
The Language Technologies Institute (LTI) of the School of Computer Science at Carnegie Mellon University (CMU) offers a dual degree Ph.D.
Program in Language and Information Technologies in cooperation with Portuguese Universities. This Ph.D. program is part of the activities of the recently created Information and Communication Technologies Institute (ICTI), resulting from the Portugal-CMU Partnership.
The Language Technologies Institute, a world leader in the areas of speech processing, language processing, information retrieval, machine translation, machine learning, and bio-informatics, has been formed 20 years ago. The breadth of language technologies expertise at LTI enablesnew research in combinations of the core subjects, for example, inspeech-to-speech translation, spoken dialog systems, language-based tutoring systems, and question/answering systems.
The Portuguese consortium of Universities includes the Spoken LanguageSystems Lab (L2F) of INESC-ID Lisbon/IST, the Center of Linguistics of the University of Lisbon (CLUL/FLUL), the Centre for Human Language Technology and Bioinformatics at the University of Beira Interior(HULTIG/UBI) and the linguistics group at the University of Algarve (UALG). These four research centers (and the corresponding Universities), share expertise in the same language technologies as LTI, although with a strong focus on processing the Portuguese language.
Each Ph.D. student will receive a dual degree from LTI and the selected Portuguese University, being co-supervised by one advisor from each institute, and spending approximately half of the 5-year doctoral program at each institute. Most of the academic part will take place at LTI, during the first 2 years, where most of the required 8 courses will be taken, with a proper balance of focus areas (Linguistic, Computer Science, Statistical/Learning, Task Orientation). The remaining 3 years of the doctoral program will be dedicated to research, mostly spent at the Portuguese institute, with one or two visits to CMU per year.
The thesis topic will be in one of the research areas of the cooperation program, defined by the two advisors. Two multilingual topics have been identified as priority research areas: computer aided language learning (CALL) and speech-to-speech machine translation (S2SMT).
The doctoral students will be involved in one of these two projects aimed at building real HLT systems. These projects will involve at least two languages, one of them being Portuguese, the target language for the CALL system to be developed and either the source or target language (or both) for the S2SMT system. These two projects provide a focus for the proposed research; through them the collaboration will explore the maincore areas in language technology.
The scholarship will be funded by the Foundation for Science and Technology (FCT), Portugal.
How to Apply
The application deadline for all Ph.D. programs in the scope of the CMU-Portugal partnership is December 15, 2007.
Students interested in the dual doctoral program must apply by filling the corresponding form at the CMU webpage http://www.lti.cs.cmu.edu/About/how-to-apply.html
The application form will be forwarded to the Portuguese University and to the Foundation for Science and Technology. Simultaneously, they should send an email to the coordinators of the Portuguese consortium and of the LTI admissions (Isabel Trancoso/Lori Levin):
Isabel.Trancoso at inesc-id dot pt
lsl at cs dot cmu dot edu
All questions about the joint degree doctoral program should be directed to these two addresses.
The applications will be screened by a joint committee formed by representatives of LTI and representatives of the Portuguese Universities involved in the joint degree program. The candidates should indicate their scores in GRE and TOEFL tests.
Letters of recommendation are due by January 3rd.
Despite this particular focus on the Portuguese language, applications are not restricted to native or non-native speakers of Portuguese.
Back to Top -
Vitaver-Recruiting agency: Sr Manager Audio algorithm development
Position Code, Title and Location: 1515 - 0AK0 - Sr Manager Audio Algorithm Development -
Irvine, CA (Orange County)
Start Date: ASAP Remote or Onsite: On location. No telecommuting or remote work.
Travel required: Occasionally. Additional Information: Below is all the information I have
from the Client. Once I setup your interview, you will have the chance to ask them
directly anything I do not include here.
Description: Our Client is currently seeking a talented Engineering Manager to
manage a group of engineers developing and productizing audio algorithms
for Bluetooth audio devices (BT headsets, PNDs, car kits, etc)
- Work closely with marketing to define the product roadmaps, priorities, and schedules
- Work closely with tier 1 customers to describe and promote specific audio algorithms
- Work closely with application engineering and customers groups to ensure successful
adoption of designed algorithm
Requirements: - Masters/BSc in engineering with a minimum of 10 years of industrial
experience in speech and audio processing
- Experience with implementation of speech enhancement devices (examples: VoIP,
speech compression, echo cancellation, noise suppression, beamforming, blind source
separation) on signal processing devices
- Experience working closely with marketing to define product directions and priorities
- Experience with the definition of requirements on telephony devices (for example,
BT headsets, cellular handsets, or cordless telephones)
- Some experience with optimization/evaluation of terminal (headset/ handset/ handsfree)
acoustics
- Experience with programming on RISC CPUs and/or DSPs
- Experience with managing groups of engineers
- Strong analytical and problem solving skills
- Experience with working closely with customers in both a customer support and
technical marketing role
Please tell me if you feel confident with the requirements and comfortable delivering
on them. As soon as you send me your reply to all questions below, and your resume
attached, we will talk on the phone. A technical interview with the Client will follow
as the last step.
1. Are you a US citizen/Green Card/H visa holder (please, specify)?
2. When will you become available?
3. What is the yearly salary you expect? $ _______ per year on
W2 (Fulltime Employee)
4. Very important: this is designed to save us both time and expedite a decision
by my Client.
Please complete the following Skill Matrix, answering each question with
A) number of years of experience with the skill,
B) your skill level on a scale from 1 to 5 (highest),
C) the last time you applied it. It has to be consistent with your resume. I.e.:
Experience 4 years - Skill level 5 - Last Applied January 07
1) Industrial experience in speech and audio processing (min 10 years): __ years;
Skill level: __; Last applied: __.
2) Experience implementing speech enhancement devices on signal processing
devices: __ years; Skill level: __; Last applied: __.
3) Experience with optimization/evaluation of terminal
(headset/ handset/ handsfree) acoustics: __ years; Skill level: __; Last applied: __.
4) Programming on RISC CPUs and/or DSPs: __ years; Skill level: __;
Last applied: __.
5) Do you have Bachelor's degree?
* Please attach your resume as a Word document
Alice Kondrat
Senior Staff Recruiter Vitaver & Associates, Inc. 2385 Executive Center Drive
#100 Boca Raton, FL 33431
Alice@vitaver.com
Direct Line: (561) 283-1136 Voice Line: (954) 840-3603 Fax: (866) 259 3777
http://www.vitaver.com/
-
Nuance: Software engineer speech dialog tools
In order to strengthen our Embedded ASR Research team, we are looking for a:
SOFTWARE ENGINEER SPEECH DIALOGUE TOOLS
As part of our team, you will be creating solutions for voice user interfaces for embedded applications on mobile and automotive platforms.
OVERVIEW:
- You will work in Nuance's Embedded ASR R&D team, developing technology, tools, and run-time software to enable our customers to develop and test embedded speech applications. Together with our team of speech and language experts, you will work on natural language dialogue systems for our customers in the Automotive and Mobile sector.
- You will work either at Nuance's Office in Aachen, a beautiful, old city right in the heart of Europe with great history and culture, or at Nuance's International Headquarters in Merelbeke, a small town just 5km away from the heart of the vibrant and picturesque city of Ghent, in the Flanders region of Belgium. Both Aachen and Ghent offer some of the most spectacular historic town centers in Europe, and are home to large international universities.
- You will work in an international company and cooperate with people on various locations including in Europe, America and Asia. You may occasionally be asked to travel.
RESPONSIBILITIES:
- You will work on the development of tools and solutions for cutting edge speech and language understanding technologies for automotive and mobile devices.
- You will work on enhancing various aspects of our advanced natural language dialogue system, such as the layer of connected applications, the configuration setup, inter-module communication, etc.
- In particular, you will be responsible for the design, implementation, evaluation, optimization and testing, and documentation of tools such as GUI and XML applications that are used to develop, configure, and fine-tune advanced dialogue systems.
QUALIFICATIONS:
- You have a university degree in computer science, engineering, mathematics, physics, computational linguistics, or a related field.
- You have very strong software and programming skills, especially in C/C++, ideally also for embedded applications.
- You have experience with Python or other scripting languages.
- GUI programming experience is a strong asset.
The following skills are a plus:
- Understanding of communication protocols
- Understanding of databases
- Understanding of computational agents and related frameworks (such as OAA).
- A background in (computational) linguistics, dialogue systems, speech processing, grammars, and parsing techniques, statistics and machine learning, especially as related to natural language processing, dialogue, and representation of information
- You can work both as a team player and as goal-oriented independent software engineer.
- You can work in a multi-national team and communicate effectively with people of different cultures.
- You have a strong desire to make things really work in practice, on hardware platforms with limited memory and processing power.
- You are fluent in English and you can write high quality documentation.
- Knowledge of other languages is a plus.
CONTACT:
Please send your applications, including cover letter, CV, and related documents (maximum 5MB total for all documents, please) to
Deanna Roe Deanna.roe@nuance.com
Please make sure to document to us your excellent software engineering skills.
ABOUT US:
Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world. Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, or digitally reproducing documents that can be shared and searched. With more than 3000 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about.
-
Nuance: Speech scientist London UK
Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world. Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, or digitally reproducing documents that can be shared and searched. With more than 2000 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about.
To strengthen our International Professional Services team, based in London, we are currently looking for a
Speech Scientist, London, UK
Nuance Professional Services (PS) has designed, developed, and optimized thousands of speech systems across dozens of industries, including directory search, call center automation, applications in telecom, finance, airline, healthcare, and other verticals; applications for video games, mobile dictation, enhanced search services, SMS, and in-car navigation. Nuance PS applications have automated approximately 7 billion phone conversations for some of the world's most respected companies, including British Airways, Vodafone, Amtrak, Bank of America, BellCanada, Citigroup, General Electric, NTT and Verizon.
The PS organization consists of energetic, motivated, and friendly individuals. The Speech Scientists in PS are among the best and brightest, with PhDs from universities such as Cambridge (UK), MIT, McGill, Harvard, Penn, CMU, and Georgia Tech, and having worked at research labs such Bell Labs, Motorola Labs, and ATR (Japan), culminating in over 300 years of Speech Science experience and covering well over 20 languages.
Come and join Nuance PS and work on the latest technology from one of the prominent speech recognition technology providers, and make a difference in the way the world communicates.
Job Overview
As a Speech Scientist in the Professional Services group, you will work on automated speech recognition applications, covering a broad range of activities in all project phases, including the design, development, and optimization of the system. You will:
- Work across application development teams to ensure best possible recognition performance in deployed systems
- Identify recognition challenges and assess accuracy feasibility during the design phase,
- Design, develop, and test VoiceXML grammars and create JSPs, Java, and ECMAscript grammars for dynamic contexts
- Optimize accuracy of applications by analyzing performance and tuning statistical language models, pronunciations, and acoustic models, including identifying areas for improvement by running the recognizer offline
- Contribute to the generation and presentation of client-facing reports
- Act as technical lead on more intensive client projects
- Develop methodologies, scripts, procedures that improve efficiency and quality
- Develop tools and enhance algorithms that facilitate deployment and tuning of recognition components
- Act as subject matter domain expert for specific knowledge domains
- Provide input into the design of future product releases
Required Skills
- MS or PhD in Computer Science, Engineering, Computational Linguistics, Physics, Mathematics, or related field (or equivalent)
- Strong analytical and problem solving skills and ability to troubleshoot issues
- Good judgment and quick-thinking
- Strong programming skills, preferably Perl or Python
- Excellent written and verbal communications skills
- Ability to scope work taking technical, business and time-frame constraints into consideration
- Works well in a team and in a fast-paced environment
Beneficial Skills
- Strong programming skills in either Perl, Python, Java, C/C++, or Matlab
- Speech recognition knowledge
- Strong pattern recognition, linguistics, signal processing, or acoustics knowledge
- Statistical data analysis
- Experience with XML, VoiceXML, and Wiki
- Ability to mentor or supervise others
- Additional language skills, eg French, Dutch, German, Spanish
-
Nuance: Research engineer speech engine
n order to strengthen our Embedded ASR Research team, we are looking for a:
RESEARCH ENGINEER SPEECH ENGINE
As part of our team, you will be creating solutions for voice user interfaces for embedded applications on mobile and automotive platforms.
OVERVIEW:
- You will work in Nuance's Embedded ASR R&D team, developing, improving and maintaining core ASR engine algorithms for our customers in the Automotive and Mobile sector.
- You will work either at Nuance's Office in Aachen, a beautiful, old city right in the heart of Europe with great history and culture, or at Nuance's International Headquarters in Merelbeke, a small town just 5km away from the heart of the vibrant and picturesque city of Ghent, in the Flanders region of Belgium. Both Aachen and Ghent offer some of the most spectacular historic town centers in Europe, and are home to large international universities.
- You will work in an international company and cooperate with people on various locations including in Europe, America and Asia. You may occasionally be asked to travel.
RESPONSIBILITIES:
- You will work on the developing, improving and maintaining core ASR engine algorithms for cutting edge speech and natural language understanding technologies for automotive and mobile devices.
- You will work on the design and development of more efficient, flexible ASR search algorithms with high focus on low memory and processor requirements.
QUALIFICATIONS:
- You have a university degree in computer science, engineering, mathematics, physics, computational linguistics, or a related field. PhD is a plus.
- A background in (computational) linguistics, speech processing, ASR search, confidence values, grammars, statistics and machine learning, especially as related to natural language processing.
- You have very strong software and programming skills, especially in C/C++, ideally also for embedded applications.
The following skills are a plus:
- You have experience with Python or other scripting languages.
- Broad knowledge about architectures of embedded platforms and processors.
- Understanding of databases
- You can work both as a team player and as goal-oriented independent software engineer.
- You can work in a multi-national team and communicate effectively with people of different cultures.
- You have a strong desire to make things really work in practice, on hardware platforms with limited memory and processing power.
- You are fluent in English and you can write high quality documentation.
- Knowledge of other languages is a plus.
CONTACT:
Please send your applications, including cover letter, CV, and related documents (maximum 5MB total for all documents, please) to
Deanna Roe Deanna.roe@nuance.com
Please make sure to document to us your excellent software engineering skills.
ABOUT US:
Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world. Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, or digitally reproducing documents that can be shared and searched. With more than 3000 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about.
-
Nuance RESEARCH ENGINEER SPEECH DIALOG SYSTEMS:
In order to strengthen our Embedded ASR Research team, we are looking for a:
RESEARCH ENGINEER SPEECH DIALOGUE SYSTEMS
As part of our team, you will be creating speech technologies for embedded applications varying from simple command and control tasks up to natural language speech dialogues on mobile and automotive platforms.
OVERVIEW:
-You will work in Nuance's Embedded ASR research and production team, creating technology, tools and runtime software to enable our customers develop embedded speech applications. In our team of speech and language experts, you will work on natural language dialogue systems that define the state of the art.
- You will work at Nuance's International Headquarters in Merelbeke, a small town just 5km away from the heart of the picturesque city of Ghent, in the Flanders region of Belgium. Ghent has one of the most spectacular historic town centers of Europe and is known for its unique vibrant yet cozy charm, and is home to a large international university.
- You will work in an international company and cooperate with people on various locations including in Europe, America, and Asia. You may occasionally be asked to travel.
RESPONSIBILITIES:
- You will work on the development of cutting edge natural language dialogue and speech recognition technologies for automotive embedded systems and mobile devices.
- You will design, implement, evaluate, optimize, and test new algorithms and tools for our speech recognition systems, both for research prototypes and deployed products, including all aspects of dialogue systems design, such as architecture, natural language understanding, dialogue modeling, statistical framework, and so forth.
- You will help the engine process multi-lingual natural and spontaneous speech in various noise conditions, given the challenging memory and processing power constraints of the embedded world.
QUALIFICATIONS:
- You have a university degree in computer science, (computational) linguistics, engineering, mathematics, physics, or a related field. A graduate degree is an asset.
-You have strong software and programming skills, especially in C/C++, ideally for embedded applications. Knowledge of Python or other scripting languages is a plus. [HQ1]
- You have experience in one or more of the following fields:
dialogue systems
applied (computational) linguistics
natural language understanding
language generation
search engines
speech recognition
grammars and parsing techniques.
statistics and machine learning techniques
XML processing
-You are a team player, willing to take initiative and assume responsibility for your tasks, and are goal-oriented.
-You can work in a multi-national team and communicate effectively with people of different cultures.
-You have a strong desire to make things really work in practice, on hardware platforms with limited memory and processing power.
-You are fluent in English and you can write high quality documentation.
-Knowledge of other languages is a strong asset.
CONTACT:
Please send your applications, including cover letter, CV, and related documents (maximum 5MB total for all documents, please) to
Deanna Roe Deanna.roe@nuance.com
ABOUT US:
Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world. Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, or digitally reproducing documents that can be shared and searched. With more than 3000 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about.
Journals
-
Papers accepted for FUTURE PUBLICATION in Speech
Full text available on http://www.sciencedirect.com/ for Speech Communication subscribers and subscribing institutions. Free access for all to the titles and abstracts of all volumes and even by clicking on Articles in press and then Selected papers.
-
Special Issue on Non-Linear and Non-Conventional Speech Processing-Speech Communication
Speech Communication
Call for Papers: Special Issue on Non-Linear and Non-Conventional Speech Processing
Editors: Mohamed CHETOUANI, UPMC
Marcos FAUNDEZ-ZANUY, EUPMt (UPC)
Bruno GAS, UPMC
Jean Luc ZARADER, UPMC
Amir HUSSAIN, Stirling
Kuldip PALIWAL, Griffith University
The field of speech processing has shown a very fast development in the past twenty years, thanks to both technological progress and to the convergence of research into a few mainstream approaches. However, some specificities of the speech signal are still not well addressed by the current models. New models and processing techniques need to be investigated in order to foster and/or accompany future progress, even if they do not match immediately the level of performance and understanding of the current state-of-the-art approaches.
An ISCA-ITRW Workshop on "Non-Linear Speech Processing" will be held in May 2007, the purpose of which will be to present and discuss novel ideas, works and results related to alternative techniques for speech processing departing from the mainstream approaches: http://www.congres.upmc.fr/nolisp2007
We are now soliciting journal papers not only from workshop participants but also from other researchers for a special issue of Speech Communication on "Non-Linear and Non-Conventional Speech Processing"
Submissions are invited on the following broad topic areas:
I. Non-Linear Approximation and Estimation
II. Non-Linear Oscillators and Predictors
III. Higher-Order Statistics
IV. Independent Component Analysis
V. Nearest Neighbours
VI. Neural Networks
VII. Decision Trees
VIII. Non-Parametric Models
IX. Dynamics of Non-Linear Systems
X. Fractal Methods
XI. Chaos Modelling
XII. Non-Linear Differential Equations
All fields of speech processing are targeted by the special issue, namely :
1. Speech Production
2. Speech Analysis and Modelling
3. Speech Coding
4. Speech Synthesis
5. Speech Recognition
6. Speaker Identification / Verification
7. Speech Enhancement / Separation
8. Speech Perception
Back to Top -
IEEE Transactions on Audio, Speech and Language Processing
Special Issue on Multimodal Processing in Speech -based Interactions in the IEEE Transactions on Audio, Speech and Language Processing has been extended to January 15, 2008. A URL to point to is: http://www.ewh.ieee.org/soc/sps/tap/sp_issue/special-issue-multimodal.pdf
Back to Top
Future Conferences
-
Publication policy: Hereunder, you will find very short announcements of future events. The full call for participation can be accessed on the conference websites
Back to Top
See also our Web pages (http://www.isca-speech.org/) on conferences and workshops.
Future Interspeech conferences
-
INTERSPEECH 2008-ICSLP
September 22-26, 2008, Brisbane, Queensland, Australia
Conference Website
Chairman: Denis Burnham, MARCS, University of West Sydney. -
INTERSPEECH 2009-EUROSPEECH
Brighton, UK,
Conference Website
Chairman: Prof. Roger Moore, University of Sheffield. -
INTERSPEECH 2010-ICSLP
Chiba, Japan
Conference Website
ISCA is pleased to announce that INTERSPEECH 2010 will take place in Makuhari-Messe, Chiba, Japan, September 26-30, 2010. The event will be chaired by Keikichi Hirose (Univ. Tokyo), and will have as a theme "Towards Spoken Language Processing for All - Regardless of Age, Health Conditions, Native Languages, Environment, etc."
Future ISCA Technical and Research Workshops
-
ITRW Odyssey 2008
The Speaker and Language Recognition Workshop
Back to Top
21-25 January 2008, Stellenbosch, South Africa
Topics
* Speaker recognition(identification, verification, segmentation, clustering)
* Text dependent and independent speaker recognition
* Multispeaker training and detection
* Speaker characterization and adaptation
* Features for speaker recognition
* Robustness in channels
* Robust classification and fusion
* Speaker recognition corporaand evaluation
* Use of extended training data
* Speaker recognition with speaker recognition
* Forensics, multimodality and multimedia speaker recogntion
* Speaker and language confidence estimation
* Language, dialect and accent recognition
* Speaker synthesis and transformation
* Biometrics
* Human recognition
* Commercial applications
Paper submission
Proaspective authors are invited to submit papers written in English via the Odyssey website. The style guide, templates,and submission form can be downloaded from the Odyssey website. Two members of the scientific committee will review each paper. Each accepted paper must have at least one registered author. The Proceedings will be published on CD
Schedule
Draft paper due July 15, 2007
Notification of acceptance September 15,2007
Final paper due October 30, 2007
Preliminary program November 30, 2007
Workshop January 21-25, 2008
Futher informations: venue, registation...
On the workshop website
Chairs
Niko Brummer, Spescom Data Voice, South Africa
Johan du Preez.Stellenbosch University,South Africa -
ISCA ITRW speech analysis and processing for knowledge discovery
June 4 - 6, 2008
Back to Top
Aalborg, Denmark
Workshop website
Humans are very efficient at capturing information and messages in speech, and they often perform this task effortlessly even when the signal is degraded by noise, reverberation and channel effects. In contrast, when a speech signal is processed by conventional spectral analysis methods, significant cues and useful information in speech are usually not taken proper advantage of, resulting in sub-optimal performance in many speech systems. There exists, however, a vast literature on speech production and perception mechanisms and their impacts on acoustic phonetics that could be more effectively utilized in modern speech systems. A re-examination of these knowledge sources is needed. On the other hand, recent advances in speech modelling and processing and the availability of a huge collection of multilingual speech data have provided an unprecedented opportunity for acoustic phoneticians to revise and strengthen their knowledge and develop new theories. Such a collaborative effort between science and technology is beneficial to the speech community and it is likely to lead to a paradigm shift for designing next-generation speech algorithms and systems. This, however, calls for a focussed attention to be devoted to analysis and processing techniques aiming at a more effective extraction of information and knowledge in speech.
Objectives:
The objective of this workshop is to discuss innovative approaches to the analysis of speech signals, so that it can bring out the subtle and unique characteristics of speech and speaker. This will also help in discovering speech cues useful for improving the performance of speech systems significantly. Several attempts have been made in the past to explore speech analysis methods that can bridge the gap between human and machine processing of speech. In particular, the time varying aspects of interactions between excitation and vocal tract systems during production seem to elude exploitation. Some of the explored methods include all-pole and polezero modelling methods based on temporal weighting of the prediction errors, interpreting the zeros of speech spectra, analysis of phase in the time and transform domains, nonlinear (neural network) models for information extraction and integration, etc. Such studies may also bring out some finer details of speech signals, which may have implications in determining the acoustic-phonetic cues needed for developing robust speech systems.
The Workshop:
G will present a full-morning common tutorial to give an overview of the present stage of research linked to the subject of the workshop
G will be organised as a single series of oral and poster presentations
G each oral presentation is given 30 minutes to allow for ample time for discussion
G is an ideal forum for speech scientists to discuss the perspectives that will further future research collaborations.
Potential Topic areas:
G Parametric and nonparametric models
G New all-pole and pole-zero spectral modelling
G Temporal modelling
G Non-spectral processing (group delay etc)
G Integration of spectral and temporal processing
G Biologically-inspired speech analysis and processing
G Interactions between excitation and vocal tract systems
G Characterization and representation of acoustic phonetic attributes
G Attributed-based speaker and spoken language characterization
G Analysis and processing for detecting acoustic phonetic attributes
G Language independent aspects of acoustic phonetic attributes detection
G Detection of language-specific acoustic phonetic attributes
G Acoustic to linguistic and acoustic phonetic mapping
G Mapping from acoustic signal to articulator configurations
G Merging of synchronous and asynchronous information
G Other related topics
Call for papers. Notification of review:
The submission deadline is January 31, 2008.
Registration
Fees for early and late registration for ISCA and non-ISCA members will be made available on the website during September 2007.
Venue:
The workshop will take place at Aalborg University, Department of Electronic Systems, Denmark. See the workshop website for further and latest information.
Accommodation:
There are a large number of hotels in Aalborg most of them close to the city centre. The list of hotels, their web sites and telephone numbers are given on the workshop website. Here you will also find information about transportation between the city centre and the university campus.
How to reach Aalborg:
Aalborg Airport is half an hour away from the international Copenhagen Airport. There are many daily flight connections between Copenhagen and Aalborg. Flying with Scandinavian Airlines System (SAS) or one of the Star Alliance companies to Copenhagen enables you to include Copenhagen-Aalborg into the entire ticket, and this way reducing the full transportation cost. There is also an hourly train connection between the two cities; the train ride lasts approx. five hours
Organising Committee:
Paul Dalsgaard, B. Yegnanarayana, Chin-Hui Lee, Paavo Alku, Rolf Carlson, Torbjørn Svendsen,
Important dates
Submission of full and final: January 31, 2008 on the Website
http://www.es.aau.dk/ITRW/
Notification of review results: No later than March 30., 2008. -
ITRW on Evidence-based Voice and Speech Rehabilitation in Head
May 2008, Amsterdam, The Netherlands,
Back to Top
Cancer in the head and neck area and its treatment can have debilitating effects on communication. Currently available treatment options such as radiotherapy, surgery, chemo-radiation, or a combination of these can often be curative. However, each of these options affects parts of the vocal tract and/or voice to a more or lesser degree. When the vocal tract or voice no longer functions optimally, this affects communication. For example, radiotherapy can result in poor voice quality, limiting the speaker's vocal performance (fatigue from speaking, avoidance of certain communicative situations, etc.). Surgical removal of the larynx necessitates an alternative voicing source, which generally results in a poor voice quality, but further affects intelligibility and the prosodic structure of speech. Similarly, a commando procedure (resection involving portions of the mandible / floor of the mouth / mobile tongue) can have a negative effect on speech intelligibility. This 2 day tutorial and research workshop will focus on evidence-based rehabilitation of voice and speech in head and neck oncology. There will be 4 half day sessions, 3 of which will deal with issues concerning total laryngectomy. One session will be devoted to research on rehabilitation of other head and neck cancer sites. The chairpersons of each session will prepare a work document on the specific topic at hand (together with the two keynote lecturers assigned), which will be discussed in a subsequent round table session. After this there will be a 30' poster session, allowing 9-10 short presentations. Each presentation consists of maximally 4 slides, and is meant to highlight the poster's key points. Posters will be visited in the subsequent poster visit session. The final work document will refer to all research presently available, discuss its (clinical) relevance, and will attempt to provide directions for future research. The combined work document, keynote lectures and poster abstracts/papers will be published under the auspices of ISCA.
Organizers
prof. dr. Frans JM Hilgers
prof. dr. Louis CW Pols,
dr. Maya van Rossum.
Sponsoring institutions:
Institute of Phonetic Sciences - Amsterdam Center for Language and Communication,
The Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital
Dates and submission details as well as a website address will be announced in a later issue. -
ISCA TR Workshop on Experimental Linguistics
August 2008, Athens, Greece
Back to Top
Website
Prof. Antonis Botinis -
Audio Visual Speech Processing Workshop (AVSP )
Tentative location:Queensland coast near Brisbane (most likely South Stradbroke Island)
Back to Top
Tentative date: 27-29 September 2008 (immediately after Interspeech 2008)
Following in the footsteps of previous AVSP workshops / conferences, AVSP workshop (ISCA Research and Tutorial Workshop) will be hold concomitantly to Interspeech2008, Brisbane, Australia, 22-26 September 2008. The aim of AVSP2008 is to bring together researchers and practitioners in areas related to auditory-visual speech processing. These include human and machine AVSP, linguistics, psychology, and computer science. One of the aims of the AVSP workshops is to foster collaborations across disciplines, as AVSP research is inherently multi-disciplinary. The workshop will include a number of tutorials / keynote addresses by internationally renowned researchers in the area of AVSP.
Organizers
Roland Goecke, Simon Lucey, Patrick Lucey
Australian National University,RSISE, Bldg. 115, Australian National University, Canberra, ACT 0200, Australia -
Robust ASR Workshop
Santiago, Chile
October-November 2008
Dr. Nestor Yoma
Forthcoming events supported (but not organized) by ISCA
-
3rd International Conference on Large-scale Knowledge Resources (LKR 2008)
3-5 March, 2008, Tokyo Institute of Technology, Tokyo Japan
Back to Top
Website
Sponsored by: 21st Century Center of Excellence (COE) Program "Framework for Systematization and Application of Large-scale",Tokyo Institute of Technology
In the 21st century, we are now on the way to the knowledge-intensive society in which knowledge plays ever more important roles. Research interest should inevitably shift from information to knowledge, namely how to build, organize, maintain and utilize knowledge are the central issues in a wide variety of fields. The 21st Century COE program, "Framework for Systematization and Application of Large-scale Knowledge Resources (COE-LKR)" conducted by Tokyo Institute of Technology is one of the attempts to challenge these important issues. Inspired by this project, LKR2008 aims at bringing together diverse contribution in cognitive science, computer science, education and linguistics to explore design, construction, extension, maintenance, validation, and application of knowledge.
Topics of Interest to the conference includes:
Infrastructure for Large-scale Knowledge
Grid computing
Network computing
Software tools and development environments
Database and archiving systems
Mobile and ubiquitous computing
Systematization for Large-scale Knowledge
Language resources
Multi-modal resources
Classification, Clustering
Formal systems
Knowledge representation and ontology
Semantic Web
Cognitive systems
Collaborative knowledge
Applications and Evaluation of Large-scale Knowledge
Archives for science and art
Educational media
Information access
Document analysis
Multi-modal human interface
Web applications
Organizing committee General conference chair: Furui, Sadaoki (Tokyo Institute of Technology)
Program co-chairs: Ortega, Antonio (University of Southern California)
Tokunaga, Takenobu (Tokyo Institute of Technology)
Publication chair: Yonezaki, Naoki (Tokyo Institute of Technology)
Publicity chair: Yokota, Haruo (Tokyo Institute of Technology)
Local organizing chair: Shinoda, Koichi (Tokyo Institute of Technology)
Submission
Since we are aiming at an interdisciplinary conference covering wide range of topics concerning large-scale knowledge resources, authors are requested to add general introductory description in the beginning of the paper so that readers of other research area can understand the importance of the work. Note that one of the reviewers of each paper is assigned from other topic area to see if this requirement is fulfilled.
There are two categories of paper presentation: oral and poster. The category of the paper should be stated at submission. Authors are invited to submit original unpublished research papers, in English, up to 12 pages for oral presentation and 4 pages for poster presentation, strictly following the LNCS/LNAI format guidelines available at the Springer LNCS Web page. . Details of the submission procedure will be announced later on.
Reviewing
The reviewing of the papers will be blind and managed by an international Conference Program Committee consisting of Area Chairs and associated Program Committee Members. Final decisions on the technical program will be made by a meeting of the Program Co-Chairs and Area Chairs. Each submission will be reviewed by at least three program committee members, and one of the reviewers is assigned from a different topic area.
Publication
The conference proceedings will be published by Springer-Verlag in their Lecture Notes in Artificial Intelligence (LNAI), which will be available at the conference.
Important dates
Paper submission deadline: 30 August, 2007
Notification of acceptance: 10 October, 2007
Camera ready papers due: 10 November, 2007
e-mail correspondence -
Call for Papers (Preliminary version) Speech Prosody 2008
Campinas, Brazil, May 6-9, 2008
Speech Prosody 2008 will be the fourth conference of a series of international events of the Special Interest Groups on Speech Prosody (ISCA), starting by the one held in Aix-en Provence, France, in 2002. The conferences in Nara, Japan (2004), and in Dresden, Germany (2006) followed the proposal of biennial meetings, and now is the time of changing place and hemisphere by trying the challenge of offering a non-stereotypical view of Brazil. It is a great pleasure for our labs to host the fourth International Conference on Speech Prosody in Campinas, Brazil, the second major city of the State of São Paulo. It is worth highlighting that prosody covers a multidisciplinary area of research involving scientists from very different backgrounds and traditions, including linguistics and phonetics, conversation analysis, semantics and pragmatics, sociolinguistics, acoustics, speech synthesis and recognition, cognitive psychology, neuroscience, speech therapy, language teaching, and related fields. Information: sp2008_info@iel.unicamp.br. Web site: http://sp2008.org. We invite all participants to contribute with papers presenting original research from all areas of speech prosody, especially, but nor limited to the following.
Scientific Topics
Prosody and the Brain
Long-Term Voice Quality
Intonation and Rhythm Analysis and Modelling
Syntax, Semantics, Pragmatics and Prosody
Cross-linguistic Studies of Prosody
Prosodic variability
Prosody in Discourse
Dialogues and Spontaneous Speech
Prosody of Expressive Speech
Perception of Prosody
Prosody in Speech Synthesis
Prosody in Speech Recognition and Understanding
Prosody in Language Learning and Acquisition
Pathology of Prosody and Aids for the Impaired
Prosody Annotation in Speech Corpora
Others (please, specify)
Organising institutions
Speech Prosody Studies Group, IEL/Unicamp | Lab. de Fonética, FALE/UFMG | LIACC, LAEL, PUC-SP
Important Dates
Call for Papers: May 15, 2007
Full Paper Submission: Nov. 2nd, 2007
Notif. of Acceptance: Dec. 14th, 2007
Early Registration: Jan. 14th, 2008
Conference: May 6-9, 2008
-
CFP:The International Workshop on Spoken Languages Technologies for Under- The International Workshop on Spoken Languages Technologies for Under-resourced languages (SLTU)
The International Workshop on Spoken Languages Technologies for Under-resourced languages (SLTU)
languages (SLTU) Hanoi University of Technology, Hanoi, Vietnam,
May 5 - May 7, 2008.
Workshop Web Site : http://www.mica.edu.vn/sltu
The STLU meeting is a technical conference focused on spoken language processing for
under-resourced languages. This first workshop will focus on Asian languages, and
the idea is to mainly (but not exclusively) target languages of the area (Vietnamese,
Khmer, Lao, Chinese dialects, Thai, etc.). However, all contributions on other
under-resourced languages of the world are warmly welcomed. The workshop aims
at gathering researchers working on:
* ASR, synthesis and speech translation for under-resourced languages
* portability issues * fast resources acquisition (speech, text, lexicons, parallel corpora)
* spoken language processing for languages with rich morphology
* spoken language processing for languages without separators
* spoken language processing for languages without writing system
Important dates
* Paper submission: January 15, 2008
* Notification of Paper Acceptance: February 20, 2008
* Author Registration Deadline: March 1, 2008 Scientific Committee
* Pr Tanja Schultz, CMU, USA
* Dr Yuqing Gao, IBM, USA
* Dr Lori Lamel, LIMSI, France
* Dr Laurent Besacier, LIG, France
* Dr Pascal Nocera, LIA, France
* Pr Jean-Paul Haton, LORIA, France
* Pr Luong Chi Mai, IOIT, Vietnam
* Pr Dang Van Chuyet, HUT, Vietnam
* Pr Pham Thi Ngoc Yen, MICA, Vietnam
* Dr Eric Castelli, MICA, Vietnam
* Dr Vincent Berment, LIG Laboratory, France
* Dr Briony Williams, University of Wales, UK
Local Organizing Committee
* Pr Nguyen Trong Giang, HUT/MICA
* Pr Ha Duyen Tu, HUT
* Pr Pham Thi Ngoc Yen, HUT/MICA
* Pr Geneviève Caelen-Haumont, MICA
* Dr Trinh Van Loan, HUT
* Dr Mathias Rossignol, MICA
* M. Hoang Xuan Lan, HUT
Back to Top
Future Speech Science and Technology Events
-
2ndes Journees de Phonetique clinique
Les inscriptions au colloque JPC2 "Deuxièmes Journees de Phonetique Clinique" qui se tiendra les 13 et 14 decembre 2007 a Grenoble sont ouvertes.
Le formulaire d'inscription est disponible sur le site a l'adresse suivante: http://www.icp.inpg.fr/JPC2/
Attention, Grenoble accueille tout au long de l'annee de nombreux colloques et seminaires; Nous vous conseillons donc de reserver un hotel dans les plus brefs delais (Une liste d'hebergements vous est proposee sur le site).
Back to Top -
5th International Workshop on Models and Analysis of Vocal Emissions for Biomedical Applications MAVEBA 2007
December 13 - 15, 2007
Back to Top
Conference Hall - Ente Cassa di Risparmio di Firenze
Via F. Portinari 5r, Firenze, Italy
DEADLINES:
EXTENDED DEADLINE: 15 June 2007 - Submission of extended abstracts (1-2 pages, 1 column), special session proposal
30 July, 2007 - Notification of paper acceptance
30 September 2007 - Final full paper submission (4 pages, 2 columns, pdf format) and early registration
13-15 December 2007
- Conference venue
CONTACT:
Dr. Claudia Manfredi - Conference Chair
Dept. of Electronics and Telecommunications
Universita degli Studi di Firenze
Via S. Marta 3
50139 Firenze, Italy
Phone: +39-055-4796410
Fax: +39-055-494569 -
e-Forensics 2008: The 1st International Conference on Forensic Applications and Techniques in Telecommunications, Information and Multimedia
with associated Workshops:
Back to Top
WKDD: First International Workshop on Knowledge Discovery and Data Mining
and WFST: First International Workshop on Forensics Sensor Technology
Adelaide, Australia, January 21-23, 2008
Website
The birth of the Internet as a commercial entity and the development of increasingly sophisticated digital consumer technology has led to new opportunities for criminals and new challenges for law enforcement. Those same technologies also offer new tools for scientific investigation of evidence.
In telecommunications, Voice over IP raises significant challenges for call intercept and route tracing. Information systems are becoming overwhelmingly large and the challenges of associating relevant information from one source with another require new sophisticated tools. And consumer multimedia devices, especially video and still cameras, are increasingly becoming the tools of choice to create potentially illegal content. At the same time, the scientific gathering and investigation of evidence at a crime scene is being pushed towards digital techniques, raising questions of the veracity and completeness of evidence.
PAPERS: The aim of this conference is to bring together state of the art research contributions in the development of tools, protocols and techniques which assist in the investigation of potentially illegal activity associated with electronic communication and electronic devices. Investigative practice and requirements for presentation of evidence in court are to be considered key underlying themes. This might include discovery, analysis, handling and storage of digital evidence; meeting the legal burden of proof; and the establishment of the forensic chain of evidence.
Technical papers describing original, previously unpublished research, not currently under review by another conference or journal, are solicited. Technical papers which clearly identify how the specific contributions fit to an overall working solution are particularly of interest.
Topics include, but are not limited to, the following:
* Voice over IP call tracing and intercept
* Records tracing and data mining
* Fraud management in commercial transactions
* Techniques for addressing identity theft
* Geo-location techniques for cellular, ad-hoc, wireless and IP network communications
* Distributed data association across massive, disparate database systems
* Data carving
* Multimedia source identification
* Image tamper identification
* Image association and recognition
* Motion analysis
* Voice analysis
* Watermarking and applications
* Transaction tracking
* Digital evidence storage and handling protocols
General papers on electronic security will be considered where there is a clear application to the underlying topic of forensic investigation.
Important Dates:
Paper Registration Deadline: September 28, 2007
Notification of Acceptance: November 16, 2007
Final paper submission and author's registration: December 3, 2007
Conference Dates: January 21-23, 2008
Committee
General Chair - Matthew Sorell, University of Adelaide, Australia
Technical Program Committee Chair - Chang-Tsun Li, University of Warwick, UK
Publicity Chair - Gale Spring, RMIT University, Australia
For details of the Workshops, and further information regarding paper submission and the conference, please refer to the conference website
Please direct all enquiries by e-mail -
LangTech2008
The language and speech technology conference.
Rome, 28-29 February 2008
San Michele a Ripa conference centre.
Website
We are most delighted to welcome you to join us at the LangTech2008 conference which will be held at the San Michele a Ripa convention center in Rome, February 28-29, 2008. After two successful national conferences on speech and language technology (2002, 2006), the ForumTal decided to promote an international event in the field. A follow up of the previous LangTech conferences (Berlin, Paris), LangTech2008 aims at giving a chance to the industrial and research communities, and public administration, to share and discuss language and speech technologies. The conference will feature world-class speakers, exhibits, lecture and poster sessions.
PAPERS SUBMISSION DEADLINE: 30th November 2007
EXHIBITION BOOTHS RESERVATION: Reduced Fares until 15th November 2007
REGISTRATION: Reduced Fees until 31st December 2007
A golden promotional opportunity for all language technology SMEs!
LangTech 2008, http://www.langtech.it/en/, the language technology business conference, is featuring a special elevator session for small and medium sized enterprises, SMEs.
An elevator session is a session with very short presentations.
If you seek business partners, you are invited to participate in LangTech 2008 in Rome, February 28-29, and make yourself known to the audience.
A committee of European experts shall choose a total of 10 SMEs from anywhere in Europe and beyond to give a 5 min self-promotional presentation in English before a floor of venture capitalists, business peers, large technology corporations and other interested parties.
A jury will select three of the presenting companies, and award the first, second and third LangTech Prize.
Submissions must be received by 30 December 2007.
The lucky candidates will be informed by 15 January 2008.
We will offer a reduced fee to LangTech 2008 to all SMEs selected to present at the elevator session.
If you wish to submit a request to present your SME for this unique opportunity, please contact sme@langtech.it immediately, and visit the web site dedicated to LangTech 2008, http://www.langtech.it/en/, where you can download a short slide set with guidelines for preparing your candidature.
Dr Calzolari would be pleased if you could spread the Conference Announcement and the Call for SMEs
Presentations to anyone you consider potentially interested in the event.
Dr PAOLA BARONI
Researcher
Consiglio Nazionale delle Ricerche
Istituto di Linguistica Computazionale
Area della Ricerca di Pisa
Via Giuseppe Moruzzi
56124 Pisa
ITALY
Phone: [+39] 050 315 2873
Fax: [+39] 050 315 2834
e-Mail: paola.baroni@ilc.cnr.it
URL: http://www.ilc.cnr.it
Skype: paola.baroni
Back to Top -
AVIOS
San Diego, March 10 - 12, 2008
Back to Top
The defining conference on Voice Search
From the Applied Voice Input Output Society and Bill Meisel's TMA Associates
Voice Search 2008 will be held at the San Diego Marriott Hotel and Marina, San Diego, California, March 10 - 12, 2008. Voice Search is a rapidly evolving technology and market. AVIOS (the Applied Voice Input Output Society) and Bill Meisel (president of TMA Associates and Editor of Speech Strategy News) are joining together to launch this new conference as a definitive resource for companies that will be impacted by this important trend.
"Voice Search" suggests an analogy to "Web Search," which has been a runaway success for both users and providers. The maturing of speech recognition and text-to-speech synthesis--and the recent involvement of large companies--has validated the availability of the core functionality necessary to support this analogy. The conference explores the possibilities, limitations, and differences of Voice Search and Web search.
Web search made the Web an effective and required marketing tool. Will Voice Search do the same for the telephone channel? The potential impact on call centers is another key issue covered by the conference.
The agenda covers:
§ What Voice Search is and will become
§ Applications of Voice Search
§ The appropriate use of speech technology to support voice search
§ Insight for service providers, enterprises, Web services, and call centers that want to take advantage of this new resource
§ Marketing channels and business models in Voice Search
§ Emerging supporting technology, development tools, and delivery platforms supporting Voice Search
§ Dealing with the surge of calls created by Voice Search.
Specific topics that will be covered at Voice Search 2008 include:
Applications
- Automated directory assistance and local search
- Voice information searches by telephone
- Ad-supported information access by phone
- Audio/Video searches on the Web and enterprises
- Speech analytics-extracting business intelligence from audio files
- Converting voicemail to searchable text
- Other new applications and services
- Application examples and demonstrations
Markets
- How the voice search market is developing
- The changing role of the telephone in marketing
- Business models
- The right way to deliver audio ads
- Justifying subscriber fees
Delivery
- Platforms, tools, and services for effectively delivering these applications
- Implementation examples and demonstrations
- Hosted versus customer-premises solutions
- Supporting multiple modes of interaction
- Key sources of technology and service
Contact centers
- The impact of Voice Search on contact centers
- Speech automation to handle the increased call flow
- Moving from handling problems to building customer relationships
Technology
- Speech recognition methods supporting voice search
- Text-to-speech quality and alternatives
- Supporting multimodal solutions
- Supporting standards
- Delivering responsive applications
- Voice User Interface issues and solutions in voice search
Sponsorships are available:
http://www.voicesearchconference.com/sponsor.htm
We're interested in proposals for speaking (available slots are limited):
http://www.voicesearchconference.com/talk.htm
Registration is open with an early-registration discount:
http://www.voicesearchconference.com/registration.htm
Other information:
What is Voice Search?
Voice Search News
About AVIOS
About Bill Meisel and TMA Associates
Or contact cONTACT. -
CfP-2nd INTERNATIONAL CONFERENCE ON LANGUAGE AND AUTOMATA THEORY AND APPLICATIONS (LATA 2008)
Tarragona, Spain, March 13-19, 2008
Back to Top
Website http://www.grlmc.com
AIMS:
LATA is a yearly conference in theoretical computer science and its applications. As linked to the International PhD School in Formal Languages and Applications that is being developed at the host institute since 2001, LATA 2008 will reserve significant room for young computer scientists at the beginning of their career. It will aim at attracting scholars from both classical theory fields and application areas (bioinformatics, systems biology, language technology, artificial intelligence, etc.). SCOPE:
Topics of either theoretical or applied interest include, but are not limited to:
- words, languages and automata
- grammars (Chomsky hierarchy, contextual, multidimensional, unification, categorial, etc.)
- grammars and automata architectures
- extended automata
- combinatorics on words
- language varieties and semigroups
- algebraic language theory
- computability
- computational and structural complexity
- decidability questions on words and languages
- patterns and codes
- symbolic dynamics
- regulated rewriting
- trees, tree languages and tree machines
- term rewriting
- graphs and graph transformation
- power series
- fuzzy and rough languages
- cellular automata
- DNA and other models of bio-inspired computing
- symbolic neural networks
- quantum, chemical and optical computing
- biomolecular nanotechnology
- automata and logic
- algorithms on automata and words
- automata for system analysis and programme verification
- automata, concurrency and Petri nets
- parsing
- weighted machines
- transducers
- foundations of finite state technology
- grammatical inference and algorithmic learning
- text retrieval, pattern matching and pattern recognition
- text algorithms
- string and combinatorial issues in computational biology and bioinformatics
- mathematical evolutionary genomics
- language-based cryptography
- data and image compression
- circuits and networks
- language-theoretic foundations of artificial intelligence and artificial life
- digital libraries
- document engineering
STRUCTURE:
LATA 2008 will consist of:
- 3 invited talks (to be announced in the second call for papers)
- 2 tutorials (to be announced in the second call for papers)
- refereed contributions
- open sessions for discussion in specific subfields or on professional issues
SUBMISSIONS:
Authors are invited to submit papers presenting original and unpublished research. Papers should not exceed 12 pages and should be formatted according to the usual LNCS article style. Submissions have to be sent through the web page.
PUBLICATION:
A volume of proceedings (expectedly LNCS) will be available by the time of the conference. A refereed volume of selected proceedings containing extended papers will be published soon after it as a special issue of a major journal.
REGISTRATION:
The period for registration will be open since January 7 to March 13, 2008. Details about how to register will be provided through the website of the conference.
Early registration fees: 250 euros
Early registration fees (PhD students): 100 euros
Registration fees: 350 euros
Registration fees (PhD students): 150 euros
FUNDING:
25 grants covering partial-board accommodation will be available for nonlocal PhD students. To apply, the candidate must e-mail her/his CV together with a copy of the document proving her/his status as a PhD student.
IMPORTANT DATES:
Paper submission: November 16, 2007
Application for funding (PhD students): December 7, 2007
Notification of funding acceptance or rejection: December 21, 2007
Notification of paper acceptance or rejection: January 18, 2008
Early registration: February 1, 2008
Final version of the paper for the proceedings: February 15, 2008
Starting of the conference: March 13, 2008
Submission to the journal issue: May 23, 2008
FURTHER INFORMATION:
E-mail
Website http://www.grlmc.com
ADDRESS:
LATA 2008
Research Group on Mathematical Linguistics
Rovira i Virgili University
Plaza Imperial Tarraco, 1
43005 Tarragona, Spain
Phone: +34-977-559543
Fax: +34-977-559597 -
CfP- LREC 2008 - 6th Language Resources and Evaluation Conference
Palais des Congrès Mansour Eddahbi, MARRAKECH - MOROCCO
MAIN CONFERENCE: 28-29-30 MAY 2008
WORKSHOPS and TUTORIALS: 26-27 MAY and 31 MAY- 1 JUNE 2008
Conference web site
The sixth international conference on Language Resources and Evaluation (LREC) will be organised in 2008 by ELRA in cooperation with a wide range of international associations and organisations.
CONFERENCE TOPICS
Issues in the design, construction and use of Language Resources (LRs): text, speech, multimodality
- Guidelines, standards, specifications, models and best practices for LRs
- Methodologies and tools for LRs construction and annotation
- Methodologies and tools for the extraction and acquisition of knowledge
- Ontologies and knowledge representation
- Terminology
- Integration between (multilingual) LRs, ontologies and Semantic Web technologies
- Metadata descriptions of LRs and metadata for semantic/content markup
Exploitation of LRs in different types of systems and applications
- For: information extraction, information retrieval, speech dictation, mobile communication, machine translation, summarisation, web services, semantic search, text mining, inferencing, reasoning, etc.
- In different types of interfaces: (speech-based) dialogue systems, natural language and multimodal/multisensorial interactions, voice activated services, etc.
- Communication with neighbouring fields of applications, e.g. e-government, e-culture, e-health, e-participation, mobile applications, etc.
- Industrial LRs requirements, user needs
Issues in Human Language Technologies evaluation
- HLT Evaluation methodologies, protocols and measures
- Validation, quality assurance, evaluation of LRs
- Benchmarking of systems and products Usability evaluation of HLT-based user interfaces, interactions and dialog systems
- Usability and user satisfaction evaluation
General issues regarding LRs & Evaluation
- National and international activities and projects
- Priorities, perspectives, strategies in national and international policies for LRs
- Open architectures
- Organisational, economical and legal issues
Special Highlights
LREC targets the integration of different types of LRs - spoken, written, and other modalities - and of the respective communities. To this end, LREC encourages submissions covering issues which are common to different types of LRs and language technologies.
LRs are currently developed and deployed in a much wider range of applications and domains. LREC 2008 recognises the need to encompass all those data that interact with language resources in an attempt to model more complex human processes and develop more complex systems, and encourages submissions on topics such as:
- Multimodal and multimedia systems, for Human-Machine interfaces, Human-Human interactions, and content processing
- Resources for modelling language-related cognitive processes, including emotions
- Interaction/Association of language and perception data, also for robotic systems
The Scientific Programme will include invited talks, oral presentations, poster and demo presentations, and panels. There is no difference in quality between oral and poster presentations. Only the appropriateness of the type of communication (more or less interactive) to the content of the paper will be considered.
SUBMISSIONS AND DATES
Submitted abstracts of papers for oral and poster or demo presentations should consist of about 1500-2000 words.
- Submission of proposals for oral and poster/demo papers: 31 October 2007
- Submission of proposals for panels, workshops and tutorials: 31 October 2007
The Proceedings on CD will include both oral and poster papers, in the same format. In addition a Book of Abstracts will be printed. Back to Top -
CALL for JEP/TALN/RECITAL 2008 - Avignon France
JEP-TALN-RECITAL'08
- XXVIIemes Journees d'Etude sur la Parole (JEP'08)
- 15eme conference sur le Traitement Automatique des Langues Naturelles (TALN'08)
- 10eme Rencontre des Etudiants Chercheurs en Informatique pour le Traitement Automatique des Langues (RECITAL'08)
Universite d'Avignon et des Pays de Vaucluse
Avignon du 9 au 13 Juin 2008Pour la troisieme fois, apres Nancy en 2002 et Fes en 2004, l'AFCP(Association Francophone pour la Communication Parlee) et l'ATALA (Association pour le Traitement Automatique des Langues) organisent conjointement leur principale conference afin de reunir en un seul lieu les deux communautes du traitement de la langue orale et ecrite.
Le colloque jeunes chercheurs RECITAL'08 est egalement associe a cet evenement.
Les conferences invitees ainsi qu'une session orale thematique seront organisees sous forme de sessions pleinieres communes aux trois conferences. L'inscription est commune aux trois evenements et les participants recevront l'ensemble des actes sur CDROM. La langue officielle est le francais.
Organise par l e LIA (Laboratoire Informatique d'Avignon), cet evenement se tiendra, du 9 au 13 juin 2008, a l'Universite d'Avignon et des Pays du Vaucluse (site centre ville - sainte Marthe).
Les JEP 2008 sont organisees sous l'egide de l'AFCP (Association Francophone pour la Communication Parlee), avec le soutien de l'ISCA (International Speech Communication Association).
CALENDRIER
Date limite de soumission: 11 fevrier 2008
Notification aux auteurs: 28 mars 2008
Confeence: 9-13 juin 2008
THEMES
Les communications porteront sur la communication parlee et le traitement de la parole dans leurs differents aspects. Les themes de la conference incluent, de facon non limitative:
Production de parole
Acoustique de la parole
Perception de parole
Phonetique et phonologie
Prosodie
Reconnaissance et comprehension de la parole
Reconnaissance de la langue et du locuteur
Modeles de langage
Synthese de la parole
Analyse, codage et compression de la parole
Applications a composantes orales (dialogue, indexation...)
Evaluation, corpus et ressources
Psycholinguistique
Acquisition de la parole et du langage
Apprentissage d'une langue seconde
Pathologies de la parole
CRITERES DE SELECTION
Les auteurs sont invites a soumettre des travaux de recherche
originaux, n'ayant pas fait l'objet de publications anterieures. Les
contributions proposees seront examinees par au moins deux
specialistes du domaine. Seront considerees en particulier:
- l'importance et l'originalite de la contribution,
- la correction du contenu scientifique et technique,
- la discussion critique des resultats, en particulier par rapport aux autres travaux du domaine
- la situation des travaux dans le contexte de la recherche internationale,
- l'organisation et la clarte de la presentation
- l'adequation aux themes de la conference.
Les articles selectionnes seront publies dans les actes de la conference.
MODALITES DE SOUMISSION
Les articles soumis ne devront pas depasser 4 pages en Times 10, sur deux colonnes, format A4. Une feuille de style LaTeX et un modee Word sont disponibles sur le site de la conference www.lia.univ-avignon.fr/jep-taln08/.
Les articles devront etre soumis - sous forme electronique - sur le site web de la conference avant le 11 fevrier 2008. Les documents devront etre envoyes exclusivement au format PDF.
BOURSES
L'AFCP offre un certain nombre de bourses pour les doctorants et jeunes chercheurs desireux de prendre part a la conference, voir : www.afcp-parole.org/doc/bourses.htm
L'ISCA apporte egalement un soutien financier aux jeunes chercheurs participant a des manifestations scientifiques sur la parole et le langage, voir : www.isca-speech.org/grants.html
CALL FOR WORKSHOPS AND TUTORIALSFor the third time, after Nancy in 2002 and Fes in 2004, the French speech association AFCP and the French NLP association ATALA are jointly organising their main conference in order to group together the two research community working in the fields of Speech and Natural Language Processing.
The conference will include oral and poster communications, invited conferences, workshops and tutorials. Workshop and tutorials will be held on June 13, 2008.
The official languages are French and English.
IMPORTANT DATES
Deadline for proposals: November 22nd 2007
Approval by the TALN committee: November 30th 2007
Final version for inclusion in the proceedings: April 4th 2008
Workshop and tutorials: June 13th 2008
OBJECTIVES
Workshops can be organized on any specific aspect of NLP. The aim of these sessions is to facilitate an in-depth discussion of this theme.
A workshop has its own president and its own program committee. The president is responsible for organizing a call for paper/participation and for the coordination of his program committee. The organizers ofthe main TALN conference will only take in charge the organization of the usual practical details (rooms, coffee breaks, proceedings).
Workshops will be organized in parallel sessions on the last day of the conference (2 to 4 sessions of 1:30).
Tutorials will be held on the same day.
HOW TO SUBMIT
Workshop and Tutorial proposals will be sent by email to taln08@atala.org before November 22nd, 2007.
** Workshop proposals will contain an abstract presenting the proposed theme, the program committee list and the expected length of the session.
** Tutorial proposals will contain an abstract presenting the proposed theme, a list of all the speakers and the expected length of the session (1 or 2 sessions of 1:30).
The TALN program committee will make a selection of the proposals and announce it on November 30th, 2008.
FORMAT
Conferences will be given in French or English (for non French native speakers). Papers to be published in the proceedings will conform to the TALN style sheet which is available on the conference web site. Worshop papers should not be longer than 10 pages in Times 12 (references included).
Contact: taln08@atala.org
-
4th IEEE Tutorial and Research Workshop on PERCEPTION AND INTERACTIVE TECHNOLOGIES FOR SPEECH-BASED SYSTEMS
June 16 - 18, 2008
Back to Top
Kloster Irsee, Germany
The 4th IEEE Tutorial and Research Workshop on PERCEPTION AND INTERACTIVE TECHNOLOGIES FOR SPEECH-BASED SYSTEMS (PIT08)will be held at the Kloster Irsee in southern Germany from June 16 to June 18, 2008.
The workshop focuses on advanced speech-based human-computer interaction where various contextual factors are modelled and taken into account when users communicate with computers. This includes mechanisms, architectures, design issues, applications, evaluation and tools. Prototype and product demonstrations will be very welcome.
The workshop will bring together researchers from various disciplines such as, for example, computer science and engineering sciences, medical, psychological and neurosciences, as well as mathematics. It will provide a forum for the presentation of research and applications and for lively discussions among researchers as well as industrialists in different fields.
WORKSHOP THEMES
Papers may discuss theories, applications, evaluation, limitations, general tools and techniques. Discussion papers that critically evaluate approaches or processing strategies and prototype demonstrations are especially welcome.
- Speech recognition and semantic analysis
- Dialogue management models
- Adaptive dialogue modelling
- Recognition of emotions from speech, gestures, facial expressions and physiological data
- User modelling
- Planning and reasoning capabilities for coordination and conflict description
- Conflict resolution in complex multi-level decisions
- Multi-modality such as graphics, gesture and speech for input and output
- Fusion and information management
- Computer-supported collaborative work
- Attention selection and guidance
- Learning and adaptability
- Visual processing and recognition for advanced human-computer interaction
- Databases and corpora
- Psychophysiological evaluation and usability analysis
- Evaluation strategies and paradigms
- Prototypes and products
WORKSHOP PROGRAMME
The format of the workshop will be a non-overlapping mixture of oral and poster sessions. A number of tutorial lectures will be given by internationally recognised experts from the area of Perception and Interactive Technologies for Speech-Based Systems.
All poster sessions will be opened by an oral summary by the session chair. A number of poster sessions will be succeeded by a discussion session focussing on the subject of the session. It is our belief that this general format will ensure a lively and valuable workshop.
The organisers would like to encourage researchers and industrialists to take the opportunity to bring their applications as well as their demonstrator prototypes and design tools for demonstration to the workshop. If sufficient interest is shown, a special demonstrator/poster session will be organised and followed by a discussion session.
The official language of the workshop is English. At the opening of the workshop hardcopies of the proceedings, published in the LNCS/LNAI/LNBI Series by Springer, will be available.
TIMING AND DATES
February 10, 2008: Deadline for Long, Short and Demo Papers
March 15, 2008: Author notification
April 1, 2008: Deadline for final submission of accepted paper
April 18, 2008: Deadline for early bird registration
June 7, 2008: Final programme available on web
June 16 - 18, 2008: Workshop
Further information may be found on our workshop website.
CONTACT
Wolfgang Minker,/a>
University of Ulm
Department of Information Technology
Albert-Einstein-Allee 43
D-89081 Ulm
Phone: +49 731 502 6254/-6251
Fax: +49 691 330 3925516 -
JHU Summer Workshop on Language Engineering
JHU Summer Workshops
CALL FOR TEAM RESEARCH PROPOSALS
Deadline: Wednesday, October 17, 2007
The Center for Language and Speech Processing at Johns Hopkins University invites one-page research proposals for a Summer Workshop on Language Engineering, to be held in Baltimore, MD, USA, July 7 to August 14, 2008.
Workshop proposals should be suitable for a six-week team exploration, and should aim to advance the state of the art in any of the various fields of Language Engineering including speech recognition, machine translation, information retrieval, text summarization and question answering. Research topics selected for investigation by teams in previous workshops may serve as good examples for your proposal. (See http://www.clsp.jhu.edu/workshops.)
This year's workshop will be sponsored by NSF and supported in part by the newly established Human Language Technology Center of Excellence (CoE). All relevant topics of scientific interest are welcomed. Proposals can receive special priority if they contribute to one of the following long-term challenges:
* AUTOMATIC POPULATION OF A KNOWLEDGE BASE FROM TEXT: Devise and develop technology to automatically populate a large knowledge base (KB) by accumulating entities, events, and relations from vast quantities of text from various formal and informal genres in multiple languages. Devise methods to do this effectively for resource rich and/or resource poor languages. The aim is to disambiguate and normalize entities, events, and relations in such a way that the KB could represent changes over time thus reflecting text sources.
* ROBUST TECHNOLOGY FOR SPEECH: Technologies like speech-to-text, speaker identification, and language identification share a common weakness: accuracy degrades disproportionately with changes in input (microphone, genre, speaker, etc.). Seemingly small amounts of noise or diverse data sources cause machines to break where humans would quickly and effectively adapt. The aim is to develop technology whose performance would be minimally degraded by input signal variations.
* PARALLEL PROCESSING FOR SPEECH AND LANGUAGE: A broad variety of pattern recognition problems in speech and language require a large amount of computation and must be run on a large amount of data. There is a need to optimize these algorithms to increase throughput and improve cost effectiveness. Proposals are invited both for novel parallelizable algorithms and for hardware configurations that achieve higher throughput or lower speed-power product than can be achieved by optimizing either alone.
An independent panel of experts will screen all received proposals for suitability. Results of this screening will be communicated no later than October 19, 2007. Authors passing this initial screening will be invited to Baltimore to present their ideas to a peer-review panel on November 2-4, 2007. It is expected that the proposals will be revised at this meeting to address any outstanding concerns or new ideas. Two or three research topics and the teams to tackle them will be selected for the 2008 workshop.
We attempt to bring the best researchers to the workshop to collaboratively pursue the selected topics for six weeks. Authors of successful proposals typically become the team leaders. Each topic brings together a diverse team of researchers and students. The senior participants come from academia, industry and government. Graduate student participants familiar with the field are selected in accordance with their demonstrated performance, usually by the senior researchers. Undergraduate participants, selected through a national search, will be rising seniors who are new to the field and have shown outstanding academic promise.
If you are interested to participate in the 2008 Summer Workshop we ask that you submit a one-page research proposal for consideration, detailing the problem to be addressed. If your proposal passes the initial screening, we will invite you to join us for the organizational meeting in Baltimore (as our guest) for further discussions aimed at consensus. If a topic in your area of interest is chosen as one of the two or three to be pursued next summer, we expect you to be available for participation in the six-week workshop. We are not asking for an ironclad commitment at this juncture, just a good faith understanding that if a project in your area of interest is chosen, you will actively pursue it.
Proposals should be submitted via e-mail to clsp@jhu.edu by 5PM ET on Wed, October 17, 2007.
-
4TH TUTORIAL AND RESEARCH WORKSHOP
4TH TUTORIAL AND RESEARCH WORKSHOP
PERCEPTION AND INTERACTIVE TECHNOLOGIES FOR SPEECH-BASED SYSTEMS
(PIT08)
will be held at the Kloster Irsee in southern Germany from June 16 to
June 18, 2008.
Please follow this link to visit our workshop website
http://it.e-technik.uni-ulm.de/World/Research.DS/irsee-workshops/pit08/introduction.html
Submissions will be short/demo or full papers of 4-10 pages.
Important dates:
February 10, 2008: Deadline for Long, Short and Demo Papers
March 15, 2008: Author notification
April 1, 2008: Deadline for final submission of accepted paper
April 18, 2008: Deadline for advance registration
June 7, 2008: Final programme available on the web
The workshop will be technically co-sponsored by the IEEE Signal
Processing Society. It is envisioned to publish the proceedings in the
LNCS/LNAI Series by Springer.
We welcome you to the workshop.
Elisabeth André, Laila Dybkjaer, Wolfgang Minker, Heiko Neumann,
Michael Weber, Roberto Pieraccini
PIT'08 Organising Committee
Back to Top -
eNTERFACE 2008 Orsay Paris
eNTERFACE'08 the next international summer workshop on multimodal
interfaces will take place at LIMSI, in Orsay (near Paris), France,
during four weeks, August 4th-29th, 2008.
Please consider proposing projects and participate to the workshop (see
the Call for Projects proposal on the web site or attached to this mail).
eNTERFACE08 is the next of a series of successful workshops initiated by
SIMILAR, the European Network of Excellence (NoE) on Multimodal
interfaces. eNTERFACE'08 will follow the fruitful path opened by
eNTERFACE05 in Mons, Belgium, continued by eNTERFACE06 in Dubrovnik,
Croatia and eNTERFACE07 in Istambul, Turkey. SIMILAR came to an end in
2007, and the eNTERFACE http://www.enterface.org workshops are now
under the aegis of the OpenInterface http://www.openinterface.org
Foundation.
eNTERFACE'08 Important Dates
. December 17th, 2007: Reception of the complete Project proposal in
the format provided by the Author's kit
. January 10rd, 2008: Notification of project acceptance
. February 1st, 2008: Publication of the Call for Participation
. August 4th -- August 29th, 2008: eNTERFACE 08 Workshop
Christophe d'Alessandro
CNRS-LIMSI, BP 133 - F91403 Orsay France
tel +33 (0) 1 69 85 81 13 / Fax -- 80 88
-
Calls for EUSIPCO 2008-Lausanne Switzerland
CALL FOR PAPERS
CALL FOR SPECIAL SESSIONS AND CALL FOR TUTORIALS
EUSIPCO-2008 - 16th European Signal Processing Conference
August 25-29, 2008, Lausanne, Switzerland - http://www.eusipco2008.org
The 2008 European Signal Processing Conference (EUSIPCO-2008) is the sixteenth in a series of conferences promoted by EURASIP, the European Association for Signal, Speech, and Image Processing (www.eurasip.org ). Formerly biannual, this conference is now ayearly event. This edition will take place in Lausanne, Switzerland, organized by the Swiss Federal Institute of Technology, Lausanne (EPFL).
EUSIPCO-2008 will focus on the key aspects of signal processing theory and applications. Exploration of new avenues and methodologies of signal processing will also be encouraged. Accepted papers will be published in the Proceedings of EUSIPCO-2008. Acceptance will be based on quality, relevance and originality. Proposals for special sessions and tutorials are also invited.
For the first time, access to the tutorials will be free to all registered participants!
IMPORTANT DATES:
Proposals for Special Sessions: December 7, 2007
Proposals for Tutorials: February 8, 2008
Electronic submission of Full papers (5 pages A4): February 8, 2008
Notification of Acceptance: April 30, 2008
Conference: August 25-29, 2008
More details on how to submit papers and proposals for special sessions and tutorials can be found on the conference web site http://www.eusipco2008.org
Prof. Jean-Philippe Thiran
EPFL - Signal Processing Institute
EUSIPCO-2008 General Chair
Back to Top -
International Seminar on Speech Production ISSP 2008
The International Seminar on Speech Production (ISSP-2008) will be held in Strasbourg (Haguenau, 25 km north of Strasbourg-France) in December 2008, from Monday the 8th to Friday the 12th.
Please take note of the following important dates:
(1) Submission of a 2 page abstract (Times 12): March 28th, 2008
(2) Notification of acceptance: April 21st, 2008
(3) Early registration: 5th May, 2008
(4) Full paper submission: September 19th, 2008
(5) Late registration: September 22nd, 2008
(6) Conference dates: Monday, December 8 to Friday, December 12.
The conference website is under construction, and will de operational
before the end of this year.
We are looking forward to seeing you in Alsace.
Kind regards,
The Organizers
Rudolph Sock (IPS), Susanne Fuchs (ZAS Phonetik, Berlin) & Yves Laprie
(INRIA- LORIA, Nancy)
Back to Top