Contents

1 . Editorial

Dear Members,

The board and myself present our best wishes to all members and expect to serve the best they can, the interests of the speech community.

Please don't ever forget that YOU are ISCA and your contribution to the activities of the association is more than welcome,... it is expected!   You have a good opportunity to contribute by proposing candidates to ISCA fellowship (see below under ISCA News).

Have  brilliant research results and wonderful conferences!

Sorry for sending this issue a bit late....be sure it is not due to excessive New Year celebration!

Prof. em. Chris Wellekens 

Institut Eurecom

Sophia Antipolis
France 

public@isca-speech.org

 

 
Back to Top

2 . ISCA News

 

Back to Top

2-1 . Nomination of new ISCA fellows

New ISCA Fellows will be soon elected. Informations on ISCA fellowship can be found on our webpage www.isca-speech.org/fellows.html

We urge ISCA members to recommend  candidates by sending  nomination forms.  New nominations forms will be soon available on our website www.isca-speech.org. 

 

 

Back to Top

2-2 . Announcements from the Student Advisory Committee (SAC)

 Announcement #1: ISCA-SAC Call for Volunteers
 
The ISCA Student Advisory Committee (ISCA-SAC) is seeking student volunteers to help with several interesting projects such as transcribe interviews from the Saras Institute, plan/organize student events at ISCA-sponsored conferences/workshops, increase awareness of speech and language research to undergraduate and junior graduate students, assist with website redesign to facilitate interaction with Google Scholar, as well as collect resources (e.g., conferences, theses, job listings, speech labs, etc.) for the isca-students.org website, to name a few.
 
There are many small tasks that can be done, each of which would only take up a few hours. Unless it is of your interest to become a long term volunteer, no further commitment is required. If interested, please contact the ISCA-SAC Volunteer Coordinator at: vo lun te er [at] isca-students [dot] org.
 
Announcement #2: ISCA-SAC Logo Contest
 
The ISCA Student Advisory Committee is in the search for a new logo. This is your chance to release your artistic side and enter the ISCA-SAC Logo Competition. All students are invited to participate and a prize (still to be determined) will be awarded to the winner; not to mention the importance of having your logo posted on the isca-students.org website for the world to see.
 
The deadline for submissions is March 31st, 2009. The new Logo will be unveiled during the Interspeech 2009 conference in the form of merchandise embedded with the new logo (e.g., mugs, pens, etc.).
If interested, please send your submissions to: logocontest [at] isca-students.org  
Back to Top

3 . SIG's activities CSLP

 2008 SIG-CSLP Report

 

The main goals of SIG-CSLP (Special Interest Group on Chinese Spoken Language Processing) are to promote research interests and activities on Chinese Spoken Language Processing among ISCA members; provide a platform to exchange idea exchange, share experiences and facilitate developments to those who are interested in Chinese Spoken Language Processing; sponsor meetings, workshops and other events related to Chinese Spoken Language Processing; and whenever possible make available resources relevant to research in Chinese Spoken Language Processing, such as text and speech corpora, analysis tools, research papers and generated data to the community. There are currently four working groups, Corpora, Website and Membership, Community Relations, and Events, to involve more colleagues in SIG affairs and aid SIG business in general.

 

The flagship event of SIG-CSLP is ISCSLP, a biennial international conference on the same theme of the SIG. The 6th International Symposium on Chinese Spoken Language (ISCSLP 2008) was successfully held in Kunming, China from December 16 and 19, 2008. This conference covers a broad range of research topics, including speech analysis, coding, enhancement, perception, production, recognition, synthesis; language modelling and understanding; speaker and language recognition; spoken dialog systems; spoken language translation; phonetics and phonology; as well as computer-assisted language learning, multimodal interfaces, and speech data mining and document retrieval. There are total 137 full paper submissions received from 12 countries and regions, including China, Taiwan, Hong Kong, Singapore, Japan, USA, Iran, Germany, England, Spain, India and Vietnam. The Technical Committee has received strong support from a group of 94 expert reviewers. A total of 95 papers (69%) was accepted for presentation at the conference, and were included in the proceedings published in IEEE Xplorer. There were four keynote speakers: (1) Dr. Yuqing Gao, IBM Watson Research Canter, USA, on “Speech-To-Speech Translation Technologies for Real-World Applications”; (2) Professor Shigeki Sagayama, The University of Tokyo, Japan on “What Can Speech Researchers Bring to Music Processing?”; (3) Dr. Vincent Vanhoucke, Google Inc., USA on “Speech and Search: Bridging The Gap”; and (4) Dr. Qiang Huo, Microsoft Research Asia, China on “Towards Robust Speech Recognition: Structured Modelling, Irrelevant Variability Normalization and Unsupervised Online Adaptation”. In addition, there were also two tutorials, one panel discussion and one exhibition. The two tutorials were on (1) “Looking into the past: Power spectral representation of periodic signals, sampling theories and fundamental frequency estimation for remaking speech” by Professor Hideki Kawahara, Faculty of Systems Engineering, Wakayama University, Japan; and (2) “A Tutorial on How to Construct and Improve Automatic Pronunciation Proficiency Evaluation System--take PSC test as an example”, by Drs. Yu Hu, Si Wei and Guoping Hu, iFLYTEK Research, Hefei, China. The theme of the panel discussion was CALL (Computer Aided Language Learning) with distinguished panellists, Professors Hiroya Fujisaki and Nobuaki Minemats of The University of Tokyo, Rushen Shi of The University of Quebec, and Dr. Yu Hu of iFLYTEK Research, Hefei, China.  The exhibition session features a variety of systems, especially those related to CALL applications.  There were approximately 150 participants.

 

The SIG-CSLP also took the opportunity of its flagship event to hold a general meeting of its attending members and turn over service posts to newly elected officers. There were progress reports from the incumbent Chairperson, Chiu-yu Tseng, Vice Chairperson, Jianhua Tao, Secretary, Bin Ma, and four leaders of SIG-CSLP working groups, Thomas Fang Zheng for CORPORA, Minghui Dong for Website and Membership, Helen Meng for Community Relations, and Yu Hu for Events. The newly elected officers for the next 2-year term, 2008-2010, are Chairperson, Chiu-yu Tseng, Vice Chairperson, Jianhua Tao, and Secretary, Bin Ma.

 

 

For more information on SIG-CSLP activities, please visit the site: http://www.SIG-CSLP.org/

 

 

Back to Top

4 . Future ISCA Conferences and Workshops(ITRW)

4-1 . (2009-06-25) ISCA Tutorial and Research Workshop on NON-LINEAR SPEECH PROCESSING

An ISCA Tutorial and Research Workshop on NON-LINEAR SPEECH PROCESSING (NOLISP'09) 25/06/2009 - // DeadLine: 20090315 Vic Catalonia Espagne http://nolisp2009.uvic.cat  After the success of NOLISP'03 held in Le Croisic, NOLISP'05 in Barcelona and NOLISP'07 in Paris, we are pleased to present NOLISP'09 to be held at the University of Vic (Catalonia, Spain) on June 25-27, 2009. The workshop will feature invited lectures by leading researchers as well as contributed talks. The purpose of NOLISP'09 is to present and discuss novel ideas, works and results related to alternative techniques for speech processing, which depart from mainstream approaches.  Prospective authors are invited to submit a 3 to 4 page paper proposal in English, which will be evaluated by the Scientific C! ommittee. Final papers will be due one month after the workshop to be included in the CD-ROM proceedings. Contributions are expected (but not restricted to) the following areas:  Non-linear approximation and estimation Non-linear oscillators and predictors Higher-order statistics Independent component analysis Nearest neighbours Neural networks Decision trees Non-parametric models Dynamics of non-linear systems Fractal methods Chaos modelling Non-linear differential equations  All fields of speech processing are targeted by the workshop, namely: Speech production, speech analysis and modelling, speech coding, speech synthesis, speech recognition, speaker identification/verification, speech enhancement/separation, speech perception, etc.  ADDITIONAL INFORMATION  Proceedings will be published in Springer-Verlag's Lecture Notes Series in Computer Science (LNCS). LNCS is published, in parallel to the printed books, in full-text electronic form. All contributions should  be original, and must not have been previously published, nor be under review for presentation elsewhere. A special issue of Speech Communication (Elsevier) on “Non-Linear and Non-Conventional Speech Processing” will be also published after the workshop  Detailed instructions for submission to NOLISP'09 and further informations will be available at the conference Web site (http://nolisp2009.uvic.cat).  IMPORTANT DATES: * March 15, 2009 - Submission (full papers) * April 30, 2009 - Notification of acceptance * September 30, 2009 - Final (revised) paper
Back to Top

4-2 . (2009-09-06) INTERSPEECH 2009 Brighton UK

                           * INTERSPEECH 2009 *

 

10th Annual Conference of the International Speech Communication

Association

 

September 6-10, 2009

Brighton, United Kingdom

http://www.interspeech2009.org

 

Theme: Speech and Intelligence

 

* CALL FOR TUTORIALS *

* CALL FOR SPECIAL SESSIONS *

* CALL FOR COMPETITION ENTRIES *

* CALL FOR PUBLIC EXHIBITS *

 

--------------------------------------------------------------

* TUTORIALS *

 

Proposals are invited for three hour tutorials to be held on

September 6th 2009 at INTERSPEECH 2009.

 

Prospective tutorial presenters should submit proposals by email

to Thomas Hain (t.hain@dcs.shef.ac.uk) before 1 January 2009.

Each tutorial proposal should contain the following information:

 

- Title of the proposed tutorial

- Names/affiliation of the presenters (including a brief bio and

contact information)

- An introduction (maximum 1 page), including the importance of

the topic, the objectives of the proposed tutorial, and a brief

outline of the main target audience.

- A presentation outline (maximum 2 pages), and an indication if

there are multiple presenters are proposed, how the presentation

will be shared.

 

- A list of any special equipment necessary for the presentation,

including information if such equipment is required to be supplied

by the conference organisers.

 

Proposals will be evaluated by the Organisation Committee based on

the relevance/significance of the topic and potential interest to

the conference attendees. Notification of acceptance of proposals

is scheduled for 16 January 2009.

 

--------------------------------------------------------------

* SPECIAL SESSIONS *

 

Proposals are also invited for Special Sessions at INTERSPEECH 2009.

Special Sessions are aimed at providing an opportunity for the

presentation of case studies, emerging technologies and research

directions in specific technical fields and application areas of

speech and language processing.

 

Prospective organisers should submit proposals by email to Ji Ming

(j.ming@qub.ac.uk) before 1 January 2009. Each special session

proposal should contain the following information:

 

- Title of the proposed Session

- Names/affiliation of organisers (including brief bio and contact

information)

- An introduction (maximum 2 pages) stating the importance of the

topic and the objectives of the proposed Session

- Tentative/confirmed list of papers (titles/affiliations/authors)

to be submitted to the proposed Session (optional)

 

Papers for approved Special Sessions should be submitted following

the same schedule/procedure as for regular papers, and will undergo

the same reviewing process by anonymous and independent reviewers

as regular papers.

 

Proposals will be evaluated by the Organisation Committee based on

the relevance/significance of the topic and potential interest to

the conference attendees. Notification of acceptance of proposals

is scheduled for 16 January 2009.

 

--------------------------------------------------------------

* SPEECH-BASED LOEBNER COMPETITION & PUBLIC EXHIBITION *

 

The theme for INTERSPEECH 2009 is 'Speech and Intelligence', and

on 6th September 2009 we will be hosting the annual contest for

the Loebner Prize.

 

In parallel with this event, we are also planning to mount (for the

very first time) a speech-based version of the competition.

 

Also, in order to raise awareness of speech science and technology

in the general public, we are putting together an exhibition of

speech-related demonstrations/displays. We envisage that these will

provide hands-on interaction with our current technologies, and will

serve to promote our field to the wider population.

 

If you are interested in entering a system for a speech-based version

of the Loebner competition, or interested in providing an exhibit for

a public display, then please contact Simon Worgan

(simon.worgan@gmail.com) who will contact you with further

information.

 

--------------------------------------------------------------

 

For further information on INTERSPEECH 2009, see

http://www.interspeech2009.org/

 

Conference Chair:

Prof. Roger K. Moore

Speech and Hearing Research Group (SPandH)

Department of Computer Science, University of Sheffield,

Regent Court, 211 Portobello,

Sheffield, S1 4DP, UK

 

e-mail: general_chair@interspeech2009.org

                                   

 

 

Back to Top

4-3 . (2010-09-26) INTERSPEECH 2010 Chiba Japan

Chiba, Japan
Conference Website
ISCA is pleased to announce that INTERSPEECH 2010 will take place in Makuhari-Messe, Chiba, Japan, September 26-30, 2010. The event will be chaired by Keikichi Hirose (Univ. Tokyo), and will have as a theme "Towards Spoken Language Processing for All - Regardless of Age, Health Conditions, Native Languages, Environment, etc."

Back to Top

4-4 . (2011-08-27) INTERSPEECH 2011 Florence Italy

Interspeech 2011

Palazzo dei Congressi,  Italy, August 27-31, 2011.

Organizing committee

Piero Cosi (General Chair),

Renato di Mori (General Co-Chair),

Claudia Manfredi (Local Chair),

Roberto Pieraccini (Technical Program Chair),

Maurizio Omologo (Tutorials),

Giuseppe Riccardi (Plenary Sessions).

More information www.interspeech2011.org

Back to Top

5 . Books, databases and softwares

 

Back to Top

5-1 . Books

 This section shows recent books whose titles been have communicated by the authors or editors.

Also some advertisement for recent books in speech are included.

Book presentation is written by the authors and not by this newsletter editor or any  voluntary reviewer.

Back to Top

5-1-1 . La production de parole

La production de la parole
Author: Alain Marchal, Universite d'Aix en Provence, France
Publisher: Hermes Lavoisier
Year: 2007
 
 
Back to Top

5-1-2 . Speech enhancement-Theory and Practice

 
 Speech enhancement-Theory and Practice
Author: Philipos C. Loizou, University of Texas, Dallas, USA
Publisher: CRC Press
Year:2007
 
 
Back to Top

5-1-3 . Speech and Language Engineering

 
 
Speech and Language Engineering
Editor: Martin Rajman
Publisher: EPFL Press, distributed by CRC Press
Year: 2007
 
 
Back to Top

5-1-4 . Human Communication Disorders/ Speech therapy

 
 
Human Communication Disorders/ Speech therapy
This interesting series can be listed on Wiley website
 
 
Back to Top

5-1-5 . Incursoes em torno do ritmo da fala

 
Incursoes em torno do ritmo da fala
Author: Plinio A. Barbosa 
Publisher: Pontes Editores (city: Campinas)
Year: 2006 (released 11/24/2006)
Website:http://www.ponteseditores.com.br/verproduto.php?id=301 
 
Back to Top

5-1-6 . Speech Quality of VoIP: Assessment and Prediction

 
Speech Quality of VoIP: Assessment and Prediction
Author: Alexander Raake
Publisher: John Wiley & Sons, UK-Chichester, September 2006
Website
 
 
Back to Top

5-1-7 . Self-Organization in the Evolution of Speech, Studies in the Evolution of Language

 

Self-Organization in the Evolution of Speech, Studies in the Evolution of Language
Author: Pierre-Yves Oudeyer
Publisher:Oxford University Press
Website
 
 

 

Back to Top

5-1-8 . Speech Recognition Over Digital Channels

 
Speech Recognition Over Digital Channels
Authors: Antonio M. Peinado and Jose C. Segura
Publisher: Wiley, July 2006
Website
 
 
Back to Top

5-1-9 . Multilingual Speech Processing

 
Multilingual Speech Processing
Editors: Tanja Schultz and Katrin Kirchhoff ,
Elsevier Academic Press, April 2006
Website
 
 
Back to Top

5-1-10 . Reconnaissance automatique de la parole: Du signal a l'interpretation

 
 Reconnaissance automatique de la parole: Du signal a l'interpretation
Authors: Jean-Paul Haton
Christophe Cerisara
Dominique Fohr
Yves Laprie
Kamel Smaili
392 Pages Publisher: Dunod
 
 
 
 
Back to Top

5-1-11 . Automatic Speech Recognition on Mobile Devices and over Communication Networks

 
 Automatic Speech Recognition on Mobile Devices and over Communication 
Networks
*Editors: Zheng-Hua Tan and Børge Lindberg
Publisher: Springer, London, March 2008
website <http://asr.es.aau.dk/>
 
About this book
The remarkable advances in computing and networking have sparked an 
enormous interest in deploying automatic speech recognition on mobile 
devices and over communication networks. This trend is accelerating.
This book brings together leading academic researchers and industrial 
practitioners to address the issues in this emerging realm and presents 
the reader with a comprehensive introduction to the subject of speech 
recognition in devices and networks. It covers network, distributed and 
embedded speech recognition systems, which are expected to co-exist in 
the future. It offers a wide-ranging, unified approach to the topic and 
its latest development, also covering the most up-to-date standards and 
several off-the-shelf systems.
 
 
Back to Top

5-1-12 . Latent Semantic Mapping: Principles & Applications

Latent Semantic Mapping: Principles & Applications
Author: Jerome R. Bellegarda, Apple Inc., USA
Publisher: Morgan & Claypool
Series: Synthesis Lectures on Speech and Audio Processing
Year: 2007
Website: http://www.morganclaypool.com/toc/sap/1/1
 
 
 
Back to Top

5-1-13 . The Application of Hidden Markov Models in Speech Recognition

 
The Application of Hidden Markov Models in Speech Recognition By Mark Gales and Steve Young (University of Cambridge)
http://dx.doi.org/10.1561/2000000004
 
in Foundations and Tr=nds in Signal Processing (FnTSIG)
www.nowpublishers.com/SIG 
 
 
 
Back to Top

5-1-14 . Proc.of the IEEE Special Issue on ADVANCES IN MULTIMEDIA INFORMATION RETRIEVAL

Proceedings of the IEEE
 
Special Issue on ADVANCES IN MULTIMEDIA INFORMATION RETRIEVAL
 
Volume 96, Number 4, April 2008
 
Guest Editors:
 
Alan Hanjalic, Delft University of Technology, Netherlands
Rainer Lienhart, University of Augsburg, Germany
Wei-Ying Ma, Microsoft Research Asia, China
John R. Smith, IBM Research, USA
 
Through carefully selected, invited papers written by leading authors and research teams, the April 2008 issue of Proceedings of the IEEE (v.96, no.4) highlights successes of multimedia information retrieval research, critically analyzes the achievements made so far and assesses the applicability of multimedia information retrieval results in real-life scenarios. The issue provides insights into the current possibilities for building automated and semi-automated methods as well as algorithms for segmenting, abstracting, indexing, representing, browsing, searching and retrieving multimedia content in various contexts. Additionally, future challenges that are likely to drive the research in the multimedia information retrieval field for years to come are also discussed.
 
 
 
Back to Top

5-1-15 . Computeranimierte Sprechbewegungen in realen Anwendungen

Computeranimierte Sprechbewegungen in realen Anwendungen
Authors: Sascha Fagel and Katja Madany
102 pages
Publisher: Berlin Institute of Technology
Year: 2008
Website http://www.ub.tu-berlin.de/index.php?id=1843
To learn more, please visit the corresponding IEEE Xplore site at
http://ieeexplore.ieee.org/xpl/tocresult.jsp?isYear=2008&isnumber=4472076&Submit32=Go+To+Issue
Usability of Speech Dialog Systems
 
 
Back to Top

5-1-16 . Usability of Speech Dialog Systems:Listening to the Target Audience

Usability of Speech Dialog Systems
Listening to the Target Audience
Series: Signals and Communication Technology
 
Hempel, Thomas (Ed.)
 
2008, X, 175 p. 14 illus., Hardcover
 
ISBN: 978-3-540-78342-8
 
 
Back to Top

5-1-17 . Speech and Language Processing

Speech and Language Processing, 2nd Edition
 
By Daniel Jurafsky, James H. Martin
 
Published May 16, 2008 by Prentice Hall.
More Info
Copyright 2009
Dimensions 7" x 9-1/4"
Pages: 1024
Edition: 2nd.
ISBN-10: 0-13-187321-0
ISBN-13: 978-0-13-187321-6
Request an Instructor or Media review copy
Sample Content
An explosion of Web-based language techniques, merging of distinct fields, availability of phone-based dialogue systems, and much more make this an exciting time in speech and language processing. The first of its kind to thoroughly cover language technology – at all levels and with all modern technologies – this book takes an empirical approach to the subject, based on applying statistical and other machine-learning algorithms to large corporations. KEY TOPICS: Builds each chapter around one or more worked examples demonstrating the main idea of the chapter, usingthe examples to illustrate the relative strengths and weaknesses of various approaches. Adds coverage of statistical sequence labeling, information extraction, question answering and summarization, advanced topics in speech recognition, speech synthesis. Revises coverage of language modeling, formal grammars, statistical parsing, machine translation, and dialog processing. MARKET: A useful reference for professionals in any of the areas of speech and language processing.
  
 
 
Back to Top

5-1-18 . Sprachverarbeitung -- Grundlagen und Methoden der Sprachsynthese und Spracherkennung

Title: Sprachverarbeitung -- Grundlagen und Methoden
       der Sprachsynthese und Spracherkennung
Authors: Beat Pfister, Tobias Kaufmann
Publisher: Springer
Year: 2008
Website: http://www.springer.com/978-3-540-75909-6 

Back to Top

5-1-19 . Advances in Digital Speech Communications

Advances in Digital Speech Transmission
Editors: Rainer Martin, Ulrich Heute and Christiane Antweiler
Publisher: Wiley&Sons
Year: 2008
Back to Top

5-1-20 . Digital Speech Transmission

Digital Speech Transmission
Authors: Peter Vary and Rainer Martin
Publisher: Wiley&Sons
Year: 2006
Back to Top

5-1-21 . Distant Speech Recognition

Distant Speech Recognition, Matthias Wölfel and John McDonough (2009), J. Wiley & Sons.

 Please link the title to http://www.distant-speech-recognition.com 

In the very recent past, automatic speech recognition (ASR) systems have attained acceptable performance when used with speech captured with a head-mounted or close-talking microphone (CTM). The performance of conventional ASR systems, however, degrades dramatically as soon as the microphone is moved away from the mouth of the speaker. This degradation is due to a broad variety of effects that are not found in CTM speech, including background noise, overlapping speech from other speakers, and reverberation. While conventional ASR systems underperform for speech captured with far-field sensors, there are a number of techniques developed in other areas of signal processing that can mitigate the deleterious effects of noise and reverberation, as well as separating speech from overlapping speakers. Distant Speech Recognition presents a contemporary and comprehensive description of both theoretic abstraction and practical issues inherent in the distant ASR problem.

Back to Top

5-2 . Database providers

 

Back to Top

5-2-1 . ELRA - Language Resources Catalogue - Update

ELRA - Language Resources Catalogue - Update
*****************************************************************

New!
A more robust search engine has been integrated to our catalogues. You may now use either the new standard search from the catalogues' main page or the extended search controlled thanks to three drop-down lists ("Match", "Format" and "Sort").  Direct access to the controlled search is given below:
- ELRA Catalogue: http://catalog.elra.info/search.php
-
Universal Catalogue: http://universal.elra.info/search.php


ELRA is happy to announce that 1 new WordNet database and  2 new Speech Resources are now available in its catalogue:

ELRA-M0047 Czech WordNet
The Czech WordNet captures nouns, verbs, adjectives, and partly adverbs, and contains 28,201 word senses (synsets). Every synset encodes the equivalence relation between several literals (at least one is present), having a unique meaning, belonging to one and the same part of speech, and expressing the same lexical meaning. Each Czech synset is related to the corresponding synset in the Princeton WordNet 2.0. via its identification number ID. There is at least one language-internal relation between a synset and another synset in the database.
For more information, see: http://catalog.elra.info/product_info.php?products_id=1089

ELRA-S0294 CHIEDE Corpus: a spontaneous child language corpus of Spanish
The spontaneous child language corpus, CHIEDE, consists of 58,163 words, in 30 texts, with 7 hours and 53 minutes of recordings and 59 child participants. About a third of the whole corpus is formed by child language and the remaining two thirds by adult speech. The main feature of CHIEDE is the interactions spontaneity: texts are recordings of communicative situations in their natural context.
For more information, see: http://catalog.elra.info/product_info.php?products_id=1090

ELRA-S0295 LILA Korean database
The LILA Korean database comprises 1,000 Korean speakers (500 males and 500 females) recorded over the Korean mobile telephone network. Each speaker uttered around 60 read and spontaneous items.
For more information, see: http://catalog.elra.info/product_info.php?products_id=1091

For more information on the catalogue, please contact Valérie Mapelli mailto:mapelli@elda.org

Visit our On-line Catalogue: http://catalog.elra.info
Visit the Universal Catalogue: http://universal.elra.info
Archives of ELRA Language Resources Catalogue Updates: http://www.elra.info/LRs-Announcements.html
 
 
 
Back to Top

5-2-2 . LDA Newsletter

 

 
-  Membership Mailbag - LDC Corpus Catalog Naming Conventions  -

 

 

-  Announcing FLaReNet (Fostering Language Resources Network)  -
 
LDC2008T25

 

 

LDC2008L03

 

LDC Offices to Close for Winter Break  -


 

 

Invitation to Join for Membership Year 2009 (MY 2009)

 

Membership Year (MY) 2008 proved to be a record setting year not only for membership and publication activity, but for all research focuses of the consortium: annotation, data collection, human subjects collection, and tools and standards development.  Your membership helps support our mission to develop and share language-technology resources.

Publications for MY 2009 are still being planned but it will be another productive year with a broad selection of publications.  The working titles of data sets we intend to provide include:

2007 ACE Spanish Data

CoNLL Shared Task Data

Arabic Gigaword Fourth Edition

English Gigaword Fourth Edition

Arabic Treebank Update

GALE Phase 1 Arabic Broadcast Conversation Parallel Text

Chinese Gigaword Fourth Edition

GALE Phase 1 Chinese Broadcast Conversation Parallel Text


In addition to receiving new publications, current year members of the LDC also enjoy the benefit of licensing older data at reduced costs; current year for-profit members may use most data for commercial applications.

The LDC appreciates the important contribution LDC members make through their continued support of the consortium.  In recognition of that loyalty and commitment, membership fees for MY 2009 will not increase over MY2008. In addition, LDC will extend discounts to members who keep their membership current and who join early in the year.

The details are as follows:

  • Organizations who joined for Membership Year (MY) 2008, will receive a 5% discount when renewing. This discount will apply throughout 2009, regardless of time of renewal. MY 2008 members renewing before March 1, 2009 will receive an additional 5% discount, for a total 10% discount off the membership fee.
  • New members as well as members who did not join for MY 2008, but who held membership in any of the previous MY's (1993-2007), will also be eligible for a 5% discount provided that they join/renew before March 1, 2009.

The Membership Fee Table provides exact pricing information.

 

MY 2009 Fee

MY 2009 Fee with 5% Discount *

MY 2009 Fee with 10% Discount **

Not-for-Profit

 

 

 

 

Standard

US$2400

US$2280

US$2160

 

Subscription

US$3850

US$3657.50

US$3465

For-Profit

 

 

 

 

Standard

US$24000

US$22800

US$21600

 

Subscription

US$27500

US$26125

US$24750

*   For MY 2008 Members renewing for MY 2009 and any previous year Member who renews before March 1, 2009

** For MY 2008 Members renewing before March 1, 2009

 

All organizations which have held LDC membership have been sent an invitation to join letter for MY 2009. If you have not received a letter and plan on joining or renewing your membership, please contact the LDC Membership Office.  As a reminder, MY 2007 will be open for joining until December 31, 2008.

 

 

Membership Mailbag - LDC Corpus Catalog Naming Conventions

 


The LDC Membership Office responds to a few thousand emailed queries a year, and, over time, we've noticed that some questions tend to crop up with regularity.  To address the questions that you, our data users, have asked, we'd like to continue our Membership Mailbag series of newsletter articles.  This month we will focus on LDC corpus catalog naming conventions, specifically on distinguishing among corpora with similar names.

The LDC Corpus Catalog contains corpora which have been used in research projects, in some cases including benchmark tests carried out under government sponsorship.  These include data from older projects such as TIPSTER and HUB4 to current programs like ACE and GALE.  Additionally, our catalog contains data donated to the LDC by a sizable group of corpus authors from organizations around the globe.   This varied origin of LDC data can make for a potentially confusing array of corpus catalog names.  However, a few general rules are observed to differentiate among corpora with similar names.  

The corpus names for data from one collection or research effort that contains unique data will include terms such as 'part', 'phase', 'set', and 'volume'.  For example,  Arabic Treebank: Part 1 v 3. will contain different source data than Arabic Treebank: Part 4 v 1.0.  Likewise, Fisher English Training Speech Part 1 Speech contains different data than Fisher English Training Part 2, Speech.  Occasionally, the catalog will contain a database such as Levantine Arabic QT Training Data Set 5, Speech which has a corresponding Set 4 and Set 3, but not a Set 2 or Set 1.  In such cases, the earlier data may have only been released for evaluation purposes or may have been incorporated into later publications.

The corpus names of enhanced data sets or those with previously unreleased data will include 'version', 'release', 'edition', or simply, a number such as '2.0'.  For example, English Gigaword Third Edition  contains all of the data from both English Gigaword Second Edition and English Gigaword.  Likewise, Chinese Treebank 6.0 should be considered an update of all previous Chinese Treebanks - additional data has been included and known errors corrected.  Not all newer corpora will also have a corresponding earlier corpus in the LDC catalog, an example being Switchboard-1 Release 2.  Such earlier data sets may have not been available for general licensing through the LDC or may have been completely superseded by a later corpus.

Got a question?  About LDC data?  Forward it to ldc@ldc.upenn.edu.  The answer may appear in a future Membership Mailbag article.

Announcing FLaReNet (Fostering Language Resources Network)


The LDC is pleased to pass along the following announcement for FLaReNet:

FLaReNet, an eContentPlus project funded by the European Commission, officially started on September 1, 2008. It is intended to promote the consolidation of a European strategy in the field of Language Resources (LR's) and Language Technologies and to be a European forum to facilitate interaction among LR stakeholders.

To learn more about FLaReNet, please visit its website.  The site is still under construction and will be frequently updated with more content.  FLaReNet is open to new subscribers who are invited to join the Network; more details on the website.

*Important Upcoming Events*


FLaReNet Launching Event: an open workshop will be organized in Vienna, Austria on February 12-13, 2009. For information on the event, see the FLaReNet website. 

Language Technology Days: in Luxembourg on January 14-15 2009.  Click here for outline agenda and registration.  This event is being organized by the new EC Unit E1-Language Technologies and Machine Translation.

Chris Cieri and Stephanie Strassel of the LDC will continue their work on FlaReNet as external consultants.

 

New Publications

(1) AQUAINT-2 Information-Retrieval Text Research Collection was developed by LDC for NIST's (National Institute for Standards and Technology) AQUAINT 2007 Question-Answer (QA) track. It consists of approximately 2.5 GB of English news text from six distinct sources collected by LDC (Agence France Presse, Associated Press, Central News Agency (Taiwan), Los Angeles Times-Washington Post, New York Times and Xinhua News Agency) covering the period from October 2004 through March 2006. The AQUAINT-2 collection is the second part of a series intended to provide data useful for developing, evaluating and testing information extraction and retrieval systems. It follows the publication of The AQUAINT Corpus of English News Text (LDC2002T31).

The AQUAINT (Advanced Question-Answering for Intelligence)  program addresses interactivity with scenarios or tasks. The scenario provides a context in which questions will be asked and answered, and the task reflects the overall assignment. The program is committed to solve a single problem: how to find topically relevant, semantically related, timely information in massive amounts of data in diverse languages, formats, and genres.

For each source, all of the usable data collected by LDC was processed into a consistent XML format in which the stories for a given month are concatenated in chronological order into a single "DOCSTREAM" element; each story is a single "DOC" element within that stream and has a globally unique "id" attribute.

AQUAINT-2 Information-Retrieval Text Research Collection is distributed on 1 DVD.

2008 Subscription Members will automatically receive two copies of this corpus. 2008 Standard Members may request a copy as part of their 16 free membership corpora. Nonmembers may license this data for US$500.

*

 

The Global Yoruba Lexical Database v. 1.0 is a set of related dictionaries providing definitions and translations for over 450,000 words from the Yoruba language and its variants: Standard Yoruba (over 368,000 words), Gullah (over 3,600 words), Lucumí (over 8,000 words) and Trinidadian (over 1,000 words).

Yoruba is a Niger-Congo language (sub classification: Kwa > Yoruboid) spoken natively by nearly 20 million people, the vast majority of them in southwestern Nigeria.  The  Yoruba language diaspora is wide, stretching from southwestern Nigeria and Benin westward to the Caribbean and islands along the southeastern United States coast.  Throughout the region, Yoruba dialects blended with each other and with languages like Spanish and French to form a variety of creoles such as Gullah in the United States and Nagô in Brazil.  The ultimate goal of this dictionary is to provide coverage for all Yoruba dialects across the globe. For that reason, it will continue to be a work in progress.

The Yoruba dialect continuum consists of over fifteen varieties, with considerable phonological and lexical differences among them and some grammatical ones as well. Peripheral areas of dialectal regions often have some similarities to adjoining dialects. Standard Yoruba is a koine used for education, writing, broadcasting, and contact between speakers of different dialects.

The dictionaries in this publication are presented in two formats, Toolbox databases and XML. Short for The Field Linguist's Toolbox, Toolbox is a lexicographical database system published by SIL. SIL makes Toolbox freely available for download. In order to use the Global Yoruba Lexical Database v. 1.0, Toolbox must first be installed on the user's local computer.

Global Yoruba Lexical Database v. 1.0 is distributed via web download.

2008 Subscription Members will automatically receive two copies of this corpus on disc. 2008 Standard Members may request a copy as part of their 16 free membership corpora. Nonmembers may license this data for US$800.
 

LDC Offices to Close for Winter Break

The LDC would like to inform our customers that we will be closed from Thursday, December 25, 2008 through Friday, January 2, 2008 in accordance with the University of Pennsylvania Winter Break Policy.  Our offices will reopen on Monday, January 5, 2009.  Requests received for membership renewals and corpora will be processed at that time.

Best wishes for a happy and safe holiday season!


 

Ilya Ahtaridis

Membership Coordinator
 
--------------------------------------------------------------------

 

Linguistic Data Consortium                     Phone: (215) 573-1275
University of Pennsylvania                       Fax: (215) 573-2175
3600 Market St., Suite 810                         ldc@ldc.upenn.edu
 Philadelphia, PA 19104 USA                   http://www.ldc.upenn.edu

 

Back to Top

5-3 . MusicSpeech group

Music and speech share numerous aspects (language, structural, acoustics, cognitive), as long in their production, that in their representation and their perception. This list has for object to warn its users, various events dealing with the study of the links between music and speech. It thus intends to connect several communities, their allowing each to take advantage of a stimulating interaction.

As a member of the speech or music community, you are invited to
subscribe to musicspeech group. The group will be moderated and
maintained by IRCAM.

Group details:
* Name: musicspeech
* Home page: http://listes.ircam.fr/wws/info/musicspeech
* Email address: musicspeech@ircam.fr

Greg Beller, IRCAM,
moderator, musicspeech list

Back to Top

5-4 . Softwares

 

TOWARD IALL (INTERNET AIDED LANGUAGE LEARNING)

 

An  CALL (Computer Aided Language Learning) demo for second language learning (English) has been designed and implemented at  LPTV (Laboratorio de Procesamiento y Transmisión de Voz -Speech Processing and Transmission Lab), Universidad de Chile. Go to:

 http://www.die.uchile.cl/LPTV

 and then click “On-line, try our second language learning demo”.

 It offers five activities based on speech technology: pronunciation evaluation assessment; text-comprehension questions; word-meaning association; spelling practice with dictation exercises; and, intonation evaluation. The proposed distributed architecture is based on a fully centralized processing: the speech signal is recorded locally on user´s host and sent over to our server by employing TCP/IP protocols. The requirements for the local host are very low: only Java version 1.6 (or more recent) and a low-cost microphone.

Please, feel free to use our demo. We would like know any comment about the system itself and about the Internet delay (in seconds) between the click to check a recorded audio and the reception of the feed-back result. The intonation evaluation may take a few seconds to be deployed in the user´s screen. As part of our research, we also need to collect speech data. Please, send any suggestion or comment to:

 

  nbecerra@ing.uchile.cl

Prof. Néstor Becerra Yoma, PhD

Speech Processing and Transmission Lab

Dept. of Electrical Engineering

Universidad de Chile

Tel. +56-2-978 4205

Fax. +56-2-695 3881

E-mail: nbecerra@ing.uchile.cl

http://www.die.uchile.cl/LPTV

 

Back to Top

6 . Jobs openings

We invite all laboratories and industrial companies which have job offers to send them to the ISCApad editor: they will appear in the newsletter and on our website for free. (also have a look at http://www.isca-speech.org/jobs.html as well as http://www.elsnet.org/ Jobs)


Back to Top

6-1 . (2008-07-01) Nuance: Junior Research Engineer for Embedded Automatic Speech Recognition

Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world.  Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, controlling their mobile phone or digitally reproducing documents that can be shared and searched.  With more than 2000 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about. To strengthen our international team we are currently looking for a

 

 

Junior Research Engineer for Embedded Automatic Speech Recognition

 

 

Work Environment

·         You will work in the Embedded ASR research and production team in Merelbeke, Belgium or Aachen, Germany, working with state-of -he-art speech technology, tools and runtime software. Both Gent and Aachen are nice, historical European university cities.

·         You will work in an international company and cooperate with people and research teams on various locations across the globe. You may occasionally be asked to travel.

·         You will work  with our natural language understanding and dialogue research teams as well support our professional services teams.

·         You will work on the development of cutting edge speech recognition products for automotive platforms and mobile devices. You will help the engine cope with multi-lingual speech in various noise conditions, and this while respecting strong limitations on the usage of memory and processing power.

 

Key Responsibilities

·         Design, implementation, evaluation, optimization and testing of new algorithms and tools, with a strong focus on speech signal processing and acoustic modeling in adverse, noisy environments.

·         Activities are targeted at the creation of commercial products for resource limited platforms.

·         Focus on creating efficient production and development processes to bring the technology to marketable products in a wide range of languages.

·         Occasional application of the developed algorithms and tools for producing systems for a specific language.

·         Specification and follow-up of projects to make the system work with third party components, such as beam formers, echo cancellers or content data providers.

 

Your Profile

  • You have a University degree in engineering, mathematics or physics.
  • A PhD degree in speech processing or equivalent relevant experience is a strong asset.
  • Experience in speech recognition research, especially acoustic modeling or signal processing, is required.
  • Experience in speech processing, machine learning techniques or statistical modeling is required.
  • Knowledge about small platforms and experience in developing software for them is a plus.
  • Strong software skills are required, especially C/C++ and a scripting language like Perl or Python in a Linux/Unix environment. Knowledge of Matlab is a plus.
  • Additional background in computational linguistics is a plus.
  • You are a team player, willing to take initiative, and are goal oriented.
  • You have a strong desire to make things “really work” in practice, on hardware platforms with limited memory and processing power.
  • You are fluent in English and at least one other language, and you can write high quality English documentation.  

 

Interested?

 

Please send your CV to Deanna Roe at deanna.roe@nuance.com. If you have any questions, please contact her at +44 207 922 5757.

 

We are looking forward to receiving your application!

 

Back to Top

6-2 . (2008-07-01) Nuance SOFTWARE ENGINEER SPEECH DIALOGUE TOOLS

In order to strengthen our Embedded ASR Research team, we are looking for a:

 

       SOFTWARE ENGINEER SPEECH DIALOGUE TOOLS

 

As part of our team, you will be creating solutions for voice user interfaces for embedded applications on mobile and automotive platforms.

 

 

OVERVIEW:

 

- You will work in Nuance's Embedded ASR (automatic speech recognition) research and development team, developing technology, tools, and run-time software to enable our customers to develop and test embedded speech applications. Together with our team of speech and language experts, you will work on natural language dialogue systems for our customers in the Automotive and Mobile sector.

- You will work on fascinating technology that has now reached the maturity to enable new generations of powerful and natural user interfaces. Your code is crucial to the research in speech and language technology that defines the state of the art in this field It is equally important for the products that you will find in the market, in speech-enabled cars, navigation devices, and cell phones.

- You will work in a large international software company that is the leading provider of speech and imaging solutions for businesses and consumers around the world. You will cooperate with people on various locations including in Europe, America and Asia. You may occasionally be asked to travel.

 

 

RESPONSIBILITIES:

 

- You will work on the development of tools and solutions for cutting edge speech and language understanding technologies for automotive and mobile devices.

- You will work on enhancing various aspects of our advanced natural language dialogue system, such as the layer of connected applications, the configuration setup, inter-module communication, etc.

- In particular, you will be responsible for the design, implementation, evaluation, optimization and testing, and documentation of tools such as GUI and XML applications that are used to develop, configure, and fine-tune advanced dialogue systems.

 

 

 

QUALIFICATIONS:

 

- You have a university degree in computer science, engineering, mathematics, physics, computational linguistics, or a related field.

- You have very strong software and programming skills, especially in C/C++, ideally also for embedded applications.

- You have experience with Python or other scripting languages.

- GUI programming experience is an asset.

 

The following skills are a plus:

- Understanding of communication protocols

- Understanding of databases

- A background in (computational) linguistics, dialogue systems, speech processing, grammars, and parsing techniques, statistics, pattern recognition, and machine learning, especially as related to natural language processing, dialogue, and representation of information

- Understanding of computational agents and related frameworks (such as OAA).

- You can work both as a team player and as goal-oriented independent software engineer.

- You can work in a multi-national team and communicate effectively with people of different cultures.

- You have a strong desire to make things really work in practice, on hardware platforms with limited memory and processing power.

- You are fluent in English and you can write high quality documentation.

- Knowledge of other languages is a plus.

 

 

 

CONTACT:

 

Please send your applications, including cover letter, CV, and related documents (maximum 5MB total for all documents, please) to

 

Benjamin Campued       Benjamin.Campued@nuance.com

 

Please make sure to document to us your excellent software engineering skills.

 

 

 

ABOUT US:

 

Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world.  Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, or digitally reproducing documents that can be shared and searched.  With more than 3500 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about.

 

Back to Top

6-3 . (2008-07-01) Nuance-Speech Scientist for Embedded Automatic Speech Recognition

Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world.  Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, controlling their mobile phone or digitally reproducing documents that can be shared and searched.  With more than 2000 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about. To strengthen our international team we are currently looking for a

 

 Speech Scientist for Embedded Automatic Speech Recognition

 

 

Work Environment

·          You will work in the Embedded ASR research and production team in Merelbeke, Belgium or Aachen, Germany, working with state-of-the-art speech technology, tools and runtime software. Both Gent and Aachen are nice, historical European university cities.

·          You will work in an international company and cooperate with people on various locations, from the USA up to Japan. You may occasionally be asked to travel.

·          You will work on the localization and production of language variants for our cutting edge speech recognition products targeted at automotive platforms and mobile devices. You will help the engine cope with multi-lingual speech in various noise conditions.

·          Initially, you will work on the production of language variants of our acoustic models, later extending your knowledge towards production of statistical language models and natural language dialogue systems.

 

Key Responsibilities

·          Training of  acoustic models or statistical language models for new languages.

·          Localizing natural language dialogue systems towards a specific market.

·          Contributing to the improvement, design, implementation, evaluation, optimization and testing of new algorithms, tools and processes.

·          Supporting our professional services teams to contribute to customer project success.

·          Assisting senior team members in research tasks.

 

Your Profile

  • You have a University degree in linguistics, engineering, mathematics or physics.
  • A PhD or similar experience in a relevant field is a plus.
  • Experience in acoustic modeling, NLU or statistical language modeling is recommended.
  • Additional background in computational linguistics is a plus.
  • Working in Windows and Linux environments comes naturally to you. Experience with computing farms and grid software is welcome.
  • You are knowledgable about small, embedded platforms and requirements of software applications designed for them.
  • Good software skills are required, especially scripting language like Perl or Python in a Linux/Unix environment, and knowledge of C/C++.
  • Experience in speech processing or machine learning techniques is an asset.
  • You are a team player, willing to take initiative, and are goal oriented.
  • You have a strong sense of precision and quality in your daily job.
  • You are fluent in English and you can write high quality documentation.  
  • You illustrate your interest in languages by speaking at least two other languages.

 

Interested?

 

Please send your CV to Deanna Roe at deanna.roe@nuance.com. If you have any questions, please contact her at +44 207 922 5757.

 

We are looking forward to receiving your application!

 

The experience speaks for itself™

 

Back to Top

6-4 . (2008-07-01) Nuance-Senior Research Engineer for Embedded Automatic Speech Recognition

Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the world.  Every day, millions of users and thousands of businesses experience Nuance by calling directory assistance, requesting account information, dictating patient records, telling a navigation system their destination, controlling their mobile phone or digitally reproducing documents that can be shared and searched.  With more than 2000 employees worldwide, we are committed to make the user experience more enjoyable by transforming the way people interact with information and how they create, share and use documents. Making each of those experiences productive and compelling is what Nuance is about. To strengthen our international team we are currently looking for a

 

 

Senior Research Engineer for Embedded Automatic Speech Recognition

 

 

Work Environment

·          You will work in the Embedded ASR research and production team in Merelbeke, Belgium or Aachen, Germany, working with state-of -he-art speech technology, tools and runtime software. Both Gent and Aachen are nice, historical European university cities.

·          You will work in an international company and cooperate with people and research teams on various locations across the globe. You may occasionally be asked to travel.

·          You will work  with our natural language understanding and dialogue research teams as well support our professional services teams.

·          You will work on the development of cutting edge speech recognition products for automotive platforms and mobile devices. You will help the engine cope with multi-lingual speech in various noise conditions, and this while respecting strong limitations on the usage of memory and processing power.

 

Key Responsibilities

·          Design, implementation, evaluation, optimization and testing of new algorithms and tools, with a strong focus on speech signal processing and acoustic modeling in adverse, noisy environments.

·          Activities are targeted at the creation of commercial products for resource limited platforms.

·          Focus on creating efficient production and development processes to bring the technology to marketable products in a wide range of languages.

·          Occasional application of the developed algorithms and tools for producing systems for a specific language.

·          Specification and follow-up of projects to make the system work with third party components, such as beam formers, echo cancellers or content data providers.

 

Your Profile

  • You have a University degree in engineering, mathematics or physics.
  • A PhD degree in speech processing or equivalent relevant experience is a strong asset.
  • Experience in speech recognition research, especially acoustic modeling or signal processing, is required.
  • Experience in speech processing, machine learning techniques or statistical modeling is required.
  • Knowledge about small platforms and experience in developing software for them is a plus.
  • Strong software skills are required, especially C/C++ and a scripting language like Perl or Python in a Linux/Unix environment. Knowledge of Matlab is a plus.
  • Additional background in computational linguistics is a plus.
  • You are a team player, willing to take initiative, and are goal oriented.
  • You have a strong desire to make things “really work” in practice, on hardware platforms with limited memory and processing power.
  • You are fluent in English and at least one other language, and you can write high quality English documentation. 

 

Interested?

 

Please send your CV to Deanna Roe at deanna.roe@nuance.com. If you have any questions, please contact her at +44 207 922 5757.

 

We are looking forward to receiving your application!

 

The experience speaks for itself™

Back to Top

6-5 . (2008-07-01) Nuance- jr. Speech Scientist

Title: jr. Speech Scientist

 

Location: Aachen, Germany

 

Type: Permanent

 

Job:  

 

Overview:

 

Nuance is the leading provider of speech and imaging solutions for businesses and consumers around the World. Our technologies, applications and services make the user experience more compelling by transforming the way people interact with information and how they create, share and use documents. Every day, millions of users and thousands of businesses, experience Nuance by calling directory assistance, getting account information, dictating patient records, telling a navigation system their destination, or digitally reproducing documents that can be shared and searched. Making each of those experiences productive and compelling is what Nuance is all about.

 

Responsibilities:

 

Nuance is seeking a jr. Speech Scientist who possesses a solid background in natural language technology and computational linguistics.

Candidates should enjoy working in a fast-paced, collaborative atmosphere that applies speech science in a commercial, result driven and customer oriented setting.

 

As a jr. Speech Scientist in the Embedded Professional Services group, you will work on speech recognition grammars, statistical language models, prompts and custom voice development for leading edge automotive applications across the world, covering a broad range of activities in all project phases, including the design, development, and optimization of the system.

 

Representative duties include:

  • Develop rule based grammars, train statistical language models for speech recognition and natural language understanding in commercial products in a variety of languages, according to UI Design specifications
  • Identify or gather suitable text for training language models and custom voices
  • Design, develop, and test semantic classifier rules and models
  • Develop custom voices for use with Nuance’s leading text to speech products
  • Direct voice talents for prompt recordings
  • Organize and conduct usability tests
  • Localization of speech resources for embedded speech applications
  • Optimize accuracy of applications by analyzing performance and tuning statistical language models, grammars, and pronunciations within CPU and memory constraints of embedded platforms
  • Contribute to the generation and presentation of client-facing reports

 

Qualifications:

  • University degree in computational linguistics or Software design or similar degree
  • Strong analytical and problem solving skills and ability to troubleshoot issues
  • Good judgment and quick-thinking
  • Strong programming skills, preferably Perl or Python
  • Excellent written and verbal communications skills
  • Ability to scope work taking technical, business and time-frame constraints into consideration
  • Ability and willingness to travel abroad
  • Works well independently and collaboratively in team settings in fast-paced environment
  • Mastering Office applications

 

Beneficial Skills

  • Additional language skills, eg. French, German, Spanish or other
  • Strong programming skills in either Perl, Python, C, VB
  • Speech recognition knowledge
  • Pattern recognition, linguistics, signal processing, or acoustics knowledge
Back to Top

6-6 . (2008-07-02) Microsoft: Danish Linguist (M/F)

Opened positions/internships at Microsoft: Danish Linguist (M/F)

MLDC – Microsoft Language Development Center, a branch of the Microsoft Product Group that develops Speech Recognition and Synthesis Technologies, situated in Porto Salvo, Portugal (http://www.microsoft.com/portugal/mldc), is seeking a full-time temporary language expert in the Danish language, for a 3-4 month contract, to work in speech technology related development projects. The successful candidate should have the following requirements:

·         Be native or near native Danish speaker

·         Have a university degree in Linguistics or related field (preferably in Danish Linguistics)

·         Have an advanced level of English

·         Have some experience in working with Speech Technology/Natural Language Processing/Linguistics, either in academia or in industry

·         Have some computational ability – no programming is required, but he/she should be comfortable working with MS Windows and MS Office tools

·         Have team work experience

·         Willing to work in Porto Salvo (near Lisbon) for the duration of the contract

·         Willing to start in September 2008

To apply, please submit your resume and a brief statement describing your experience and abilities to Daniela Braga: i-dbraga@microsoft.com

We will only consider electronic submissions. 

Back to Top

6-7 . (2008-07-02) Microsoft: Catalan Linguist (M/F)

Opened positions/internships at Microsoft: Catalan Linguist (M/F)

MLDC – Microsoft Language Development Center, a branch of the Microsoft Product Group that develops Speech Recognition and Synthesis Technologies, situated in Porto Salvo, Portugal (http://www.microsoft.com/portugal/mldc), is seeking a full-time temporary language expert in the Catalan language, for a 3-4 month contract, to work in speech technology related development projects. The successful candidate should have the following requirements:

·         Be native or near native Catalan speaker

·         Have a university degree in Linguistics or related field (preferably in Catalan Linguistics)

·         Have an advanced level of English

·         Have some experience in working with Speech Technology/Natural Language Processing/Linguistics, either in academia or in industry

·         Have some computational ability – no programming is required, but he/she should be comfortable working with MS Windows and MS Office tools

·         Have team work experience

·         Willing to work in Porto Salvo (near Lisbon) for the duration of the contract

·         Willing to start in September 2008

To apply, please submit your resume and a brief statement describing your experience and abilities to Daniela Braga: i-dbraga@microsoft.com

We will only consider electronic submissions. 

Back to Top

6-8 . (2008-07-10) Bourse de these IRISA Lannion (in french)



titre du sujet : Synthèse vocale de haute qualité

Introduction
=========

Ces dernières années ont vu l'émergence de systèmes de synthèse de la parole construits autour de base de données de parole de taille importante qui correspondent le plus souvent à quelques heures d'enregistrement de parole. A des degrés divers, ces systèmes considèrent qu'il est possible de produire une parole de qualité en allant chercher des fragments de sons dans une base de données enregistrée au préalable par un locuteur. Ce type d'approche pousse à l'extrême l'hypothèse fonctionnelle des systèmes fondés sur la concaténation d'unités acoustiques. Avec une base de données suffisamment importante, il doit être possible de couvrir statistiquement les cas les plus fréquents de coarticulation sonore.

Des systèmes récents comme Festival (Black 1995), CHATR (Campbell 1996), Whistler (Huang 1996), XIMERA (Toda 2006), IBM \citethese{Eid06}, prouvent que cette approche méthodologique permet de construire des systèmes de synthèse de très bonne qualité.

En suivant cette méthodologie, les modèles ne sont plus utilisés pour produire des valeurs de paramètres qui serviront à la génération d'un signal de parole. Ils sont en revanche utilisés pour rechercher dans la base d'exemples sonores un extrait de parole qui sera le plus proche possible des paramètres modélisés et conformes à une élocution humaine. Concernant la problématique de recherche d'une séquence d'unités acoustiques, différentes solutions sont possibles. Les plus connues appliquent des solutions de recherche de meilleurs chemins (Sagisaka 1992) (Hunt 1996) en proposant une hypothèse de programmation dynmique. D'autres travaux (Donovan 1995) ont défini des modèles acoustiques permettant de guider le choix d'une séquence d'unités.

L'enjeu du procédé de sélection est double, (Iwahashi 1992). Il s'agit d'une part de trouver une correspondance entre une sous-séquence de la chaîne phonémique à synthétiser et un exemplaire plausible dans le corpus de référence. On parle alors d'une \emph{discrimination par critères de cible}, (Hune 1996). Une correspondance à la cible ne suffit pas puisque cette décision est prise unité par unité. Il faut un mécanisme supplémentaire garantissant que l'enchaînement du séquencement proposé réponde à des critères de continuité acoustique (de nature segmentale ou supra-segmentale). On parle dans ce cas de critères de concaténation. La difficulté du problème réside dans le fait que les deux critères sont combinés. Le choix d'une sous-séquence en correspondance avec une unité du corpus dépend de son contexte passé (contexte de la séquence à gauche) et à venir (contexte à droite). Il s'agit encore une fois d'un problème de nature combinatoire qui peut formellement être posé comme un problème de recherche d'un meilleur chemin dans un graphe.

La grande majorité des systèmes de synthèse appliquent un algorithme de Viterbi. Cet algorithme, efficace en complexité spatiale et temporelle, tire sa justification du fait que l'expression du coût global d'une séquence d'unités s'écrit, par hypothèse, sous la forme d'une suite récurrente additive. Cette justification est largement partagée par l'ensemble de la communauté pour ce qui est de l'expression des coûts de concaténation et des coûts de proximité à la cible. En revanche pour ce qui concerne la prise en compte de coûts de nature prosodique, une mise en forme recurrente est plus délicate et difficilement justifiable puisque ces phénomènes ont lieu à l'échelle du groupe intonatif et de la phrase.

Nous considérons qu'il est possible de dépasser la qualité des systèmes de synthèse actuels par la prise en compte de critère prosodiques lors de la recherche de la séquence optimale des unités. Tenir compte de ces critières proposidiques n'est pas une chose simple, puiqu'il faut définir de nouveaux modèles de description des coûts acoustiques et prosodiques d'une séquence. Ces nouvelles techniques de sélection devraient être acapables de proposer des voix avec d'une part plus de relief ou d'expressivité tout en maintenant une très bonne qualité sonore.

(Sagisake 1992) : Sagisaka, Y. and Kaiki, N. and Iwahashi, N. and Mimura, K., ATR mu-TALK speech synthesis system, proceedings of the International Conference on Spoken Language Processing (ICSLP'92)", 1992, pp. 483-486.
(Hunt 1996) :  Hunt, A. and Black, A.W., Unit selection in a concatenative speech synthesis system using a large speech database, proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP'96), 1996, pp. 373-376.
(Donovan 1995) :  Donovan, R. and P. Woodland, P., Automatic speech synthesizer parameter estimation using HMMs, proceedings of the IEEE International Conference on Acoustics and Signal Processing (ICASSP'95), 1995, pp. 640-643.
(Iwahashi 1992) : Iwahashi, N. and Kaiki, N. and Sagisaka, Y., Concatenative speech synthesis by minimum distortion criteria, proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP'92), 1992, pp. 65-68.
(Black 1995) :  Alan W. Black and Nick Campbell, Optimizing selection of units from speech databases for concatenative sysnthesis, proceedings of the IEEE International Conference on Acoustics and Signal Processing (ICASSP'95), vol. 1, pp. 581-584.
(Campbell 1996) :  Campbell, N. and Black, A., CHART: A High-definition speech re-sequencing system, in Progress in Speech Synthesis,
eds. van Santen, J. and Sproat, R. and Olive, J. and Hirschberg, J., 1996, pp. 365-381,
(Huang 1996) : Huang, X. and Acero, A. and Adcock, J. and Hon, H.-W. and Goldsmith, J. and Liu, J. and Plumpe, M., Whistler: A trainable text-to-speech system, proceedings of the International Conference on Spoken Language Processing (ICSLP'96), 1996, pp. 2397-2390.
(Toda 2006) :  Tomoki Toda and Hisashi Kawai and Toshio Hirai and Jinfu Ni and Nobuyuki Nishizawa and Junichi Yamagishi and Minoru Tsuzaki and Keiichi Tokuda and Satoshi Nakamura, Developing a test bed of English text-to-speech system XIMERA for the Blizzard challenge 2006,  Blizzard Challenge, 2006.
(Eide 2006) :  Ellen Eide and Raul Fernandez and Ron Hoory and Wael Hamza and Zvi Kons and Michael Picheny and Ariel Sagi and Slava Shechtman and Zhi Wei Shuang, The IBM submission to the 2006 Blizzard text-to-speech challenge, Blizzard Challenge, 2006.

Proposition d'un travail de thèse
======================

Nous proposons de nous intéresser à de nouvelles méthodologies de sélection d'unités acoustiques pour la synthèse de la parole à partir du texte. La proposition de thèse comporte deux volets: un axe de propositions scientifiques permettant de lever certains verrous notamment dans la formulation du coût d'une séquence d'unités, et un axe expérimental par la proposition d'une évolution du système de synthèse du groupe Cordial permettant de mettre en place des évaluations perceptuelles qui permettront de valider ou d'invalider les hypothèses de travail qui auront été choisies. Le travail expérimental sera réalisé sur le français. Nous souhaitons doubler les expérimentations sur l'anglais et participer ainsi au challenge Blizzard qui est une compétition internationale en synthèse de la parole.

Le travail de thèse prendra comme point d'appui la proposition suivante:
  * Mise en place et évaluation d'un premier système reposant sur l'état de l'art actuel en synthèse de la parole par corpus de parole continue. Prise en compte des niveaux acoustiques. Utilisation d'une base de parole expressive,  "chronic",  issue du projet ANR Vivos.
  * Proposition de modèles de sélection de nature prosodique.
  * Propositions algorithmiques, définition d'heuristiques pour une solution acceptable en temps de calcul.
  * Intégration des propositions prosodiques au système de synthèse de référence et évaluation.

Contexte du travail de thèse
===================

L'étudiant sera accueilli au sein de l'équipe Cordial de l'irisa : http://www.irisa.fr/cordial dont les
principaux travaux concernent le traitement de la parole : synthèse, transformation de parole, annotation de corpus.
L'équipe de recherche est hébergée dans les locaux de l'Ecole Nationale Supérieure des Sciences Appliquées et de Technologie,
http://www.enssat.fr, à Lannion. La thèse est financée sur trois ans par une bourse du conseil général des Côtes d'Armor.

__________________________________________________________________________________________________________

Olivier BOEFFARD
IRISA/ENSSAT - Université de Rennes 1
6 rue de Kerampont - BP 80518
F-22305 Lannion Cedex, France
Tel: +33 2 96 46 90 91
Fax: +33 2 96 37 01 99
e-mail: olivier.boeffard@irisa.fr, Olivier.Boeffard@univ-rennes1.fr
web: http://www.irisa.fr/cordial, http://www.enssat.fr
Back to Top

6-9 . (2008-07-24) Microsoft: Norwegian Linguist (M/F)

Opened positions/internships at Microsoft: Norwegian Linguist (M/F)

MLDC – Microsoft Language Development Center, a branch of the Microsoft Product Group that develops Speech Recognition and Synthesis Technologies, situated in Porto Salvo, Portugal (http://www.microsoft.com/portugal/mldc), is seeking a full-time temporary language expert in the Norwegian language (Bokmal), for a 4-6 month contract, to work in speech technology related development projects. The successful candidate should have the following requirements:

·         Be native or near native Norwegian Bokmal speaker

·         Have a university degree in Linguistics or related field (preferably in Norwegian Linguistics)

·         Have an advanced level of English

·         Have some experience in working with Speech Technology/Natural Language Processing/Linguistics, either in academia or in industry

·         Have some computational ability – no programming is required, but he/she should be comfortable working with MS Windows and MS Office tools

·         Have team work experience

·         Willing to work in Porto Salvo (near Lisbon) for the duration of the contract

·         Willing to start in October 2008

To apply, please submit your resume and a brief statement describing your experience and abilities to Daniela Braga: i-dbraga@microsoft.com

We will only consider electronic submissions.

Deadline for submissions: August 10, 2008 

Back to Top

6-10 . (2008-07-24) Microsoft: Finnish Linguist (M/F)

Opened positions/internships at Microsoft: Finnish Linguist (M/F)

MLDC – Microsoft Language Development Center, a branch of the Microsoft Product Group that develops Speech Recognition and Synthesis Technologies, situated in Porto Salvo, Portugal (http://www.microsoft.com/portugal/mldc), is seeking a full-time temporary language expert in the Finnish language, for a 6 month contract, to work in speech technology related development projects. The successful candidate should have the following requirements:

·         Be native or near native Finnish speaker

·         Have a university degree in Linguistics or related field (preferably in Norwegian Linguistics)

·         Have an advanced level of English (oral and written)

·         Have some experience in working with Speech Technology/Natural Language Processing/Linguistics, either in academia or in industry

·         Have some computational ability – no programming is required, but he/she should be comfortable working with MS Windows and MS Office tools

·         Have team work experience

·         Willing to work in Porto Salvo (near Lisbon) for the duration of the contract

·         Willing to work in a multicultural and multinational team across the globe

·         Willing to start in September 1, 2008

To apply, please submit your resume and a brief statement describing your experience and abilities to Daniela Braga: i-dbraga@microsoft.com

We will only consider electronic submissions.

Deadline for submissions: August 10, 2008

Back to Top

6-11 . (2008-08-12)International Internship Program “Speech and Language Technology” at digital publishing, Munich, Germany

International Internship Program “Speech and Language Technology” at digital publishing, Munich, Germany
digital publishing AG is one of Europe’s leading producers of interactive software for foreign language training. The e-learning courses of digital publishing place an emphasis on speaking and spoken language understanding.
Internships are usually organized as 2 – 6 month projects. Candidates are expected to work on site at the digital publishing R&D Lab. People in the lab speak English and German. We especially welcome applications by native speakers of the languages German, English, French, Spanish, Italian, Russian and Chinese.
Projects will be in the fields of:
- computer-aided language learning (CALL)
- computer-aided pronunciation teaching (CAPT)
- training, configuration or evaluation of speech recognizers for CALL and CAPT systems
- grammar writing for syntactic and semantic parsers
- programming projects in C or C++ involving speech recognition
We offer
- a creative working atmosphere in an international team of software engineers, linguists and editors working on challenging research projects in speech recognition and speech dialogue systems
- a workplace in the center of Munich, in the neighborhood where Albert Einstein spent his childhood
- with beautiful lakes and the mountains of the Alps nearby, Munich is the ideal starting point for activities like swimming, sailing, moutaineering and skiing
- flexible working times, fair compensation, and arguably the best espresso in town
We expect
- good knowledge of C or C++ (for projects that involve programming)
- knowledge of scripting languages
- knowledge of HTK or other speech recognition toolkits
- a background in speech technology, (computational) linguistics, computer science or machine learning
- knowledge about grammar writing
- good knowledge of English or German
Interested? We look forward to your application:
(preferably by e-mail and a preferred project area)
digital publishing AG Karl Weilhammer k.weilhammer@digitalpublishing.de Tumblinger Straße 32 D-80337 München
Germany
Back to Top

6-12 . (2008-08-12) C or C++ Coder for Speech Technology Software at digital publishing AG Munchen Germany

C or C++ Coder for Speech Technology Software
digital publishing AG is one of Europe’s leading producers of interactive software for foreign language training. The e-learning courses of digital publishing place an emphasis on speaking and spoken language understanding.
In order to strengthen our Research & Development Team in Munich, Germany, we are looking for experienced C or C++ programmers for design and coding of desktop applications under Windows. We are looking forward to applications from experienced professionals and recent graduates with excellent coding skills.
We offer
- a creative working atmosphere in an international team of software engineers, linguists and editors working on challenging research projects in speech recognition and speech dialogue systems
- participation in all phases of a product life cycle, as we are interested in the fast transfer of research results into products
- the possibility to participate in international scientific conferences.
- a permanent job in the center of Munich
- excellent possibilities for development within our fast growing company
- flexible working times, competitive compensation and arguably the best espresso in town
We expect
- practical experience in software development in C or C++.
- experience with object-oriented design
- experience with parallel algorithms and thread programming
- good knowledge of English or German
Desirable is
- experience in commercial software development
- experience with optimization of algorithms
- experience in statistical speech or language processing, preferably speech recognition, speech synthesis, speech dialogue systems or chatbots
- experience with Delphi or Turbo Pascal
Interested? We look forward to your application:
(preferably by e-mail)
digital publishing AG Karl Weilhammer k.weilhammer@digitalpublishing.de Tumblinger Straße 32 D-80337 München
Germany
Back to Top

6-13 . (2008-08-27) Language experts at Microsoft Development Center (PORTUGAL)

Opened positions/internships at Microsoft: Norwegian, Finnish, Italian, French, English, Polish Linguists (M/F)

MLDC – Microsoft Language Development Center, a branch of the Microsoft Product Group that develops Speech Recognition and Synthesis Technologies, situated in Porto Salvo, Portugal (http://www.microsoft.com/portugal/mldc), is seeking full-time temporary language experts in the following languages:

-          Norwegian language (Bokmal variety)

-          Finnish

-          Italian

-          English (UK)

-          French (France)

-          Polish

The contracts range from 2-6 months and the scope of them is to work in speech technology related development projects. The successful candidate should have the following requirements:

·         Be a native or near native speaker (for each of the required language)

·         Have a university degree in Linguistics or related field

·         Have an advanced level of English (oral and written)

·         Have some experience in working with Speech Technology/Natural Language Processing/Linguistics, either in academia or in industry

·         Have some computational ability – no programming is required, but he/she should be comfortable working with MS Windows and MS Office tools

·         Have team work experience

·         Willing to work in Porto Salvo (near Lisbon) for the duration of the contract

·         Willing to work in a multicultural and multinational team across the globe

·         Willing to start immediately

To apply, please submit your resume and a brief statement describing your experience and abilities to Daniela Braga: i-dbraga@microsoft.com

We will only consider electronic submissions.

Deadline for submissions: Opened until filled.

 

 

Back to Top

6-14 . (2008-09-02) Assistant professor at Institut Eurecom Sophia Antipolis France

Title:      Assistant Professor Position at EURECOM
            in Multimedia content analysis and processing
Department: Multimedia Communications
URL:        http://www.eurecom.fr/research/
Start date: September 2008

Description:
Research in the Department currently focuses on several aspects of the processing of Multimedia Content:
• Video analysis and information filtering,
• Image Processing with application to 3D Face Cloning, watermarking and
biometrics,
• Speech and sound processing.
The new faculty is expected to undertake research in close collaboration with the existing activities and to participate in the teaching program for Master students. Extensions of the current research areas are encouraged.

Requirements:
The candidates must have a Ph.D. in computer science or electrical engineering with a solid background in signal processing and statistical analysis. The ideal candidate will have an established record of conducting research activities at the international level, and a proven record of collaboration with academic and industrial partners in national and European programs or equivalent. Excellence in research is a constant requirement for EURECOM. A strong commitment to excellence in research and teaching is essential.

Applications:
Send, by email, a letter of motivation, a resume including a list of publications, the names of 3 references and a copy of the three most important publications.

Contact:        Prof. Bernard Merialdo
Postal address: 2229 route des Crêtes
                BP 193
                06904 Sophia Antipolis cedex
                France
Email address:  Bernard.Merialdo@eurecom.fr
Web address:    http://www.eurecom.fr/main/institute/job.fr.htm
Phone:          +33/0 4 93 00 81 29
Fax:            +33/0 4 93 00 82 00

Located in the heart of Sophia Antipolis technology park, EURECOM is a graduate school and a Research center in Communication Systems, founded in 1991 by TELECOM ParisTech (Ecole Nationale Supérieure des Télécommunications) and EPFL (Swiss federal institute of Lausanne) in a consortium form, combining academic and industrial partners. Teaching and research activities of EURECOM focus on three areas: networking and security, mobile communications and multimedia communications. EURECOM has a strong international scope and strategy. 

Back to Top

6-15 . (2008-09-11) Post doc at IRISA Brittany France

Dans le cadre du projet RAPSODIS, l'IRISA recrute un post-doc pour une période de 12 mois. Le sujet porte sur l'intégration de connaissances syntaxiques et sémantiques dans un système de transcription automatique de la parole (voir les détails en fin de message). Le poste est à pourvoir d'ici la fin de l'année. Les personnes intéressés sont invités à contacter Guillaume Gravier (guig@irisa.fr) et/ou Pascale Sébillot (pascale.sebillot@irisa.fr).

Back to Top

6-16 . (2008-09-18) Senior staff position at ICSI Berkeley CA USA

SENIOR STAFF OPENING AT THE ICSI SPEECH GROUP

 

The International Computer Science Institute (ICSI) invites applications for a senior staff position in its speech research group. Successful applicants must have significant post-PhD experience, and a world-class research reputation.  Candidates must also have demonstrated ability to grow and manage a strong research effort. A successful track record with obtaining funding for the chosen area is essential. 

 

The ICSI Speech Group (including its predecessor, the ICSI Realization Group) has been a source of novel approaches to speech processing since 1988. ICSI’s speech group is well known for its efforts in speech recognition (particularly in neural network approaches and novel forms of feature extraction), as well as in speaker recognition, diarization, and speech understanding. It has close ties to research efforts in machine translation on the Berkeley campus, and to the STAR lab at SRI for the complete range of its research. It also works closely with several European labs, particularly IDIAP in Switzerland and to the University of Edinburgh.

 

ICSI is an independent not-for-profit Institute located a few blocks from the Berkeley campus of the University of California. It is closely affiliated with the University, and particularly with the Electrical Engineering and Computer Science (EECS) Department. Students, faculty, and administrative colleagues from the University all play a key role at the Institute. In addition to its Speech Group, areas of current strength in the Institute include: Artificial Intelligence (primarily in natural language), Internet research (primarily in architecture and internet security), and Algorithms (primarily associated with problems in bioinformatics and networking). We also have new activities in Computer Vision and Computer Architecture. See http://www.icsi.berkeley.edu to learn more about ICSI.

 

Applications should include a cover letter, a vita, the names of at least 3 references (with both postal and email addresses), and a research statement. Applications should be sent by email to speechjob@icsi.berkeley.edu and by postal mail to

 

Nelson Morgan (re Senior Search)

ICSI

1947 Center Street, Suite 600

Berkeley, CA 94704

 

ICSI is an Affirmative Action/Equal Opportunity Employer. Applications from women and minorities are especially encouraged. Hiring is contingent on eligibility to work in the United States.

Back to Top

6-17 . (2008-10-30) Programmer Analyst Position at LDC

                                                          Programmer Analyst Position at LDC
The Linguistic Data Consortium (LDC) at the University of Pennsylvania, Philadelphia, PA has an immediate opening for a full-time programmer analyst.
 
Programmer Analyst – Publications Programmer (#081025790)
 
Duties: Position will have primary responsibility for developing, implementing and managing data processing systems required to coordinate and prepare publications of language resources used for human language technology research and technology development.  Such resources include video, computer-readable speech, software and text data that are distributed via media and internet.  Position will  communicate with external data providers and internal project managers to acquire raw source material and to schedule releases; perform quality assessment of large data collections and render analyses/descriptions of their formats; create or adapt software tools to condition data to a uniform format and level of quality (e.g., eliminating corrupted data, normalizing data, etc.); validate quality control standards to published data and verify results; document initial and final data formats; review author documentation and supporting materials; create additional documentation as needed; and master and replicate publications. Position will also maintain the publications catalog system, the publications inventory, the archive of publishable and published data and the publication equipment, software and licenses.  Position requires attention to detail and is responsible for managing multiple short-term projects.
 
For further information on the duties and qualifications for this position, or to apply online please visit http://jobs.hr.upenn.edu/; search postings for the reference number indicated above.
 
Penn offers an excellent benefits package including medical/dental, retirement plans, tuition assistance and a minimum of 3 weeks paid vacation per year. The University of Pennsylvania is an affirmative action/equal opportunity employer.
 
Position contingent upon funding. For more information about LDC and the programs we support, visit http://www.ldc.upenn.edu/.
 
Back to Top

6-18 . (2008-11-17) Volunteering at ISCA Student advisory committee

 Announcement #1: ISCA-SAC Call for Volunteers
 
The ISCA Student Advisory Committee (ISCA-SAC) is seeking student volunteers to help with several interesting projects such as transcribe interviews from the Saras Institute, plan/organize student events at ISCA-sponsored conferences/workshops, increase awareness of speech and language research to undergraduate and junior graduate students, assist with website redesign to facilitate interaction with Google Scholar, as well as collect resources (e.g., conferences, theses, job listings, speech labs, etc.) for the isca-students.org website, to name a few.
 
There are many small tasks that can be done, each of which would only take up a few hours. Unless it is of your interest to become a long term volunteer, no further commitment is required. If interested, please contact the ISCA-SAC Volunteer Coordinator at: vo lun te er [at] isca-students [dot] org.
 
Announcement #2: ISCA-SAC Logo Contest
 
The ISCA Student Advisory Committee is in the search for a new logo. This is your chance to release your artistic side and enter the ISCA-SAC Logo Competition. All students are invited to participate and a prize (still to be determined) will be awarded to the winner; not to mention the importance of having your logo posted on the isca-students.org website for the world to see.
 
The deadline for submissions is March 31st, 2009. The new Logo will be unveiled during the Interspeech 2009 conference in the form of merchandise embedded with the new logo (e.g., mugs, pens, etc.).

If interested, please send your submissions to: logocontest [at] isca-students.org  

Back to Top

6-19 . (2008-11-20) 12 PhD Positions and 2 Post Doc Positions available in SCALE (EU Marie Curie)

12 PhD Positions and 2 Post Doc Positions available in

 

the Marie Curie International Training Network on

 

Speech Communication with Adaptive LEarning (SCALE)

 

SCALE is a cooperative project between

 

·        IDIAP Research Institute in Martigny, Switzerland (Prof Herve Bourlard)

·        Radboud University Nijmegen, The Netherlands (Prof Lou Boves, Dr Louis ten Bosch, Dr-ir Bert Cranen, Dr O. Scharenborg)

·        RWTH Aachen, Germany (Prof Hermann Ney, Dr Ralf Schlüter)

·        Saarland University, Germany (Prof Dietrich Klakow, Dr John McDonough)

·        University of Edinburgh, UK (Prof Steve Renals, Dr Simon King, Dr Korin Richmond, Dr Joe Frankel)

·        University of Sheffield, UK (Prof Roger Moore, Prof Phil Green, Dr Thomas Hain, Dr Guido Sanguinetti) .

 

Companies like Motorola or Philips Speech Recognition Systems/Nuance are associated partners of the program.

 

Each PhD position is funded for three years and degrees can be obtained from the participating academic institutions. 

 

Distinguishing features of the cooperation include:

 

·        Joint supervision of dissertations by lecturers from two partner institutions

·        While staying with one institution for most of the time, the program includes a stay at a second partner institution either from academic or industry for three to nine month 

·        An intensive research exchange program between all participating institutions

 

PhD and Post Doc projects will be in the area of

 

·        Automatic Speech Recognition

·        Machine learning

·        Speech Synthesis

·        Signal Processing

·        Human speech recognition

 

The salary of a PhD position is roughly 33.800€ per year. There are additional mobility (up to 800€/month) and travel allowances (yearly allowance). Applicants should hold a strong university degree which would entitle them to embark on a doctorate (Masters/diploma or equivalent) in a relevant discipline, and should be in the first four years of their research careers. As the project is funded by a EU mobility scheme, there are also certain mobility requirements.

 

Each Post Doc position is funded for two years. The salary is approximately 52000€ per year. Applicants must have a doctoral degree at the time of recruitment or equivalent research experience. The research experience may not exceed 5 years at the time of appointment.

 

Women are particularly encouraged to apply.

 

Deadlines for applications:

 

January 1, 2009

April 1, 2009

July 1, 2009

September 1, 2009.

 

After each deadline all submitted applications will be reviewed and positions awarded until all positions are filled.

 

Applications should be submitted at http://www.scale.uni-saarland.de/ .

 

To be fully considered, please include:

 

- a curriculum vitae indicating degrees obtained, disciplines covered

(e.g. list of courses ), publications, and other relevant experience

 

- a sample of written work (e.g. research paper, or thesis,

preferably in English)

 

- copies of high school and university certificates, and transcripts

 

- two references (e-mailed directly to the SCALE office

(Diana.Schreyer@LSV.Uni-Saarland.De) before the deadline)

 

- a statement of research interests, previous knowledge and activities

in any of the relevant research areas.

 

In case an application can only be submitted by regular post, it should

be sent to:

 

SCALE office

Diana Schreyer

Spoken Language Systems, FR 7.4

C 71 Office 0.02

Saarland University

P.O. Box 15 11 50

D-66041 Saarbruecken

Germany

 

If you have any questions, please contact Prof. Dr. Dietrich Klakow

(Dietrich.Klakow@LSV.Uni-Saarland.De).

 

Back to Top

6-20 . (2009-01-08) Assistant Professor Toyota Technological Institute at Chicago

Assistant Professor Toyota Technological Institute at Chicago  ########################################################  Toyota Technological Institute at Chicago (http://www.tti-c.org) is a philanthropically endowed academic computer science institute, dedicated to basic research and graduate education in computer science.  TTI-C opened for operation in 2003 and by 2010 plans to have 12 tenured and tenure track faculty and 18 research (3-year) faculty. Regular faculty will have a teaching load of at most one course per year and research faculty will have no teaching responsibilities.  Applications are welcome in all areas of computer science, but TTI-C is currently focusing on a number of areas including speech and language processing.  For all positions we require a Ph.D. degree or Ph.D. candidacy, with the degree conferred prior to date of hire.  Applications received after December 31 may not get full consideration.  Applications can be submitted online at http://www.tti-c.org/facultyapplication
Back to Top

6-21 . (2009-01-09) Poste d'ingénieur CDD : environnement intelligent

Poste d'ingénieur CDD : environnement intelligent

Ingenieur - CDD

DeadLine: 15/02/2008

olivier.pietquin@supelec.fr

http://ims.metz.supelec.fr/spip.php?article99

 

Un poste d'ingénieur CDD de 18 mois est ouvert sur le campus de Metz de Supélec. Le candidat s’intégrera au sein de l’équipe « Information, Multimodalité & Signal » (http://ims.metz.supelec.fr). Cette équipe composée de 15 personnes est active dans les domaines du traitement numérique du signal et de l’information (traitement statistique du signal, apprentissage numérique, méthodes d’inspiration biologique), de la représentation des connaissances (fouille de données, analyse et apprentissage symbolique) et du calcul intensif et distribué. Le poste vise un profil permettant l’implémentation matérielle intégrée des méthodes développées au sein de l’équipe dans des applications liées aux environnements intelligents ainsi que leur maintenance. Le campus de Metz s’est en effet doté d’une plateforme en vraie grandeur reproduisant une pièce intelligente intégrant caméras, microphones, capteurs infrarouges, interfaces homme-machine (interface vocale, interface cerveau-machine), robo!

 ts et moyens de diffusion d’information. Il s’agira de réaliser une plateforme intégrée permettant de déployer des démonstrations rapidement dans cet environnement et de les maintenir.

 

 

Profil recherché :

– diplôme d’ingénieur en informatique, ou équivalent universitaire

– expérience de travail dans le cadre d’équipes multidisciplinaires,

– une bonne pratique de l’anglais est un plus.

 

Plus d'informations sont disponibles sur le site de l'équipe (http://ims.metz.supelec.fr)

 

Faire acte de candidature (CV+lettre) auprès de O. Pietquin : olivier.pietquin@supelec.fr.

 

 

 

http://gdr-isis.org/rilk/gdr/Kiosque/poste.php?jobid=3010

Back to Top

6-22 . (2009-01-13) 2009 PhD Research Fellowships at the University of Trento (Italy)

2009 PhD Research Fellowships 

=============================

 

The Adaptive Multimodal Information and  Interface  Research Lab

(casa.disi.unitn.it) at University of Trento (Italy) has several

PhD Research fellowships in the following areas:

 

                Statistical Machine Translation                

                Natural Language Processing    

                Automatic Speech Recognition

                Machine Learning

                Spoken/Multimodal Conversational Systems

                   

We are looking for students with _excellent_ academic records

and relevant technical background. Students with EE, CS Master degrees

( or equivalent ) are welcome as well other related disciplines will

be considered. Prospective students are encouraged to look at the lab

website to search for current and past research projects.

 

PhD research fellowships benefits are described in the graduate school

website (http://ict.unitn.it/disi/edu/ict).

The  applicants should be fluent in _English_. The Italian language

competence is optional and applicants are encouraged to acquire

this skill during training. All applicants should have very good

programming skills. University of Trento is an equal opportunity employer.

 

The selection of candidates will be open until positions are filled.

Interested applicants should send their CV along with their

statement of research interest, transcript records and three reference

letters to :

 

 

                Prof. Dr.-Ing. Giuseppe Riccardi

        Email: riccardi@disi.unitn.it

 

 

-------------------

About University of Trento and Information Engineering and Computer

 Science Department

 

The University of Trento is constantly ranked as

 premiere Italian graduate university institution (see www.disi.unitn.it).

 

Please visit the DISI Doctorate school website at http://ict.unitn.it/edu/ict

 

DISI Department

DISI has a strong focus on Interdisciplinarity with professors from

different faculties of the University (Physical Science, Electrical

Engineering, Economics, Social Science, Cognitive Science, Computer Science)

 with international background.

DISI aims at exploiting the complementary experiences present in the

various research areas in order to develop innovative methods and

technologies, applications and advanced services.

English is the official language.

 

 

--

Prof. Ing. Giuseppe Riccardi

Marie Curie Excellence Leader

Department of Information Engineering and Computer Science

University of Trento

Room D11, via Sommarive 14

38050 Povo di Trento, Italy

tel  : +39-0461 882087

email: riccardi@dit.unitn.it

   

Back to Top

7 . Journals

 

Back to Top

7-1 . IEEE Signal Processing Magazine: Special Issue on Digital Forensics

Guest Editors:
Edward Delp, Purdue University (ace@ecn.purdue.edu)
Nasir Memon, Polytechnic University (memon@poly.edu)
Min Wu, University of Maryland, (minwu@eng.umd.edu)

We find ourselves today in a "digital world" where most information
is created, captured, transmitted, stored, and processed in digital 
form. Although, representing information in digital form has many 
compelling technical and economic advantages, it has led to new 
issues and significant challenges when performing forensics analysis 
of digital evidence.  There has been a slowly growing body of 
scientific techniques for recovering evidence from digital data.  
These techniques have come to be loosely coupled under the umbrella 
of "Digital Forensics." Digital Forensics can be defined as "The 
collection of scientific techniques for the preservation, collection, 
validation, identification, analysis, interpretation, documentation 
and presentation of digital evidence derived from digital sources for 
the purpose of facilitating or furthering the reconstruction of 
events, usually of a criminal nature."

This call for papers invites tutorial articles covering all aspects 
of digital forensics with an emphasis on forensic methodologies and 
techniques that employ signal processing and information theoretic 
analysis. Thus, focused tutorial and survey contributions are 
solicited from topics, including but not limited to, the following:

 . Computer Forensics - File system and memory analysis. File carving.
 . Media source identification - camera, printer, scanner, microphone
identification.
 . Differentiating synthetic and sensor media, for example camera vs.
computer graphics images.
 . Detecting and localizing media tampering and processing.
 . Voiceprint analysis and speaker identification for forensics.
 . Speech transcription for forensics. Analysis of deceptive speech.
 . Acoustic processing for forensic analysis - e.g. acoustical gunshot
analysis, accident reconstruction.
 . Forensic musicology and copyright infringement detection.
 . Enhancement and recognition techniques from surveillance video/images.
Image matching techniques for auto-matic visual evidence
extraction/recognition.
 . Steganalysis - Detection of hidden data in images, audio, video. 
Steganalysis techniques for natural language steganography. Detection of covert
channels.
 . Data Mining techniques for large scale forensics.
 . Privacy and social issues related to forensics.
 . Anti-forensics. Robustness of media forensics methods against counter
measures.
 . Case studies and trend reports.

White paper submission: Prospective authors should submit white 
papers to the web based submission system at http://
www.ee.columbia.edu/spm/ according to the timetable. given below.  
White papers, limited to 3 single-column double-spaced pages, should 
summarize the motivation, the significance of the topic, a brief 
history, and an outline of the content.  In all cases, prospective 
contributors should make sure to emphasize the signal processing in 
their submission.

Schedule
 . White Paper Due: April 7, 2008
 . Notification of White paper Review Results: April 30, 2008
 . Full Paper Submission: July 15, 2008
 . Acceptance Notification: October 15, 2008
 . Final Manuscript Due: November 15, 2008
 . Publication Date: March 2009.


Back to Top

7-2 . Special Issue on Integration of Context and Content for Multimedia Management

IEEE Transactions on Multimedia            
 Special Issue on Integration of Context and Content for Multimedia Management
=====================================================================

Guest Editors:

Alan Hanjalic, Delft University of Technology, The Netherlands
Alejandro Jaimes, IDIAP Research Institute, Switzerland
Jiebo Luo, Kodak Research Laboratories, USA
        Qi Tian, University of Texas at San Antonio, USA

---------------------------------------------------
URL: http://www.cs.utsa.edu/~qitian/cfp-TMM-SI.htm
---------------------------------------------------
Important dates:

Manuscript Submission Deadline:       April 1, 2008
        Notification of Acceptance/Rejection: July 1, 2008
        Final Manuscript Due to IEEE:         September 1, 2008
        Expected Publication Date:            January 2009

---------------------
Submission Procedure
---------------------
Submissions should follow the guidelines set out by IEEE Transaction on Multimedia.
Prospective authors should submit high quality, original manuscripts that have not
appeared, nor are under consideration, in any other journals.

-------
Summary
-------
Lower cost hardware and growing communications infrastructure (e.g., Web, cell phones,
etc.) have led to an explosion in the availability of ubiquitous devices to produce,
store, view and exchange multimedia (images, videos, music, text). Almost everyone is
a producer and a consumer of multimedia in a world in which, for the first time,
tremendous amount of contextual information is being automatically recorded by the
various devices we use (e.g., cell ID for the mobile phone location, GPS integrated in
a digital camera, camera parameters, time information, and identity of the producer).

In recent years, researchers have started making progress in effectively integrating
context and content for multimedia mining and management. Integration of content and
context is crucial to human-human communication and human understanding of multimedia:
without context it is difficult for a human to recognize various objects, and we
become easily confused if the audio-visual signals we perceive are mismatched. For the
same reasons, integration of content and context is likely to enable  (semi)automatic
content analysis and indexing methods to become more powerful in managing multimedia
data. It can help narrow part of the semantic and sensory gap that is difficult or
even impossible to bridge using approaches that do not explicitly consider context for
(semi)automatic content-based analysis and indexing.

The goal of this special issue is to collect cutting-edge research work in integrating
content and context to make multimedia content management more effective. The special
issue will unravel the problems generally underlying these integration efforts,
elaborate on the true potential of contextual information to enrich the content
management tools and algorithms, discuss the dilemma of generic versus narrow-scope
solutions that may result from "too much" contextual information, and provide us
vision and insight from leading experts and practitioners on how to best approach the
integration of context and content. The special issue will also present the state of
the art in context and content-based models, algorithms, and applications for
multimedia management.

-----
Scope
-----

The scope of this special issue is to cover all aspects of context and content for
multimedia management.

Topics of interest include (but are not limited to):
        - Contextual metadata extraction
        - Models for temporal context, spatial context, imaging context (e.g., camera
          metadata), social and cultural context and so on
- Web context for online multimedia annotation, browsing, sharing and reuse
- Context tagging systems, e.g., geotagging, voice annotation
- Context-aware inference algorithms
        - Context-aware multi-modal fusion systems (text, document, image, video,
          metadata, etc.)
- Models for combining contextual and content information
        - Context-aware interfaces
- Context-aware collaboration
- Social networks in multimedia indexing
- Novel methods to support and enhance social interaction, including
          innovative ideas integrating context in social, affective computing, and
          experience capture.
- Applications in security, biometrics, medicine, education, personal
          media management, and the arts, among others
- Context-aware mobile media technology and applications
- Context for browsing and navigating large media collections
- Tools for culture-specific content creation, management, and analysis

------------
Organization
------------
Next to the standard open call for papers, we will also invite a limited number of
papers, which will be written by prominent authors and authorities in the field
covered by this Special Issue. While the papers collected through the open call are
expected to sample the research efforts currently invested within the community on
effectively combining contextual and content information for optimal analysis,
indexing and retrieval of multimedia data, the invited papers will be selected to
highlight the main problems and approaches generally underlying these efforts.

All papers will be reviewed by at least 3 independent reviewers. Invited papers will
be solicited first through white papers to ensure the quality and relevance to the
special issue. The accepted invited papers will be reviewed by the guest editors and
expect to account for about one fourth of the papers in the special issue.

---------
Contacts
---------
Please address all correspondences regarding this special issue to the Guest Editors
Dr. Alan Hanjalic (A.Hanjalic@ewi.tudelft.nl), Dr. Alejandro Jaimes
(alex.jaimes@idiap.ch), Dr. Jiebo Luo (jiebo.luo@kodak.com), and Dr. Qi Tian
(qitian@cs.utsa.edu).
-------------------------------------------------------------------------------------

Guest Editors:
Alan Hanjalic, Alejandro Jaimes, Jiebo Luo, and Qi Tian


Back to Top

7-3 . Speech Communication: Special Issue On Spoken Language Technology for Education

CALL FOR PAPERS
Special Issue of Speech Communication

on *Spoken Language Technology for Education*


*Guest-editors:*
Maxine Eskenazi, Associate Teaching Professor, Carnegie Mellon University
Abeer Alwan, Professor, University of California at Los Angeles
Helmer Strik, Assistant Professor, University of Nijmegen
 

Language technologies have evolved to the stage where they are reliable
enough, if their strong and weak points are properly dealt with, to be
used for education. The creation of an application for education
presents several challenges: making the language technology sufficiently
reliable (and thus advancing our knowledge in the language
technologies), creating an application that actually enables students to
learn, and engaging the student. Papers in this special issue should
deal with several of these issues. Although language learning is the
primary target of research at present, papers on the use of language
technologies for other education applications are encouraged. The scope
of acceptable topic interests includes but is not limited to:

 

- Use of speech technology for education

- Use of spoken language dialogue for education

- Applications using speech and natural language processing for education

- Intelligent tutoring systems using speech and natural language

- Pedagogical issues in using speech and natural language technologies
for education

- Assessment of tutoring software

- Assessment of student performance

 

*Tentative schedule for paper submissions, review, and revision**: ** *

Deadline for submissions: June 1, 2008.

Deadline for decisions and feedback from reviewers and editors: August
31, 2008.

Deadline for revisions of papers: November 31, 2008.

 

*Submission instructions:*

Authors should consult the "Guide for Authors", available online, at
http://www.elsevier.com/locate/specom for information about the
preparation of their manuscripts. Authors, please submit your paper via
_http://ees.elsevier.com/specom_, choosing *Spoken Language Tech. *as
the Article Type, and  Dr. Gauvain as the handling E-i-C.

Back to Top

7-4 . Special Issue on Processing Morphologically Rich Languages IEEE Trans ASL

Call for Papers for a Special Issue on
                Processing Morphologically Rich Languages 
          IEEE Transactions on Audio, Speech and Language Processing
 
This is a call for papers for a special issue on Processing Morphologically
Rich Languages, to be published in early 2009 in the IEEE Transactions on 
Audio, Speech and Language Processing. 
 
Morphologically-rich languages like Arabic, Turkish, Finnish, Korean, etc.,
present significant challenges for speech processing, natural language 
processing (NLP), as well as speech and text translation. These languages are 
characterized by highly productive morphological processes (inflection, 
agglutination, compounding) that may produce a very large number of word 
forms for a given root form.  Modeling each form as a separate word leads 
to a number of problems for speech and NLP applications, including: 1) large 
vocabulary growth, 2) poor language model (LM) probability estimation, 
3) higher out-of-vocabulary (OOV) rate, 4) inflection gap for machine 
translation:  multiple different forms of  the same underlying baseform 
are often treated as unrelated items, with negative effects on word alignment 
and translation accuracy.  
 
Large-scale speech and language processing systems require advanced modeling 
techniques to address these problems. Morphology also plays an important 
role in addressing specific issues of “under-studied” languages such as data 
sparsity, coverage and robust modeling. We invite papers describing 
previously unpublished work in the following broad areas: Using morphology for speech recognition and understanding, speech and text translation, 
speech synthesis, information extraction and retrieval, as well as 
summarization . Specific topics of interest include:
- methods addressing data sparseness issue for morphologically rich 
  languages with application to speech recognition, text and speech 
  translation, information extraction and retrieval, speech   
  synthesis, and summarization
- automatic decomposition of complex word forms into smaller units 
- methods for optimizing the selection of units at different levels of 
  processing
- pronunciation modeling for morphologically-rich languages
- language modeling for morphologically-rich languages
- morphologically-rich languages in speech synthesis
- novel probability estimation techniques that avoid data sparseness 
  problems
- creating data resources and annotation tools for morphologically-rich 
  languages
 
Submission procedure:  Prospective authors should prepare manuscripts 
according to the information available at 
http://www.signalprocessingsociety.org/periodicals/journals/taslp-author-in=ormation/. 
Note that all rules will apply with regard to submission lengths, 
mandatory overlength page charges, and color charges. Manuscripts should 
be submitted electronically through the online IEEE manuscript submission 
system at http://sps-ieee.manuscriptcentral.com/. When selecting a 
manuscript type, authors must click on "Special Issue of TASLP on 
Processing Morphologically Rich Languages". 
 
Important Dates:
Submission deadline:  August 1, 2008               
Notification of acceptance: December 31, 2008
Final manuscript due:  January 15, 2009    
Tentative publication date: March 2009
 
Editors
Ruhi Sarikaya (IBM T.J. Watson Research Center) sarikaya@us.ibm.com
Katrin Kirchhoff (University of Washington) katrin@ee.washington.edu
Tanja Schultz (University of Karlsruhe) tanja@ira.uka.de
Dilek Hakkani-Tur (ICSI) dilek@icsi.berkeley.ed
Back to Top

7-5 . Special Issue Analysis and Signal Processing of of Oesophageal and Pathological Voices

Special Issue of EURASIP Journal on Advances in Signal Processing
on Analysis and Signal Processing of Oesophageal and Pathological Voices
 
 
Call for Papers
Speech is the most important means of communication
among humans. Speech, however, is not limited only to the
process of communication, but is also very important for
transferring emotions, expressing our personality, and reflecting
situations of stress. Modern lifestyles have increased
the risk of experiencing some kind of voice alteration. It is
estimated that around 19% of the population suffer or have
suffered from dysphonic voicing. Thismotivates new and objective
ways to evaluate speech, its quality, and its connection
with other phenomena.
Speech research to date has favored areas such as synthesis,
recognition, and speaker verification. The last few years have
witnessed the emerging topic of processing and evaluation
of disordered speech. Acoustic analysis is a noninvasive technique
providing an efficient tool for the objective diagnosis,
the screening of voice diseases, the objective determination
of vocal function alterations, and the evaluation of surgical
treatment and rehabilitation. Its application extends beyond
medicine, and now includes forensic analysis as well as voice
quality control for voice professionals. Acoustic analysis may
also be seen as complementary to other methods of evaluation
based on the direct observation of the vocal folds using
videoendoscopy.
This special issue aims to foster an interdisciplinary forumfor
presenting new work in the analysis andmodeling of
voice signals and videoendoscopic images, with applications
in pathological and oesophageal voices.
Topics of interest include (but are not limited to):
• Automatic detection of voice disorders
• Automatic assessment and classification of voice quality
• New strategies for the parameterization and modeling
of normal and pathological voices (biomechanicalbased
parameters, chaos modeling, etc.)
• Databases of vocal disorders
• Inverse filtering
• Signal processing for remote diagnosis
• Speech enhancement for pathological and oesophageal
voices
• Objective parameters extraction from vocal fold
images using videolaryngoscopy, videokymography,
fMRI, and other emerging techniques
• Multimodal analysis of disordered speech
• Robust pitch extraction algorithms for pathological
and oesophageal voices Robust pitch extraction algorithms
for pathological and oesophageal voices
Since speech communication is fundamental to human interaction,
we are moving towards a new scenario where speech
is gaining greater importance in our daily lives, and many
common speech disorders and dysfunctions would be treated
using computer-based or physical prosthetics.
Authors should follow the EURASIP Journal on Advances
in Signal Processing manuscript format described
at http://www.hindawi.com/journals/asp/. Prospective authors
should submit an electronic copy of their complete
manuscript through the journalManuscript Tracking System
at http://mts.hindawi.com/ according to the following tentative
timetable:
Manuscript Due November 1, 2008
First Round of Reviews February 1, 2009
Publication Date May 1, 2009
Guest Editors
Juan I. Godino-Llorente, Department of Circuits and
Systems Engineering, Polytechnic University of Madrid
(UPM), Ctra. de Valencia, 28031 Madrid, Spain;
igodino@ics.upm.es
Pedro Gómez-Vilda, Department of Computer Science
and Engineering, Polytechnic University of Madrid (UPM),
Boadilla del Monte, 28660 Madrid, Spain;
pedro@pino.datsi.fi.upm.es
Tan Lee, Department of Electronic Engineering, The
Chinese University of Hong Kong, Shatin, New Territories,
Hong Kong; tanlee@ee.cuhk.edu.hk
Hindawi Publishing Corporation
http://www.hindawi.com
Back to Top

7-6 . Issue of Speech Communication on ‘‘Silent Speech’’ Interfaces

Special Issue on ‘‘Silent Speech’’ Interfaces
Guest Editors
Professor Bruce Denby (denby@ieee.org)
Prof. Dr. Ing. Tanja Schultz (tanja@ira.uka.de)
Dr. Kiyoshi Honda, MD, DMsc. (honda@atr.jp)
A ‘‘Silent Speech’’ Interface (SSI) allows to process a speech signal which a user outputs without actually vocalizing any
sound. Based on sensors of different types, such systems would provide normal speech to those for whom vocalization is
difficult or impossible due to age or illness, and as such are a complement to surgical solutions, vocal prostheses, and touchscreen
synthesis systems. More recently, the advent of the cellular telephone has created interest in SSIs from quite a
different perspective. The electronic representation of the speech signal created by an SSI can be injected directly into a
digital transmission system, leaving synthesis to be carried out only at the distant user’s handset. This opens the way to
telecommunications systems operating in total silence, thus assuring the privacy and security of users’ communications,
while at the same time protecting the acoustic environment of those not participating in the exchange. As a further benefit,
since SSIs do not use standard acoustic capture techniques, they will also be very interesting in terms of speech processing
in noisy environments. Quite naturally, the ‘‘silent communication’’ and high-noise environment capabilities of SSIs have
attracted the interest of the defense and security communities, as well.
Prototype SSI systems have already appeared in the research literature, including: imaging-based solutions such as
ultrasound and standard video capture; inertial approaches which translate articulator movement directly into electrical
signals, for example electromagnetic articulography; electromyographic techniques, which capture the minute electrical
signals associated with articulator movement; systems exploiting non-audible acoustic signals produced by articulator
movement, such as ‘‘non-acoustic murmur’’ microphones; all the way to ‘‘brain computer interfaces’’ in which neural
speech command signals are captured before they reach the articulators, thus obviating the need for movement of any kind
on the part of the speaker.
The goal of the special issue on ‘‘Silent Speech’’ Interfaces is to provide to the speech community an introduction to this
exciting, emergent field. Contributions should therefore cover as broad an area as possible, but at the same time, be of
sufficient depth to encourage the critical evaluations and reflections that will lead to further advances in the field, and
hopefully to new collaborations. To obtain the necessary quality, breadth, and balance, a limited number of invited articles
will be complemented by a call for submission of 1-page paper proposals. The final issue will be compiled from the invited
contributions and the follow-up full articles from accepted 1-page proposals. There will also be a comprehensive review
article, to which some article authors may be asked to contribute. All papers, both invited and submitted, will undergo the
usual Speech Communication peer review process.
Proposals for contributions (1-page only, in .pdf format), outlining the originality of the approach, current status of
the research work, as well as benefits and potential drawbacks of the method, should be sent to denby@ieee.org by
9 September 2008. A list of important dates is given below.
Important dates
Invited articles: Invitations are sent concurrently with the Call for Papers.
Deadline for submission of 1-page proposals: 9 September 2008 (submit .pdf directly to denby@ieee.org).
Notification of acceptance for 1-page proposals: 30 September 2008.
Deadline of submission for full papers, both proposed and invited: 30 November 2008. All authors are asked to prepare their
full papers according to the guidelines set in the Guide for Authors, located at http://www.elsevier.com/locate/specom, and
to submit their papers to the online submission and reviewing tool, at http://ees.elsevier.com/specom. They should select
Special Issue: ‘‘Silent Speech’’ Interfaces, as the article type, and Professor Kuldip Paliwal as the handling Editor-in-Chief.
Journal publication: Second quarter 2009.
Back to Top

7-7 . Special issue of EURASIP Journal of Advances in Signal Processing on Biometrics

Call for Papers

Recent Advances in Biometric Systems: A Signal Processing Perspective



Biometrics a digital recognition technology that relies on highly distinctive physical and physiological characteristics of an individual is potentially a powerful and reliable method for personal authentication. The increasing importance of biometrics is underscored by the rapidly growing number of educational and research activities devoted to this field; and by a large number of annually organized Conferences and Symposia exclusively devoted to biometrics. Biometrics is a multidisciplinary field with researchers from signal processing, pattern recognition, computer vision, and statistics. Recently, a number of new important directions have been identified for biometric research, including processing and encoding of nonideal data, biometrics at a distance, and data quality assessment. Problems in nonideal biometric data include off-angle, occluded, blurred, and noisy images. Biometrics at a distance is concerned with recognition from video or snapshots of a biometric samples captured from a noncooperative moving individual. The goal of this special issue is to focus on recent advances in signal processing of biometric data that allow improved recognition performance through novel restoration, processing, and encoding; matching techniques capable of dealing with complexity and distortions in data acquired from a distance; recognition from biometric data acquired from unconstrained environments or complex experimental set ups; and the characterization of quality and its relationship with performance.

Topics of interest include, but are not limited to:

Biometric-based recognition under unconstrained presentation and/or complex environment using the following:
    o Face
    o Iris
    o Fingerprint
    o Voice
    o Hand
    o Soft biometrics

Multimodal biometric recognition using nonideal data

Biometric image/signal quality assessment:
    o Face
    o Iris
    o Fingerprint
    o Voice
    o Hand
    o Soft biometrics

Biometric security and privacy
    o Liveness detection
    o Encryption
    o Cancelable biometrics

The special issue will focus both on the development and comparison of novel signal/image processing approaches and on their expanding range of applications.

Authors should follow the EURASIP Journal on Advances in Signal Processing manuscript format described at the journal site http://www.hindawi.com/journals/asp/. Prospective authors should submit an electronic copy of their complete manuscript through the journal Manuscript Tracking System at http://mts.hindawi.com/, according to the following timetable:

Manuscript Due                 October 1, 2008
First Round of Reviews         January 1, 2009
Publication Date               April 1, 2009

Guest Editors

o Natalia A. Schmid, Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV 26506, USA; natalia.schmid@mail.wvu.edu
o Stephanie Schuckers, Electrical &amp; Computer Engineering, Clarkson University, Potsdam, NY 13699, USA; sschucke@clarkson.edu
o Jonathon Phillips, National Institute of Standard and Technology, Gaithersburg, MD 20899, USA; jonathon@nist.gov
o Kevin Bowyer, University of Notre Dame, Notre Dame, IN 46556, USA; kwb@cse.nd.edu 

Back to Top

7-8 . Special issue of Eurasip Journal on Advanced signal processing

Call for Papers

Special issue of Eurasip journal on advanced signal processing

Analysis and Signal Processing of Oesophagial and Pathological Voices



Speech is the most important means of communication among humans. Speech, however, is not limited only to the process of communication, but is also very important for transferring emotions, expressing our personality, and reflecting situations of stress. Modern lifestyles have increased the risk of experiencing some kind of voice alteration. It is estimated that around 19% of the population suffer or have suffered from dysphonic voicing. This motivates new and objective ways to evaluate speech, its quality, and its connection with other phenomena.

Speech research to date has favored areas such as synthesis, recognition, and speaker verification. The last few years have witnessed the emerging topic of processing and evaluation of disordered speech. Acoustic analysis is a noninvasive technique providing an efficient tool for the objective diagnosis, the screening of voice diseases, the objective determination of vocal function alterations, and the evaluation of surgical treatment and rehabilitation. Its application extends beyond medicine, and now includes forensic analysis as well as voice quality control for voice professionals. Acoustic analysis may also be seen as complementary to other methods of evaluation based on the direct observation of the vocal folds using videoendoscopy.

This special issue aims to foster an interdisciplinary forum for presenting new work in the analysis and modeling of voice signals and videoendoscopic images, with applications in pathological and oesophageal voices.

Topics of interest include (but are not limited to):

o Automatic detection of voice disorders
o Automatic assessment and classification of voice quality
o New strategies for the parameterization and modeling of normal and pathological voices (biomechanical-based parameters, chaos modeling, etc.)
o Databases of vocal disorders
o Inverse filtering
o Signal processing for remote diagnosis
o Speech enhancement for pathological and oesophageal voices
o Objective parameters extraction from vocal fold images using videolaryngoscopy, videokymography, fMRI, and other emerging techniques
o Multimodal analysis of disordered speech
o Robust pitch extraction algorithms for pathological and oesophageal voices

Since speech communication is fundamental to human interaction, we are moving towards a new scenario where speech is gaining greater importance in our daily lives, and many common speech disorders and dysfunctions would be treated using computer-based or physical prosthetics.

Authors should follow the EURASIP Journal on Advances in Signal Processing manuscript format described at http://www.hindawi.com/journals/asp/. Prospective authors should submit an electronic copy of their complete manuscript through the journal Manuscript Tracking System at http://mts.hindawi.com/ according to the following tentative timetable:

Manuscript Due                 November 1, 2008
First Round of Reviews         February 1, 2009
Publication Date               May 1, 2009

Guest Editors

o Juan I. Godino-Llorente, Department of Circuits and Systems Engineering, Polytechnic University of Madrid (UPM), Ctra. de Valencia, 28031 Madrid, Spain; igodino@ics.upm.es
o Pedro Gómez-Vilda, Department of Computer Science and Engineering, Polytechnic University of Madrid (UPM), Boadilla del Monte, 28660 Madrid, Spain; pedro@pino.datsi.fi.upm.es
o Tan Lee, Department of Electronic Engineering, The Chinese University of Hong Kong, Shatin, New Territories, Hong Kong; tanlee@ee.cuhk.edu.hk 

Back to Top

7-9 . Special issue of CSL on Emergent Artificial Intelligence Approaches for Pattern Recognition in Speech and Language Processing

Special Issue on "Emergent Artificial Intelligence Approaches for Pattern Recognition in Speech and Language Processing"
      Computer Speech and Language, Elsevier       Deadline for paper submission: September 26, 2008.  http://speechlab.ifsc.usp.br/call/csl.pdf                        =
Back to Top

7-10 . CfP Special issue IEEE Trans. ASL Signal models and representation of musical and environmental sounds

Special Issue of IEEE Transactions on Audio, Speech and Language Processing **SIGNAL MODELS AND REPRESENTATION OF MUSICAL AND ENVIRONMENTAL SOUNDS**  http://www.ewh.ieee.org/soc/sps/tap http://www.ewh.ieee.org/soc/sps/tap/sp_issue/audioCFP.pdf * *-- Submission deadline: 15 December, 2008 --  Notification of acceptance: 15 June, 2009 Final manuscript due: 1st July, 2009 Tentative publication date: 1st September, 2009   Guest editors Dr. Bertrand David (Telecom ParisTech, France)  bertrand.david@telecom-paristech.fr Dr. Laurent Daudet (UPMC University Paris 06, France) daudet@lam.jussieu.fr Dr. Masataka Goto (National Institute of Advanced Industrial Science and  Technology, Japan) m.goto@aist.go.jp Dr. Paris Smaragdis (Adobe Systems, Inc, USA) paris@adobe.com ------------------------------------------------------------------------------    The non-stationary nature, the richness of the spectra and the mixing of  diverse sources are common characteristics shared by musical and  environmental audio scenes. It leads to specific challenges of audio  processing tasks such as information retrieval, source separation,  analysis-transformation-synthesis and coding. When seeking to extract  information from musical or environmental audio signals, the  time-varying waveform or spectrum are often further analysed and  decomposed into sound elements. Two aims of this decomposition can be  identified, which are sometimes antagonist: to be together adapted to  the particular properties of the signal and to the targeted application.  This special issue is focused on how the choices of a low level  representation (typically a time-frequency distribution with or without  probabilistic framework, with or without perceptual considerations), a  source model or a decomposition technique may influence the overall  performance.  Specific topics of interest include but are not limited to: * factorizations of time-frequency distribution * sparse representations * Bayesian frameworks * parametric modeling * subspace-based methods for audio signals * representations based on instrument or/and environmental sources  signal models * sinusoidal modeling of non-stationary spectra (sinusoids, noise,  transients)  Typical applications considered are (non exclusively): * source separation/recognition * mid or high level features extraction (metrics, onsets, pitches, …) * sound effects * audio coding * information retrieval * audio scene structuring, analysis or segmentation * ...  B. David, L.Daudet, P. Smaragdis, M. Goto.
Back to Top

7-11 . "Speech Communication" special issue on "Speech and Face to Face Communication

 "Speech Communication" special issue on "Speech and Face to Face Communication 
http://www.elsevier.com/wps/find/journaldescription.cws_home/505597/description

Speech communication is increasingly studied in a face to face perspective:
- It is interactive: the speaking partners build a complex communicative act together
involving linguistic, emotional, expressive, and more generally cognitive and social
dimensions;
- It involves multimodality to a large extent: the “listener” sees and hears the speaker who
produces sounds as well as facial and more generally bodily gestures;
- It involves not only linguistic but also psychological, affective and social aspects of
interaction. Gaze together with speech contribute to maintain mutual attention and to
regulate turn-taking for example. Moreover the true challenge of speech communication is
to take into account and integrate information not only from the speaker but also from the
entire physical environment in which the interaction takes place.

The present issue proposes to synthetize the most recent developments in
this topic considering its various aspects from complementary perspectives: cognitive and
neurocognitive (multisensory and perceptuo-motor interactions), linguistic (dialogic face to
face interactions), paralinguistic (emotions and affects, turn-taking, mutual attention),
computational (animated conversational agents, multimodal interacting communication
systems).

There will be two stages in the submission procedure.

- First stage (by DECEMBER 1ST): submission of a one-to-two page abstract describing the
contents of the work and its relevance to the "Speech and Face to Face Communication" topic
by DECEMBER 1ST. The guest editors will then make a selection of the most relevant
proposals in December.

- Second stage (by MARCH 1ST): the selected contributors will be invited to submit a full
paper by MARCH 1ST. The submitted papers will then be peer reviewed through the regular
Speech Communication journal process (two independent reviews). Accepted papers will then
be published in the special issue.

Abstracts should be directly sent to the guest editors:
Marion.Dohen@gipsa-lab.inpg.fr, Gerard.Bailly@gipsa-lab.inpg.fr, Jean-Luc.Schwartz@gipsa-lab.inpg.fr

Back to Top

7-12 . SPECIAL ISSUE of the EURASIP Journal on Audio, Speech, and Music Processing. ON SCALABLE AUDIO-CONTENT ANALYSIS

SPECIAL ISSUE ON SCALABLE AUDIO-CONTENT ANALYSIS

The amount of easily-accessible audio, whether in the form of large
collections of audio or audio-video recordings, or in the form of
streaming media, has increased exponentially in recent times.
However this audio is not standardized: much of it is noisy,
recordings are frequently not clean, and most of it is not labelled.
The audio content covers a large range of categories including
sports, music and songs, speech, and natural sounds. There is
therefore a need for algorithms that allow us make sense of these
data, to store, process, categorize, summarize, identify and
retrieve them quickly and accurately.

In this special issue we invite papers that present novel approaches
to problems such as (but not limited to):

Audio similarity
Audio categorization
Audio classification
Indexing and retrieval
Semantic tagging
Audio event detection
Summarization
Mining

We are especially interested in work that addresses real-world
issues such as:

Scalable and efficient algorithms
Audio analysis under noisy and real-world conditions
Classification with uncertain labeling
Invariance to recording conditions
On-line and real-time analysis of audio.
Algorithms for very large audio databases.

We encourage theoretical or application-oriented papers that
highlight exploitation of such techniques in practical systems/products.

Authors should follow the EURASIP Journal on Audio, Speech, and Music
Processing manuscript format described at the journal site
http://www.hindawi.com/journals/asmp/. Prospective authors should
submit an electronic copy of their complete manuscript through the
journal Manuscript Tracking System at http://mts.hindawi.com/,
according to the following timetable:

Manuscript Due: June 1st, 2009
First Round of Reviews: September 1, 2009
Publication Date: December 1st, 2009


Guest Editors:

1) Bhiksha Raj
Associate professor


School of computer science
Carnegie Mellon university


2) Paris Smaragdis
Senior Research Scientist
Advanced Technology Labs, Adobe Systems Inc.
Newton, MA, USA

3) Malcolm Slaney
Principal Scientist
Yahoo! Research
Santa Clara, CA
and
(Consulting) Professor
Stanford CCRMA

4) Chung-Hsien Wu
Distinguished Professor
Dept. of Computer Science & Infomation Engineering
National Cheng Kung University,
Tainan, TAIWAN

5) Liming Chen
Professor and head of the Dept. Mathematics & Informatics
Ecole Centrale de Lyon
University of Lyon
Lyon, France

6) Professor Hyoung-Gook Kim
Intelligent Multimedia Signal Processing Lab.
Kwangwoon University, Republic of Korea 

Back to Top

7-13 . Special issue of the EURASIP Journal on Audio, Speech, and Music Processing.on Atypical Speech

Atypical Speech
Call for Papers

Research in speech processing (e.g., speech coding, speech enhancement, speech recognition, speaker recognition, etc.) tends to concentrate on speech samples collected from normal adult talkers. Focusing only on these “typical speakers” limits the practical applications of automatic speech processing significantly. For instance, a spoken dialogue system should be able to understand any user, even if he or she is under stress or belongs to the elderly population. While there is some research effort in language and gender issues, there remains a critical need for exploring issues related to “atypical speech”. We broadly define atypical speech as speech from speakers with disabilities, children's speech, speech from the elderly, speech with emotional content, speech in a musical context, and speech recorded through unique, nontraditional transducers. The focus of the issue is on voice quality issues rather than unusual talking styles.

In this call for papers, we aim to concentrate on issues related to processing of atypical speech, issues that are commonly ignored by the mainstream speech processing research. In particular, we solicit original, previously unpublished research on:
• Identification of vocal effort, stress, and emotion in speech
• Identification and classification of speech and voice disorders
• Effects of ill health on speech
• Enhancement of disordered speech
• Processing of children's speech
• Processing of speech from elderly speakers
• Song and singer identification
• Whispered, screamed, and masked speech
• Novel transduction mechanisms for speech processing
• Computer-based diagnostic and training systems for speech dysfunctions
• Practical applications

Authors should follow the EURASIP Journal on Audio, Speech, and Music Processing manuscript format described at the journal site

http://www.hindawi.com/journals/asmp/. Prospective authors should submit an electronic copy of their complete manuscript through the journal Manuscript Tracking System at

http://mts.hindawi.com/, according to the following timetable:
Manuscript Due
April 1, 2009
First Round of Reviews
July 1, 2009
Publication Date
October 1, 2009

Guest Editors

Georg Stemmer, Siemens AG, Corporate Technology, 80333 Munich, Germany

Elmar Nöth, Department of Pattern Recognition, Friedrich-Alexander University of Erlangen-Nuremberg, 91058 Erlangen, Germany

Vijay Parsa, National Centre for Audiology, The University of Western Ontario, London, ON, Canada N6G 1H1 

Back to Top

7-14 . Special issue of the EURASIP Journal on Audio, Speech, and Music Processing on Animating virtual speakers or singers from audio: lip-synching facial animation

 

Special issue on 
Animating virtual speakers or singers from audio: lip-synching facial animation 

Call for PapersLip synchronization (lip-synch) is the term used to describe matching lip movements to a pre-recorded speaking or singing voice. This often is used in the production of films, cartoons, television programs, and computer games. 
We focus here on technologies that are able to compute automatically the facial movements of animated characters given pre-recorded audio. Automating the lip-synch process, generally termed visual speech synthesis, has potential for use in a wide range of applications: from desktop agents on personal computers, to language translation tools, to providing a means for generating and displaying stimuli in speech perception experiments.A visual speech synthesizer comprises at least three modules: a control model that computes articulatory trajectories from the input signal; a shape model that animates the facial geometry from computed trajectories and an appearance model for rendering the animation by varying the colors of pixels. There are numerous solutions proposed in the literature for each of these modules. Control models exploit either direct signal-to-articulation mappings, or more complex trajectory formation systems that utilize a phonetic segmentation of the acoustic signal. Shape models vary from ad-hoc parametric deformations of a 2D mesh to sophisticated 3D biomechanical models. Appearance models exploit morphing of natural images, texture blending or more sophisticated texture models.The aim of this special issue is to provide a detailed description of state-of-the-art systems and identify new techniques that have recently emerged from both the audiovisual speech and computer graphics research communities. 
In particular, we solicit original, previously unpublished research on: 

 

Audiovisual synthesis from text

Facial animation from audio

Trajectory formation systems

Evaluation methods for audiovisual synthesis

Perception of audiovisual asynchrony in speech and music

Control of speech and facial expressions

 

 

This special issue follows the first visual speech synthesis challenge (LIPS’2008) that took place as a special session at INTERSPEECH 2008 in Brisbane, Australia. The aim of the challenge was to stimulate discussion about the subjective quality assessment of synthesized visual speech, with a view to developing standardized evaluation procedures.For this special issue, all papers selected for publication should include a description of a subjective evaluation experiment that outlines the impact of the proposed synthesis scheme on some subjective measure, such as audiovisual intelligibility, cognitive load or perceived naturalness. This evaluation metric could be assessed either by participation in the LIPS’2008 challenge, or by an independent perceptual experiment.Technical organization 
The issue is coordinated by three guest editors: G. Bailly, B.-J. Theobald & S. Fagel. These editors co-organized the LIPS’2008 challenge, and they cover a large spectrum of scientific backgrounds coherent with the theme: audiovisual speech processing, facial animation & computer graphics. They are assisted by a scientific committee. The members of the scientific committee are also invited to submit papers, and promote papers by helping in the communication process around the issue.The special issue will be introduced by a paper written by the editors, with a critical review of the selected papers and with a discussion of the results obtained by the systems participating to the LIPS’2008 challenge. 
ScheduleAuthors should follow the EURASIP Journal on Audio, Speech, and Music Processing manuscript format described at the journal site http://www.hindawi.com/journals/asmp/. Prospective authors should submit an electronic copy of their complete manuscript through the journal Manuscript Tracking System at http://mts.hindawi.com/, according to the following timetable: 

 


One page abstractJanuary 1, 2009
Preselection of papersFebruary 1, 2009
Manuscript dueMarch 1, 2009
First round of reviewsMay 1, 2009
Camera-ready papersJuly 1, 2009
Publication dateSeptember 1,2009

Guest Editors

 

Gérard Bailly, GIPSA-Lab, Speech & Cognition Dept., Grenoble-France; 

gerard.bailly@gipsa-lab.inpg.fr 

Sascha Fagel, Speech & Communication Institute, TU Berlin, Germany; 
sascha.fagel@tu-berlin.de 

Barry-John Theobald, School of Computing Sciences, University of East Anglia, UK; 
b.theobald@uea.ac.uk 

 

Back to Top

8 . Future Speech Science and Technology Events

8-1 . (2009-01-14) Language Technology Days Luxemburg

We would like to invite you to participate in the Language Technology Days organised by Unit E1 - Language Technologies and Machine Translation, on 14-15 January 2009 in Luxembourg.

Our Unit has been established in July 2008 as part of Directorate E - Digital Content and Cognitive Systems of the Directorate-General Information Society and Media. In parallel, some 40 Meuro have been earmarked for multinational projects and networks to be established in 2009 under 

       the FP7-ICT research programme (http://cordis.europa.eu/fp7/ict/programme/home_en.html)
       The ICT Policy Support Programme (ICT-PSP), an innovation programme which supports actions aimed at fostering technology uptake and best practices  (http://ec.europa.eu/information_society/activities/ict_psp/about/index_en.htm)

The fourth FP7-ICT call for proposals, due to be launched within a few days, features Objective 2.2: Language-based interaction. 

The third ICT-PSP call is expected to be published in the first quarter of 2009, and will include Theme 5: Multilingual Web.

The respective work programmes and calls are in the process of being adopted. They will be presented at the ICT 2008 conference and exhibition held in Lyon on 25-27 November 2008 (http://ec.europa.eu/information_society/events/ict/2008/index_en.htm). Our session is scheduled for 26 November 2008 at 16:00.

Further information is or will soon become available on

FP7-ICT: http://cordis.europa.eu/fp7/ict/participating/calls_en.html

ICT-PSP: http://ec.europa.eu/information_society/activities/ict_psp/participating/calls/index_en.htm

Our Website (about to be launched): http://cordis.europa.eu/fp7/ict/language-technologies/fp7-call4_en.htm

The LT Days will take place in the conference complex of the Jean Monnet building in Luxembourg. Participation is free of charge but subject to prior registration and confirmation.

The particularities of each call will be described in detail and there will be Q&A opportunities after each presentation. We are also offering the chance to meet with EC project officers and discuss your ideas. The skills marketplace is a chance for you to present your organisation and perhaps find the ideal partner for your proposal.

Preliminary agenda:

14 January, 2009        LT and Innovation
12:30 - 14:00           Registration
14:00 - 15:30           Presentation of the Multilingual Web theme within the ICT-PSP call
16:00 - 18:00           Themed session I with the participation of high-level speakers:
Broadening the coverage. How to ensure a more balanced coverage of EU official languages, how to provide each language with an adequate starter kit, how to reduce fragmentation and mobilize or coordinate national efforts. Methods, tools and resources aimed at making LT more scalable and portable across languages, domains and tasks.

In parallel:
16:00 18:00           Proposals clinic; meet EC project officers and discuss your ideas

15 January, 2009        LT and Research
08:30 - 09:15           Registration
09:15 - 10:45           Presentation of the Language-based Interaction objective within the FP7-ICT call
11:00 - 13:00           Themed session II with the participation of high-level speakers:
Machine Translation and multilingual LT: Progress and prospects. Presentations will address topics such as Research avenues, towards a paradigm shift? Global communication and content management; Web, translation and LT: business drivers and success factors; LT within and beyond Web 2.0.

13:00 - 14:15           Lunch break & networking
14:15 - 15:30           Skills marketplace; present your organisation, skills and ideas
In parallel:
14:15 - 15:30           Proposals clinic; meet EC project officers and discuss your ideas
15:30 16:00           Wrap-up session: Two programmes, one broad vision. Interactive session with speakers and delegates

Participation is subject to prior registration and confirmation. Delegates will have to pay for their own travel and accommodation. The number of participants is limited due to space restrictions. With this in mind, and to help us ensure that the widest possible range of participants is represented, organisations should send only one delegate whenever possible. Please therefore do not definitely book your travel arrangements before your registration has been confirmed.

You are invited to register to the event by e-mail to our functional mailbox. The registration form is attached.

Please register by 19 December 2008 at the latest. For further information please contact: Mrs Susan Fraser (susan.fraser@ec.europa.eu)

Looking forward to seeing soon in Luxembourg,

Roberto Cencioni
Head of unit
DG Information Society and Media
Unit INFSO.E1 - Language technologies and Machine translation

Please fill the registration form below and return it to infso-e1@ec.europa.eu with the mention "LT Days" in the subject field.

Inquiries regarding the imminent FP7-ICT call can be submitted via the above mailbox.
Guidance regarding FP7-ICT pre-proposals and the ICT-PSP call will be provided on the upcoming unit's website.

----------------------------------------------------------------------------------------------------------------------

Registration Form

[ ]  Tick the box to confirm that you have read and agreed to the Privacy Statement

[ ]  Tick the box if you do NOT wish your personal information to be published

Personal Information
Title (Mr/Mrs):
Last Name:     
First Name:    
Organisation:  
Department:    
Country of Organisation:
Email address:
You may choose to attend one or both of the days. 
Please tick your preferences:
14 January 2009
[ ] Multilingual Web within the ICT-PSP Programme

15 January 2009

[ ] Language based Interaction within the FP7-ICT Programme

Breakout sessions

[ ] Skills marketplace  your submission (one MS-PowerPoint foil) must reach us no later than 6 Jan 2009
[ ] Proposals clinic    your submission (up to three MS-Word pages) must reach us no later than 6 Jan 2009
Other information
[ ] I wish to bring a Poster

Have you previously participated in EU funded (IST, ICT, eContent+) projects?   [ ] Yes [ ] No 
Category of delegate:
[ ] Researcher          [ ] Developer/Vendor    [ ] User        [ ] EC or Gov official  [ ] Other (specify):

----------------------------------------------------------------------------------------------------------------------
Privacy Statement
1. Registration relating to the meeting Language Technology Information Days, 14-15 January 2009
While registering your personal data will be collected and further processed for the purposes detailed hereafter under point 2. The relevant processing operations thereof and the meeting are under the responsibility of the Head of the Unit “Language technologies, machine translation”, Directorate “Digital Content & Cognitive Systems”, Directorate-General Information Society and Media, acting as the Controller.

As this service collects and further processes personal data, Regulation (EC) 45/2001, of the European Parliament and of the Council of 18 December 2000 on the protection of individuals with regard to the processing of personal data by the Community institutions and bodies and on the free movement of such data, is applicable.

2. What personal information do we collect, for what purpose and through which technical means?
Identification Data
The personal data collected and further processed are:
data necessary for the organisation and management of the meeting, such as Gender (needed for the right title)/name/surname/country/postal & e-mail addresses/phone number/fax number... 

The purpose of the processing of personal data performed in the context of the meeting is the management/organisation of the meeting, including management of: lists for contacts, invitations, participants, distribution of reports, feedback on reports, follow-on actions and meetings, and publication on EUROPA and/or CORDIS websites.

The processing operations on personal data linked to the organisation and management of this meeting are necessary for the management and functioning of the Commission, as mandated by the treaties, and more specifically articles 5, 7 and 211 - 219 of the EC Treaty.

Technical information
3. Who has access to your information and to whom is it disclosed?
Some of your personal data (e.g. name, organisation, email address) will be published on the EC website(s).  If you do not agree with this publication of your personal data, please opt-out where required or inform the data controller by using the Contact Information below and by explicitly specifying your request).

The access to all personal data as well as all information collected in the context of this meeting, and the organisation thereof, is only granted through User Id/Password to a defined population of users, without prejudice to a possible transmission to the bodies in charge of a monitoring or inspection task in accordance with Community legislation. These users typically are: the chairman and the secretary of the meeting, both supported by a team in the operational services inside DG Information Society and Media, acting as processor(s).

In order to keep the network of the participants operational a contact list of participants is shared between attendees. The report including the presence list is shared between participants and distributed to all organisations represented in the meeting.

As reports of these kind of official meetings are Commission documents, they fall under the scope of the Regulation (EC) 1049/2001 of the European Parliament and of the Council of 30 May 2001 regarding public access to European Parliament, Council and Commission documents.

No personal data is transmitted to parties which are outside the recipients and the legal framework mentioned.
4. How do we protect and safeguard your information?
The collected personal data and all information related to the above mentioned meeting are stored on the servers of DG Information Society and Media, the operations of which abide by the Commission's security decisions and provisions established by the Directorate of Security for this kind of servers and services.

Access to all collected personal data and all information related to the above mentioned meeting is only possible to the above described populations with a User Id/Password.

5. How can you verify, modify or delete your information?
In case you want to verify which personal data is stored on your behalf by the responsible controller, have it modified respectively corrected or deleted, please contact the Controller by using the Contact Information below and by explicitly specifying your request.

Special attention is drawn to the consequences of a request for deletion, in which case any trace to be able to contact you will be lost.

6. How long do we keep your data?
Your personal data will be part of a list of contact details shared internally amongst the staff of Directorate General Information Society & Media for the purpose of contacting you in the future in the context of the Directorate General's activities. If you do not agree with this, please contact the Controller by using the Contact Information below and by explicitly specifying your request.

7. Contact Information
In case you wish to verify which personal data is stored on your behalf by the responsible controller, have it modified, corrected, or deleted, or if you have questions regarding the organisation of the meeting, or concerning any information processed in the context of the meeting, or on your rights, feel free to contact the support team, operating under the responsibility of the Controller, using the following contact information:

Mrs Susan Fraser
Communication Officer
DG Information Society and Media
Unit INFSO.E1 - Language technologies and Machine translation
Tel.: +352-4301 33721 / (Fax: +352-4301 32100)
E-mail address : susan.fraser@ec.europa.eu

8. Recourse
Complaints, in case of conflict, can be addressed to the European Data Protection Supervisor. 

Back to Top

8-2 . (2009-01-22) Conferences au GIPSA (Grenoble France) (in french)

 Jeudi 22 janvier 2009, 13h30 – Séminaire externe
========================================
Rachid RIDOUANE
Laboratoire de Phonétique et Phonologie, Paris

Titre à préciser

Résumé à préciser

Salle de réunion du Département Parole et Cognition (B314)
3ème étage Bâtiment B ENSE3
961 rue de la Houille Blanche
Domaine Universitaire

Jeudi 5 février 2009, 13h30 – Séminaire externe
========================================
Daniele SCHON
Institut de Neurosciences Cognitives de la Méditerranée, Marseille

Titre à préciser

Résumé à préciser

Salle de réunion du Département Parole et Cognition (B314)
3ème étage Bâtiment B ENSE3
961 rue de la Houille Blanche
Domaine Universitaire

Jeudi 12 février 2009, 13h30 – Séminaire externe
========================================
Thierry NAZZI
Laboratoire Psychologie de la Perception
Paris

Titre à préciser

Résumé à préciser

Salle de réunion du Département Parole et Cognition (B314)
3ème étage Bâtiment B ENSE3
961 rue de la Houille Blanche
Domaine Universitaire

--  _______________________________________________________________________   Please note our NEW ADDRESS:  Helene Loevenbruck, CR CNRS     | Helene.Loevenbruck@gipsa-lab.inpg.fr  Dept Parole & Cognition (ex ICP)|  GIPSA-lab                       | Tel : (33) 4 76 57 47 14  961 rue de la Houille Blanche   | Fax : (33) 4 76 82 63 84  Domaine Universitaire Grenoble  |  F38402 St Martin d'Hères cedex  | http://www.gipsa-lab.inpg.fr/
Back to Top

8-3 . (2009-02-08)CfP Workshop on Multimodal Interfaces for Automotive Applications (MIAA)

Workshop on
 
** Multimodal Interfaces for Automotive Applications (MIAA) **
 
In conjunction with IUI 2009, 
the International Conference on Intelligent User Interfaces
 
8th -- 11th of February 2009, Sanibel Island, Florida
 
organized by Christian Müller (DFKI) and Gerald Friedland (ICSI)
==================================================
 
 
 
Important dates: 
 
Paper submissions    Monday, November 27th 2008 11:59pm US PDT
Paper notification    Monday, December 19th 2009
Paper camera-ready due  Monday, January 12th 2009
IUI conference    February 8th -- February 11th 2009
 
 
Multimodal interaction constitutes a key technology for intelligent user interfaces (IUI). The possibility to control devices and applications in a natural way enables an easier access to complex functionality as well as infotainment contents. This kind of interaction is particularly suited for use in automotive scenarios where additional restrictions with respect to input and output modalities have to be taken into account. In recent years, the complexity of on-board and accessory devices, infotainment services, and driver assistance systems in cars has experienced an enormous increase. This development emphasizes the need for new concepts for advanced human-machine interfaces that support the seamless, intuitive and efficient use of this large variety of devices and services.
 
This workshop is intended to gather novel, innovative interaction concepts for automotive applications with the aim to foster collaborations in the field and to establish a IUI-wide consciousness for the specific user interface requirements in the area of car- centered applications.
 
The topics include but are not restricted to:
 
    * novel interfaces on any sort (e.g. see-trough displays)
    * speech in the car
    * tangible (haptic) interfaces, e.g. novel means of interaction with switches, knobs, levers, etc
    * multi-party interaction: there's more people in the car than only the driver
    * sensor networks (car2car, car2X) involving user interaction (i.e. information-seeking dialogs)
    * access to services
    * referring to the outside world
    * beyond directions
 
MIAA Career Day:
Bachelor and Master students are especially encouraged to submit their theses or student research projects. A special concern of the MIAA workshop is to serve as a career day. A limited number of travel grants and reduced registration fees will be given to high-class students contributing to the workshop. Also, representatives of the automotive industry will be invited to the workshop. Opportunities for Master and PHD programs will be presented and connections to industrial employers will be arranged.
 
Mode of Presentations:
To emphasize the innovative character of the MIAA workshop, the mode of presentation will be different. There will be one session with brief oral presentations of the contributions. Afterwards, focus groups will be formed to discuss individual topics of interest. The topics will be selected and the participants will be assigned to focus groups by the organizers beforehand. The pre-selection of topics will be announced on this website. However, new topics can be proposed during the workshop and everyone is free to join the focus group that she or he is interested in. A designated part of workshop will be used to discuss career opportunities. 
 
------------------------------------------------------------------
Dr. Christian Müller
DFKI GmbH, , Campus D3 2
Stuhlsatzenhausweg 3
D-66123 Saarbrücken, Germany
+49 (681) 302 -5269 (office) -5020 (fax)
------------------------------------------------------------------
Deutsches Forschungszentrum fuer Kuenstliche Intelligenz GmbH
Firmensitz: Trippstadter Strasse 122, D-67663 Kaiserslautern
Geschaeftsfuehrung:
Prof. Dr. Dr. h.c. mult. Wolfgang Wahlster (Vorsitzender)
Dr. Walter Olthoff
Vorsitzender des Aufsichtsrats: Prof. Dr. h.c. Hans A. Aukes
Amtsgericht Kaiserslautern, HRB 2313
-------------------------------------------------------------
Back to Top

8-4 . (2009-03-02) Voice Search Conference San Diego

Early discounted registration for the Voice Search Conference

 

Save $200 on registration for the Voice Search Conference, to be held in San Diego, March 2 - 4, 2009, by registering before October 15 at www.voicesearchconference.com.

Back to Top

8-5 . (2009-03-26) International Workshop on Pharyngeals & Pharyngealisation

International Workshop on Pharyngeals & Pharyngealisation Co-organised by the Centre for Research in Linguistics and Language Science (CRiLLS), Newcastle University and Praxiling Centre National de la Recherche Scientifique (CNRS), Université Montpellier III  Date: 26 March - 27 March 2009 Venue: Research Beehive, Newcastle University This is the first call for posters and for participants for this Pharyngeals and Pharyngealisation Workshop.  Background on Pharyngeals and Pharyngealisation Over the last 50 years, the study of pharyngeal and pharyngealised sounds has fascinated many a linguist due to their varied phonological representations, their complex articulation and co-articulation patterns, their late development in the process of language acquisition, and their sociolinguistic and crosslinguistic patterning. Approaches to the study of pharyngeal and pharyngealised sounds have included: a) phonological work looking at identifying consistent and uniform characteristics for this class of sounds b) experimental work looking at auditory, acoustic, and articulatory manifestations of pharyngeal constrictions c) acquisition studies looking at pharyngeals in babies’ early vocalisations regardless of language and later development in languages which contain these sounds d) socioliguistic work identifying cross-linguistic and cross-dialectal differences in the realisation of these sounds as well as the influence of social factors such gender, ethnicity, and social class. For more information please click here.  Aims of the Workshop The purpose of this workshop is to bring together researchers from around the world who have worked on pharyngeal and pharyngealised sounds in order to share expertise in different methodologies and theoretical approaches to the study of these sounds and attempt to answer various outstanding questions regarding:  Language universals: why are pharyngeal sounds present in only 1% of languages surveyed in the UPSID data base when they are present in children’s early vocalisations? How have these sounds evolved in languages that have lost the pharyngeal distinctions? Is there a relationship between a dense consonantal system and the existence of pharyngeal/pharyngealised sounds in a language? Production and perception: what are the acoustic, articulatory, and perceptual correlates of pharyngeal and pharyngealised sounds? What role do visual cues (e.g. lip rounding) play in processing pharyngeal articulations? The sociolinguistic indices of pharyngeal/pharyngealised articulations: How does pharyngealisation manifest itself in different languages/dialects? Is the gender-correlated patterning that has been documented in urban areas in the Arab world with respect to de-emphasis found in other varieties/languages with pharyngeal/pharyngealised articulations? How are pharyngeal articulations affected in language contact situations? Acquisition: at what age are pharyngeal and pharyngealised sounds acquired and what are the developmental manifestations across languages and/or dialects? Invited Speakers Dr John Esling, University of Victoria (Canada) Dr Zeki Hassan, University of Gottenbörg (Sweden) Dr Barry Heselwood, University of Leeds (UK) Prof Asher Laufer, The Hebrew University (Israel Prof Ian Maddieson, University of New Mexico (USA) Dr Slim Ouni, CNRS-Université de Nancy I (France) Dr Yves Laprie, CNRS-Université de Nancy I (France) Dr Rachid Ridouane, LPP CNRS-Université Paris III (France) Dr Kimary Shahin, Qatar University (Qatar) Prof Jim Scobbie, Queen Margaret University (UK) Prof Janet Watson, University of Salford (UK) Dr Chakir Zeroual, Faculté Polydisciplinaire de Taza (Morocco) Scientific Committee Jalal Al-Tamimi (Newcastle University) Enam Al Wer (Essex University) Thomas Baer (University of Cambridge) Jean-Francois Bonnot (Université de Besançon) Nick Clements (LPP CNRS-Université Paris III) Stuart Davis (Indiana University) Susanne Fuchs (Berlin University) John McCarthy (UMASS) Mohamed Embarki (Université Montpellier III) Kenneth de Jong (Indiana University) Ghada Khattab (Newcastle University) Amanda Miller (University of British Columbia) Daniel Recasens (University Autònoma of Barcelona) Harvey Sussman (University of Texas) Nathalie Vallée (Université Stendhal, Grenoble) Organisation of the Workshop The programme consists of 11 invited oral presentations and a poster session.  Workshop Themes We welcome work on pharyngeals and pharyngealisation in any of the following fields:  Universals Phonology Production and perception Modelling Acquisition Sociolinguistic variation To Apply to Present a Poster Abstracts on any of the workshop themes are invited for the poster session. Abstracts should be no longer than two pages including illustrations and references. Please submit your abstract to Crills@ncl.ac.uk by November 30, 2008. Abstracts will be reviewed by the Scientific Committee and applicants will be notified of their acceptance by December the 15th, 2008.  To Apply to Participate Places are available for participants who are not presenting papers but who will participate in discussions. Registration for this event will open in mid-December.  Important dates: Abstract Submission for the poster session: November 30, 2008 Notification of Acceptance: December 15, 2008 Workshop Dates: March 26-27, 2009 Registration: Early deadline (£80 for staff and £40 for students):  January 30, 2009 Late registration (£100 for staff and £50 for students):  January 30- March 25 Registration on the day: £120 for staff and £60 for students Registration for this event will open mid-December so please keep checking the website. Scholarships: Four scholarships of 125 Euros (approximately £100) each will be offered to students travelling from outside the UK and presenting at the poster session. In order to apply for one of the scholarships, please fill in the scholarship form and submit it together with your abstract by the 30th of November deadline. Travel For more information on travelling to and around Newcastle please visit link to www.ncl.ac.uk/travel/info where you can also download maps of the city centre and campus.  Accommodation Coming soon  Organization Committee Jalal Al-Tamimi, CRiLLS, Newcastle University (UK). Jalal.Al-Tamimi@ncl.ac.uk  Mohamed Embarki, Praxiling UMR 5267 CNRS-Montpellier III (France). mohamed.embarki@univ-montp3.fr  Ghada Khattab, CRiLLS, Newcastle University (UK). Ghada.Khattab@ncl.ac.uk  Hussain Kriba, CRiLLS, Newcastle University (UK). Hussin.Kriba@ncl.ac.uk
Back to Top

8-6 . (2009-04-02) CfP 3rd INT. CONF. ON LANGUAGE AND AUTOMATA THEORY AND APPLICATIONS (LATA 2009)

Call for Papers  3rd INTERNATIONAL CONFERENCE ON LANGUAGE AND AUTOMATA THEORY AND APPLICATIONS (LATA 2009)  Tarragona, Spain, April 2-8, 2009  http://grammars.grlmc.com/LATA2009/  *********************************************************************  AIMS:  LATA is a yearly conference in theoretical computer science and its applications. As linked to the International PhD School in Formal Languages and Applications that was developed at the host institute in the period 2002-2006, LATA 2009 will reserve significant room for young scholars at the beginning of their career. It will aim at attracting contributions from both classical theory fields and application areas (bioinformatics, systems biology, language technology, artificial intelligence, etc.).  SCOPE:  Topics of either theoretical or applied interest include, but are not limited to:  - algebraic language theory - algorithms on automata and words - automata and logic - automata for system analysis and programme verification - automata, concurrency and Petri nets - biomolecular nanotechnology - cellular automata - circuits and networks - combinatorics on words - computability - computational, descriptional, communication and parameterized complexity - data and image compression - decidability questions on words and languages - digital libraries - DNA and other models of bio-inspired computing - document engineering - extended automata - foundations of finite state technology - fuzzy and rough languages - grammars (Chomsky hierarchy, contextual, multidimensional, unification, categorial, etc.) - grammars and automata architectures - grammatical inference and algorithmic learning - graphs and graph transformation - language varieties and semigroups - language-based cryptography - language-theoretic foundations of natural language processing, artificial intelligence and artificial life - mathematical evolutionary genomics - parsing - patterns and codes - power series - quantum, chemical and optical computing - regulated rewriting - string and combinatorial issues in computational biology and bioinformatics - symbolic dynamics - symbolic neural networks - term rewriting - text algorithms - text retrieval, pattern matching and pattern recognition - transducers - trees, tree languages and tree machines - weighted machines  STRUCTURE:  LATA 2009 will consist of:  - 3 invited talks (to be announced in the second call for papers) - 2 invited tutorials (to be announced in the second call for papers) - refereed contributions - open sessions for discussion in specific subfields or on professional issues (if requested by the participants)  PROGRAMME COMMITTEE:  Parosh Abdulla (Uppsala) Stefania Bandini (Milano) Stephen Bloom (Hoboken) John Brzozowski (Waterloo) Maxime Crochemore (London) Juergen Dassow (Magdeburg) Michael Domaratzki (Winnipeg) Henning Fernau (Trier) Rusins Freivalds (Riga) Vesa Halava (Turku) Juraj Hromkovic (Zurich) Lucian Ilie (London, Canada) Kazuo Iwama (Kyoto) Aravind Joshi (Philadelphia) Juhani Karhumaki (Turku) Jarkko Kari (Turku) Claude Kirchner (Bordeaux) Maciej Koutny (Newcastle) Kamala Krithivasan (Chennai) Martin Kutrib (Giessen) Andrzej Lingas (Lund) Aldo de Luca (Napoli) Rupak Majumdar (Los Angeles) Carlos Martin-Vide (Tarragona & Brussels, chair) Joachim Niehren (Villeneuve d'Ascq) Antonio Restivo (Palermo) Joerg Rothe (Duesseldorf) Wojciech Rytter (Warsaw) Philippe Schnoebelen (Cachan) Thomas Schwentick (Dortmund) Helmut Seidl (Muenchen) Alan Selman (Buffalo) Jeffrey Shallit (Waterloo) Frank Stephan (Singapore)  ORGANIZING COMMITTEE:  Madalina Barbaiani Gemma Bel-Enguix Cristina Bibire Adrian-Horia Dediu Szilard-Zsolt Fazekas Alexander Krassovitskiy Guangwu Liu Carlos Martin-Vide (chair) Robert Mercas Catalin-Ionut Tirnauca Bianca Truthe Sherzod Turaev Florentina-Lilica Voicu  SUBMISSIONS:  Authors are invited to submit papers presenting original and unpublished research. Papers should not exceed 12 single-spaced pages and should be formatted according to the standard format for Springer Verlag's LNCS series (see http://www.springer.com/computer/lncs/lncs+authors?SGWID=0-40209-0-0-0). Submissions have to be uploaded at:  http://www.easychair.org/conferences/?conf=lata2009  PUBLICATION:  A volume of proceedings published by Springer in the LNCS series will be available by the time of the conference. A refereed volume of extended versions of selected papers will be published after it as a special issue of a major journal. (This was Information and Computation for LATA 2007 and LATA 2008.)  REGISTRATION:  The period for registration will be open since September 1, 2008 to April 2, 2009. The registration form can be found at the website of the conference: http://grammars.grlmc.com/LATA2009/  Early registration fees: 450 euros Early registration fees (PhD students): 225 euros Registration fees: 540 euros Registration fees (PhD students): 270 euros  At least one author per paper should register. Papers that do not have a registered author by December 31, 2008 will be excluded from the proceedings.  Fees comprise free access to all sessions, one copy of the proceedings volume, and coffee breaks. For the participation in the full-day excursion and conference lunch on Sunday April 5, the amount of 70 euros is to be added to the fees above: accompanying persons are welcome at the same rate.  PAYMENT:  Early registration fees must be paid by bank transfer before December 31, 2008 to the conference account at Open Bank (Plaza Manuel Gomez Moreno 2, 28020 Madrid, Spain): IBAN: ES1300730100510403506598 - Swift code: OPENESMMXXX (account holder: LATA 2009 – Carlos Martin-Vide).  (Non-early) registration fees can be paid either by bank transfer to the same account or in cash on site.  Besides paying the registration fees, it is required to fill in the registration form at the website of the conference. A receipt for the payment will be provided on site.  FUNDING:  Up to 20 grants covering partial-board accommodation will be available for nonlocal PhD students. To apply, candidates must e-mail their CV together with a copy of the document proving their present status as a PhD student.  IMPORTANT DATES:  Paper submission: October 22, 2008 Notification of paper acceptance or rejection: December 10, 2008 Application for funding (PhD students): December 15, 2008 Notification of funding acceptance or rejection: December 19, 2008 Final version of the paper for the proceedings: December 24, 2008 Early registration: December 31, 2008 Starting of the conference: April 2, 2009 Submission to the journal special issue: June 22, 2009  FURTHER INFORMATION:  carlos.martin@urv.cat  ADDRESS:  LATA 2009 Research Group on Mathematical Linguistics Rovira i Virgili University Plaza Imperial Tarraco, 1 43005 Tarragona, Spain Phone: +34-977-559543 Fax: +34-977-559597
Back to Top

8-7 . (2009-04-19) ICASSP 2009 Taipei, Taiwan

IEEE International Conference on Acoustics, Speech, and Signal Processing

http://icassp09.com

Sponsored by IEEE Signal Processing Society

April 19 - 24, 2009

Taipei International Convention Center

Taipei, Taiwan, R.O.C.

 

The 34th International Conference on Acoustics, Speech, and Signal Processing (ICASSP) will be held at the Taipei International Convention Center in Taipei, Taiwan, April 19 - 24, 2009. The ICASSP meeting is the world’s largest and most comprehensive technical conference focused on signal processing and its applications. The conference will feature world-class speakers, tutorials, exhibits, and over 50 lecture and poster sessions on:

 

Audio and electroacoustics

 

Bio imaging and signal processing

 

Design and implementation of signal processing systems

 

Image and multidimensional signal processing

 

Industry technology tracks

 

Information forensics and security

 

Machine learning for signal processing

 

Multimedia signal processing

 

Sensor array and multichannel systems

 

Signal processing education

 

Signal processing for communications

 

Signal processing theory and methods

 

Speech and language processing

 

Taiwan: The Ideal Travel Destination. Taiwan, also referred to as Formosa – the Portuguese word for "graceful" – is situated on the western edge of the Pacific Ocean off the southeastern coast of mainland Asia, across the Taiwan Strait from Mainland China. To the north lie Okinawa and the main islands of Japan, and to the south is the Philippines. ICASSP 2009 will be held in Taipei, a city that blends traditional culture and cosmopolitan life. As the political, economic, educational, and recreational center of Taiwan, Taipei offers a dazzling array of cultural sights not seen elsewhere, including exquisite food from every corner of China and the world. You and your entire family will be able to fully experience and enjoy this unique city and island. Prepare yourself for the trip of your dreams, as Taiwan has it all: fantastic food, a beautiful ocean, stupendous mountains and lots of sunshine!

 

Submission of Papers: Prospective authors are invited to submit full-length, four-page papers, including figures and references, to the ICASSP Technical Committee. All ICASSP papers will be handled and reviewed electronically. The ICASSP 2009 website www.icassp09.com will provide you with further details. Please note that the submission dates for papers are strict deadlines.

 

Tutorial and Special Session Proposals: Tutorials will be held on April 19 and 20, 2009. Brief proposals should be submitted by August 4, 2008, to Tsuhan Chen at tutorials@icassp09.com and must include title, outline, contact information, biography and selected publications for the presenter, a description of the tutorial, and material to be distributed to participants. Special sessions proposals should be submitted by August 4, 2008, to Shih-Fu Chang at specialsessions@icassp09.com and must include a topical title, rationale, session outline, contact information, and a list of invited speakers. Tutorial and special session authors are referred to the ICASSP website for additional information regarding submissions.

 

Important Dates

Tutorial Proposals Due

August 4, 2008

Special Session Proposals Due

August 4, 2008

Notification of Special Session & Tutorial Acceptance

September 8, 2008

Submission of Regular Papers

September 29, 2008

Notification of Acceptance (by email)

December 15, 2008

Author’s Registration Deadline

February 2, 2009

 

 

 

Organizing Committee

 

 

General Chair

Lin-shan, Lee

National Taiwan University

 

General Vice-Chair

Iee-Ray Wei

Chunghwa Telecom Co.,Ltd.

 

Secretaries General

Tsungnan Lin

National Taiwan University

Fu-Hao Hsing

Chunghwa Telecom Co.,Ltd

 

Technical Program Chairs

Liang-Gee Chen

National Taiwan University

James R. Glass

Massachusetts Institute of Technology

 

Technical Program Members

Petar Djuric

Stony Brook University

Joern Ostermann

Leibniz University Hannover

Yoshinori Sagisaka

Waseda University

 

Plenary Sessions

Soo-Chang Pei (Chair)

National Taiwan University

Hermann Ney (Co-chair)

RWTH Aachen

 

Special Sessions

Shih-Fu Chang (Chair)

Columbia University

Lee Swindlehurst (Co-chair)

University of California, Irvine

 

Tutorial Chair

Tsuhan Chen

Carnegie Mellon University

 

Publications Chair

Homer Chen

National Taiwan University

 

Publicity Chair

Chin-Teng Lin

National Chiao Tung University

 

Finance Chair

Hsuan-Jung Su

National Taiwan University

 

Local Arrangements Chairs

Tzu-Han Huang

Chunghwa Telecom Co.,Ltd.

Chong-Yung Chi

National Tsing Hwa University

Jen-Tzung Chien

National Cheng Kung University

 

Conference Management

Conference Management Services

Back to Top

8-8 . (2009-05-18) 3rd Advanced Voice Function Assessment International Workshop (AVFA2009)

3rd Advanced Voice Function Assessment International Workshop (AVFA2009)

Madrid (Spain), 18th - 20th May 2009

http://www.avfa09.upm.es

     This is the first Call for Papers and Posters for the 3rd Advanced Voice Function Assessment International Workshop (AVFA2009) that will be held from May 18th to 20th at the Universidad Politécncia de Madrid, Spain.

Motivation

    Speech is the most important means of communication among humans, resulting from a complex interaction among vocal folds vibration at the larynx and voluntary movements of the articulators (i.e., mouth, tongue, velum, jaw, etc.). The function of voice, however, is not limited to speech communication. It also transfers emotions, expresses personality features and reflects situations of stress or pathology. Moreover, it has an aesthetic value in many different professional activities, affecting salesmen, managers, lawyers, singers, actors, etc.

     Although research in speech science has traditionally favoured areas such as synthesis, recognition or speaker verification, the previous facts motivate the current emerging of a new research area related to voice function assessment.

     AVFA2009 aims at fostering interdisciplinary collaboration and interactions among researchers in voice assessment beyond the framework of COST Action 2103, thus reaching the whole scientific community.

Topics

     Topics of interest include, but are not limited to: 

  • Automatic detection of voice disorders
  • Automatic assessment & rating of voice quality
  • New strategies for parameterization and modelling normal and pathological voices (biomechanical-based parameters, chaos modelling, etc.)
  • Databases of vocal disorders
  • Inverse filtering
  • Signal processing for remote diagnosis
  • Speech enhancement for pathological & oesophageal voices
  • Objective parameters extraction from vocal fold images using videolaryngoscopy, videokymography, fMRI and other emerging techniques
  • Multi-modal analysis of disordered speech
  • Robust pitch extraction algorithms for pathological & oesophageal voices
  • Emotions in speech
  • Speaker adaptation
  • Voice Physiology and Biomechanics
  • Modelling of Voice Production
  • Diagnosis and Evaluation Protocols
  • Substitution Voices
  • Evaluation of Clinical Treatments
  • Analysis of Oesophageal Voices

Submission

    Prospective authors are asked to electronically submit preliminary version of full papers with a maximum length of 4 pages, including figures and tables, in English. Preliminary papers should be submitted as pdf documents, fitted to the linked  templateby the 15th of January. The submitted documents should include the title and authors' names, affiliations and addresses. In addition, the e-mail address and phone number of the corresponding author should be given. 

    Workshop proceedings will be edited both in paper and CD-ROM. Author registration to the conference is required for accepted papers to be included in the proceedings. The best papers presented at the workshop will be eligible for publication in a referred journal.

Best student paper award

Based on the comments given by the reviewers and the presentation at the conference, the organizing committee will give a best student paper award. The awarded author will be nominated at the closing ceremony of AVFA2009.

Schedule

·        Proposal due 15th January 2009

·        Notification of acceptance 15th February 2009

·        Final papers due 28th February 2009

·        Preliminary program 1st May 2009

·        Workshop 18th May – 20th May 2009

Registration and Information

Registration will be handled via the AVFA2009 web site (http://www.avfa09.upm.es). Please contact the secretariat (avfa09@ics.upm.es) for further information.

Program Committee

  • Juan Ignacio Godino Llorente, Universidad Politécnica de Madrid, Co-Chair
  • Pedro Gómez Vilda, Universidad Politécnica de Madrid, Co-Chair
  • Rubén Fraile, Universidad Politécnica de Madrid, Scientific Secretariat
  • Bartolomé Scola Yurrita, Gregorio Marañón Hospital 

·         Phillippe H. Dejonckere, University Medical Center Utrecht

·         Yannis Stylianou, University of Crete

Back to Top

8-9 . (2009-05-31) Call for Workshop proposals EACL 2009, NAACL HLT 2009, ACL-UCNLP 2009

CALL FOR WORKSHOP PROPOSALS EACL 2009, NAACL HLT 2009, AND ACL-IJCNLP 2009

 

Joint site:  http://www.eacl2009.gr/conference/callforworkshops

The Association for Computational Linguistics invites proposals for
workshops to be held in conjunction with one of the three flagship
conferences sponsored in 2009 by the Association for Computational
Linguistics: ACL-IJCNLP 2009, EACL 2009, and NAACL HLT 2009.  We solicit
proposals on any topic of interest to the ACL community. Workshops will
be held at one of the following conference venues:

EACL 2009 is the annual meeting of the European chapter of the ACL. The
conference will be held in Athens, Greece, March 30-April 3 2009;
workshops March 30-31.

NAACL HLT 2009 is the annual meeting of the North American chapter of
the ACL.  It continues the inclusive tradition of encompassing relevant
work from the natural language processing, speech and information
retrieval communities.  The conference will be held in Boulder,
Colorado, USA, from May 31-June 5 2009; workshops will be held June 4-5.

ACL-IJCNLP 2009 combines the 47th Annual Meeting of the Association for
Computational Linguistics (ACL 2009) with the 4th International Joint
Conference on Natural Language Processing (IJCNLP).  The conference will
be held in Singapore, August 2-7 2009; workshops will be held August 6-7.


    SUBMISSION INFORMATION

In a departure from previous years, ACL-IJCNLP, EACL and NAACL HLT will
coordinate the submission and reviewing of workshop proposals for all
three ACL 2009 conferences.

Proposals for workshops should contain:

    * A title and brief (2-page max) description of the workshop topic
      and content.
    * The desired workshop length (one or two days), and an estimate
      of the audience size.
    * The names, postal addresses, phone numbers, and email addresses
      of the organizers, with one-paragraph statements of their
      research interests and areas of expertise.
    * A budget.
    * A list of potential members of the program committee, with an
      indication of which members have already agreed.
    * A description of any shared tasks associated with the workshop.
    * A description of special requirements for technical needs.
    * A venue preference specification.

The venue preference specification should list the venues at which the
organizers would be willing to present the workshop (EACL, NAACL HLT, or
ACL-IJCNLP).  A proposal may specify one, two, or three acceptable
workshop venues; if more than one venue is acceptable, the venues should
be preference-ordered.  There will be a single workshop committee,
coordinated by the three sets of workshop chairs.  This single committee
will review the quality of the workshop proposals.  Once the reviews are
complete, the workshop chairs will work together to assign workshops to
each of the three conferences, taking into account the location
preferences given by the proposers.

The ACL has a set of policies on workshops. You can find general
information on policies regarding attendance, publication, financing,
and sponsorship, as well as on financial support of SIG workshops, at
the following URL:
http://www.cis.udel.edu/~carberry/ACL/index-policies.html

Please submit proposals by electronic mail no later than September 1
2008, to acl09-workshops at acl09-workshops@uni-konstanz.de with the
subject line: "ACL 2009 WORKSHOP PROPOSAL."


    PRACTICAL ARRANGEMENTS

Notification of acceptance of workshop proposals will occur no later
than September 23, 2008.  Since the three ACL conferences will occur at
different times, the timescales for the submission and reviewing of
workshop papers, and the preparation of camera-ready copies, will be
different for the three conferences. Suggested timescales for each of
the conferences are given below.

ALL CONFERENCES
Sep 1, 2008     Workshop proposal deadline
Sep 23, 2008    Notification of acceptance of workshops

EACL 2009
Sep 30, 2008    Call for papers issued by this date
Dec 12, 2008    Deadline for paper submission
Jan 23, 2009    Notification of acceptance of papers
Feb  6, 2009    Camera-ready copies due
Mar 30-31, 2009 EACL 2009 workshops

NAACL HLT 2009
Dec 10, 2008    Call for papers issued by this date
Mar 6, 2009     Deadline for paper submissions
Mar 30, 2009    Notification of paper acceptances
Apr 12, 2009    Camera-ready copies due
June 4-5, 2009  NAACL HLT 2009 workshops

ACL-IJCNLP 2009
Feb 6, 2009     Call for papers issued issued by this date
May 1, 2009     Deadline for paper submissions
Jun 1, 2009     Notification of acceptances
Jun 14, 2009    Camera-ready copies due
Aug 6-7, 2009   ACL-IJCNLP 2009 Workshops

Workshop Co-Chairs:

    * Miriam Butt, EACL, University of Konstanz
    * Stephen Clark, EACL, Oxford University
    * Nizar Habash, NAACL HLT, Columbia University
    * Mark Hasegawa-Johnson, NAACL HLT, University of Illinois at
Urbana-Champaign
    * Jimmy Lin, ACL-IJCNLP, University of Maryland
    * Yuji Matumoto, ACL-IJCNLP, Nara Institute of Science and Technology

For inquiries, send email to: acl09-workshops at
acl09-workshops@uni-konstanz.de

 

Back to Top

8-10 . (2009-05-31) CfP NAACL HLT 2009 Bouldr CO, USA

Call for Papers for NAACL HLT 2009
http://www.naaclhlt2009.org May 31 – June 5, 2009, Boulder, Colorado
 
Deadline for full paper submission – Monday, December 1, 2008 Deadline for short paper submission – Monday, February 9, 2009 NAACL HLT 2009 combines the Annual Meeting of the North American Association for Computational Linguistics (NAACL) with the Human Language Technology Conference (HLT) of NAACL. The conference covers a broad spectrum of disciplines working towards enabling intelligent systems to interact with humans using natural language, and towards enhancing human-human communication through services such as speech recognition, automatic translation, information retrieval, text summarization, and information extraction. NAACL HLT 2009 will feature full papers, short papers, posters, demonstrations, and a doctoral consortium, as well as pre- and post-conference tutorials and workshops. The conference invites the submission of papers on substantial, original, and unpublished research in disciplines that could impact human language processing systems. We encourage the submission of short papers that can be characterized as a small, focused contribution, a work in progress, a negative result, an opinion piece or an interesting application note. A separate review form for short papers will be introduced this year.
NAACL HLT 2009 aims to hold two special sessions, Large Scale Language Processing and Speech Indexing and Retrieval.
Topics include, but are not limited to, the following areas, and are understood to be applied to speech and/or text:
- Large scale language processing
- Speech indexing and retrieval
- Information retrieval (including monolingual and CLIR)
- Information extraction
- Speech-centered applications (e.g., human-computer, human-robot interaction, education and learning systems, assistive technologies, digital entertainment)
- Machine translation
- Summarization
- Question answering
- Topic classification and information filtering
- Non-topical classification (e.g., sentiment/attribution/genre analysis)
- Topic clustering
- Text and speech mining
- Statistical and machine learning techniques for language processing
- Spoken term detection and spoken document indexing
- Language generation
- Speech synthesis
- Speech understanding
- Speech analysis and recognition
- Multilingual processing
- Phonology
- Morphology (including word segmentation)
- Part of speech tagging
- Syntax and parsing (e.g., grammar induction, formal grammar, algorithms)
- Word sense disambiguation
- Lexical semantics
- Formal semantics and logic
- Textual entailment and paraphrasing
- Discourse and pragmatics
- Dialog systems
- Knowledge acquisition and representation
- Evaluation (e.g., intrinsic, extrinsic, user studies)
- Development of language resources (e.g., lexicons, ontologies, annotated corpora)
- Rich transcription (automatic annotation of information structure and sources in speech)
- Multimodal representations and processing, including speech and gesture
Submission information will soon be available at: http://www.naaclhlt2009.org
General Conference Chair: Mari Ostendorf, University of Washington Program Co-Chairs: Michael Collins, Massachusetts Institute of Technology Shri Narayanan, University of Southern California
Douglas W. Oard, University of Maryland Lucy Vanderwende, Microsoft Research Local Arrangements: James Martin, University of Colorado at Boulder Martha Palmer, University of Colorado at Boulder
Back to Top

8-11 . (2009-05-31) NAACL HLT 09 Call for Demonstrations

NAACL HLT 09 Call for Demonstrations

The NAACL HLT 2009 Program Committee invites proposals for the Demonstration Program to be held June 1-3, 2009 at the University of Colorado at Boulder. We encourage both the exhibition of early research prototypes and interesting mature systems. Commercial sales and marketing activities are not appropriate in the Demonstration Program, and should be arranged as part of the Exhibit Program. We invite proposals for two types of demonstrations:

·        Type I: theater-style, as part of the regular program

·        Type II: poster-style, where demos are to be presented on table-tops in sessions scheduled for a specific time slot.

 

Submission of a demonstration proposal on a particular topic does not preclude or require a separate submission of a paper on that topic; it is possible that some but not all of the demonstrations will illustrate concepts that are described in companion papers.

 

Areas of Interest

Areas of interest include, but are not limited to, the following types of systems, some of which have been demonstrated at recent ACL conferences:

·        End-to-end natural language processing systems

·        User interfaces for monolingual and multilingual information access systems, including retrieval, summarization, and QA engines

·        Voice search interfaces

·        Dialogue and conversational systems

·        Multimodal systems utilizing language technology

·        Language technology on mobile devices

·        Applications using embedded language technology components

·        Meeting capture and analysis systems utilizing language technology

·        Natural language processing systems for medical informatics

·        Assistive applications of language technology

·        Visualization tools

·        Software for evaluating natural language systems and components

·        Aids for teaching computational linguistics concepts

·        Software tools for facilitating computational linguistics research Reusable components (parsers, generators, speech recognizers, etc.)

·        Tools that assist in the development of other NLP applications (e.g., error analysis)

      

Format for Submission

Demo proposals consist of the following parts, which should all be sent to the Demonstration Co-Chairs. Please use the main ACL paper formatting guidelines. Please note that no hardware or software will be provided by the local organizer.

·        An extended abstract of the technical content to be demonstrated, including title, authors, full contact information, references, and acknowledgements. Please indicate a Type I or Type II demo.

·        A "script outline" of the demo presentation, including accompanying narrative, and either a Web address for accessing the demo or visual aids (e.g., screenshots, snapshots, or diagrams).

The entire proposal must not be more than four pages.

Submissions Procedure

Proposals must be submitted by February 9, 2009 to the Demonstration Co-Chairs. Submissions must be received electronically. Please submit your proposals and any inquiries to:

Michael Johnston                                         Fred Popowich         

AT&T                                                                      Simon Fraser University

johnston “at” research “dot” att “dot” com   popowich “at” sfu “dot” ca

Submissions will be evaluated on the basis of their relevance to computational linguistics, innovation, scientific contribution, presentation, as well as potential logistical constraints.

Accepted submissions will be allocated four pages in the Companion Volume to the Proceedings of the Conference.

 

Further Details

Further details on the date, time, and format of the demonstration session(s) will be determined and provided at a later date. Please send any inquiries to the demonstration co-chairs at the email addresses listed above.

 

Important Dates

 

February 9, 2009

Submission deadline

March 27, 2009

Notification of acceptance

April 6, 2009

Submission of final demo related literature

June 1-3, 2009

Conference

 

All submissions or camera-ready copies are due by 11:59pm EST on the dates specified above.

Back to Top

8-12 . (2009-04-19) ICASSP 2009

IEEE ICASSP 2009

  International Conference on Acoustics, Speech, and Signal Processing

                           April 19-24, 2009

         Taipei International Convention Center, Taipei, Taiwan

 

4 distinguished plenary speakers have been invited:

 

(1) Dynamic Spectrum Management

    by John M. Cioffi, Stanford University & ASSIA Inc.

(2) Signal Processing Bringing on Market Growth – Its Industrial Successes and Future Expectations

    by Kazuo Murano, President, Fujitsu Laboratories Limited

(3) Cognitive User Interfaces: an Engineering Approach

    by Steve Young, University of Cambridge

(4) The Evolution and Trends of the Semiconductor Industry and Signal Processing SoC

    by Ming-Kai Tsai, Chairman of the Board and Chief Executive Officer, MediaTek Inc.

 

12 Special Sessions have been organized:

Multimedia Surveillance

Voice Transformation

Signal Processing Challenges for 4G Wireless Communications

Distributed Signal Processing and Consensus Gossiping

Handling Reverberant Speech: Methodologies and Applications

Interference Channels and Spectrum Sharing

Signal Processing for Neural Spike Trains

Multimedia Social Networks

Signal Processing Techniques and Algorithms on Robot Audition

Signal Processing for Improper and Noncircular Complex-Valued Data

The Data Deluge: the Challenges and Opportunities of Unlimited Data in Signal Processing

Video Search and Event Analysis

 

Over 140 oral and poster sessions are planned on the following fields:

Audio and electroacoustics

Bio imaging and signal processing

Design and implementation of signal processing systems

Industry technology tracks

Image and multidimensional signal processing

Information forensics and security

Machine learning for signal processing

Multimedia signal processing

Sensor array and multichannel systems

Signal processing education

Signal processing for communications

Speech and language processing

Signal processing theory and methods

 

17 Tutorials are included in the technical program:

 

 - Fundamentals -

Distributed Adaptive Filters and Networks

Sparse Sampling: Theory, Algorithms and Applications

Beyond Bandlimited Sampling: Nonideal Sampling, Smoothness and Sparsity

Signal and System Theoretic Foundations of Quantization and Data Acquisition (A/D and D/A conversion)

 

 - Speech and audio -

Automatic Recognition of Natural Speech

Applications of Psychoacoustics to Signal Processing

Special Interests in Mandarin Speech Recognition and Chinese-to-English Machine Translation

 

 - Image and video -

Distributed Video Coding for Low Cost Multimedia Communications Systems

Distributed Processing in Smart Cameras

Digital Video Image Quality and Perceptual Coding

 

 - Communications -

Distributed Source Coding: Theory, Code Designs and Applications

Peak-to-Average Power Ratio Reduction: Applications and Algorithms

A Unified Design Framework for Non-Linear MIMO Transceivers Using Majorization Theory

Game Theory and Resource Allocation in Wireless Communications

 

 - Retrieval -

Analysis and Retrieval Techniques for Music and Motion Data

Recent Developments in Content-based and Concept-based Image/Video Retrieval

 

 - Implementation -

Multimedia Signal Processing on CPU and GPU with Many Cores

 

                           Call for Proposals

                              Show & Tell

                                   at

                              ICASSP 2009

                           in Taipei, Taiwan

 

For the second year, the program committee for the International Conference on Acoustics, Speech, and Signal Processing is accepting proposals for the Show & Tell event that will be held during the conference.  The event will include demonstrations of latest innovations by research groups in industry, academia and government. Demonstrations can be related to any of the topics defined by ICASSP as shown in the call for papers (see http://www.icassp09.com/ICASSP2009.pdf).  Show & Tell sessions will run concurrently with regular ICASSP sessions in the morning and afternoon of April 22-23.

 

The deadline for submission of Show & Tell proposals is January 15, 2009 with notification of acceptance by February 13, 2009.  Submissions will be accepted via the ICASSP 2009 web site, http://www.icassp09.com.  Proposals should include the demonstration title, list of authors, and an abstract of one page or less.  The abstract should clearly explain what is novel and innovative in the proposed demonstration.  Show & Tell demonstrations should have an interactive component (i.e., should be more than a simple computer simulation or graph).  On the other hand, commercial product demonstrations will be considered more appropriate for ICASSP exhibits than Show & Tell.

 

Each demonstration will be allotted one 180x60cm table, space for a poster, and a power outlet.  Presenters are responsible for all other equipment and shipping to and from the conference.  Unlike 2008, there will be no charge for Show & Tell at ICASSP 2009 in Taipei.  Presenters must register for ICASSP 2009 in order to participate in Show & Tell demonstrations.

 

For more detailed information, please visit the ICASSP 2009 official website, http://icassp09.com/ . We look forward to seeing you in Taipei!

Back to Top

8-13 . (2009-05-31) NAACL-HLT-09: Call for Tutorial Proposals

NAACL-HLT-09: Call for Tutorial Proposals

Proposals are invited for the Tutorial Program of the North American Chapter of the Association for Computational Linguistics - Human Language Technologies (NAACL HLT) 2009 Conference. The conference is to be held from May 31 to June 5, 2009 in Boulder, Colorado. The tutorials will be held on Sunday, May 31.

Proposals for tutorials on all topics of computational linguistics and speech processing, such as processing for purposes of indexing and retrieval, processing for data mining, and so forth, are welcome. Especially encouraged are tutorials that educate the community about advancements in speech and natural language processing occurring in situ with contextual awareness, such as understanding speech, language or gesture in particular physical contexts.

Information on the tutorial instructor payment policy can be found at http://aclweb.org/aclwiki/index.php?title= Tutorial_teacher_payment_policy

PLEASE NOTE: Remuneration for Tutorial presenters is fixed according to the above policy and does not cover registration fees for the main conference.

SUBMISSION DETAILS

Proposals for tutorials should contain:

  1. A title and brief description of the tutorial content and its relevance to the NAACL-HLT community (not more than 2 pages).
  2. A brief outline of the tutorial structure showing that the tutorial's core content can be covered in a three-hour slot (including a coffee break). In exceptional cases six-hour tutorial slots are available as well.
  3. The names, postal addresses, phone numbers, and email addresses of the tutorial instructors, including a one-paragraph statement of their research interests and areas of expertise.
  4. A list of previous venues and approximate audience sizes, if the same or a similar tutorial has been given elsewhere; otherwise an estimate of the audience size.
  5. A description of special requirements for technical equipment (e.g., internet access).

Proposals should be submitted by electronic mail, in plain ASCII text no later than January 15, 2009 to tutorials.hlt09 "at" gmail "dot" com. The subject line should be: "NAACL HLT 2009: TUTORIAL PROPOSAL".

PLEASE NOTE:

  1. Proposals will not be accepted by regular mail or fax, only by email to: tutorials.hlt09 "at" gmail "dot" com.
  2. You will receive an email confirmation from us that your proposal has been received. If you do not receive this confirmation 24 hours after sending the proposal, please contact us personally using all the following emails: ciprianchelba "at" google "dot" com,
    kantor "at" scils "dot" rutgers "dot" edu, and
    roark "at" cslu "dot" ogi "dot" edu.

TUTORIAL SPEAKER RESPONSIBILITIES

Accepted tutorial speakers will be notified by February 1, 2009, and must then provide abstracts of their tutorials for inclusion in the conference registration material by March 1, 2009. The description should be in two formats: an ASCII version that can be included in email announcements and published on the conference web site, and a PDF version for inclusion in the electronic proceedings (detailed instructions will be given). Tutorial speakers must provide tutorial materials, at least containing copies of the course slides as well as a bibliography for the material covered in the tutorial, by April 15, 2009.

IMPORTANT DATES

  • Submission deadline for tutorial proposals: January 15, 2009
  • Notification of acceptance: February 1, 2009
  • Tutorial descriptions due: March 1, 2009
  • Tutorial course material due: April 15, 2009
  • Tutorial date: May 31, 2009

TUTORIALS CO-CHAIRS

  • Ciprian Chelba, Google
  • Paul Kantor, Rutgers
  • Brian Roark, Oregon Health & Science University
     

Please send inquiries concerning NAACL-HLT-09 tutorials to tutorials.hlt09 "at" gmail "dot" com  

Back to Top

8-14 . (2009-06-03) 7th International Workshop on Content-Based Multimedia Indexing

7ème Atelier International sur Indexation Multimédia Par le Contenu.
7th International Workshop on Content-Based Multimedia Indexing

Après le succès des six événements précédents (Toulouse 1999, Brescia 2001, Rennes 2003, Riga 2005, Bordeaux 2007, Londres 2008), l’atelier international  CBMI 2009 aura lieu du 3 au 5 juin 2009 dans la ville pittoresque de Chania sur l'île de Crète en Grèce. Il sera organisé par le laboratoire Image, Vidéo et Multimédia de l'Université Technique Nationale d'Athènes. Le CBMI 2009 a pour but de rassembler les différentes communautés impliquées dans les différents aspects de l'indexation multimédia basée sur le contenu, tels que le traitement d'images et la recherche d'information avec les tendances et développements actuels des industriels. L’atelier est soutenu par les sociétés savantes IEEE et EURASIP, Université d’Athènes. Le programme technique du CBMI 2009 comprend les conférences plénières invitées, des sessions spéciales ainsi que des sessions régulières.

 

Liste non exhaustive des thèmes traités:

l  Indexation et recherche multimédia (image, audio, vidéo, texte)

l  Mise en correspondance et recherche de similarité

l  Construction d'indices de haut niveau

l  Extraction du contenu multimédia

l  Identification et suivi des régions sémantiques dans les scènes

l  Indexation multi-modale et cross-modale

l  Recherche basée contenu

l  L'extraction de données multimédia

l  Génération, codage et transformation de métadonnées

l  Gestion de bases de données multimédia de grande échelle

l  Résumé, navigation et organisation du contenu multimédia

l  Outils de présentation et de visualisation

l  Interaction avec l'utilisateur et pertinence du retour

l  Personnalisation et adaptation au contenu

l  Evaluation et métriques

 

 

Soumission

            Les auteurs sont invités à soumettre des papiers sur le site web de la conférence: http://www.cbmi2009.org/submission.  Des fichiers de style (Latex et Word) seront fourni pour la commodité des auteurs.

Dates importantes


Présentation des textes complets:

8 janvier 2009

Notification d'acceptation:

23 février 2009

Soumission des versions finales:

13 mars 2009

Début de l'enregistrement:

13 mars 2009

Conférence

3 au 5 Juin 2009

 

Lieu de la manifestation

            Le CBMI 2009 aura lieu dans l'enceinte du KAM - Center méditerranéen de l'architecture, de Chania, sur l'île de la Crète, l'une des destinations les plus excitantes en Grèce. Le KAM a été créé par la commune de Chania en 1996 et est situé depuis 2002 au Grand Arsenal, le vieux port de Chania.

Following the six successful previous events (Toulouse 1999, Brescia 2001, Rennes 2003, Riga 2005, Bordeaux 2007, London 2008), 2009 International Workshop on Content-Based Multimedia Indexing (CBMI) will be held on June 3-5, 2009 at the picturesque city of Chania, in Crete Island, Greece. It will be organized by Image, Video and Multimedia Laboratory of National Technical University of Athens. CBMI 2009 aims at bringing together the various communities involved in the different aspects of content-based multimedia indexing, such as image processing and information retrieval with current industrial trends and developments. CBMI 2009 is supported by IEEE, EURASIP, University of Athens. The technical program of CBMI 2009 will include presentation of invited plenary talks, special sessions as well as regular sessions with contributed research papers.

Topics of interest include, but are not limited to:

Multimedia indexing and retrieval (image, audio, video, text)
Matching and similarity search
Construction of high level indices
Multimedia content extraction
Identification and tracking of semantic regions in scenes
Multi-modal and cross-modal indexing
Content-based search
Multimedia data mining
Metadata generation, coding and transformation
Large scale multimedia database management
Summarisation, browsing and organization of multimedia content
Presentation and visualization tools
User interaction and relevance feedback
Personalization and content adaptation
Evaluation and metrics

Paper Submission

Prospective authors are invited to submit full papers at the conference web site: http://www.cbmi2009.org/submission. Style files (Latex and Word) will be provided for the convenience of the authors.

Important Dates

Submission of full papers:

January 8, 2009

Notification of acceptance:

February 23, 2009

Submission of camera-ready papers:

March 13, 2009

Early registration due:

March 13, 2009

Main Workshop:

June 3-5, 2009

Venue

CBMI 2009 will be hosted at KAM - Mediterranean Centre of Architecture, Chania, at the island of Crete, one of the most exciting Greek destinations. KAM was settled by Chania municipality in 1996 and is situated since 2002 at Great Arsenali, the old port of Chania.

 

Back to Top

8-15 . (2009-06-05) Nasal 2009 Nasalité en phonétique et en phonologie (french)

*Nasal 2009*
_Appel à communications_
*Site internet:* http://w3.umh.ac.be/~nasal/Workshop/home.html <http://w3.umh.ac.be/%7Enasal/Workshop/home.html>
*Contact: *nasal2009@umh.ac.be <mailto:nasal2009@umh.ac.be>
L?équipe Praxiling UMR 5267 CNRS (Université Paul Valéry, Montpellier 3), et le Laboratoire des Sciences de la Parole de l'Académie Universitaire Wallonie-Bruxelles (Université de Mons-Hainaut), organisent un colloque international consacré à la *nasalité en phonétique et en phonologie.
*Le colloque aura lieu le *vendredi 5 juin 2009* de 9h00 à 18h30 au Grand Amphithéâtre de la Délégation régionale du CNRS, 1919, route de Mende, F-34293 Montpellier cedex 5.

Depuis les premières descriptions de langues signalant l?existence de sons nasals (dès Panini, au 5e siècle ACN) jusqu?aux monographies phonétiques et phonologiques des 20e et 21e siècles, en passant par les travaux effectués en grammaire comparée au 19e siècle, la nasalité a toujours été l?un des objets d?étude privilégiés pour ceux qui s?intéressent au fonctionnement des langues et de la parole humaine. Avec l?éclosion de la phonétique expérimentale, dès la fin du 19e siècle, des jalons importants ont été posés vers une meilleure compréhension des phénomènes de nasalité, notamment grâce à la mise au point de dispositifs instrumentaux novateurs (Rousselot, 1897). A partir du milieu du 20e s, les principes fondamentaux de l?acoustique et de la perception de la nasalité ont été établis (Delattre, 1954, House & Stevens, 1956, Fant, 1960). Au même moment, plusieurs médecins investiguaient les troubles de la parole liés aux fentes palatines, en particulier l?hypernasalité (Warren et Dubois, 1964). Les années 60 et 70 ont vu de nombreux progrès en ce qui concerne la description de la production de la nasalité, y compris en parole pathologique (p.ex. Björk, 1961, Fritzell, 1969, Bell-Berti , 1976, Benguerel et al., 1977, Weinberg et al., 1968). Diverses techniques ont été convoquées : cinéradiographie, électromyographie, fibroscopie, aérodynamique, tandis que des dispositifs spécifiquement dédiés à la nasalité étaient développés (p.ex. le nasographe : Ohala, 1971). Les études nasales ont ainsi massivement contribué à l?élaboration des théories et des modèles de la coarticulation (pour une revue, voir p.ex. Chafcouloff et Marchal, 1999). Plus récemment, dans les années 80 et 90, notre compréhension de la perception de la nasalité a connu d?importantes avancées, notamment grâce au développement de la synthèse de la parole et de la modélisation (Beddor, 1993, Kingston et MacMillan, 1995, Krakow et al., 1988, Maeda, 1993).

Malgré ces avancées, la nasalité demeure aujourd?hui encore l?un de ces objets linguistiques qui résistent à une complète compréhension. Les phénomènes de nasalisation et les processus qui les sous-tendent ne peuvent être décrits que via le recours à une complexité conceptuelle, méthodologique et instrumentale rarement égalée, et résistent souvent à l'articulation avec les théories et les modèles les mieux éprouvés. Ainsi, malgré de nombreux travaux sur les différents aspects de la production des sons nasals, on manque toujours d?un modèle opérationnel de la production de la nasalité qui rende compte des non linéarités entre les phases articulatoire, aérodynamique, et acoustique. Par ailleurs, malgré des résultats prometteurs obtenus en modélisation (p.ex. Maeda, 1982, 1993), on ne sait pas encore exactement dans quelle mesure l?étendue spatiale du geste d?ouverture vélo-pharyngée est reliée au percept de nasalité. La question du déploiement de la nasalité dans le domaine temporel demeure elle aussi largement non résolue. Chaque langue présente ses patrons spécifiques de coarticulation, d?ajustements et de coordination gestuels, mais il reste à déterminer les contraintes qui encadrent (ou déterminent) ces spécificités: liens avec la structure prosodique (Vaissière, 1988, Fougeron, 2001), avec le statut phonologique de la nasalité dans la langue, avec les autres éléments de l?inventaire phonologique, covariation avec d?autres traits/gestes (Solé, 2007) tels que le voisement et la friction (consonnes), l?aperture vocalique ou le lieu d?articulation (voyelles), etc. La perception de la coarticulation nasale à travers les langues est actuellement l?une des pistes de recherche les plus prometteuses pour la compréhension des phénomènes de nasalité (Beddor, 2007).

Enfin, la grande diversité des langues est génératrice d?infinie variabilité. Or, de nombreuses langues sont encore peu décrites, qui présentent des phénomènes intéressants liés à la nasalité. Qu?il s?agisse des fricatives nasales pré-nasalisées du kinyarwanda (Demolin, 2005) ou des occlusives nasales pré et post-oralisées du karitiana (Storto & Demolin, 2008), les travaux les plus récents concernant les variantes nasales dans les langues du monde permettent de confronter certaines hypothèses ou modélisations préalablement proposées (p.ex. Demolin, 2007). En effet, les bénéfices potentiels des études nasales pour la compréhension du fonctionnement du langage humain sont heureusement à la hauteur du défi que pose la nasalité au chercheur en sciences du langage. L?étude de la nasalisation permet de tester et d?améliorer nos hypothèses quant à la nature des relations entre phonétique et phonologie, que ce soit du point de vue de l?individu, au niveau cognitif, ou du point de vue de la langue, au niveau structurel. Les phénomènes de nasalité permettent également d?investiguer ce qui, dans l?organisation des sons d?une langue, relève de l?universel, et ce qui est spécifique à une langue donnée, en lien avec son système propre (voir p.ex. Maddieson, 2007). De même, les études nasales ont été, et seront sans doute à l?avenir, amenées à contribuer au développement et à la vérification d?outils, de théories et de modèles centraux en phonétique et en phonologie : modélisation articulatoire et acoustique, théories de la coarticulation, modèles d?apprentissage des langues étrangères, etc. Les études nasales sont au c?ur des enjeux majeurs des disciplines scientifiques reliées à la communication parlée.
L?objectif de ce colloque international est de permettre aux chercheurs du monde entier de se réunir et d?échanger à propos de leurs travaux récents concernant la nasalité. Toute proposition de communication concernant la nasalité est la bienvenue, en particulier les travaux concernant: la production (mesures articulatoires, études aérodynamiques, analyses acoustiques, etc.), la perception, les aspects phonologiques, les universaux phonétiques, la modélisation, les langues peu décrites, les aspects pathologiques et cliniques, l?acquisition du langage et l?apprentissage d?une langue seconde. Un intérêt tout particulier sera accordé aux communications traitant de questions transversales aux champs disciplinaires cités ci-dessus : multiinstrumentation, liens entre production et perception, comparaisons inter-langues, relations entre l?organisation des systèmes phonologiques et les contraintes phonétiques, points communs et divergences entre acquisition en L1 et apprentissage en L2, etc.
*Conférenciers invités
*Patrice S. Beddor, University of Michigan, USA
Didier Demolin, Université Libre de Bruxelles, Belgique
John Hajek, University of Melbourne, Australia
Ian Maddieson, University of Albuquerque, New Mexico, USA
Alain Marchal, Université d'Aix-en-Provence, France
Jacqueline Vaissière, Université de Paris III, France
*Comité scientifique
*Pierre Badin, Gipsa-Lab, France
Nick Clements, Université de Paris III, France
Bernard Harmegnies, Université de Mons-Hainaut, Belgique
Sarah Hawkins, University of Cambridge, UK
Marie Huffman, State University of New York Stony Brook, USA
John Kingston, University of Massachussets at Amherst, USA
Christine Matyear, University of Texas at Austin, USA
John Ohala, University of California at Berkeley, USA
Daniel Recasens, Universitat Autonoma de Barcelona, Espana
Ryan Shosted, University of Illinois at Urbana-Champaign, USA
Maria Josep Solé, Universitat Autonoma de Barcelona, Espana
Nathalie Vallée, Gipsa-Lab, France
Doug Whalen, Haskins Laboratories, USA
*Comité d'organisation
*Véronique Delvaux, FNRS, Université de Mons-Hainaut, Présidente
Mohamed Embarki, Université de Montpellier 3, Vice-Président
Christelle Dodane, Université de Montpellier 3, Actes
Josiane Clarenc, Université de Montpellier 3, Logistique
*Date limite de soumission
*15 février 2009

*Modalités de soumission
*Envoyer un message avec les coordonnées complètes du premier auteur et le nom des éventuels autres auteurs à nasal2009@umh.ac.be <mailto:nasal2009@umh.ac.be> avec, en fichier attaché, un article _anonyme_ de quatre pages A4 maximum sous la forme d'un fichier pdf.
Un modèle de fichier Word est téléchargeable sur le site du colloque: http://w3.umh.ac.be/~nasal/Workshop/appel.html <http://w3.umh.ac.be/%7Enasal/Workshop/appel.html>
*Date de notification de l'acceptation
*6 avril 2009
*Publication
*Les actes du colloque seront distribués gratuitement à toutes les personnes régulièrement inscrites et présentes le 5 juin 2009 à Montpellier. Pré-inscrivez-vous par simple mail à l'adresse: nasal2009@umh.ac.be <mailto:nasal2009@umh.ac.be>.
Les participants sont invités à soumettre une version longue de leur communication (50000 caractères) pour une éventuelle publication dans un livre à paraître chez un éditeur international. (Plus de détails bientôt). Date limite de soumission des papiers: 14 septembre 2009.
Pour le comité d'organisation,
V. Delvaux
Chargée de Recherches FNRS
Laboratoire de Phonétique
Service de Métrologie et Sciences du Langage
Université de Mons-Hainaut
18, Place du Parc
7000 Mons
Belgium
+3265373140
http://staff.umh.ac.be/Delvaux.Veronique/index.html 

Back to Top

8-16 . (2009-06-21) CfP Specom 2009- St Petersburg Russia

SPECOM 2009 - SECOND CALL FOR PAPERS

    13-th International Conference "Speech and Computer"
                             21-25 June 2009
     Grand Duke Vladimir's palace, St. Petersburg, Russia
                      http://www.specom.nw.ru


Organized by St. Petersburg Institute for Informatics and Automation of the Russian Academy of Sciences (SPIIRAS)

Dear Colleagues, we are pleased to invite you to the 13-th International Conference on Speech and Computer SPECOM'2009, which will be held in June
21-25, 2009 in St.Petersburg. The global aim of the conference is to discuss state-of-the-art problems and recent achievements in Signal Processing and
Human-Computer Interaction related to speech technologies. Main topics of SPECOM'2009 are:
- Signal processing and feature extraction
- Multimodal analysis and synthesis
- Speech recognition and understanding
- Natural language processing
- Spoken dialogue systems
- Speaker and language identification
- Text-to-speech systems
- Speech perception and speech disorders
- Speech and language resources
- Applications for human-computer interaction

The official language of the event is English. Full papers up to 6 pages will be published in printed and electronic proceedings with ISBN.

Imporatnt Dates:
- Submission of full papers: February 1, 2009 (extended)
- Notification of acceptance: March 1, 2009
- Submission of final papers: March 20, 2009
- Early registration: March 20, 2009
- Conference dates: June 21-25, 2009

Scientific Committee:
Andrey Ronzhin, Russia (conference chairman)
Niels Ole Bernsen, Denmark
Denis Burnham, Australia
Jean Caelen, France
Christoph Draxler, Germany
Thierry Dutoit, Belgium
Hiroya Fujisaki, Japan
Sadaoki Furui, Japan
Jean-Paul Haton, France
Ruediger Hoffmann, Germany
Dimitri Kanevsky, USA
George Kokkinakis, Greece
Steven Krauwer, Netherlands
Lin-shan Lee, Taiwan
Boris Lobanov, Belarus
Benoit Macq, Belgium
Jury Marchuk, Russia
Roger Moore, UK
Heinrich Niemann, Germany
Rajmund Piotrowski, Russia
Louis Pols, Netherlands
Rodmonga Potapova, Russia
Josef Psutka, Czech Republic
Lawrence Rabiner, USA
Gerhard Rigoll, Germany
John Rubin, UK
Murat Saraclar, Turkey
Jesus Savage, Mexico
Pavel Skrelin, Russia
Viktor Sorokin, Russia
Yannis Stylianou, Greece
Jean E. Viallet, France
Taras Vintsiuk, Ukraine
Christian Wellekens, France

The invited speakers of SPECOM'2009 are:
- Prof. Walter Kellermann (University of Erlangen-Nuremberg, Germany), lecture "Towards Natural Acoustic Interfaces for Automatic Speech Recognition"
- Prof. Mikko Kurimo (Helsinki University of Technology, Finland), lecture "Unsupervised decomposition of words for speech recognition and retrieval"

The conference venue is House of Scientists (former Grand Duke Vladimir's palace) located in the very heart of the city, in the neighborhood
of the Winter Palace (Hermitage), the residence of Russian emperor, and the Peter's and Paul's Fortress. Independently of the scientific actions
we will provide essential possibilities for acquaintance with cultural and historical valuables of  Saint-Petersburg, the conference will be hosted
during a unique and wonderful period known as the White Nights.

Contact Information:
SPECOM'2009 Organizing Committee,
SPIIRAS, 39, 14-th line, St.Petersburg, 199178, RUSSIA
E-mail: specom@iias.spb.su
Web: http://www.specom.nw.ru 

 

 

 

Back to Top

8-17 . (2009-06-22) Summer workshop at Johns Hopkins University

                                            The Center for Language and Speech Processing

 

at Johns Hopkins University invites one page research proposals for a

NSF-sponsored, Six-week Summer Research Workshop on

Machine Learning for Language Engineering

to be held in Baltimore, MD, USA,

June 22 to July 31, 2009.

CALL FOR PROPOSALS

Deadline: Wednesday, October 15, 2008.

One-page proposals are invited for the 15th annual NSF sponsored JHU summer workshop.  Proposals should be suitable for a six-week team exploration, and should aim to advance the state of the art in any of the various fields of Human Language Technology (HLT) including speech recognition, machine translation, information retrieval, text summarization and question answering.  This year, proposals in related areas of Machine Intelligence, such as Computer Vision (CV), that share techniques with HLT are also being solicited.  Research topics selected for investigation by teams in previous workshops may serve as good examples for your proposal. (See http://www.clsp.jhu.edu/workshops.)

Proposals on all topics of scientific interest to HLT and technically related areas are encouraged.  Proposals that address one of the following long-term challenges are particularly encouraged.

Ø  ROBUST TECHNOLOGY FOR SPEECH:  Technologies like speech transcription, speaker identification, and language identification share a common weakness: accuracy degrades disproportionately with seemingly small changes in input conditions (microphone, genre, speaker, dialect, etc.), where humans are able to adapt quickly and effectively. The aim is to develop technology whose performance would be minimally degraded by input signal variations.

Ø  KNOWLEDGE DISCOVERY FROM LARGE UNSTRUCTURED TEXT COLLECTIONS: Scaling natural language processing (NLP) technologies—including parsing, information extraction, question answering, and machine translation—to very large collections of unstructured or informal text, and domain adaptation in NLP is of interest.

Ø  VISUAL SCENE INTERPRETATION: New strategies are needed to parse visual scenes or generic (novel) objects, analyzing an image as a set of spatially related components.  Such strategies may integrate global top-down knowledge of scene structure (e.g., generative models) with the kind of rich bottom-up, learned image features that have recently become popular for object detection.  They will support both learning and efficient search for the best analysis.

Ø  UNSUPERVISED AND SEMI-SUPERVISED LEARNING: Novel techniques that do not require extensive quantities of human annotated data to address any of the challenges above could potentially make large strides in machine performance as well as lead to greater robustness to changes in input conditions.  Semi-supervised and unsupervised learning techniques with applications to HLT and CV are therefore of considerable interest.

An independent panel of experts will screen all received proposals for suitability. Results of this screening will be communicated no later than October 22, 2008. Authors passing this initial screening will be invited to Baltimore to present their ideas to a peer-review panel on November 7-9, 2008.  It is expected that the proposals will be revised at this meeting to address any outstanding concerns or new ideas. Two or three research topics and the teams to tackle them will be selected for the 2009 workshop.

We attempt to bring the best researchers to the workshop to collaboratively pursue the selected topics for six weeks.  Authors of successful proposals typically become the team leaders.  Each topic brings together a diverse team of researchers and students.  The senior participants come from academia, industry and government.  Graduate student participants familiar with the field are selected in accordance with their demonstrated performance, usually by the senior researchers. Undergraduate participants, selected through a national search, will be rising seniors who are new to the field and have shown outstanding academic promise.

If you are interested in participating in the 2009 Summer Workshop we ask that you submit a one-page research proposal for consideration, detailing the problem to be addressed.  If your proposal passes the initial screening, we will invite you to join us for the organizational meeting in Baltimore (as our guest) for further discussions aimed at consensus.  If a topic in your area of interest is chosen as one of the two or three to be pursued next summer, we expect you to be available for participation in the six-week workshop. We are not asking for an ironclad commitment at this juncture, just a good faith understanding that if a project in your area of interest is chosen, you will actively pursue it.

Proposals should be submitted via e-mail to clsp@jhu.edu by 4PM EST on Wed, October 15, 2008.

Back to Top

8-18 . (2009-06-22) Third International Conference on Intelligent Technologies for Interactive Entertainment (Intetain 2009)

Intetain 2009, Amsterdam, 22-24th June 2009

Third International Conference on Intelligent Technologies for Interactive Entertainment

 http://intetain.org/

**********************************************************************

 

Call for Papers

 

==================

==== OVERVIEW ====

==================

The Human Media Interaction (HMI) department of the University of Twente in the Netherlands and the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering (ICST) are pleased to announce the Third International Conference on Intelligent Technologies for Interactive Entertainment to be held on June 22-24, 2009 in Amsterdam, the Netherlands.

 

INTETAIN 09 intends to stimulate interaction among academic researchers and commercial developers of interactive entertainment systems. We are seeking long (full) and short (poster) papers as well as proposals for interactive demos. In addition, the conference organisation aims at an interactive hands-on session along the lines of the Design Garage that was held at INTETAIN 2005. Individuals who want to organise special sessions during INTETAIN 09 may contact the General Chair, Anton Nijholt  (anijholt@cs.utwente.nl). 

 

The global theme of this third edition of the international conference is “Playful interaction, with others and with the environment”.

 

Contributions may, for example, contribute to this theme by focusing on the Supporting Device Technologies underlying interactive systems (mobile devices, home entertainment centers, haptic devices, wall screen displays, information kiosks, holographic displays, fog screens, distributed smart sensors, immersive screens and wearable devices), on the Intelligent Computational Technologies used to build the interactive systems, or by discussing the Interactive Applications for Entertainment themselves.

 

We seek novel, revolutionary, and exciting work in areas including but not limited to:

 

== Supporting Technology ==

 * New hardware technology for interaction and entertainment

 * Novel sensors and displays

 * Haptic devices

 * Wearable devices

 

== Intelligent Computational Technologies ==

 * Animation and Virtual Characters

 * Holographic Interfaces

 * Adaptive Multimodal Presentations

 * Creative language environments

 * Affective User Interfaces

 * Intelligent Speech Interfaces

 * Tele-presence in Entertainment

 * (Collaborative) User Models and Group Behavior

 * Collaborative and virtual Environments

 * Brain Computer Interaction

 * Cross Domain User Models

 * Augmented, Virtual and Mixed Reality

 * Computer Graphics & Multimedia

 * Pervasive Multimedia

 * Robots

 * Computational humor

 

== Interactive Applications for Entertainment ==

 * Intelligent Interactive Games

 * Emergent games

 * Human Music Interaction

 * Interactive Cinema

 * Edutainment

 * Urban Gaming

 * Interactive Art

 * Interactive Museum Guides

 * Evaluation

 * City and Tourism Explorers Assistants

 * Shopping Assistants

 * Interactive Real TV

 * Interactive Social Networks

 * Interactive Story Telling

 * Personal Diaries, Websites and Blogs

 * Comprehensive assisting environments for special populations

     (handicapped, children, elderly)

 * Exertion games

 

===========================

==== SUBMISSION FORMAT ====

===========================

INTETAIN 09 accepts long papers and short poster papers as well as demo proposals accompanied by a two page extended abstract. Accepted long and short papers will be published in the new Springer series LNICST: Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering. The organisation of INTETAIN 09 is currently working to secure a special edition of a journal, as happened previously for the 2005 edition of the Intetain conference.

 

Submissions should adhere to the LNICST instructions for authors, available from the INTETAIN 09 web site.

 

== Long papers ==

Submissions of a maximum of 12 pages that describe original research work not submitted or published elsewhere. Long papers will be orally presented at the conference.

 

== Short papers ==

Submissions of a maximum of 6 pages that describe original research work not submitted or published elsewhere. Short papers will be presented with a poster during the demo and poster session at the conference.

 

== Demos ==

Researchers are invited to submit proposals for demonstrations to be held during a special demo and poster session at the INTETAIN 09. For more information, see the Call for Demos below. Demo proposals may either be accompanied by a long or short paper submission, or by a two page extended abstract describing the demo. The extended abstracts will be published in a supplementary proceedings distributed during the conference.

 

=========================

==== IMPORTANT DATES ====

=========================

Submission deadline:

Monday, Februari 16, 2009

 

Notification:

Monday, March 16, 2009

 

Camera ready submission deadline:

Monday, March 30, 2009

 

Late demo submission deadline (extended abstract only!):

Monday, March 30, 2009

 

Conference:

June 22-24, 2009, Amsterdam, the Netherlands

 

===================

==== COMMITTEE ====

===================

General Program Chair:

Anton Nijholt, Human Media Interaction, University of Twente, the Netherlands

 

Local Chair:

Dennis Reidsma, Human Media Interaction, University of Twente, the Netherlands

 

Web Master and Publication Chair:

Hendri Hondorp, Human Media Interaction, University of Twente, the Netherlands

 

Steering Committee Chair:

Imrich Chlamtac, Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering

 

========================

==== CALL FOR DEMOS ====

========================

We actively seek proposals from both industry and academia for interactive demos to be held during a dedicated session at the conference. Demos may accompany a long or short paper. Also, demos may be submitted at a later deadline instead, with a short, two page extended abstract explaining the demo and showing why the demo would be a worthwhile contribution the INTETAIN 09's demo session.

 

== Format ==

Demo submissions should be accompanied by the following additional information:

 * A short description of the setup and demo (2 alineas)

 * Requirements (hardware, power, network, space,

     sound conditions, etc, time needed for setup)

 * A sketch or photo of the setup

 

Videos showing the demonstration setup in action are very welcome.

 

== Review ==

Demo proposals will be reviewed by a review team that will take into account aspects such as novelty, relevance to the conference, coverage of topics and available resources.

 

== Topics ==

 * Topics for demo submissions include, but are not limited to:

 * New technology for interaction and entertainment

 * (serious) gaming

 * New entertainment applications

 * BCI

 * Human Music Interaction

 * Music technology

 * Edutainment

 * Exertion interfaces

 

============================

==== PROGRAM COMMITTEE ====

============================

Stefan Agamanolis Distance Lab, Forres, UK
Elisabeth Andre Augsburg University, Germany
Lora Aroyo Vrije Universiteit Amsterdam, the Netherlands
Regina Bernhaupt University of Salzburg, Austria
Kim Binsted University of Hawai, USA
Andreas Butz University of Munich, Germany
Yang Cai Visual Intelligence Studio, CYLAB, Carnegie Mellon, USA
Antonio Camurri University of Genoa, Italy
Marc Cavazza University of Teesside, UK
Keith Cheverst University of Lancaster, UK
Drew Davidson CMU, Pittsburgh, USA
Barry Eggen University of Eindhoven, the Netherlands
Arjan Egges University of Utrecht, the Netherlands
Anton Eliens Vrije Universiteit Amsterdam, the Netherlands
Steven Feiner Columbia University, New York
Alois Ferscha University of Linz, Austria
Matthew Flagg Georgia Tech, USA
Jaap van den Herik University of Tilburg, the Netherlands
Dirk Heylen University of Twente, the Netherlands
Frank Kresin Waag Society, Amsterdam, the Netherlands
Antonio Krueger University of Muenster, Germany
Tsvi Kuflik University of Haifa, Israel
Markus Löckelt DFKI Saarbrücken, Germany
Henry Lowood University of Stanford, USA 
Mark Maybury MITRE, Boston, USA
Oscar Mayora Create-Net Research Consortium, Italy
John-Jules Meijer University of Utrecht, the Netherlands
Louis-Philippe Morency Institute for Creative Technologies, USC, USA
Florian 'Floyd' Mueller University of Melbourne, Australia
Patrick Olivier University of Newcastle, UK
Paolo Petta Medical University of Vienna, Austria
Fabio Pianesi ITC-irst, Trento, Italy
Helmut Prendinger National Institute of Informatics, Tokyo, Japan
Matthias Rauterberg University of Eindhoven, the Netherlands
Isaac Rudomin Monterrey Institute of Technology, Mexico
Pieter Spronck University of Tilburg, the Netherlands
Oliviero Stock ITC-irst, Trento, Italy
Carlo Strapparava ITC-irst, Trento, Italy
Mariet Theune University of Twente, the Netherlands
Thanos Vasilikos University of Western Macedonia, Greece
Sean White Columbia University, USA
Woontack Woo Gwangju Institute of Science and Technology, Korea
Wijnand IJsselstein University of Eindhoven, the Netherlands
Massimo Zancanaro ITC-irst, Trento, Italy


Back to Top

8-19 . (2009-06-24) DIAHOLMIA 2009: THE 13TH WORKSHOP ON THE SEMANTICS AND PRAGMATICS OF DIALOGUE

DIAHOLMIA 2009: THE 13TH WORKSHOP ON THE SEMANTICS AND PRAGMATICS OF DIALOGUE

KTH, Stockholm, Sweden, 24-26 June, 2009

The SemDial series of workshops aims to bring together researchers working on the semantics and pragmatics of dialogue in fields such as artificial intelligence, computational linguistics, formal semantics/pragmatics, philosophy, psychology, and neural science. DiaHolmia will be the 13th workshop in the SemDial series, and will be organized at the Department of Speech Music and Hearing, KTH (Royal Institute of Technology). KTH is Scandinavia's largest institution of higher education in technology and is located in central Stockholm (Holmia in Latin).

WEBSITE: www.diaholmia.org

DATES AND DEADLINES:

Full 8-page papers:
Submission due: 22 March 2009
Notification of acceptance: 25 April 2009
Final version due: 7 May 2009

2-page poster or demo descriptions:
Submission due: 25 April 2009
Notification of acceptance: 7 May 2009

DiaHolmia 2009: 24-26 June 2009 (Wednesday-Friday)

SCOPE:

We invite papers on all topics related to the semantics and pragmatics of dialogues, including, but not limited to:

- common ground/mutual belief
- turn-taking and interaction control
- dialogue and discourse structure
- goals, intentions and commitments
- natural language understanding/semantic interpretation
- reference, anaphora and ellipsis
- collaborative and situated dialogue
- multimodal dialogue
- extra- and paralinguistic phenomena
- categorization of dialogue phenomena in corpora
- designing and evaluating dialogue systems
- incremental, context-dependent processing
- reasoning in dialogue systems
- dialogue management

Full papers will be in the usual 8-page, 2-column format. There will also be poster and demo presentations. The selection of posters and demos will be based on 2-page descriptions. Selected descriptions will be included in the proceedings.

Details on the invited speakers, submission format and procedure and local arrangements will be announced at a later date.

PROGRAMME COMMITTEE:

Jan Alexandersson, Srinivas Bangalore, Ellen Gurman Bard, Anton Benz, Johan Bos, Johan Boye, Harry Bunt, Donna Byron, Jean Carletta, Robin Cooper, Paul Dekker, Giuseppe Di Fabbrizio, Raquel Fernández, Claire Gardent, Simon Garrod, Jonathan Ginzburg, Pat Healey, Peter Heeman, Joris Hulstijn, Michael Johnston, Kristiina Jokinen, Arne Jönsson, Alistair Knott, Ivana Kruijff-Korbayova, Staffan Larsson, Oliver Lemon, Ian Lewin, Diane Litman, Susann Luperfoy, Colin Matheson, Nicolas Maudet, Michael McTear, Wolfgang Minker, Philippe Muller, Fabio Pianesi, Martin Pickering, Manfred Pinkal, Paul Piwek, Massimo Poesio, Alexandros Potamianos, Matthew Purver, Manny Rayner, Hannes Rieser, Laurent Romary, Alex Rudnicky, David Schlangen, Stephanie Seneff, Ronnie Smith, Mark Steedman, Amanda Stent, Matthew Stone, David Traum, Marilyn Walker and Mats Wirén

ORGANIZING COMMITTEE:

Jens Edlund
Joakim Gustafson
Anna Hjalmarsson
Gabriel Skantze 

Back to Top

8-20 . (2009-07-09) MULTIMOD 2009 Multimodality of communication in children: gestures, emotions, language and cognition

The Multimod 2009 conference - Multimodality of communication in children:
gestures, emotions, language and cognition is being organized jointly by
psychologists and linguists from the Universities of Toulouse (Toulouse II)
and Grenoble (Grenoble III) and will take place in Toulouse (France) from
Thursday 9th July to Saturday 11th July 2009.

The aim of the conference will be to assess research on theories, concepts
and methods relating to multimodality in children.

The invited speakers are :
- Susan Goldin-Meadow (University of Chicago, USA),
- Jana Iverson (University of Pittsburg, USA),
- Paul Harris (Harvard University, USA),
- Judy Reilly (San Diego State University, USA),
- Gwyneth Doherty-Sneddon (University of Stirling, UK),
- Marianne Gullberg (MPI Nijmegen, The Netherlands).

We invite you to submit proposals for symposia, individual papers or posters
of original, previously unpublished research on all aspects of multimodal
communication in children, including:

- Gestures and language development, both typical and atypical
- Emotional development, both typical and atypical
- Multimodality of communication and bilingualism
- Gestural and/or emotional communication in non-human and human primates
- Multimodality of communication and didactics
- Multimodality of communication in the classroom
- Multimodality of communication and brain development
- Prosodic (emotional) aspects of language and communication development
- Pragmatic aspects of multimodality development

Please visit the conference website
http://w3.eccd.univ-tlse2.fr/multimod2009/ to find all useful Information
about submissions (individual papers, posters and symposia); the deadline
for submissions is December 15th, 2008. 

Back to Top

8-21 . (2009-08-02) ACL-IJCNLP 2009 1st Call for Papers

ACL-IJCNLP 2009 1st Call for Papers

Joint Conference of
the 47th Annual Meeting of the Association for Computational Linguistics
and
the 4th International Joint Conference on Natural Language Processing of
the Asian Federation of Natural Language Processing

August 2 - 7, 2009
Singapore

http://www.acl-ijcnlp-2009.org

Full Paper Submission Deadline:  February 22, 2009 (Sunday)
Short Paper Submission Deadline:  April 26, 2009 (Sunday)

For the first time, the flagship conferences of the Association of
Computational Linguistics (ACL) and the Asian Federation of Natural
Language Processing (AFNLP) --the ACL and IJCNLP -- are jointly
organized as a single event. The conference will cover a broad
spectrum of technical areas related to natural language and
computation. ACL-IJCNLP 2009 will include full papers, short papers,
oral presentations, poster presentations, demonstrations, tutorials,
and workshops. The conference invites the submission of papers on
original and unpublished research on all aspects of computational
linguistics.

Important Dates:

* Feb 22, 2009    Full paper submissions due;
* Apr 12, 2009    Full paper notification of acceptance;
* Apr 26, 2009    Short paper submissions due;
* May 17, 2009    Camera-ready full papers due;
* May 31, 2009    Short Paper notification of acceptance;
* Jun 7, 2009       Camera-ready short papers due;
* Aug 2-7, 2009   ACL-IJCNLP 2009

Topics of interest:

Topics include, but are not limited to:

* Phonology/morphology, tagging and chunking, and word segmentation
* Grammar induction and development
* Parsing algorithms and implementations
* Mathematical linguistics and grammatical formalisms
* Lexical and ontological semantics
* Formal semantics and logic
* Word sense disambiguation
* Semantic role labeling
* Textual entailment and paraphrasing
* Discourse, dialogue, and pragmatics
* Language generation
* Summarization
* Machine translation
* Information retrieval
* Information extraction
* Sentiment analysis and opinion mining
* Question answering
* Text mining and natural language processing applications
* NLP in vertical domains, such as biomedical, chemical and legal text
* NLP on noisy unstructured text, such as email, blogs, and SMS
* Spoken language processing
* Speech recognition and synthesis
* Spoken language understanding and generation
* Language modeling for spoken language
* Multimodal representations and processing
* Rich transcription and spoken information retrieval
* Speech translation
* Statistical and machine learning methods
* Language modeling for text processing
* Lexicon and ontology development
* Treebank and corpus development
* Evaluation methods and user studies
* Science of annotation

Submissions:

Full Papers: Submissions must describe substantial, original,
completed and unpublished work. Wherever appropriate, concrete
evaluation and analysis should be included. Submissions will be judged
on correctness, originality, technical strength, significance,
relevance to the conference, and interest to the attendees. Each
submission will be reviewed by at least three program committee
members.

Full papers may consist of up to eight (8) pages of content, plus one
extra page for references, and will be presented orally or as a poster
presentation as determined by the program committee.  The decisions as
to which papers will be presented orally and which as poster
presentations will be based on the nature rather than on the quality
of the work. There will be no distinction in the proceedings between
full papers presented orally and those presented as poster
presentations.

The deadline for full papers is February 22, 2009 (GMT+8). Submission
is electronic using paper submission software at:
https://www.softconf.com/acl-ijcnlp09/papers

Short papers: ACL-IJCNLP 2009 solicits short papers as well. Short
paper submissions must describe original and unpublished work. The
short paper deadline is just about three months before the conference
to accommodate the following types of papers:

* A small, focused contribution
* Work in progress
* A negative result
* An opinion piece
* An interesting application nugget

Short papers will be presented in one or more oral or poster sessions,
and will be given four pages in the proceedings. While short papers
will be distinguished from full papers in the proceedings, there will
be no distinction in the proceedings between short papers presented
orally and those presented as poster presentations. Each short paper
submission will be reviewed by at least two program committee members.
The deadline for short papers is April 26, 2009 (GMT + 8).  Submission
is electronic using paper submission software at:
https://www.softconf.com/acl-ijcnlp09/shortpapers

Format:

Full paper submissions should follow the two-column format of
ACL-IJCNLP 2009 proceedings without exceeding eight (8) pages of
content plus one extra page for references.  Short paper submissions
should also follow the two-column format of ACL-IJCNLP 2009
proceedings, and should not exceed four (4) pages, including
references. We strongly recommend the use of ACL LaTeX style files or
Microsoft Word style files tailored for this year's conference, which
are available on the conference website under Information for Authors.
Submissions must conform to the official ACL-IJCNLP 2009 style
guidelines, which are contained in the style files, and they must be
electronic in PDF.

As the reviewing will be blind, the paper must not include the
authors' names and affiliations. Furthermore, self-references that
reveal the author's identity, e.g., "We previously showed (Smith,
1991) ...", must be avoided. Instead, use citations such as "Smith
previously showed (Smith, 1991) ...". Papers that do not conform to
these requirements will be rejected without review.

Multiple-submission policy:

Papers that have been or will be submitted to other meetings or
publications must provide this information at submission time. If
ACL-IJCNLP 2009 accepts a paper, authors must notify the program
chairs by April 19, 2009 (full papers) or June 7, 2009 (short papers),
indicating which meeting they choose for presentation of their work.
ACL-IJCNLP 2009 cannot accept for publication or presentation work
that will be (or has been) published elsewhere.

Mentoring Service:

ACL is providing a mentoring (coaching) service for authors from
regions of the world where English is less emphasized as a language of
scientific exchange. Many authors from these regions, although able to
read the scientific literature in English, have little or no
experience in writing papers in English for conferences such as the
ACL meetings. The service will be arranged as follows. A set of
potential mentors will be identified by Mentoring Service Chairs Ng,
Hwee Tou (NUS, Singapore) and Reeder, Florence (Mitre, USA), who will
organize this service for ACL-IJCNLP 2009. If you would like to take
advantage of the service, please upload your paper in PDF format by
January 14, 2009 for long papers and March 18 2009 for short papers
using the paper submission software for mentoring service which will
be available at conference website.

An appropriate mentor will be assigned to your paper and the mentor
will get back to you by February 8 for long papers or April 12 for
short papers, at least 2 weeks before the deadline for the submission
to the ACL-IJCNLP 2009 program committee.

Please note that this service is for the benefit of the authors as
described above. It is not a general mentoring service for authors to
improve the technical content of their papers.

If you have any questions about this service please feel free to send
a message to Ng, Hwee Tou (nght[at]comp.nus.edu.sg) or Reeder,
Florence (floreederacl[at]yahoo.com).

General Conference Chair:
Su, Keh-Yih (Behavior Design Corp., Taiwan; kysu[at]bdc.com.tw)

Program Committee Chairs:
Su, Jian (Institute for Infocomm Research, Singapore;
sujian[at]i2r.a-star.edu.sg)
Wiebe, Janyce (University of Pittsburgh, USA; janycewiebe[at]gmail.com)

Area Chairs:
Agirre, Eneko (University of Basque Country, Spain; e.agirre[at]ehu.es)
Ananiodou, Sophia (University of Manchester, UK;
      sophia.ananiadou[at]manchester.ac.uk)
Belz, Anja (University of Brighton, UK; a.s.belz[at]itri.brighton.ac.uk)
Carenini, Giuseppe (University of British Columbia, Canada;
carenini[at]cs.ubc.ca)
Chen, Hsin-Hsi (National Taiwan University, TaiWan, hh_chen[at]csie.ntu.edu.tw)
Chen, Keh-Jiann (Sinica, Tai Wan, kchen[at]iis.sinica.edu.tw)
Curran, James (University of Sydney, Australia; james[at]it.usyd.edu.au)
Gao, Jian Feng (MSR, USA; jfgao[at]microsoft.com)
Harabagiu, Sanda (University of Texas at Dallas, USA, sanda[at]hlt.utdallas.edu)
Koehn, Philipp (University of Edinburgh, UK; pkoehn[at]inf.ed.ac.uk)
Kondrak, Grzegorz (University of Alberta, Canada; kondrak[at]cs.ualberta.ca)
Meng, Helen Mei-Ling (Chinese University of Hong Kong, Hong Kong;
      hmmeng[at]se.cuhk.edu.hk )
Mihalcea, Rada (University of Northern Texas, USA; rada[at]cs.unt.edu)
Poesio, Massimo(University of Trento, Italy; poesio[at]disi.unitn.it)
Riloff, Ellen (University of Utah, USA; riloff[at]cs.utah.edu)
Sekine, Satoshi (New York University, USA; sekine[at]cs.nyu.edu)
Smith, Noah (CMU, USA; nasmith[at]cs.cmu.edu)
Strube, Michael (EML Research, Germany; strube[at]eml-research.de)
Suzuki, Jun (NTT, Japan; jun[at]cslab.kecl.ntt.co.jp)
Wang, Hai Feng (Toshiba, China; wanghaifeng[at]rdc.toshiba.com.cn) 

Back to Top

8-22 . (2009-09-11) SIGDIAL 2009 CONFERENCE

SIGDIAL 2009 CONFERENCE 
10th Annual Meeting of the Special Interest Group on Discourse and Dialogue 
Queen Mary University of London, UK September 11-12, 2009 
(right after Interspeech 2009) 
 
 
Submission Deadline: April 24, 2009 
 
 
PRELIMINARY CALL FOR PAPERS 
 
Following a series of nine successful workshops, SIGDIAL is transitioning into a conference. The 
SIGDIAL venue provides a regular forum for the presentation of cutting edge research in discourse and 
dialogue to both academic and industry researchers. The conference is sponsored by the SIGDIAL 
organization, which serves as the Special Interest Group in discourse and dialogue for both ACL and 
ISCA. SIGDIAL 2009 will be co-located with Interspeech 2009 as a satellite event. 
 
TOPICS OF INTEREST 
 
We welcome formal, corpus-based, implementation, experimental, or analytical work on discourse and 
dialogue including, but not restricted to, the following themes: 
 
1. Discourse Processing and Dialogue Systems
 
Discourse semantic and pragmatic issues in NLP applications such as text summarization, 
question answering, information retrieval including topics like:
* Discourse structure, temporal structure, information structure ;
* Discourse markers, cues and particles and their use;
* (Co-)Reference and anaphora resolution, metonymy and bridging resolution;
* Subjectivity, opinions and semantic orientation; 
 
 
Spoken, multi-modal, and text/web based dialogue systems including topics such as:
* Dialogue management models;
* Speech and gesture, text and graphics integration;
* Strategies for preventing, detecting or handling miscommunication (repair and correction 
types, clarification and under-specificity, grounding and feedback strategies);
* Utilizing prosodic information for understanding and for disambiguation; 
 
2. Corpora, Tools and Methodology 
 
Corpus-based and experimental work on discourse and spoken, text-based and multi-modal 
dialogue including its support, in particular:
* Annotation tools and coding schemes;
* Data resources for discourse and dialogue studies;
* Corpus-based techniques and analysis (including machine learning);
* Evaluation of systems and components, including methodology, metrics and case studies;
 
3. Pragmatic and/or Semantic Modeling
 
The pragmatics and/or semantics of discourse and dialogue (i.e. beyond a single sentence) 
including the following issues: 
 
* The semantics/pragmatics of dialogue acts (including those which are less studied in the 
semantics/pragmatics framework);
* Models of discourse/dialogue structure and their relation to referential and relational 
structure;
* Prosody in discourse and dialogue;
* Models of presupposition and accommodation; operational models of conversational 
implicature.
 
SUBMISSIONS 
 
The program committee welcomes the submission of long papers for full plenary presentation as well as 
short papers and demonstrations. Short papers and demo descriptions will be featured in short plenary 
presentations, followed by posters and demonstrations.
* Long papers must be no longer than 8 pages, including title, examples, references, etc. In 
addition to this, two additional pages are allowed as an appendix which may include extended 
example discourses or dialogues, algorithms, graphical representations, etc.
* Short papers and demo descriptions should be 4 pages or less (including title, examples, 
references, etc.). 
 
Please use the official ACL style files: http://ufal.mff.cuni.cz/acl2007/styles/ 
 
Papers that have been or will be submitted to other meetings or publications must provide this 
information (see submission format). SIGDIAL 2009 cannot accept for publication or presentation work 
that will be (or has been) published elsewhere.  Any questions regarding submissions can be sent to the 
General Co-Chairs. 
 
Authors are encouraged to make illustrative materials available, on the web or otherwise. For example, 
excerpts of recorded conversations, recordings of human-computer dialogues, interfaces to working 
systems, etc. 
 
NEW INITIATIVE 
 
In order to recognize significant advancements in dialog and discourse science and technology, SIGDIAL 
will recognize a BEST PAPER AWARD and a BEST STUDENT PAPER AWARD. A selection committee 
consisting of prominent researchers in the fields of interest will select the recipients of the awards. 
 
IMPORTANT DATES (SUBJECT TO CHANGE) 
 
Submission: April 24, 2009 
Workshop: September 11-12, 2009 
 
WEBSITES 
 
SIGDIAL organization website: http://www.sigdial.org
Interspeech 2009 website: http://www.interspeech2009.org/
 
ORGANIZING COMMITTEE 
 
For any questions, please contact the appropriate members of the organizing committee: 
 
GENERAL CO-CHAIRS
Pat Healey, Queen Mary University of London, ph@dcs.qmul.ac.uk
Roberto Pieraccini, SpeechCycle, roberto@speechcycle.com
TECHNICAL PROGRAM CO-CHAIRS
Donna Byron, Ohio State University, dbyron@cse.ohio-state.edu
Steve Young, University of Cambridge, sjy@eng.cam.ac.uk
LOCAL CHAIR
       Matt Purver, mpurver@dcs.qmul.ac.uk
SIGDIAL PRESIDENT
Tim Paek, Microsoft Research, timpaek@microsoft.com
SIGDIAL VICE PRESIDENT
Amanda Stent, amanda.stent@gmail.com
Back to Top

8-23 . (2009-09-14) 7th International Conference on Recent Advances in Natural Language Processing

First Call for Papers

================================================

7th International Conference on Recent Advances in Natural Language Processing

RANLP-2009

14-16 September 2009   Borovets, Bulgaria

http://www.lml.bas.bg/ranlp2009

================================================

 

 

The conference will take the form of addresses from invited keynote speakers plus peer-reviewed individual papers. All accepted papers will be published in the conference proceedings. Selected papers are traditionally published by John Benjamins Publishers. There will also be an exhibition area for poster and demo sessions.

 

The conference will be preceded by two days of tutorials (12-13 September 2009). Post-conference workshops will be held during 17-18 September 2009.

 

 

TOPICS

 

We invite papers reporting on recent advances in all aspects of Natural Language Processing (NLP). We encourage the representation of a broad range of areas including but not limited to: phonetics, phonology, and morphology; syntax; semantics; discourse, dialogue and pragmatics;  terminology; electronic dictionaries; mathematical models and complexity; corpus-based language processing; text understanding and generation; POS tagging; parsing; semantic processing; temporal processing; textual entailment; knowledge acquisition; word-sense disambiguation; mention detection; anaphora resolution/coreference; relation extraction; meronymy identification; parallel NLP algorithms; speech recognition; text-to-speech synthesis; multilingual NLP; machine translation, machine-aided translation, translation memory systems, translation aids and tools; information retrieval; text and web mining; information extraction; text summarisation; term recognition; text categorisation; question answering; opinion mining and sentiment analysis; demographic analysis; dialogue systems; computer-aided language learning; biomedical NLP; language resources; evaluation; visualisation; and theoretical and application-orientated papers related to NLP of every kind.

 

 

KEYNOTE SPEAKERS

 

-- Shalom Lappin (King's College, London)

-- Massimo Poesio (University of Essex)

-- Mirella Lapata (University of Edinburgh)

 

 

TUTORIALS

 

Tutorial lecturers will be listed in the 2nd CFP.

 

 

PROGRAMME COMMITTEE

 

The PC members are distinguished experts from all over the world. The list of PC members will be announced in the 2nd CFP.

 

 

WORKSHOPS

 

Several workshops will be organised after the main conference. A call for workshop proposals will be sent out shortly.

 

 

SUBMISSION (PAPERS, POSTERS, DEMOS)

 

Submissions should be A4, two-column format and should not exceed 7 pages (poster and demo submissions should be no longer than 5 pages), including main body, figures, tables and bibliography. Times New Roman 12 font is preferred. The first page should state the title of the paper followed by keywords and an abstract and continue with the first section of the paper. Articles should be submitted electronically in **PDF** format. For free conversions to PDF see https://createpdf.adobe.com/index.pl?BP=NS The review process is annonymous. Submissions will be reviewed by 3 members of the Programme Committee. Submission and camera ready copy guidelines will be also available at the conference web site.

 

 

IMPORTANT DATES

 

Workshop proposals due: 16 February 2009

Workshop selection: 2 March 2009

 

Conference paper registration (abstracts): 6 April 2009

Conference paper submission deadline: 13 April 2009

Conference paper acceptance notification: 1 June 2009

Conference camera ready papers due: 13 July 2009

 

Workshop paper submission deadline (tentative): 3 June 2009

Workshop paper acceptance notification (tentative): 20 July 2009

Workshop camera ready papers due (tentative): 24 August 2009

 

Tutorials: 12-13 September 2009 (Sat-Sun)

Main conference: 14-16 September 2009 (Mon-Wed)

Workshops: 17-18 September 2009 (Thu-Fri)

 

 

VENUE

 

Hotel Samokov, Borovets. The picturesque resort of Borovets is located in the Rila mountains and is one of the best known winter resorts in South-East Europe, a frequent meeting place for the elite in world skiing. The resort is 1350m above sea level, at the foot of the highest peak on the Balkan Peninsula - Moussala (2925m). The resort of Borovets is 73km from Sofia and the Sofia International Airport can serve as arrival/departure point. In addition to regular public transport, the organisers will provide daily shuttle buses from Sofia airport to the conference location at an inexpensive rate. A taxi from Sofia to Borovets is relatively cheap.

 

 

THE TEAM BEHIND RANLP-2009

 

-- Galia Angelova, Bulgarian Academy of Sciences, Bulgaria (Organising Committee Chair)

-- Kalina Bontcheva, University of Sheffield, U.K.

-- Ruslan Mitkov, University of Wolverhampton, U.K. (Programme Committee Chair)

-- Nicolas Nicolov, J.D.Power and Associates [McGraw-Hill company], U.S.A. (Editor of the volume with selected papers)

-- Nikolai Nikolov, INCOMA Ltd., Shoumen, Bulgaria

-- Kiril Simov, Bulgarian Academy of Sciences, Bulgaria (Workshop Coordinator)

 

The main local organiser is the Linguistic Modelling Department, Institute for Parallel Processing, Bulgarian Academy of Sciences.

 

CONTACT: ranlp2009[AT]lml[dot]bas[dot]bg

Dr Nicolas Nicolov

Senior Director

J.D. Power and Associates [McGraw-Hill company]

Web Intelligence REsearch Division (WIRED)

4888 Pearl East Circle, Suite 300W

Boulder, CO 80301, U.S.A.

Back to Top

8-24 . (2009-10-18) 2009 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics

Call for Papers

2009 IEEE Workshop on Applications of Signal Processing to Audio and

Acoustics

 

Mohonk Mountain House

New Paltz, New York

October 18-21, 2009

http://www.waspaa2009.com

 

The 2009 IEEE Workshop on Applications of Signal Processing to Audio and

Acoustics (WASPAA'09) will be held at the Mohonk Mountain House in New

Paltz, New York, and is sponsored by the Audio & Electroacoustics committee

of the IEEE Signal Processing Society. The objective of this workshop is to

provide an informal environment for the discussion of problems in audio and

acoustics and the signal processing techniques leading to novel solutions.

Technical sessions will be scheduled throughout the day. Afternoons will be

left free for informal meetings among workshop participants.

 

Papers describing original research and new concepts are solicited for

technical sessions on, but not limited to, the following topics:

 

* Acoustic Scenes

- Scene Analysis: Source Localization, Source Separation, Room Acoustics

- Signal Enhancement: Echo Cancellation, Dereverberation, Noise Reduction,

Restoration

- Multichannel Signal Processing for Audio Acquisition and Reproduction

- Microphone Arrays

- Eigenbeamforming

- Virtual Acoustics via Loudspeakers

 

* Hearing and Perception

- Auditory Perception, Spatial Hearing, Quality Assessment

- Hearing Aids

 

* Audio Coding

- Waveform Coding and Parameter Coding

- Spatial Audio Coding

- Internet Audio

- Musical Signal Analysis: Segmentation, Classification, Transcription

- Digital Rights

- Mobile Devices

 

* Music

- Signal Analysis and Synthesis Tools

- Creation of Musical Sounds: Waveforms, Instrument Models, Singing

- MEMS Technologies for Signal Pick-up

 

 

Submission of four-page paper: April 15, 2009

Notification of acceptance: June 26, 2009

Early registration until:  September 1, 2009

 

Workshop Committee

 

General Co-Chair:

Jacob Benesty

Université du Québec

INRS-EMT

Montréal, Québec, Canada

benesty@emt.inrs.ca

 

General Co-Chair:

Tomas Gaensler

mh acoustics

Summit, NJ, USA

tfg@mhacoustics.com

 

Technical Program Chair:

Yiteng (Arden) Huang

WeVoice Inc.

Bridgewater, NJ, USA

arden_huang@ieee.org

 

Technical Program Chair:

Jingdong Chen

Bell Labs

Alcatel-Lucent

Murray Hill, NJ, USA

jingdong@research.bell-labs.com

 

Finance Chair:

Michael Brandstein

Information Systems

Technology Group

MIT Lincoln Lab

Lexington, MA, USA

msb@ll.mit.edu

 

Publications Chair:

Eric J. Diethorn

Multimedia Technologies

Avaya Labs Research

Basking Ridge, NJ, USA

ejd@avaya.com

 

Publicity Chair:

Sofiène Affes

Université du Québec

INRS-EMT

Montréal, Québec, Canada

affes@emt.inrs.ca

 

Local Arrangements Chair:

Heinz Teutsch

Multimedia Technologies

Avaya Labs Research

Basking Ridge, NJ, USA

teutsch@avaya.com

 

Far East Liaison:

Shoji Makino

NTT Communication Science

Laboratories, Japan

maki@cslab.kecl.ntt.co.jp

Back to Top

8-25 . (2009-12-13) ASRU 2009

IEEE ASRU 2009, December 13-17, Merano, Italy
Back to Top

8-26 . (2009-12-14) 6th International Workshop on Models and Analysis of Vocal Emissions for Biomedical Applications MAVEBA 2009

University degli Studi di Firenze Italy
Department of Electronics and Telecommunications
6th International Workshop
on
Models and Analysis of Vocal Emissions for Biomedical
Applications
MAVEBA 2009
December 14 - 16, 2009
Firenze, Italy
http://maveba.det.unifi.it
Speech is the primary means of communication among humans, and results from
complex interaction among vocal folds vibration at the larynx and voluntary articulators
movements (i.e. mouth tongue, jaw, etc.). However, only recently has research
focussed on biomedical applications. Since 1999, the MAVEBA Workshop is
organised every two years, aiming to stimulate contacts between specialists active in
clinical, research and industrial developments in the area of voice signal and images
analysis for biomedical applications. This sixth Workshop will offer the participants
an interdisciplinary platform for presenting and discussing new knowledge in the field
of models, analysis and classification of voice signals and images, as far as both
adults, singing and children voices are concerned. Modelling the normal and
pathological voice source, analysis of healthy and pathological voices, are among the
main fields of research. The aim is that of extracting the main voice characteristics,
together with their deviation from “healthy conditions”, ranging from fundamental
research to all kinds of biomedical applications and related established and advanced
technologies.
SCIENTIFIC PROGRAM
linear and non-linear models of voice
signals;
physical and mechanical models;
aids for disabled;
measurement devices (signal and image); prostheses;
robust techniques for voice and glottal
analysis in time, frequency, cepstral,
wavelet domain;
neural networks, artificial intelligence and
other advanced methods for pathology
classification;
linguistic and clinical phonetics; new-born infant cry analysis;
neurological dysfunction; multiparametric/multimodal analysis;
imaging techniques (laryngography,
videokymography, fMRI);
voice enhancement;
protocols and database design;
Industrial applications in the biomedical
field;
singing voice;
speech/hearing interactions;
DEADLINES
30 May 2009 Submission of extended abstracts (1-2 pages, 1 column)
/special session proposal
30 July 2009 Notification of paper acceptance
30 September 2009 Final full paper submission (4 pages, 2 columns, pdf
format) and early registration
14-16 December 2009 Conference venue
SPONSORS
ENTE CRF Ente Cassa di Risparmio di Firenze
IEEE EMBS
IEEE Engineering in Medicine and Biology
Society
ELSEVIER Eds.
Biomedical Signal Processing and Control
ISCA
International Speech and Communication
Association
A.I.I.M.B.
Associazione Italiana di Ingegneria Medica e
Biologica
COST Action
2103 Europ. COop. in Science & Tech. Research
FURTHER INFORMATION
Claudia Manfredi – Conference Chair
Department of Electronics and
Telecommunications
Via S. Marta 3, 50139 Firenze, Italy
Phone: +39-055-4796410
Fax: +39-055-494569
E-mail: claudia.manfredi@unifi.it
Piero Bruscaglioni
Department of Physics
Polo Scientifico Sesto Fiorentino, 50019
Firenze,Italy
Phone: +39-055-4572038
Fax: +39-055-4572356
E-mail: piero.bruscaglioni@unifi.it
Back to Top