| Preface On behalf of the program committee for the LREC 
                    2004 "Workshop on the Processing of Sign 
                    Languages", we are pleased to present you 
                    with the proceedings which contain the papers 
                    accepted for presentation at the Lisbon meeting 
                    on May 30th, 2004.
 
 This volume, full of eye-catching signs, symbols, 
                    robots and screen-shots may charmingly attract 
                    readers who, although having a sound knowledge 
                    of Natural Language Processing, might be confused 
                    by the great variety of topics and approaches. 
                    How do SignWriting, avatars, XML, videos and 
                    image recognition fit together? Are they competitive 
                    approaches or different solutions to different 
                    problems? Where will future research lead us, 
                    which endeavours answer real social needs and 
                    which scenarios are still illusionary - or congenially 
                    visionary?
 
 As always, the answers to these questions lie 
                    between slow and quick, up and down, straight 
                    and curbed. It is by drawing analogies to the 
                    processing of spoken languages that we might 
                    better understand the contribution and benefits 
                    of the different approaches, span the space 
                    of possible research and identify future tendencies 
                    in the research on the processing of sign languages.
 
 Trivially speaking, spoken languages are spoken 
                    and heard. Sign languages are signed and seen. 
                    Spoken languages have been written on stone, 
                    wood, paper and electronic media. The technical 
                    support ranged from a chisel to a keyboard. 
                    The writing systems which developed have been 
                    under the influence of the particular language 
                    and the technical support. Having a hammer in 
                    your right and a chisel in the left makes it 
                    difficult to write from left to right. Having 
                    stable vowels motivates their representation 
                    in the written form. So how can sign languages 
                    be written for love letters, poems, verdicts 
                    and recipes?
 
 One possible answer is SignWriting. SignWriting 
                    does not decompose a sign into phonemes, syllables 
                    or morphemes but body-parts, movements and face 
                    expressions and assigns a representation to 
                    each of them. Given such representations - e.g. 
                    an alphabet for potentially all sign languages 
                    -how may a keyboard, the input system, look 
                    like? How are the simple elements (body-parts, 
                    movements and face expressions) to be encoded 
                    in the computer and how the composed signs? 
                    As pictures, in Unicode or XML? How will this 
                    influence the input of signs, the layout and 
                    formatting of SignWriting documents, the possibilities 
                    to perform fuzzy matches on texts, in dictionaries, 
                    in the Internet? The papers written by Richard 
                    Gleaves, Valerie Sutton (SignWriter), Antônio 
                    Carlos da Rocha Costa, Graçaliz Pereira 
                    Dimuro, Juliano Baldez de Freitas (A Sign Matching 
                    Technique to Support Searches in Sign Language 
                    Texts), Angel Herrero (A Practical Writing System 
                    for Sign Languages), Steven Aerts, Bart Braem, 
                    Katrien Van Mulders, Kristof De Weerdt (Searching 
                    SignWriting Signs), Daniel Thomas Ulrich Noelpp 
                    (Development of a new 'SignWriter' Program) 
                    discuss these and related questions.
 
 SignWriting, however, is by no means the only 
                    possible way of writing signs. Thomas Hanke 
                    in his invited talk “HamNoSys – 
                    Representing Sign Language Data in Language 
                    Resources and Language Processing Contexts” 
                    introduces an alternative approach, the Hamburg 
                    Notation System for Sign Languages. The purpose 
                    of HamNoSys has never been a usage in everyday 
                    communication. It was designed to comply with 
                    research requirements, e.g. for corpus annotation, 
                    sign generation, machine translation and dictionary 
                    construction. It thus differs from SignWriting 
                    in its scope and granularity. Unicode and XML 
                    solutions are available for HamNoSys, c.f. Ralph 
                    Elliott, John Glauert, Vince Jennings and Richard 
                    Kennaway in their contribution “An Overview 
                    of the SiGML Notation and SiGMLSigning Software 
                    System”.
 
 Once these fundamental questions regarding the 
                    writing of sign languages will be settled, derived 
                    notions such as word n-grams and character n-grams, 
                    important for computational approaches, may 
                    be used for applications such as language recognition, 
                    document classification and information retrieval. 
                    Spelling checking, syntax checking and parsing 
                    are obvious further developments once these 
                    more fundamental questions about the writing 
                    of signs will have been agreed upon.
 It is a matter of fact, however, that most signers 
                    have not been trained in reading or writing 
                    in SignWriting. What is known as “text-to-speech” 
                    in the processing of spoken languages would 
                    seem a possible solution: a front-end to web-pages, 
                    mail boxes etc. would sign out the written text. 
                    As shown by Maria Papadogiorgaki, Nikos Grammalidis, 
                    Nikos Sarris, Michael G. Strintzis in “Synthesis 
                    of virtual Reality Animations from SWML using 
                    MPEG-4 Body Animation Parameters” and 
                    Yiqiang Chen, Wen Gao, Changshui Yang, Dalong 
                    Jiang and Cunbao Ge in “Chinese Sign Language 
                    Synthesis and Its Applications”, avatars, 
                    i.e. virtual signers, may be constructed which 
                    translate a written form of a sign language 
                    or spoken language into signs, just like translating 
                    "d" into the corresponding sound wave.
 
 A front-end on the input side of the system 
                    might translate signs into a written representation. 
                    Speech Recognition becomes Sign Recognition. 
                    Two different techniques are introduced. The 
                    recognition with the help of a data glove precedes 
                    from the signer's perspective and his/her articulations, 
                    c.f. Jose L. Hernandez-Rebollar’s contribution 
                    “Phonetic Model for Automatic Recognition 
                    of Hand Gestures”. This approach may seem 
                    in line with the definition of phonemes in terms 
                    of their articulation and not their acoustic 
                    properties. On the other hand, it does not match 
                    our every-day experience in which we use a microphone 
                    and not electronic contact points at our vocal 
                    cords, tongue, velum, teeth and lips when using 
                    a telephone. The recognition of signs with the 
                    help of cameras, the second alternative, leads 
                    to the description of signs from the observer's 
                    point of view, in terms of formants and f0, 
                    so to say. However, the articulation can be 
                    reconstructed and might be a better representation 
                    for the signs than the ‘phonetic’ 
                    description, as suggested by Boris Lenseigne, 
                    Frédérik Gianni, and Patrice Dalle 
                    in “A New Gesture Representation for Sign 
                    Language Analysis”.
 
 Both modules, sign recognition and sign generation, 
                    may serve MT systems with a sign language as 
                    source or target language respectively. A sign 
                    language as target language is used in translation 
                    experiments described by Jan Bungeroth and Hermann 
                    Ney in “Statistical Sign Language Translation”. 
                    This corpus-based approach to Machine Translation, 
                    by the way, raises the question of sign language 
                    corpora. The only paper which really tackles 
                    the question of signed corpora in this collection 
                    is that of Onno Crasborn, Els van der Kooij, 
                    Daan Broeder, Hennie Brugman “Sharing 
                    sing language corpora online. Proposals for 
                    transcription and metadata”. Matt Huenerfauth 
                    in his contribution “Spatial Representations 
                    for Generating Classifiers Predicates in an 
                    English to American Sign Language Machine Translation 
                    System”, focuses on a particularly difficult 
                    aspect of sign language generation, the classifier 
                    predicates. Thus, when signing "leaves 
                    are falling", it is not enough to generate 
                    the sign "leave" and "falling", 
                    e.g. a downward movement. Instead the hand shape 
                    of "falling" should indicate the kind 
                    of object that is falling, e.g. with a flat 
                    hand.
 
 The usage of classifiers leads us directly to 
                    the question of how to construct dictionaries 
                    for sign languages. Learners' dictionaries, 
                    reference dictionaries, dictionaries of NLP 
                    applications all need information about part 
                    of speech, lexical functions, idioms, subcategorization 
                    and semantics, which by no means is the same 
                    as in the national spoken language. How do we 
                    search in a sign language dictionary? Have you 
                    ever looked up a Chinese or Japanese Dictionary? 
                    Paola Laterza and Claudio Baj in their paper 
                    “Progetto e-LIS@” propose an at 
                    least partially equivalent approach to the ordering 
                    of signs in a sign language dictionary. How 
                    do you present the dictionary content to a learner? 
                    In the national spoken language or in SignWriting? 
                    The complexity of the question can be gauged 
                    from Elana Ochse’s contribution “A 
                    Language via Two Others, Learning English through 
                    LIS”. Should we use videos, photos, animations 
                    or drawings to represent the entries in dictionaries? 
                    A number of authors discuss these and related 
                    topics in the context of specific dictionary 
                    projects: for static presentations, i.e. paper 
                    dictionaries, Inge Zwitserlood and Doeko Hekstra 
                    propose the “Sign Printing System – 
                    SignPS” to compose pictures of signs. 
                    Eleni Efthimiou, Anna Vacalopoulou, Stavroula-Evita 
                    Ftinea, Gregory Steinhauer focus in their paper 
                    “Multipurpose Design and Creation of GSL 
                    Dictionaries” on the content, i.e. the 
                    types of information to be included in a sign 
                    language dictionary. Chiara Vettori, Oliver 
                    Streiter and Judith Knapp focus on different 
                    user requirements and the possible role of SignWriting 
                    in a sign language dictionary. Rubén 
                    Nogueira, Jose M. Martínez and present 
                    a dictionary project of a particular kind: “19th 
                    Century Signs in the Online Spanish Sign Language 
                    Library: the Historical Dictionary Project.” 
                    Ingvild Roald finally gives a practical account 
                    on the history of techniques for the creation 
                    of sign language dictionaries, discussing advantages 
                    and drawbacks of the respective approaches.
 
 When writing these lines, the preparation of 
                    the workshop and the proceedings is almost finished. 
                    This workshop wouldn’t have been possible 
                    without the energy many people have invested 
                    in their spare time. First of all we would like 
                    to thank the authors who have done their best 
                    and provided superb papers. Our thank goes also 
                    to the reviewers for their detailed and inspiring 
                    reviews. Last but not least we want to thank 
                    Sara Goggi who accompanied the workshop on behalf 
                    of the LREC Programme Committee.
 
 In closing we would like to thank you for attending 
                    the workshop, and we wish you will have a great 
                    time.
 
 Oliver Streiter
 ostreiter@eurac.edu
 
 Antônio Carlos da Rocha Costa
 rocha@ATLAS.UCPEL.TCHE.BR
 
 April 22, 2004
 |