Overview of the 2015 Workshop on Speech, Language and Audio in Multimedia

Guillaume Gravier, Gareth J.F. Jones, Martha Larson, Roeland J.F. Ordelman

    Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

    4 Downloads (Pure)

    Abstract

    The Workshop on Speech, Language and Audio in Multimedia (SLAM) positions itself at at the crossroad of multiple scientific fields - music and audio processing, speech processing, natural language processing and multimedia - to discuss and stimulate research results, projects, datasets and benchmarks initiatives where audio, speech and language are applied to multimedia data. While the first two editions were collocated with major speech events, SLAM’15 is deeply rooted in the multimedia community, opening up to computer vision and multimodal fusion. To this end, the workshop emphasizes video hyperlinking as an showcase where computer vision meets speech and language. Such techniques provide a powerful illustration of how multimedia technologies incorporating speech, language and audio can make multimedia content collections better accessible, and thereby more useful, to users.
    Original languageEnglish
    Title of host publicationProceedings of the 23rd ACM International Conference on Multimedia (MM 2015)
    Place of PublicationNew York
    PublisherAssociation for Computing Machinery (ACM)
    Pages1347-1348
    Number of pages2
    ISBN (Print)978-1-4503-3459-4
    DOIs
    Publication statusPublished - Oct 2015
    Event23rd ACM Multimedia Conference, MM 2015 - Brisbane, Australia
    Duration: 26 Oct 201530 Oct 2015
    Conference number: 23

    Conference

    Conference23rd ACM Multimedia Conference, MM 2015
    Abbreviated titleMM
    CountryAustralia
    CityBrisbane
    Period26/10/1530/10/15

    Keywords

    • HMI-MR: MULTIMEDIA RETRIEVAL
    • EWI-26657
    • Language
    • METIS-315141
    • IR-98930
    • Audio
    • Multi-media
    • Speech

    Fingerprint Dive into the research topics of 'Overview of the 2015 Workshop on Speech, Language and Audio in Multimedia'. Together they form a unique fingerprint.

    Cite this