ARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (2024)

ARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (2)

Advanced Search

Browse

Article

Free Access

  • Authors:
  • Agnes Lisowska ISSCO/TIM/ETI, University of Geneva, Geneva, Switzerland

    ISSCO/TIM/ETI, University of Geneva, Geneva, Switzerland

    View Profile

    ,
  • Martin Rajman CGC/IC, LIA/IIF/IC, Swiss Federal Institute of Technology Lausanne, Bat. INR, Lausanne, Switzerland

    CGC/IC, LIA/IIF/IC, Swiss Federal Institute of Technology Lausanne, Bat. INR, Lausanne, Switzerland

    View Profile

    ,
  • Trung H. Bui CGC/IC, LIA/IIF/IC, Swiss Federal Institute of Technology Lausanne, Bat. INR, Lausanne, Switzerland

    CGC/IC, LIA/IIF/IC, Swiss Federal Institute of Technology Lausanne, Bat. INR, Lausanne, Switzerland

    View Profile

MLMI'04: Proceedings of the First international conference on Machine Learning for Multimodal InteractionJune 2004Pages 291–304https://doi.org/10.1007/978-3-540-30568-2_25

Published:21 June 2004Publication History

  • 7citation
  • 0
  • Downloads

Metrics

Total Citations7Total Downloads0

Last 12 Months0

Last 6 weeks0

  • Get Citation Alerts

    New Citation Alert added!

    This alert has been successfully added and will be sent to:

    You will be notified whenever a record that you have chosen has been cited.

    To manage your alert preferences, click on the button below.

    Manage my Alerts

    New Citation Alert!

    Please log in to your account

  • Publisher Site

MLMI'04: Proceedings of the First international conference on Machine Learning for Multimodal Interaction

ARCHIVUS: a system for accessing the content of recorded multimodal meetings

Pages 291–304

PreviousChapterNextChapter

ARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (3)

ABSTRACT

This paper describes a multimodal dialogue driven system, ARCHIVUS, that allows users to access and retrieve the content of recorded and annotated multimodal meetings. We describe (1) a novel approach taken in designing the system given the relative inapplicability of standard user requirements elicitation methodologies, (2) the components of ARCHIVUS, and (3) the methodologies that we plan to use to evaluate the system.

References

  1. Dix, A., J. Finlay, G. Abowd and R. Beale, Human Computer Interaction Second Edition, Prentice Hall, England, 1998. Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (4)Digital Library
  2. IM2 webpage http://www.im2.ch/e/home.htmlGoogle ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (6)
  3. IM2.MDM webpage http://issco-www.unige.ch/projects/im2/mdm/Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (7)
  4. A. Lisowska, "Multimodal Interface Design for the Multimodal Meeting Domain: Preliminary Indications from a Query Analysis Study", Report IM2.MDM-11, Nov. 2003.Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (8)
  5. D. Mekhaldi, D. Lalanne and R. Ingold. "Thematic Alignment of recorded speech with documents", DocEng 2003, ACM Symposium on Document Engineering, Grenoble, 2003. Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (9)Digital Library
  6. T. H. Bui and M. Rajman "Rapid Dialogue Prototyping Methodology", Technical Report No. 200401, Swiss Federal Institute of Technology, Lausanne (Switzerland), January, 2004.Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (11)
  7. E. Bilange, Dialogue personne-machine, modélisation et réalisation informatique, Langue, Raisonnement, Calcul, Hermès, Paris, France, 1992.Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (12)
  8. IM2 newsletter, May 2004.Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (13)
  9. D. Moore, "The IDIAP Smart Meeting Room", IDIAP-Com 02-07, 2002Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (14)
  10. A. Popescu-Belis, "Dialogue act tagsets for meeting understanding: an abstraction based on the DAMSL, Switchboard and ICSI-MR tagsets", Report IM2.MDM-09, September 2003.Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (15)
  11. N. Dahlbäck, A. Jönsson and L. Ahrenberg, "Wizard of Oz Studies - Why and How", in W.D. Gray, W.E. Helfley and Murray, D. (eds). Proceedings of the 1993 Workshop on Intelligent User Interfaces (pp. 193/200) Orlando, FL. New York, ACM Press, 1993. Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (16)Digital Library
  12. D. Salber, and J Coutaz, "Applying the Wizard of Oz technique to the study of Multimodal Systems", 3rd International Conference EWHCI'93, East/West Human Computer Interaction, Moscow. L. Bass, J. Gornostaev, C. Unger Eds. Springer Verlag Publ. Lecture notes in Computer Science, Vol. 73. pp. 219-230. 1993. Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (18)Digital Library
  13. Ferret Meeting Browser http://rhonedata.idiap.ch/documentation/Ferret_User_Guide/help.htmlGoogle ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (20)
  14. M. Flynn and P. Wellner, "In Search of a Good BET", IDIAP-Com 03-11, 2003.Google ScholarARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (21)

Cited By

View all

ARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (22)

    Index Terms

    1. ARCHIVUS: a system for accessing the content of recorded multimodal meetings

      1. Human-centered computing

        1. Human computer interaction (HCI)

      Index terms have been assigned to the content through auto-classification.

      Recommendations

      • Archivus: a multimodal system for multimedia meeting browsing and retrieval

        COLING-ACL '06: Proceedings of the COLING/ACL on Interactive presentation sessions

        This paper presents Archivus, a multi-modal language-enabled meeting browsing and retrieval system. The prototype is in an early stage of development, and we are currently exploring the role of natural language for interacting in this relatively ...

      • Dialog acts in greeting and leavetaking in social talk

        ISIAA 2017: Proceedings of the 1st ACM SIGCHI International Workshop on Investigating Social Interactions with Artificial Agents

        Conversation proceeds through dialogue moves or acts, and dialog act annotation can aid the design of artificial dialog. While many dialogs are task-based or instrumental, with clear goals, as in the case of a service encounter or business meeting, ...

        Read More

      • Natural Language, Mixed-initiative Personal Assistant Agents

        IMCOM '18: Proceedings of the 12th International Conference on Ubiquitous Information Management and Communication

        The increasing popularity and use of personal voice assistant technologies, such as Siri and Google Now, is driving and expanding progress toward the long-term and lofty goal of using artificial intelligence to build human-computer dialog systems ...

        Read More

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      Get this Publication

      • Information
      • Contributors
      • Published in

        ARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (23)

        MLMI'04: Proceedings of the First international conference on Machine Learning for Multimodal Interaction

        June 2004

        359 pages

        ISBN:354024509X

        • Editors:
        • Samy Bengio

          IDIAP Research Institute, Martigny, Switzerland

          ,
        • Hervé Bourlard

          IDIAP Research Institute, Martigny, Switzerland

        Sponsors

          In-Cooperation

            Publisher

            Springer-Verlag

            Berlin, Heidelberg

            Publication History

            • Published: 21 June 2004

            Qualifiers

            • Article

            Conference

            Funding Sources

            • ARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (26)

              Other Metrics

              View Article Metrics

            • Bibliometrics
            • Citations7
            • Article Metrics

              • 7

                Total Citations

                View Citations
              • Total Downloads

              • Downloads (Last 12 months)0
              • Downloads (Last 6 weeks)0

              Other Metrics

              View Author Metrics

            • Cited By

              View all

              Digital Edition

              View this article in digital edition.

              View Digital Edition

              • Figures
              • Other

                Close Figure Viewer

                Browse AllReturn

                Caption

                View Table of Contents

                Export Citations

                  Your Search Results Download Request

                  We are preparing your search results for download ...

                  We will inform you here when the file is ready.

                  Download now!

                  Your Search Results Download Request

                  Your file of search results citations is now ready.

                  Download now!

                  Your Search Results Download Request

                  Your search export query has expired. Please try again.

                  ARCHIVUS | Proceedings of the First international conference on Machine Learning for Multimodal Interaction (2024)

                  FAQs

                  What is the multimodal approach in machine learning? ›

                  The concept of multimodal in machine learning plays a pivotal role in advancing the capabilities of AI systems. By integrating data from various modalities, such as images, text, and speech, AI models gain a holistic understanding of the input, leading to enhanced decision-making and predictive abilities.

                  What is multimodal representation learning for real world applications? ›

                  Multimodality Representation Learning, as a technique of learning to embed information from different modalities and their correlations, has achieved remarkable success on a variety of applications, such as Visual Question Answering (VQA), Natural Language for Visual Reasoning (NLVR), and Vision Language Retrieval (VLR ...

                  Is ChatGPT multimodal? ›

                  ChatGPT's upgrade is a noteworthy example of a multimodal AI system. Instead of using a single AI model designed to work with a single form of input, like a large language model (LLM) or speech-to-voice model, multiple models work together to create a more cohesive AI tool.

                  Is GPT-4 multimodal? ›

                  Hence, multimodality in models, like GPT-4, allows them to develop intuition and understand complex relationships not just inside single modalities but across them, mimicking human-level cognizance to a higher degree.

                  What is an example of multimodal learning? ›

                  For example, a video shown in class should involve captions, images, narration, music and examples to be multimodal. Students today regularly interact with many different forms of text, so educators should reflect this in their classroom lessons.

                  Is multimodal learning good? ›

                  Research has shown that learning in multiple ways reinforces knowledge comprehension, underlining the need for a multimodal learning strategy in classrooms. From a more qualitative standpoint, multimodal learning creates a more exciting and all-encompassing learning environment for students.

                  Why is multimodal interaction important? ›

                  Specifically, multimodal systems can offer a flexible, efficient and usable environment allowing users to interact through input modalities, such as speech, handwriting, hand gesture and gaze, and to receive information by the system through output modalities, such as speech synthesis, smart graphics and other ...

                  What is the definition of multimodal approach? ›

                  Multimodal learning uses multiple modes or methodologies to teach a concept. Instructors create materials for different learning styles like visual, reading, auditory, writing, and kinesthetic. Multimodal learning includes teaching methods that engage multiple sensory systems simultaneously.

                  What is the multimodal analysis approach? ›

                  Multimodal analysis traditionally involves conceptualising abstract frameworks for language, images, and other resources and their intersemiotic relations (e.g. text and image relations) and then demonstrating these frameworks with some examples.

                  What does multimodal mean in AI? ›

                  Think of multimodal AI as a multilingual translator. It's an AI system that can comprehend and communicate in multiple 'languages'—in this case, data formats like text, visuals, or speech. It combines the strengths of different types of AI models to process various data formats.

                  What is the Multimodal learning style? ›

                  Multimodal learning engages the brain in multiple learning styles at once using various media. For example, a video lesson with subtitles and a downloadable information sheet leverages visual, auditory, and written learning styles.

                  References

                  Top Articles
                  Latest Posts
                  Article information

                  Author: Prof. Nancy Dach

                  Last Updated:

                  Views: 6050

                  Rating: 4.7 / 5 (57 voted)

                  Reviews: 80% of readers found this page helpful

                  Author information

                  Name: Prof. Nancy Dach

                  Birthday: 1993-08-23

                  Address: 569 Waelchi Ports, South Blainebury, LA 11589

                  Phone: +9958996486049

                  Job: Sales Manager

                  Hobby: Web surfing, Scuba diving, Mountaineering, Writing, Sailing, Dance, Blacksmithing

                  Introduction: My name is Prof. Nancy Dach, I am a lively, joyous, courageous, lovely, tender, charming, open person who loves writing and wants to share my knowledge and understanding with you.