SlideShare ist ein Scribd-Unternehmen logo
1 von 77
Downloaden Sie, um offline zu lesen
Creating a Data Collection for Evaluating Rich Speech Retrieval




            Creating a Data Collection
       for Evaluating Rich Speech Retrieval

                 Maria Eskevich1 , Gareth J.F. Jones1
                Martha Larson 2 , Roeland Ordelman 3

1   Centre for Digital Video Processing, Centre for Next Generation Localisation
          School of Computing, Dublin City University, Dublin, Ireland
            2   Delft University of Technology, Delft, The Netherlands
                       3   University of Twente, The Netherlands
Creating a Data Collection for Evaluating Rich Speech Retrieval


Outline


          MediaEval benchmark
          MediaEval 2011 Rich Speech Retrieval Task
          What is crowdsourcing?
          Crowdsourcing in Development of Speech and
          Language Resources
          Development of effective crowdsourcing task
          Comments on results
          Conclusion
          Future Work: Brave New Task at MediaEval 2012
Creating a Data Collection for Evaluating Rich Speech Retrieval


  ediaEval
Multimedia Evaluation benchmarking inititative




         Evaluate new algorithms for multimedia access and
        retrieval.
        Emphasize the ”multi” in multimedia: speech, audio,
        visual content, tags, users, context.
        Innovates new tasks and techniques focusing on the
        human and social aspects of multimedia content.
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention




        Transcript 1                             Transcript 2
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention




        Transcript 1                             Transcript 2
        Meaning 1                                Meaning 2
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention




        Transcript 1                  =          Transcript 2
        Meaning 1                     =          Meaning 2
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention




        Transcript 1                  =          Transcript 2
        Meaning 1                     =          Meaning 2
              Conventional retrieval
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention




        Transcript 1                 =           Transcript 2
        Meaning 1                    =           Meaning 2
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention




        Transcript 1                 =           Transcript 2
        Meaning 1                    =           Meaning 2
        Speech act 1                 =           Speech act 2
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task

       Task Goal:
         Information to be found - combination of required
         audio and visual content, and speaker’s intention




        Transcript 1                 =           Transcript 2
        Meaning 1                    =           Meaning 2
        Speech act 1                 =           Speech act 2
              Extended speech retrieval
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task


       ME10WWW dataset:
         Videos from Internet video sharing platform blip.tv
         (1974 episodes, 350 hours)
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task


       ME10WWW dataset:
         Videos from Internet video sharing platform blip.tv
         (1974 episodes, 350 hours)
         Automatic Speech Recognition (ASR) transcript provided
         by LIMSI and Vocapia Research
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task


       ME10WWW dataset:
         Videos from Internet video sharing platform blip.tv
         (1974 episodes, 350 hours)
         Automatic Speech Recognition (ASR) transcript provided
         by LIMSI and Vocapia Research
         No queries and relevant items
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task


        ME10WWW dataset:
          Videos from Internet video sharing platform blip.tv
          (1974 episodes, 350 hours)
          Automatic Speech Recognition (ASR) transcript provided
          by LIMSI and Vocapia Research
          No queries and relevant items

      − > Collect for Retrieval Experiment:
          user-generated queries
          user-generated relevant items
Creating a Data Collection for Evaluating Rich Speech Retrieval


   ediaEval 2011
Rich Speech Retrieval (RSR) Task


        ME10WWW dataset:
          Videos from Internet video sharing platform blip.tv
          (1974 episodes, 350 hours)
          Automatic Speech Recognition (ASR) transcript provided
          by LIMSI and Vocapia Research
          No queries and relevant items

      − > Collect for Retrieval Experiment:
          user-generated queries
          user-generated relevant items
      − > Collect via crowdsourcing technology
Creating a Data Collection for Evaluating Rich Speech Retrieval


What is crowdsourcing?


        Crowdsourcing is a form of human computation.
         Human computation is a method of having people do
      things that we might consider assigning to a computing
      device, e.g. a language translation task.
        A crowdsourcing system facilitates a crowdsourcing
      process.
Creating a Data Collection for Evaluating Rich Speech Retrieval


What is crowdsourcing?


        Crowdsourcing is a form of human computation.
         Human computation is a method of having people do
      things that we might consider assigning to a computing
      device, e.g. a language translation task.
        A crowdsourcing system facilitates a crowdsourcing
      process.

        Factors to take into account:
Creating a Data Collection for Evaluating Rich Speech Retrieval


What is crowdsourcing?


        Crowdsourcing is a form of human computation.
         Human computation is a method of having people do
      things that we might consider assigning to a computing
      device, e.g. a language translation task.
        A crowdsourcing system facilitates a crowdsourcing
      process.

        Factors to take into account:
          Sufficient number of workers
Creating a Data Collection for Evaluating Rich Speech Retrieval


What is crowdsourcing?


        Crowdsourcing is a form of human computation.
         Human computation is a method of having people do
      things that we might consider assigning to a computing
      device, e.g. a language translation task.
        A crowdsourcing system facilitates a crowdsourcing
      process.

        Factors to take into account:
          Sufficient number of workers
          Level of payment
Creating a Data Collection for Evaluating Rich Speech Retrieval


What is crowdsourcing?


        Crowdsourcing is a form of human computation.
         Human computation is a method of having people do
      things that we might consider assigning to a computing
      device, e.g. a language translation task.
        A crowdsourcing system facilitates a crowdsourcing
      process.

        Factors to take into account:
          Sufficient number of workers
          Level of payment
          Clear instructions
Creating a Data Collection for Evaluating Rich Speech Retrieval


What is crowdsourcing?


        Crowdsourcing is a form of human computation.
         Human computation is a method of having people do
      things that we might consider assigning to a computing
      device, e.g. a language translation task.
        A crowdsourcing system facilitates a crowdsourcing
      process.

        Factors to take into account:
          Sufficient number of workers
          Level of payment
          Clear instructions
          Possible cheating
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing in Development of Speech and
Language Resources
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing in Development of Speech and
Language Resources

       Suitability of crowdsourcing for simple/straightforward
     natural language processing tasks:
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing in Development of Speech and
Language Resources

       Suitability of crowdsourcing for simple/straightforward
     natural language processing tasks:
         Work by non-experts crowdsource workers is of similar
         standard to that performed by expert workers:
              translation/translation assessment
              transcription of native language
              word sense disambiguation
              temporal annotation
                                                                    [Snow et al., 2008]
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing in Development of Speech and
Language Resources

       Suitability of crowdsourcing for simple/straightforward
     natural language processing tasks:
         Work by non-experts crowdsource workers is of similar
         standard to that performed by expert workers:
              translation/translation assessment
              transcription of native language
              word sense disambiguation
              temporal annotation
                                                                    [Snow et al., 2008]
       Research question at collection creation stage:
         Can untrained crowdsource workers undertake
         extended tasks which require them to be creative?
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing with Amazon Mechanical Turk
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing with Amazon Mechanical Turk




     Task is referred to as a ‘Human Intelligence Task’ or HIT.
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing with Amazon Mechanical Turk




     Task is referred to as a ‘Human Intelligence Task’ or HIT.
     Crowdsourcing procedure:
         HIT initiation: Requester uploads a HIT.
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing with Amazon Mechanical Turk




     Task is referred to as a ‘Human Intelligence Task’ or HIT.
     Crowdsourcing procedure:
         HIT initiation: Requester uploads a HIT.
         Work: Workers carry out the HIT
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing with Amazon Mechanical Turk




     Task is referred to as a ‘Human Intelligence Task’ or HIT.
     Crowdsourcing procedure:
         HIT initiation: Requester uploads a HIT.
         Work: Workers carry out the HIT
         Review: Requester reviews the completed work and
         confirms payment to the worker with a previously set
         payment.
         *Requester has an option of paying more (”Bonus”)
Creating a Data Collection for Evaluating Rich Speech Retrieval


Information expected from the worker
to create a test collection for RSR Task
Creating a Data Collection for Evaluating Rich Speech Retrieval


Information expected from the worker
to create a test collection for RSR Task


        Speech act type:
          ’expressives’: apology, opinion
          ’assertives’: definition
          ’directives’: warning
          ’commissives’: promise
Creating a Data Collection for Evaluating Rich Speech Retrieval


Information expected from the worker
to create a test collection for RSR Task


        Speech act type:
          ’expressives’: apology, opinion
          ’assertives’: definition
          ’directives’: warning
          ’commissives’: promise
        Time of the labeled speech act: beginning and end
Creating a Data Collection for Evaluating Rich Speech Retrieval


Information expected from the worker
to create a test collection for RSR Task


        Speech act type:
          ’expressives’: apology, opinion
          ’assertives’: definition
          ’directives’: warning
          ’commissives’: promise
        Time of the labeled speech act: beginning and end
        Accurate transcript of the labeled speech act
Creating a Data Collection for Evaluating Rich Speech Retrieval


Information expected from the worker
to create a test collection for RSR Task


        Speech act type:
          ’expressives’: apology, opinion
          ’assertives’: definition
          ’directives’: warning
          ’commissives’: promise
        Time of the labeled speech act: beginning and end
        Accurate transcript of the labeled speech act
        Queries to refind this speech act:
          a full sentence query
          a short web style query
Creating a Data Collection for Evaluating Rich Speech Retrieval


Data management for Amazon MTurking



     ME10WWW videos vary in length:
Creating a Data Collection for Evaluating Rich Speech Retrieval


Data management for Amazon MTurking



     ME10WWW videos vary in length:

     − > Starting points for longer videos at a distance of
     approximately 7 minutes apart are calculated:

                 Data set           Episodes              Starting points
                  Dev                 247                       562
                  Test               1727                      3278
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment



     Worker expectations:
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment



     Worker expectations:
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment



     Worker expectations:




       Reward vs Work
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment



     Worker expectations:




       Reward vs Work
       Per hour Rate
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment


                                                 Requester uploads the HIT:
     Worker expectations:




       Reward vs Work
       Per hour Rate
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment


                                                 Requester uploads the HIT:
     Worker expectations:




       Reward vs Work                                           Pilot wording
       Per hour Rate
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment


                                                 Requester uploads the HIT:
     Worker expectations:




       Reward vs Work                                           Pilot wording
       Per hour Rate                                        0.11 $ + bonus per
                                                         speech act type
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment


      Workers feedback:                          Requester uploads the HIT:




       Reward is not worth
         the Work                                               Pilot wording
       Task is                                              0.11 $ + bonus per
         too complicated                                 speech act type
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment

                                                 Requester updates the HIT:
      Workers feedback:




                                                                Rewording
       Reward is not worth
         the Work
       Task is
         too complicated
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment

                                                 Requester updates the HIT:
      Workers feedback:




                                                                Rewording
       Reward is not worth                                      Examples
         the Work
       Task is
         too complicated
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment

                                                 Requester updates the HIT:
      Workers feedback:




                                                                Rewording
       Reward is not worth                                      Examples
         the Work                                            0.19 $ + bonus (0-21$)
       Task is                                           Workers suggest bonus
         too complicated                                 size (Mention to be a
                                                         non-profit organization)
Creating a Data Collection for Evaluating Rich Speech Retrieval


Crowdsourcing experiment

                                                 Requester updates the HIT:
     Workers feedback:




       Reward is worth
                                                                Rewording
         the Work
                                                                Examples
       Task is
         comprehensible                                      0.19 $ + bonus (0-21$)
                                                         Workers suggest bonus
       Workers are
                                                         size (Mention that we are a
        not greedy!
                                                         non-profit organization)
Creating a Data Collection for Evaluating Rich Speech Retrieval


HIT example

        Pilot:
     “Please watch the video and find a short portion of the
     video (a segment) that contains an interesting quote. The
     quote must fall into one of these six categories”
Creating a Data Collection for Evaluating Rich Speech Retrieval


HIT example

        Pilot:
     “Please watch the video and find a short portion of the
     video (a segment) that contains an interesting quote. The
     quote must fall into one of these six categories”

        Revised:
     “Imagine that you are watching videos on YouTube.
     When you come across something interesting you might
     want to share it on Facebook, Twitter or your favorite
     social network. Now please watch this video and search
     for an interesting video segment that you would like to
     share with others because it is (an apology, a definition,
     an opinion, a promise, a warning)”.
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results:
Number of collected queries per speech act




       Prices:
         Dev set: 40 $ per 30 queries
         Test set: 80 $ per 50 queries
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries




       No overlap of workers in dev and test sets
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries




       No overlap of workers in dev and test sets
       Creative work - Creative Cheating:
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries




       No overlap of workers in dev and test sets
       Creative work - Creative Cheating:
         Copy and paste provided examples
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries




       No overlap of workers in dev and test sets
       Creative work - Creative Cheating:
        Copy and paste provided examples
       − > Examples should be pictures, not texts
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries




       No overlap of workers in dev and test sets
       Creative work - Creative Cheating:
        Copy and paste provided examples
       − > Examples should be pictures, not texts
        Choose the option of no speech act found in the video
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries




       No overlap of workers in dev and test sets
       Creative work - Creative Cheating:
        Copy and paste provided examples
       − > Examples should be pictures, not texts
        Choose the option of no speech act found in the video
       − > Manual assessment by requester needed
Creating a Data Collection for Evaluating Rich Speech Retrieval


Results assessment


       Number of accepted HITs = number of collected queries




       No overlap of workers in dev and test sets
       Creative work - Creative Cheating:
         Copy and paste provided examples
        − > Examples should be pictures, not texts
         Choose the option of no speech act found in the video
        − > Manual assessment by requester needed
        Workers rarely find noteworthy content later than the
     third minute from the start of playback point in the video
Creating a Data Collection for Evaluating Rich Speech Retrieval


Conclusions



        It is possible to crowdsource extensive and complex
        tasks to support speech and language resources
Creating a Data Collection for Evaluating Rich Speech Retrieval


Conclusions



        It is possible to crowdsource extensive and complex
        tasks to support speech and language resources
        Use concepts and vocabulary familiar to the workers
Creating a Data Collection for Evaluating Rich Speech Retrieval


Conclusions



        It is possible to crowdsource extensive and complex
        tasks to support speech and language resources
        Use concepts and vocabulary familiar to the workers
        Pay attention to technical issues of watching the video
Creating a Data Collection for Evaluating Rich Speech Retrieval


Conclusions



        It is possible to crowdsource extensive and complex
        tasks to support speech and language resources
        Use concepts and vocabulary familiar to the workers
        Pay attention to technical issues of watching the video
        Video preprocessing into smaller segments
Creating a Data Collection for Evaluating Rich Speech Retrieval


Conclusions



        It is possible to crowdsource extensive and complex
        tasks to support speech and language resources
        Use concepts and vocabulary familiar to the workers
        Pay attention to technical issues of watching the video
        Video preprocessing into smaller segments
        Creative work demands higher reward level, or just
        more flexible system
Creating a Data Collection for Evaluating Rich Speech Retrieval


Conclusions



        It is possible to crowdsource extensive and complex
        tasks to support speech and language resources
        Use concepts and vocabulary familiar to the workers
        Pay attention to technical issues of watching the video
        Video preprocessing into smaller segments
        Creative work demands higher reward level, or just
        more flexible system
        High level of wastage due to task complexity
Creating a Data Collection for Evaluating Rich Speech Retrieval


  ediaEval 2012 Brave New Task:
Search and Hyperlinking

        Use Scenario: a user is searching for a known segment
     in a video collection. Furthermore, because the information
     in the segment might not be sufficient for his information
     need, s/he wants to have links to other related video
     segments, which may help to satisfy information need
     related to this video.
Creating a Data Collection for Evaluating Rich Speech Retrieval


  ediaEval 2012 Brave New Task:
Search and Hyperlinking

        Use Scenario: a user is searching for a known segment
     in a video collection. Furthermore, because the information
     in the segment might not be sufficient for his information
     need, s/he wants to have links to other related video
     segments, which may help to satisfy information need
     related to this video.

       Sub-tasks:
Creating a Data Collection for Evaluating Rich Speech Retrieval


  ediaEval 2012 Brave New Task:
Search and Hyperlinking

        Use Scenario: a user is searching for a known segment
     in a video collection. Furthermore, because the information
     in the segment might not be sufficient for his information
     need, s/he wants to have links to other related video
     segments, which may help to satisfy information need
     related to this video.

       Sub-tasks:
         Search: finding suitable video segments based on a short
         natural language query,
Creating a Data Collection for Evaluating Rich Speech Retrieval


  ediaEval 2012 Brave New Task:
Search and Hyperlinking

        Use Scenario: a user is searching for a known segment
     in a video collection. Furthermore, because the information
     in the segment might not be sufficient for his information
     need, s/he wants to have links to other related video
     segments, which may help to satisfy information need
     related to this video.

       Sub-tasks:
         Search: finding suitable video segments based on a short
         natural language query,
         Linking: defining links to other relevant video segments in
         the collection.
Creating a Data Collection for Evaluating Rich Speech Retrieval


ediaEval 2012


           Thank you for your attention!

     Welcome to MediaEval 2012! http://multimediaeval.org

Weitere ähnliche Inhalte

Ähnlich wie Creating a Data Collection for Evaluating Rich Speech Retrieval (LREC 2012)

NLP in Practice - Part I
NLP in Practice - Part INLP in Practice - Part I
NLP in Practice - Part IDelip Rao
 
Cross language information retrieval (clir)slide
Cross language information retrieval (clir)slideCross language information retrieval (clir)slide
Cross language information retrieval (clir)slideMohd Iqbal Al-farabi
 
Natural Language Processing for Irish
Natural Language Processing for IrishNatural Language Processing for Irish
Natural Language Processing for IrishTeresa Lynn
 
Mastering the Art of Conversation
Mastering the Art of ConversationMastering the Art of Conversation
Mastering the Art of ConversationOliver Cox
 
What are the basics of Analysing a corpus? chpt.10 Routledge
What are the basics of Analysing a corpus? chpt.10 RoutledgeWhat are the basics of Analysing a corpus? chpt.10 Routledge
What are the basics of Analysing a corpus? chpt.10 RoutledgeRajpootBhatti5
 
Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...
Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...
Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...Simplilearn
 
A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...
A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...
A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...kevig
 
Gareth Jones 1209
Gareth Jones 1209Gareth Jones 1209
Gareth Jones 1209Yandex
 
Umd draft-2010 jun22
Umd draft-2010 jun22Umd draft-2010 jun22
Umd draft-2010 jun22Ed Bice
 
"Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found...
"Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found..."Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found...
"Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found...Dataconomy Media
 
Automatic and rapid generation of massive knowledge repositories from data
Automatic and rapid generation of massive knowledge repositories from dataAutomatic and rapid generation of massive knowledge repositories from data
Automatic and rapid generation of massive knowledge repositories from dataSIKM
 
Towards Universal Semantic Understanding of Natural Languages
Towards Universal Semantic Understanding of Natural LanguagesTowards Universal Semantic Understanding of Natural Languages
Towards Universal Semantic Understanding of Natural LanguagesYunyao Li
 
Forum Tal 2014: Celi company presentation
Forum Tal 2014: Celi company presentationForum Tal 2014: Celi company presentation
Forum Tal 2014: Celi company presentationCELI
 
Conversational AI from CAT I to III
Conversational AI  from CAT I to IIIConversational AI  from CAT I to III
Conversational AI from CAT I to IIIHuangmao(Homer) Quan
 
AWS Artificial Intelligence Day - Toronto
AWS Artificial Intelligence Day - TorontoAWS Artificial Intelligence Day - Toronto
AWS Artificial Intelligence Day - TorontoAmazon Web Services
 

Ähnlich wie Creating a Data Collection for Evaluating Rich Speech Retrieval (LREC 2012) (20)

NLP in Practice - Part I
NLP in Practice - Part INLP in Practice - Part I
NLP in Practice - Part I
 
Artificial Intelligence on AWS
Artificial Intelligence on AWS Artificial Intelligence on AWS
Artificial Intelligence on AWS
 
NISO Forum, Denver, September 24, 2012: ResourceSync: Web-Based Resource Sync...
NISO Forum, Denver, September 24, 2012: ResourceSync: Web-Based Resource Sync...NISO Forum, Denver, September 24, 2012: ResourceSync: Web-Based Resource Sync...
NISO Forum, Denver, September 24, 2012: ResourceSync: Web-Based Resource Sync...
 
Cross language information retrieval (clir)slide
Cross language information retrieval (clir)slideCross language information retrieval (clir)slide
Cross language information retrieval (clir)slide
 
Natural Language Processing for Irish
Natural Language Processing for IrishNatural Language Processing for Irish
Natural Language Processing for Irish
 
Mastering the Art of Conversation
Mastering the Art of ConversationMastering the Art of Conversation
Mastering the Art of Conversation
 
What are the basics of Analysing a corpus? chpt.10 Routledge
What are the basics of Analysing a corpus? chpt.10 RoutledgeWhat are the basics of Analysing a corpus? chpt.10 Routledge
What are the basics of Analysing a corpus? chpt.10 Routledge
 
Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...
Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...
Hive Tutorial | Hive Architecture | Hive Tutorial For Beginners | Hive In Had...
 
A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...
A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...
A NOVEL APPROACH FOR NAMED ENTITY RECOGNITION ON HINDI LANGUAGE USING RESIDUA...
 
Gareth Jones 1209
Gareth Jones 1209Gareth Jones 1209
Gareth Jones 1209
 
2015 CIC: #EdTech Forum - LRMI
2015 CIC: #EdTech Forum - LRMI2015 CIC: #EdTech Forum - LRMI
2015 CIC: #EdTech Forum - LRMI
 
Umd draft-2010 jun22
Umd draft-2010 jun22Umd draft-2010 jun22
Umd draft-2010 jun22
 
"Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found...
"Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found..."Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found...
"Updates on Semantic Fingerprinting", Francisco Webber, Inventor and Co-Found...
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
Icwl2015 wahl
Icwl2015 wahlIcwl2015 wahl
Icwl2015 wahl
 
Automatic and rapid generation of massive knowledge repositories from data
Automatic and rapid generation of massive knowledge repositories from dataAutomatic and rapid generation of massive knowledge repositories from data
Automatic and rapid generation of massive knowledge repositories from data
 
Towards Universal Semantic Understanding of Natural Languages
Towards Universal Semantic Understanding of Natural LanguagesTowards Universal Semantic Understanding of Natural Languages
Towards Universal Semantic Understanding of Natural Languages
 
Forum Tal 2014: Celi company presentation
Forum Tal 2014: Celi company presentationForum Tal 2014: Celi company presentation
Forum Tal 2014: Celi company presentation
 
Conversational AI from CAT I to III
Conversational AI  from CAT I to IIIConversational AI  from CAT I to III
Conversational AI from CAT I to III
 
AWS Artificial Intelligence Day - Toronto
AWS Artificial Intelligence Day - TorontoAWS Artificial Intelligence Day - Toronto
AWS Artificial Intelligence Day - Toronto
 

Mehr von Maria Eskevich

Video Hyperlinking (LNK) Task at TRECVid 2016
Video Hyperlinking (LNK) Task at TRECVid 2016Video Hyperlinking (LNK) Task at TRECVid 2016
Video Hyperlinking (LNK) Task at TRECVid 2016Maria Eskevich
 
Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014Maria Eskevich
 
Focus on spoken content in multimedia retrieval
Focus on spoken content in multimedia retrievalFocus on spoken content in multimedia retrieval
Focus on spoken content in multimedia retrievalMaria Eskevich
 
Audio/Video Search: Why? What? How?
Audio/Video Search: Why? What? How?Audio/Video Search: Why? What? How?
Audio/Video Search: Why? What? How?Maria Eskevich
 
DCU at the NTCIR-9 SpokenDoc Passage Retrieval Task
DCU at the NTCIR-9 SpokenDoc Passage Retrieval TaskDCU at the NTCIR-9 SpokenDoc Passage Retrieval Task
DCU at the NTCIR-9 SpokenDoc Passage Retrieval TaskMaria Eskevich
 
Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...
Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...
Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...Maria Eskevich
 
Search and Hyperlinking Task at MediaEval 2012
Search and Hyperlinking Task at MediaEval 2012Search and Hyperlinking Task at MediaEval 2012
Search and Hyperlinking Task at MediaEval 2012Maria Eskevich
 
New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...
New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...
New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...Maria Eskevich
 

Mehr von Maria Eskevich (8)

Video Hyperlinking (LNK) Task at TRECVid 2016
Video Hyperlinking (LNK) Task at TRECVid 2016Video Hyperlinking (LNK) Task at TRECVid 2016
Video Hyperlinking (LNK) Task at TRECVid 2016
 
Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014
 
Focus on spoken content in multimedia retrieval
Focus on spoken content in multimedia retrievalFocus on spoken content in multimedia retrieval
Focus on spoken content in multimedia retrieval
 
Audio/Video Search: Why? What? How?
Audio/Video Search: Why? What? How?Audio/Video Search: Why? What? How?
Audio/Video Search: Why? What? How?
 
DCU at the NTCIR-9 SpokenDoc Passage Retrieval Task
DCU at the NTCIR-9 SpokenDoc Passage Retrieval TaskDCU at the NTCIR-9 SpokenDoc Passage Retrieval Task
DCU at the NTCIR-9 SpokenDoc Passage Retrieval Task
 
Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...
Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...
Towards Methods for Efficient Access to Spoken Content in the AMI Corpus (SSC...
 
Search and Hyperlinking Task at MediaEval 2012
Search and Hyperlinking Task at MediaEval 2012Search and Hyperlinking Task at MediaEval 2012
Search and Hyperlinking Task at MediaEval 2012
 
New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...
New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...
New Metrics for Meaningful Evaluation of Informally Structured Speech Retriev...
 

Kürzlich hochgeladen

08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...Martijn de Jong
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CVKhem
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessPixlogix Infotech
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?Igalia
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024Results
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfsudhanshuwaghmare1
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUK Journal
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 

Kürzlich hochgeladen (20)

08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CV
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your Business
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 

Creating a Data Collection for Evaluating Rich Speech Retrieval (LREC 2012)

  • 1. Creating a Data Collection for Evaluating Rich Speech Retrieval Creating a Data Collection for Evaluating Rich Speech Retrieval Maria Eskevich1 , Gareth J.F. Jones1 Martha Larson 2 , Roeland Ordelman 3 1 Centre for Digital Video Processing, Centre for Next Generation Localisation School of Computing, Dublin City University, Dublin, Ireland 2 Delft University of Technology, Delft, The Netherlands 3 University of Twente, The Netherlands
  • 2. Creating a Data Collection for Evaluating Rich Speech Retrieval Outline MediaEval benchmark MediaEval 2011 Rich Speech Retrieval Task What is crowdsourcing? Crowdsourcing in Development of Speech and Language Resources Development of effective crowdsourcing task Comments on results Conclusion Future Work: Brave New Task at MediaEval 2012
  • 3. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval Multimedia Evaluation benchmarking inititative Evaluate new algorithms for multimedia access and retrieval. Emphasize the ”multi” in multimedia: speech, audio, visual content, tags, users, context. Innovates new tasks and techniques focusing on the human and social aspects of multimedia content.
  • 4. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention
  • 5. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention
  • 6. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention
  • 7. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention Transcript 1 Transcript 2
  • 8. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention Transcript 1 Transcript 2 Meaning 1 Meaning 2
  • 9. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention Transcript 1 = Transcript 2 Meaning 1 = Meaning 2
  • 10. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention Transcript 1 = Transcript 2 Meaning 1 = Meaning 2 Conventional retrieval
  • 11. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention Transcript 1 = Transcript 2 Meaning 1 = Meaning 2
  • 12. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention Transcript 1 = Transcript 2 Meaning 1 = Meaning 2 Speech act 1 = Speech act 2
  • 13. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task Task Goal: Information to be found - combination of required audio and visual content, and speaker’s intention Transcript 1 = Transcript 2 Meaning 1 = Meaning 2 Speech act 1 = Speech act 2 Extended speech retrieval
  • 14. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task ME10WWW dataset: Videos from Internet video sharing platform blip.tv (1974 episodes, 350 hours)
  • 15. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task ME10WWW dataset: Videos from Internet video sharing platform blip.tv (1974 episodes, 350 hours) Automatic Speech Recognition (ASR) transcript provided by LIMSI and Vocapia Research
  • 16. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task ME10WWW dataset: Videos from Internet video sharing platform blip.tv (1974 episodes, 350 hours) Automatic Speech Recognition (ASR) transcript provided by LIMSI and Vocapia Research No queries and relevant items
  • 17. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task ME10WWW dataset: Videos from Internet video sharing platform blip.tv (1974 episodes, 350 hours) Automatic Speech Recognition (ASR) transcript provided by LIMSI and Vocapia Research No queries and relevant items − > Collect for Retrieval Experiment: user-generated queries user-generated relevant items
  • 18. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2011 Rich Speech Retrieval (RSR) Task ME10WWW dataset: Videos from Internet video sharing platform blip.tv (1974 episodes, 350 hours) Automatic Speech Recognition (ASR) transcript provided by LIMSI and Vocapia Research No queries and relevant items − > Collect for Retrieval Experiment: user-generated queries user-generated relevant items − > Collect via crowdsourcing technology
  • 19. Creating a Data Collection for Evaluating Rich Speech Retrieval What is crowdsourcing? Crowdsourcing is a form of human computation. Human computation is a method of having people do things that we might consider assigning to a computing device, e.g. a language translation task. A crowdsourcing system facilitates a crowdsourcing process.
  • 20. Creating a Data Collection for Evaluating Rich Speech Retrieval What is crowdsourcing? Crowdsourcing is a form of human computation. Human computation is a method of having people do things that we might consider assigning to a computing device, e.g. a language translation task. A crowdsourcing system facilitates a crowdsourcing process. Factors to take into account:
  • 21. Creating a Data Collection for Evaluating Rich Speech Retrieval What is crowdsourcing? Crowdsourcing is a form of human computation. Human computation is a method of having people do things that we might consider assigning to a computing device, e.g. a language translation task. A crowdsourcing system facilitates a crowdsourcing process. Factors to take into account: Sufficient number of workers
  • 22. Creating a Data Collection for Evaluating Rich Speech Retrieval What is crowdsourcing? Crowdsourcing is a form of human computation. Human computation is a method of having people do things that we might consider assigning to a computing device, e.g. a language translation task. A crowdsourcing system facilitates a crowdsourcing process. Factors to take into account: Sufficient number of workers Level of payment
  • 23. Creating a Data Collection for Evaluating Rich Speech Retrieval What is crowdsourcing? Crowdsourcing is a form of human computation. Human computation is a method of having people do things that we might consider assigning to a computing device, e.g. a language translation task. A crowdsourcing system facilitates a crowdsourcing process. Factors to take into account: Sufficient number of workers Level of payment Clear instructions
  • 24. Creating a Data Collection for Evaluating Rich Speech Retrieval What is crowdsourcing? Crowdsourcing is a form of human computation. Human computation is a method of having people do things that we might consider assigning to a computing device, e.g. a language translation task. A crowdsourcing system facilitates a crowdsourcing process. Factors to take into account: Sufficient number of workers Level of payment Clear instructions Possible cheating
  • 25. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing in Development of Speech and Language Resources
  • 26. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing in Development of Speech and Language Resources Suitability of crowdsourcing for simple/straightforward natural language processing tasks:
  • 27. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing in Development of Speech and Language Resources Suitability of crowdsourcing for simple/straightforward natural language processing tasks: Work by non-experts crowdsource workers is of similar standard to that performed by expert workers: translation/translation assessment transcription of native language word sense disambiguation temporal annotation [Snow et al., 2008]
  • 28. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing in Development of Speech and Language Resources Suitability of crowdsourcing for simple/straightforward natural language processing tasks: Work by non-experts crowdsource workers is of similar standard to that performed by expert workers: translation/translation assessment transcription of native language word sense disambiguation temporal annotation [Snow et al., 2008] Research question at collection creation stage: Can untrained crowdsource workers undertake extended tasks which require them to be creative?
  • 29. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing with Amazon Mechanical Turk
  • 30. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing with Amazon Mechanical Turk Task is referred to as a ‘Human Intelligence Task’ or HIT.
  • 31. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing with Amazon Mechanical Turk Task is referred to as a ‘Human Intelligence Task’ or HIT. Crowdsourcing procedure: HIT initiation: Requester uploads a HIT.
  • 32. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing with Amazon Mechanical Turk Task is referred to as a ‘Human Intelligence Task’ or HIT. Crowdsourcing procedure: HIT initiation: Requester uploads a HIT. Work: Workers carry out the HIT
  • 33. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing with Amazon Mechanical Turk Task is referred to as a ‘Human Intelligence Task’ or HIT. Crowdsourcing procedure: HIT initiation: Requester uploads a HIT. Work: Workers carry out the HIT Review: Requester reviews the completed work and confirms payment to the worker with a previously set payment. *Requester has an option of paying more (”Bonus”)
  • 34. Creating a Data Collection for Evaluating Rich Speech Retrieval Information expected from the worker to create a test collection for RSR Task
  • 35. Creating a Data Collection for Evaluating Rich Speech Retrieval Information expected from the worker to create a test collection for RSR Task Speech act type: ’expressives’: apology, opinion ’assertives’: definition ’directives’: warning ’commissives’: promise
  • 36. Creating a Data Collection for Evaluating Rich Speech Retrieval Information expected from the worker to create a test collection for RSR Task Speech act type: ’expressives’: apology, opinion ’assertives’: definition ’directives’: warning ’commissives’: promise Time of the labeled speech act: beginning and end
  • 37. Creating a Data Collection for Evaluating Rich Speech Retrieval Information expected from the worker to create a test collection for RSR Task Speech act type: ’expressives’: apology, opinion ’assertives’: definition ’directives’: warning ’commissives’: promise Time of the labeled speech act: beginning and end Accurate transcript of the labeled speech act
  • 38. Creating a Data Collection for Evaluating Rich Speech Retrieval Information expected from the worker to create a test collection for RSR Task Speech act type: ’expressives’: apology, opinion ’assertives’: definition ’directives’: warning ’commissives’: promise Time of the labeled speech act: beginning and end Accurate transcript of the labeled speech act Queries to refind this speech act: a full sentence query a short web style query
  • 39. Creating a Data Collection for Evaluating Rich Speech Retrieval Data management for Amazon MTurking ME10WWW videos vary in length:
  • 40. Creating a Data Collection for Evaluating Rich Speech Retrieval Data management for Amazon MTurking ME10WWW videos vary in length: − > Starting points for longer videos at a distance of approximately 7 minutes apart are calculated: Data set Episodes Starting points Dev 247 562 Test 1727 3278
  • 41. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment
  • 42. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Worker expectations:
  • 43. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Worker expectations:
  • 44. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Worker expectations: Reward vs Work
  • 45. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Worker expectations: Reward vs Work Per hour Rate
  • 46. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Requester uploads the HIT: Worker expectations: Reward vs Work Per hour Rate
  • 47. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Requester uploads the HIT: Worker expectations: Reward vs Work Pilot wording Per hour Rate
  • 48. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Requester uploads the HIT: Worker expectations: Reward vs Work Pilot wording Per hour Rate 0.11 $ + bonus per speech act type
  • 49. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Workers feedback: Requester uploads the HIT: Reward is not worth the Work Pilot wording Task is 0.11 $ + bonus per too complicated speech act type
  • 50. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Requester updates the HIT: Workers feedback: Rewording Reward is not worth the Work Task is too complicated
  • 51. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Requester updates the HIT: Workers feedback: Rewording Reward is not worth Examples the Work Task is too complicated
  • 52. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Requester updates the HIT: Workers feedback: Rewording Reward is not worth Examples the Work 0.19 $ + bonus (0-21$) Task is Workers suggest bonus too complicated size (Mention to be a non-profit organization)
  • 53. Creating a Data Collection for Evaluating Rich Speech Retrieval Crowdsourcing experiment Requester updates the HIT: Workers feedback: Reward is worth Rewording the Work Examples Task is comprehensible 0.19 $ + bonus (0-21$) Workers suggest bonus Workers are size (Mention that we are a not greedy! non-profit organization)
  • 54. Creating a Data Collection for Evaluating Rich Speech Retrieval HIT example Pilot: “Please watch the video and find a short portion of the video (a segment) that contains an interesting quote. The quote must fall into one of these six categories”
  • 55. Creating a Data Collection for Evaluating Rich Speech Retrieval HIT example Pilot: “Please watch the video and find a short portion of the video (a segment) that contains an interesting quote. The quote must fall into one of these six categories” Revised: “Imagine that you are watching videos on YouTube. When you come across something interesting you might want to share it on Facebook, Twitter or your favorite social network. Now please watch this video and search for an interesting video segment that you would like to share with others because it is (an apology, a definition, an opinion, a promise, a warning)”.
  • 56. Creating a Data Collection for Evaluating Rich Speech Retrieval Results: Number of collected queries per speech act Prices: Dev set: 40 $ per 30 queries Test set: 80 $ per 50 queries
  • 57. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment
  • 58. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries
  • 59. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries
  • 60. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries No overlap of workers in dev and test sets
  • 61. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries No overlap of workers in dev and test sets Creative work - Creative Cheating:
  • 62. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries No overlap of workers in dev and test sets Creative work - Creative Cheating: Copy and paste provided examples
  • 63. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries No overlap of workers in dev and test sets Creative work - Creative Cheating: Copy and paste provided examples − > Examples should be pictures, not texts
  • 64. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries No overlap of workers in dev and test sets Creative work - Creative Cheating: Copy and paste provided examples − > Examples should be pictures, not texts Choose the option of no speech act found in the video
  • 65. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries No overlap of workers in dev and test sets Creative work - Creative Cheating: Copy and paste provided examples − > Examples should be pictures, not texts Choose the option of no speech act found in the video − > Manual assessment by requester needed
  • 66. Creating a Data Collection for Evaluating Rich Speech Retrieval Results assessment Number of accepted HITs = number of collected queries No overlap of workers in dev and test sets Creative work - Creative Cheating: Copy and paste provided examples − > Examples should be pictures, not texts Choose the option of no speech act found in the video − > Manual assessment by requester needed Workers rarely find noteworthy content later than the third minute from the start of playback point in the video
  • 67. Creating a Data Collection for Evaluating Rich Speech Retrieval Conclusions It is possible to crowdsource extensive and complex tasks to support speech and language resources
  • 68. Creating a Data Collection for Evaluating Rich Speech Retrieval Conclusions It is possible to crowdsource extensive and complex tasks to support speech and language resources Use concepts and vocabulary familiar to the workers
  • 69. Creating a Data Collection for Evaluating Rich Speech Retrieval Conclusions It is possible to crowdsource extensive and complex tasks to support speech and language resources Use concepts and vocabulary familiar to the workers Pay attention to technical issues of watching the video
  • 70. Creating a Data Collection for Evaluating Rich Speech Retrieval Conclusions It is possible to crowdsource extensive and complex tasks to support speech and language resources Use concepts and vocabulary familiar to the workers Pay attention to technical issues of watching the video Video preprocessing into smaller segments
  • 71. Creating a Data Collection for Evaluating Rich Speech Retrieval Conclusions It is possible to crowdsource extensive and complex tasks to support speech and language resources Use concepts and vocabulary familiar to the workers Pay attention to technical issues of watching the video Video preprocessing into smaller segments Creative work demands higher reward level, or just more flexible system
  • 72. Creating a Data Collection for Evaluating Rich Speech Retrieval Conclusions It is possible to crowdsource extensive and complex tasks to support speech and language resources Use concepts and vocabulary familiar to the workers Pay attention to technical issues of watching the video Video preprocessing into smaller segments Creative work demands higher reward level, or just more flexible system High level of wastage due to task complexity
  • 73. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2012 Brave New Task: Search and Hyperlinking Use Scenario: a user is searching for a known segment in a video collection. Furthermore, because the information in the segment might not be sufficient for his information need, s/he wants to have links to other related video segments, which may help to satisfy information need related to this video.
  • 74. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2012 Brave New Task: Search and Hyperlinking Use Scenario: a user is searching for a known segment in a video collection. Furthermore, because the information in the segment might not be sufficient for his information need, s/he wants to have links to other related video segments, which may help to satisfy information need related to this video. Sub-tasks:
  • 75. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2012 Brave New Task: Search and Hyperlinking Use Scenario: a user is searching for a known segment in a video collection. Furthermore, because the information in the segment might not be sufficient for his information need, s/he wants to have links to other related video segments, which may help to satisfy information need related to this video. Sub-tasks: Search: finding suitable video segments based on a short natural language query,
  • 76. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2012 Brave New Task: Search and Hyperlinking Use Scenario: a user is searching for a known segment in a video collection. Furthermore, because the information in the segment might not be sufficient for his information need, s/he wants to have links to other related video segments, which may help to satisfy information need related to this video. Sub-tasks: Search: finding suitable video segments based on a short natural language query, Linking: defining links to other relevant video segments in the collection.
  • 77. Creating a Data Collection for Evaluating Rich Speech Retrieval ediaEval 2012 Thank you for your attention! Welcome to MediaEval 2012! http://multimediaeval.org