{"id":4508,"date":"2023-03-28T10:41:37","date_gmt":"2023-03-28T08:41:37","guid":{"rendered":"https:\/\/vivolab-dev.i3a.es\/?p=4508"},"modified":"2023-03-28T14:03:47","modified_gmt":"2023-03-28T12:03:47","slug":"multimedia-content-retrieval-indexing","status":"publish","type":"post","link":"https:\/\/vivolab.i3a.es\/es\/multimedia-content-retrieval-indexing\/","title":{"rendered":"Multimedia Content Retrieval &#038; Indexing"},"content":{"rendered":"<div id=\"pl-gb4508-69deffea15fbd\"  class=\"panel-layout\" ><div id=\"pg-gb4508-69deffea15fbd-0\"  class=\"panel-grid panel-has-style\" ><div class=\"siteorigin-panels-stretch panel-row-style panel-row-style-for-gb4508-69deffea15fbd-0\" data-stretch-type=\"full-width-stretch\" ><div id=\"pgc-gb4508-69deffea15fbd-0-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb4508-69deffea15fbd-0-0-0\" class=\"so-panel widget widget_sow-hero panel-first-child panel-last-child\" data-index=\"0\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-hero so-widget-sow-hero-default-277c047652f8-4508 so-widget-fittext-wrapper\"\n\t\t\t data-fit-text-compressor=\"0.85\"\n\t\t>\t\t\t\t<div class=\"sow-slider-base\" style=\"display: none\" tabindex=\"0\">\n\t\t\t\t\t<ul\n\t\t\t\t\tclass=\"sow-slider-images\"\n\t\t\t\t\tdata-settings=\"{&quot;pagination&quot;:true,&quot;speed&quot;:800,&quot;timeout&quot;:8000,&quot;paused&quot;:false,&quot;pause_on_hover&quot;:false,&quot;swipe&quot;:true,&quot;nav_always_show_desktop&quot;:&quot;&quot;,&quot;nav_always_show_mobile&quot;:&quot;&quot;,&quot;breakpoint&quot;:&quot;780px&quot;,&quot;unmute&quot;:false,&quot;anchor&quot;:null}\"\n\t\t\t\t\t\t\t\t\t\tdata-anchor-id=\"\"\n\t\t\t\t>\t\t<li class=\"sow-slider-image  sow-slider-image-cover\" style=\"visibility: visible;;background-color: #ee7d00\" >\n\t\t\t\t\t<div class=\"sow-slider-image-container\">\n\t\t\t<div class=\"sow-slider-image-wrapper\">\n\t\t\t\t<h3 style=\"text-align: center\"><a href=\"..\/research-lines\/\">L\u00edneas de investigaci\u00f3n<\/a><\/h3>\n<h1 class=\"ultp-block-title\" style=\"text-align: center\">Audiovisual Information Processing<\/h1>\t\t\t<\/div>\n\t\t<\/div>\n\t\t<div class=\"sow-slider-image-overlay sow-slider-image-cover\" style=\"opacity: 0.8;background-image: url(https:\/\/vivolab.i3a.es\/wp-content\/uploads\/2021\/07\/qtq80-Uqdl79.jpeg)\"  ><\/div>\t\t<\/li>\n\t\t<\/ul>\t\t\t\t<ol class=\"sow-slider-pagination\">\n\t\t\t\t\t\t\t\t\t\t\t<li><a href=\"#\" data-goto=\"0\" aria-label=\"mostrar diapositiva 1\"><\/a><\/li>\n\t\t\t\t\t\t\t\t\t<\/ol>\n\n\t\t\t\t<div class=\"sow-slide-nav sow-slide-nav-next\">\n\t\t\t\t\t<a href=\"#\" data-goto=\"next\" aria-label=\"diapositiva siguiente\" data-action=\"next\">\n\t\t\t\t\t\t<em class=\"sow-sld-icon-thin-right\"><\/em>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\n\t\t\t\t<div class=\"sow-slide-nav sow-slide-nav-prev\">\n\t\t\t\t\t<a href=\"#\" data-goto=\"previous\" aria-label=\"diapositiva anterior\" data-action=\"prev\">\n\t\t\t\t\t\t<em class=\"sow-sld-icon-thin-left\"><\/em>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\t\t\t\t<\/div><\/div><\/div><\/div><\/div><\/div><\/div>\n\n<div id=\"pl-gb4508-69deffea17298\"  class=\"panel-layout\" ><div id=\"pg-gb4508-69deffea17298-0\"  class=\"panel-grid panel-no-style\" ><div id=\"pgc-gb4508-69deffea17298-0-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb4508-69deffea17298-0-0-0\" class=\"so-panel widget widget_sow-headline panel-first-child\" data-index=\"0\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-headline so-widget-sow-headline-default-14e39a71bb9f-4508\"\n\t\t\t\n\t\t><div class=\"sow-headline-container\">\n\t\t\t\t\t\t\t<h2 class=\"sow-headline\">\n\t\t\t\t\t\t Classification and segmentation of audiovisual documents\t\t\t\t\t\t<\/h2>\n\t\t\t\t\t\t\t\t\t\t\t<div class=\"decoration\">\n\t\t\t\t\t\t<div class=\"decoration-inside\"><\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n<\/div><\/div><div id=\"panel-gb4508-69deffea17298-0-0-1\" class=\"so-panel widget widget_sow-editor panel-last-child\" data-index=\"1\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-editor so-widget-sow-editor-base\"\n\t\t\t\n\t\t>\n<div class=\"siteorigin-widget-tinymce textwidget\">\n\t<div class=\"kc-elm kc-css-86046 kc_text_block\">\n<p>This research line aims to develop technologies that facilitate the access to huge multimedia repositories through automatic labelling and extraction of the different audiovisual documents present in it. Our prioritary working environment is the audiovisual content coming from broadcast emissions because of its interest both in commercial terms and in scientific terms, providing a variety of acoustic, semantic and emotional scenarios.<\/p>\n<p>This activity is highly influenced from basic research results from the audiovisual information processing research line. Namely, we incorporate our latest advances in deep learning in order to separate an audiovisual content in homogeneous classes such as speech, noise, music or a combination of these.<\/p>\n<\/div>\n<\/div>\n<\/div><\/div><\/div><\/div><div id=\"pg-gb4508-69deffea17298-1\"  class=\"panel-grid panel-no-style\" ><div id=\"pgc-gb4508-69deffea17298-1-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb4508-69deffea17298-1-0-0\" class=\"so-panel widget widget_sow-headline panel-first-child\" data-index=\"2\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-headline so-widget-sow-headline-default-14e39a71bb9f-4508\"\n\t\t\t\n\t\t><div class=\"sow-headline-container\">\n\t\t\t\t\t\t\t<h2 class=\"sow-headline\">\n\t\t\t\t\t\t Analysis and retrieval of audiovisual content\t\t\t\t\t\t<\/h2>\n\t\t\t\t\t\t\t\t\t\t\t<div class=\"decoration\">\n\t\t\t\t\t\t<div class=\"decoration-inside\"><\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n<\/div><\/div><div id=\"panel-gb4508-69deffea17298-1-0-1\" class=\"so-panel widget widget_sow-editor panel-last-child\" data-index=\"3\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-editor so-widget-sow-editor-base\"\n\t\t\t\n\t\t>\n<div class=\"siteorigin-widget-tinymce textwidget\">\n\t<p>Due to the huge in increase in the generation of multimedia content, systems that are able to analyze and index its content in a fast and accurate way are becoming more and more relevant nowadays. Our research group maintains stable agreements with Radio Televisi\u00f3n Espa\u00f1ola (RTVE) thanks to the <a style=\"font-size: 14px; font-style: inherit;\" href=\"http:\/\/catedrartve.unizar.es\/\" target=\"_blank\" rel=\"noopener\">\u201cC\u00e1tedra RTVE en la Universidad de&nbsp; Zaragoza\u201d<\/a> since 2017. This agreement seeks to boost the work on audiovisual content analysis with an special emphasis on the digital transformation of huge multimedia archives.<\/p>\n<p>We have also a close relationship with <a href=\"https:\/\/www.cartv.es\/\" target=\"_blank\" rel=\"noopener\">Corporaci\u00f3n Aragonesa de Radio y Televisi\u00f3n (CARTV)<\/a> since 2008 helping them to develop new technologies for enhancing accessibility to their multimedia contents. Furthermore, since 2016 we collaborate actively by means of a long-term technology transfer agreement with <a style=\"font-size: 14px; font-style: inherit;\" href=\"http:\/\/etiqmedia.com\/\" target=\"_blank\" rel=\"noopener\">ETIQMEDIA<\/a> to develop tools for audiovisual document management.<\/p>\n<\/div>\n<\/div><\/div><\/div><\/div><div id=\"pg-gb4508-69deffea17298-2\"  class=\"panel-grid panel-no-style\" ><div id=\"pgc-gb4508-69deffea17298-2-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb4508-69deffea17298-2-0-0\" class=\"so-panel widget widget_sow-headline panel-first-child\" data-index=\"4\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-headline so-widget-sow-headline-default-14e39a71bb9f-4508\"\n\t\t\t\n\t\t><div class=\"sow-headline-container\">\n\t\t\t\t\t\t\t<h2 class=\"sow-headline\">\n\t\t\t\t\t\t Multimodal person and event recognition\t\t\t\t\t\t<\/h2>\n\t\t\t\t\t\t\t\t\t\t\t<div class=\"decoration\">\n\t\t\t\t\t\t<div class=\"decoration-inside\"><\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n<\/div><\/div><div id=\"panel-gb4508-69deffea17298-2-0-1\" class=\"so-panel widget widget_sow-editor panel-last-child\" data-index=\"5\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-editor so-widget-sow-editor-base\"\n\t\t\t\n\t\t>\n<div class=\"siteorigin-widget-tinymce textwidget\">\n\t<div class=\"kc-elm kc-css-334281 kc_text_block\">\n<div class=\"kc-elm kc-css-334281 kc_text_block\">\n<p>The confluence of machine learning techniques applied to audio and image processing allow the reuse of these algorithms from a multimodal perspective. In this topic, our research group holds an open research line in multimodal person recognition, bringing together our experience in speaker recognition and recent advances in image and video processing. We recently participated in the Albayz\u00edn 2020 multimodal diarization challenge with competitive results, presenting a system that assigns speaker identities thanks to audio information and facial recognition.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div><\/div><\/div><\/div><div id=\"pg-gb4508-69deffea17298-3\"  class=\"panel-grid panel-no-style\" ><div id=\"pgc-gb4508-69deffea17298-3-0\"  class=\"panel-grid-cell\" ><div id=\"panel-gb4508-69deffea17298-3-0-0\" class=\"so-panel widget widget_sow-headline panel-first-child\" data-index=\"6\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-headline so-widget-sow-headline-default-14e39a71bb9f-4508\"\n\t\t\t\n\t\t><div class=\"sow-headline-container\">\n\t\t\t\t\t\t\t<h2 class=\"sow-headline\">\n\t\t\t\t\t\t Multimedia content summarization\t\t\t\t\t\t<\/h2>\n\t\t\t\t\t\t\t\t\t\t\t<div class=\"decoration\">\n\t\t\t\t\t\t<div class=\"decoration-inside\"><\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n<\/div><\/div><div id=\"panel-gb4508-69deffea17298-3-0-1\" class=\"so-panel widget widget_sow-editor panel-last-child\" data-index=\"7\" ><div\n\t\t\t\n\t\t\tclass=\"so-widget-sow-editor so-widget-sow-editor-base\"\n\t\t\t\n\t\t>\n<div class=\"siteorigin-widget-tinymce textwidget\">\n\t<div class=\"kc-elm kc-css-334281 kc_text_block\">\n<div class=\"kc-elm kc-css-334281 kc_text_block\">\n<div class=\"kc-elm kc-css-260247 kc_text_block\">\n<p>This research line combines both speech and language technologies and video and image processing techniques with the goal of extracting the most relevant fragments from an audiovisual document. The main idea is to automatically generate an abstract by detecting the most significant objects in the scenes, that are later described in natural language. Some of our recent work in this line has successfully created different proofs of concept for automatic summarization both in textual and audiovisual documents.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div><\/div><\/div><\/div><\/div>\n\n\n<p><\/p>","protected":false},"excerpt":{"rendered":"<ul>\n<li>Classification and segmentation of audiovisual documents<\/li>\n<li>Analysis and retrieval of audiovisual content<\/li>\n<li>Multimodal person and event recognition <\/li>\n<li>Multimedia content summarization<\/li>\n<\/ul>","protected":false},"author":1,"featured_media":3413,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[236],"tags":[],"class_list":["post-4508","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-lineas-de-investigacion"],"_links":{"self":[{"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/posts\/4508","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/comments?post=4508"}],"version-history":[{"count":11,"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/posts\/4508\/revisions"}],"predecessor-version":[{"id":4621,"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/posts\/4508\/revisions\/4621"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/media\/3413"}],"wp:attachment":[{"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/media?parent=4508"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/categories?post=4508"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/vivolab.i3a.es\/es\/wp-json\/wp\/v2\/tags?post=4508"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}