[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: LAST CALL for Papers - ICME Workshop on Broadcast and User-generated Content Recognition and Analysis (BRUREC)

Dear all,

This is just a reminder that the deadline for ICME BRUREC is one week
from now. If you work on related research topics and missed the ICME
regular paper deadline, please consider submitting your work to the
following workshop. The page limite is the same as the ICME regular
paper (6 pages). Accepted papers will be included in the conference
proceeding and published by IEEE Xplore.

Best regards,
BRUREC organizers

On Fri, Feb 22, 2013 at 9:51 PM, jinyu han <jinyuhan2008@xxxxxxxxx> wrote:
> Dear list subscriber,
> Please help us circulate this CfPs widely and forgive any cross-postings.
> The deadline for BRUREC(March 7th) is approaching. Short (up to 4
> pages) and long papers (up to 6 pages) are both welcome. All the
> accepted papers will be included in the conference proceeding and
> published by IEEE. The online submission is open:
> https://cmt.research.microsoft.com/ICMEW2013. To make sure your paper
> is submitted to BRUREC for review, please select the workshop track
> '1st IEEE International Workshop on Broadcast and User-Generated
> Content Recognition and Analysis (BRUREC)' when creating a new
> submission.
> ----------------------------------------------------------------------------------------------
> *************** LAST CALL FOR PAPERS ***************
> ----------------------------------------------------------------------------------------------
> ------------------------------------------------------------------------------------------------------------------------------
> *** 1st IEEE International Workshop on Broadcast and User-generated
> Content Recognition and Analysis (BRUREC) at ICME 2013***
> ------------------------------------------------------------------------------------------------------------------------------
> ***** July 15-19, 2013 • Fairmont Hotel, San Jose, USA |
> http://www.BRUREC.org *****
> ------------------------------------------------------------------------------------------------------------------------------
> In the past decade, we have seen great advancement in the area of
> visual and acoustic content recognition and analysis. Audio
> fingerprinting, for example, has led to many successful commercial
> applications and fundamentally changed the way people listen to, share
> and store music. In the meanwhile, research and development in visual
> content identification has reached a watershed, and large-scale
> commercial applications have started to emerge. Automated Content
> Recognition (ACR) applications have found their way into consumer
> applications. In the meanwhile, major Hollywood movie studios and TV
> networks have adapted ACR to track and manage their content at large
> scale. While broadcast quality video and audio analysis is already at
> an advanced stage, consumer-produced content analysis is not. Great
> potential exists on the media connectivity between broadcast media and
> user-generated content.   Therefore, in the last few years,
> user-generated content has attracted increasing attention from both
> academia and industry.
> This workshop aims to extend the ICME conference by focusing on
> algorithms, systems, applications, and standards for content
> recognition and analysis that can be applied across video and audio
> domains. BRUREC will cover all aspects on multimedia data generated by
> broadcast media such as TV and Radio as well as user-generated content
> such as Youtube and Vimeo. The goal of the workshop is to bring
> together researchers and practitioners in both industry and academia
> in the scientific community of content recognition and analysis across
> video, audio and multimedia domains to discuss the latest advance,
> challenges and unaddressed problems as well as exchange views, ideas
> in related technologies and applications, in which we attempt to
> advance the State of the Art of Broadcast and User-generated Content
> Recognition and Analysis.
> ---------------------------------------------------------------------
> ***** Topics include but are not limited to: *****
> ---------------------------------------------------------------------
> Content recognition and analysis algorithms and techniques:
> - Video and audio fingerprinting for content identification
> - Segmentation and classification of audio and visual content
> - Image classification and recognition
> - Features and descriptors for video and audio content
> - Audio and visual content clustering
> - Large database indexing, matching, and search
> - Machine learning for content classification
> - Evaluation of content-based identification and classification
> Content identification systems and applications:
> - Automated Content Recognition (ACR)
> - TV-centric content analysis and recognition
> - Emerging standards related to visual and audio content identification
> - Automatic content recognition from TV or Radio
> - Implementation of content recognition systems and services
> - Content identification in mobile devices
> - Other content recognition based applications (e.g., recommendation
> and ad targeting)
> ----------------------------------------------------------
> ********** Important Dates **********
> ----------------------------------------------------------
> -Paper Submission: March 7, 2013
> -Notification of Acceptance: April 15, 2013
> -Camera-Ready Paper Due: April 30, 2013
> ----------------------------------------------------------
> ********** Organizers **********
> ----------------------------------------------------------
> Jinyu Han
> Gracenote, Inc.
> jhan@xxxxxxxxxxxxx
> Gerald Friedland
> ICSI, UC Berkeley
> fractor@xxxxxxxxxxxxxxxxx
> Peter Dunker
> Gracenote, Inc.
> pdunker@xxxxxxxxxxxxx
> ------------------------------------------------------------------------
> ********** Technical Program Committee **********
> ------------------------------------------------------------------------
> Ching-Wei Chen (Gracenote, USA)
> Jingdong Chen (Northwestern Polytechnical University, China)
> Ngai-Man Cheung (Singapore University of Technology and Design, Singapore)
> Oscar Celma (Gracenote, USA)
> Trista Chen (Cognitive Networks, USA)
> Roger Dannenberg (Carnegie Mellon University, USA)
> Lixin Duan (SAP Research, Singapore)
> Yuan Dong (Orange Labs, France Telecom China)
> Zhiyao Duan (Northwestern University, USA)
> Dan Ellis (Columbia University, USA)
> Matthias Gruhne (Bach Technologies, Germany)
> Peter Grosche (Huawei European Research Center, Germany)
> Congcong Li (Google, USA)
> Lie Lu (Dolby Labs, China)
> Xiaofan Lin (A9.com, USA)
> Gautham Mysore (Adobe, USA)
> Bryan Pardo (Northwestern University, USA)
> Regu Radhakrishnan (Dolby Labs, USA)
> Paris Smaragdis (University of Illinois at Urbana-Champaign, USA)
> George Tzanetakis (University of Victoria, Canada)
> Junsong Yuan (Nanyang Technological University, Singapore)
> Honggang Zhang (Beijing University of Posts and Telecommunications, China)