History: Two-channel noisy recordings of a moving speaker within a limited area

Comparing version 13 with version 21


@@ -Lines: 1-3 changed to +Lines: 1-11 @@
!::Two-channel noisy recordings of a moving speaker within a limited area::
+
+ !! Motivation
+ This task is focused on a natural situation when the target is an uttering person whose location is limited to a specific area. For example, the target could be a speaker who is seated in a meeting (noisy) room. Its position is distant from microphones (say more than 1 meter) and is changing due to small movements of the speaker's head. The goal is to remove typical noise (e.g. babble noise) from the recorded speech. We assume that two microphones are available.
+
+ For such situation, a priori information may be provided in the form of noise-free recordings of the target from several (fixed) positions within the assumed area. For example, such recordings could be obtained during speaker-only intervals. How effectively can we exploit the a priori knowledge to enhance recordings of the speaker when the noise is present and his/her position is not perfectly known and could be changing within the limited area?
+
+ !! Results
+ The results are available [http://www.onn.nii.ac.jp/sisec13/evaluation_result/MOV/MOV2013.htm|here]

!! Scenario

@@ -Lines: 22-30 changed to +Lines: 30-38 @@
!! Tasks
The participants are encouraged to submit
- * Enhanced (de-noised) testing as well as development recordings
* Estimated trajectories of the loudspeaker in terms of sequences of indices of positions
+ * Enhanced (de-noised) testing as well as development noisy recordings
* Estimated trajectories of the loudspeaker in terms of sequences of indices of positions (mandatory)

!!Submissions
- Each participant should make his results available online in the form of a tarball called <YourName>_<dataset>.zip.
+ Each participant should make his/her results available online in the form of a tarball called <YourName>_<dataset>.zip.
The files containing the enhanced utterances should be named: <dataset>_<set>_x_x_x_x_enh.wav
where <dataset> is either dev or test, <set> is A, B, C, D, or E, and x_x_x_x are the estimated positions of the target during the movement.

@@ -Lines: 35-42 changed to +Lines: 43-53 @@
*the URL of the tarball(s)
- The submitted audio files will be made available on this website.
!! Evaluation criteria
The evaluation will be done through the perceptual evaluation toolkit [http://bass-db.gforge.inria.fr/peass/|PEASS v.2.0].
- [./tiki-index.php?page=Audio+source+separation|Back to Audio source separation top]
+ !!Licensing issues
/>All files are distributed under the terms of the [http://creativecommons.org/licenses/by-nc/3.0/|Creative Commons Attribution-Noncommercial-ShareAlike 3.0] license. The files to be submitted by participants will be made available on a website under the terms of the same license.
+ The recordings are authored by Emmanuel Vincent, Zbynek Koldovsky, and Jiri Malek.
+
+ [./tiki-index.php?page=Audio+source+separation|Back to Audio source separation top]

History

Legend: v=view, c=compare, d=diff
Date UserEdit Comment Version Action
Fri 26 of July, 2013 09:26 CEST admin   21
Current
 v
Fri 29 of Mar., 2013 10:57 CET admin   20  v  c  d  
Fri 29 of Mar., 2013 10:55 CET admin   19  v  c  d  
Fri 29 of Mar., 2013 10:54 CET admin   18  v  c  d  
Fri 29 of Mar., 2013 10:45 CET admin   17  v  c  d  
Wed 27 of Mar., 2013 10:51 CET admin   16  v  c  d  
Wed 27 of Mar., 2013 10:47 CET admin   15  v  c  d  
Wed 27 of Mar., 2013 10:45 CET admin   14  v  c  d  
Wed 27 of Mar., 2013 10:44 CET admin   13  v  c  d  
Wed 27 of Mar., 2013 10:37 CET admin   12  v  c  d  
Wed 27 of Mar., 2013 10:34 CET admin   11  v  c  d  
Wed 27 of Mar., 2013 10:32 CET admin   10  v  c  d  

Menu

Google Search

 
sisec2013.wiki.irisa.fr
WWW