Skip to main content

News about Scitation

In December 2016 Scitation will launch with a new design, enhanced navigation and a much improved user experience.

To ensure a smooth transition, from today, we are temporarily stopping new account registration and single article purchases. If you already have an account you can continue to use the site as normal.

For help or more information please visit our FAQs.

banner image
No data available.
Please log in to see this content.
You have no subscription access to this content.
No metrics data to plot.
The attempt to load metrics for this article has failed.
The attempt to plot a graph for these metrics has failed.
The full text of this article is not currently available.
/content/asa/journal/jasa/138/5/10.1121/1.4935350
1.
1. Aucouturier, J.-J. , and Defreville, B. (2009). “ Judging the similarity of soundscapes does not require categorization: Evidence from spliced stimuli,” J. Acoust. Soc. Am. 125(4), 21552161.
http://dx.doi.org/10.1121/1.3083232
2.
2. Aucouturier, J.-J. , Defreville, B. , and Pachet, F. (2007). “ The bag-of-frames approach to audio pattern recognition: A sufficient model for urban soundscapes but not for polyphonic music,” J. Acoust. Soc. Am. 122(2), 881891.
http://dx.doi.org/10.1121/1.2750160
3.
20. Aucouturier, J.-J. , Defreville, B. , and Pachet, F. (2015). https://archive.org/details/defreville-Aucouturier_urbanDb (Last viewed November 12, 2015).
4.
3. Ballas, J. A. , and Mullins, T. (1991). “ Effects of context on the identification of everyday sounds,” Hum. Perform. 4(3), 199219.
http://dx.doi.org/10.1207/s15327043hup0403_3
5.
4. Barchiesi, D. , Giannoulis, D. , Stowell, D. , and Plumbley, M. (2015). “ Acoustic scene classification: Classifying environments from the sounds they produce,” IEEE Sign. Process. Mag. 32(3), 1634.
http://dx.doi.org/10.1109/MSP.2014.2326181
6.
5. Flexer, A. (2007). “ A closer look on artist filters for musical genre classification,” in Proceedings of the 2007 International Conference on Music Information Retrieval (ISMIR).
7.
21. Giannoulis, D. , Benetos, E. , Stowell, D. , and Plumbley, M. D. (2012). http://c4dm.eecs.qmul.ac.uk/rdr/handle/123456789/29 (Last viewed November 12, 2015).
8.
6. Giannoulis, D. , Stowell, D. , Benetos, E. , Rossignol, M. , Lagrange, M. , and Plumbley, M. D. (2013a). “ A database and challenge for acoustic scene classification and event detection,” in Proceedings of EUSIPCO, pp. 15.
9.
7. Giannoulis, D. , Stowell, D. , Benetos, E. , Rossignol, M. , Lagrange, M. , and Plumbley, M. D. (2013b). “ Detection and classification of acoustic scenes and events: An IEEE AASP challenge,” in Proceedings of WASPAA: IEEE Workshop on Applications of Signal Processing to Audio and Acoustics.
10.
22. Google Scholar (2015). https://scholar.google.com/ (Last viewed November 12, 2015).
11.
8. Guastavino, C. (2007). “ Categorization of environmental sounds,” Can. J. Exp. Psychol. 61(1), 5463.
http://dx.doi.org/10.1037/cjep2007006
12.
9. Guastavino, C. , Katz, B. , Polack, J. , Levitin, D. , and Dubois, D. (2005). “ Ecological validity of soundscape reproduction,” Acta Acust. Acust. 91, 333341.
13.
10. Leech, R. , Gygi, B. , Aydelott, J. , and Dick, F. (2009). “ Informational factors in identifying environmental sounds in natural auditory scenes,” J. Acoust. Soc. Am. 126(6), 31473155.
http://dx.doi.org/10.1121/1.3238160
14.
11. McDermott, J. H. , Schemitsch, M. , and Simoncelli, E. P. (2013). “ Summary statistics in auditory perception,” Nat. Neurosci. 16(4), 493498.
http://dx.doi.org/10.1038/nn.3347
15.
12. Nelken, I. , and de Cheveigné, A. (2013). “ An ear for statistics,” Nat. Neurosci. 16(4), 381382.
http://dx.doi.org/10.1038/nn.3360
16.
13. Page, K. R. , Fields, B. , De Roure, D. , Crawford, T. , and Downie, J. S. (2013). “ Capturing the workflows of music information retrieval for repeatability and reuse,” J. Intell. Inf. Syst. 41(3), 435459.
http://dx.doi.org/10.1007/s10844-013-0260-9
17.
14. Park, T. H. , Turner, J. , Musick, M. , Lee, J. H. , Jacoby, C. , Mydlarz, C. , and Salamon, J. (2014). “ Sensing urban soundscapes,” in Proceedings of the EDBT/ICDT Workshops, pp. 375382.
18.
15. Slaney, M. (1998). “ Auditory toolbox,” interval research technical report.
19.
16. Sturm, B. L. (2012). “ An analysis of the gtzan music genre dataset,” in Proceedings of the Second International ACM Workshop on Music Information Retrieval with User-Centered and Multimodal Strategies, pp. 712.
20.
17. Sturm, B. L. (2013). “ Classification accuracy is not enough,” J. Intell. Inf. Syst. 41(3), 371406.
http://dx.doi.org/10.1007/s10844-013-0250-y
21.
18. Su, L. , Yeh, C. , Liu, J. , Wang, J. , and Yang, Y. (2014). “ A systematic evaluation of the bag-of-frames representation for music information retrieval,” IEEE Trans. Multimedia 16(5), 11881200.
http://dx.doi.org/10.1109/TMM.2014.2311016
22.
19. Tardieu, J. , Susini, P. , Poisson, F. , Lazareff, P. , and McAdams, S. (2008). “ Perceptual study of soundscapes in train stations,” Appl. Acoust. 69(12), 12241239.
http://dx.doi.org/10.1016/j.apacoust.2007.10.001
http://aip.metastore.ingenta.com/content/asa/journal/jasa/138/5/10.1121/1.4935350
Loading
/content/asa/journal/jasa/138/5/10.1121/1.4935350
Loading

Data & Media loading...

Loading

Article metrics loading...

/content/asa/journal/jasa/138/5/10.1121/1.4935350
2015-11-17
2016-12-06

Abstract

The “bag-of-frames” (BOF) approach, which encodes audio signals as the long-term statistical distribution of short-term spectral features, is commonly regarded as an effective and sufficient way to represent environmental sound recordings (soundscapes). The present paper describes a conceptual replication of a use of the BOF approach in a seminal article using several other soundscape datasets, with results strongly questioning the adequacy of the BOF approach for the task. As demonstrated in this paper, the good accuracy originally reported with BOF likely resulted from a particularly permissive dataset with low within-class variability. Soundscapemodeling, therefore, may not be the closed case it was once thought to be.

Loading

Full text loading...

/deliver/fulltext/asa/journal/jasa/138/5/1.4935350.html;jsessionid=R45N_-LVVGozNbGhYfM6yIfu.x-aip-live-03?itemId=/content/asa/journal/jasa/138/5/10.1121/1.4935350&mimeType=html&fmt=ahah&containerItemId=content/asa/journal/jasa
true
true

Access Key

  • FFree Content
  • OAOpen Access Content
  • SSubscribed Content
  • TFree Trial Content
752b84549af89a08dbdd7fdb8b9568b5 journal.articlezxybnytfddd
/content/realmedia?fmt=ahah&adPositionList=
&advertTargetUrl=//oascentral.aip.org/RealMedia/ads/&sitePageValue=asadl.org/jasa/138/5/10.1121/1.4935350&pageURL=http://scitation.aip.org/content/asa/journal/jasa/138/5/10.1121/1.4935350'
Right1,Right2,Right3,