five

Towards Using Virtual Acoustics for Evaluating Spatial Ecoacoustic Monitoring Technologies - Data|生态声学监测数据集|虚拟声学技术数据集

收藏
Mendeley Data2024-05-11 更新2024-06-28 收录
生态声学监测
虚拟声学技术
下载链接:
https://zenodo.org/records/11105332
下载链接
链接失效反馈
资源简介:
About This database contains the raw data and certain outputs used for the project 'Towards Using Virtual Acoustics for Evaluating Spatial Ecoacoustic Monitoring Technologies'. In this work, we developed an ambisonic Virtual Sound Environment (VSE) for simulating real natural soundscapes to evaluate spatial PAM technologies in a more controlled and repeatable manner. We set three objectives to validate this approach: (O1) to determine whether the VSE could replicate natural soundscapes well enough to be a test environment; (O2) to pilot the VSE as a test environment for Passive Acoustic Monitoring (PAM) hardware; and, (O3) to pilot the VSE as a test platform for PAM software. To meet these objectives, we used a recently-developed, open source six-microphone field recorder to capture recordings of six field sites and their VSE-based simulations. Sites were based at the Imperial College Silwood Park Campus (Ascot, UK). For O1, we compared field and VSE recordings using a typical suite of ecoacoustic metrics. For O2, we used the VSE to explore how orientation impacts the performance of the six-microphone array. We extended the suite of metrics from O1 to compare VSE recordings from this array at various pitch angles: vertical (as in the field), 45° pitch, and horizontal. For O3, we investigate how BirdNET and HARKBird, software for classifying and localising avian calls, respectively, perform on bird calls added to the VSE-replicated soundscapes. We compare adding calls by encoding to the ambisonics domain and by playback from individual loudspeakers. The data is organised as follows: '6mic Audio O1O2': contains the six-channel field and VSE-based recordings of all six sites. Recordings are approximately 10 minutes long. Note that there are three VSE-based recording per site, for each recording orientation used in the VSE (vertical, 45°, and horizontal). Several of our analyses used low-passed versions of these recordings (4 kHz cutoff and 12 dB roll-off); note that the recordings provided here are raw and therefore not low-passed. File names indicate whether the recording was performed in the field or VSE ('Field' vs 'ReRec') and the orientation of the 6mic array for the latter ('V', '45', or 'H'). The final number in each filename indicates the Site the recording corresponds to (1-6). '6mic Audio O3': contains the six-channel VSE recordings of five of the original sites with additional avian calls located at certain moments in space and time in each recording. 10 bird calls were added to the soundscapes, each one in its own soundscape recording (each of the five sites' soundscapes was therefore used twice). Four methods of adding the bird calls were trialled, hence there are 40 files in this folder – 10 for each of: ambisonic encoding, playback from individual loudspeakers, playback from individual loudspeakers with no reverb, and playback from individual loudspeakers with no reverb or background soundscape (i.e., 'soloed'). Our analyses focussed on comparing the first two of these methods. The filenames are structured as follows: "sXsYbirdNameEmbeddingMethod', where X corresponds to the site the background soundscape was recorded at, Y indicates the position of the sinusoidal sweep used to create simulated reverb of the bird call (1-4 for 0°, 90°, 180°, or 270° around the device used to capture the ambisonic soundscape recordings), 'birdName' is the common name of the added species, and 'EmbeddingMethod' is either: 'Ambi' (ambisonic encoding), 'LSPK' (playback from an individual loudspeaker), 'LSPK-NR' (playback from an individual loudspeaker with no reverb), or 'Solo' (soloed playback from an individual loudspeaker). Note that due to issues of spatial aliasing, we low-passed the 'Ambi' and 'LSPK' recordings in our analyses (as the two main embedding methods compared) using a 4 kHz cutoff and 12 dB roll-off. However, again the raw (not low-passed) audio is shared here. 'Acoustic Indices': contains the CSV files with matrices of the Acoustic Indices extracted from the first channel of the low-passed '6mic Audio O1O2' recordings. Each column is for a different index, rows are the values over time (indices were extracted on 30 s windows). We extracted the following 7 common acoustic indices: Acoustic Complexity Index (ACI; column 1), Acoustic Diversity Index (ADI; column 2), Acoustic Evenness (AEve; column 3), Bioacoustic Index (Bio; column 4), Normalised Difference Soundscape Index (NDSI; column 5) Acoustic Entropy (H; column 6), and Median of the Amplitude Envelope (M; column 7). Filenames indicate: the field site, whether the indices are for a field or VSE ('Lab') recording, and the orientation of the recording for the latter ('Vert', '45', or 'H'). The 'LP' suffix denotes that these indices were extracted from a low-passed version of the 6mic O1O2 recordings (4 kHz cutoff and 12 dB roll-off). 'BirdNET O1O2 Outputs': CSV files generated from avian call classifier BirdNET (using the Winows GUI version) using the first channel ('Mic 1') of the '6mic Audio O1O2' recordings. 'BirdNET O3 Outputs': outputs of BirdNET on the '6mic Audio O3' recordings. Here, rather than the raw CSV files generated by BirdNET, the BirdNET results have been filtered to just those during the added bird calls' start and end times, and have been compiled into two CSV files: one for recordings of VSE-based soundscapes with bird calls added by ambisonic encoding and another for calls added by individual loudspeaker playback. These CSV files also contain columns to specify added birds' site, sweep (used to generate reverberation, see '6mic Audio O3' above), azimuth and elevation. 'HARKBird O1O2 Outputs': outputs of avian call localisation tool HARKBird on the '6mic Audio O1O2' recordings. Note that HARKBird outputs a folder with additional results for each file, however, only the CSV files presented here. These contain the times and estimated azimuth angles of bird calls and were the only HARKBird ouput used for subsequent analyses. Filenames indicate whether recordings are from the field ('Field') or VSE ('ReRec'), and those for the latter also indicate the recording orientation ('Vert', 45' 'H'). The final number in each filename indicates the site the recording corresponds to. 'HARKBird O3 Outputs': CSV files of HARKBird outputs (as described above) for the '6mic Audio O3' recordings. Filenames are based on those for this set of recordings (see '6mic Audio O3' above); names that contain 'LP' were low-passed (with a 4 kHz cutoff frequency and 12 dB roll-off) prior to passing through HARKBird. 'Manual Labels O2': manual labels of audible bird calls' species in the omnidirectional (first) channel of the Zylia ZM-1 ambisonic recordings (see 'Zylia Recordings' below) for all sites. This data was used to calculate the precision and recall of BirdNET's outputs on the 6mic O1O2 recordings for these sites. 'VGGish Features': contains the 128-dimension feature embedding of the pre-trained VGGish convolutional neural network extracted from the '6mic Audio O1O2' recordings. Filenames indicate the site, whether the recording was made in the field or VSE ('Lab'), and the recording's orientation. Note again that recordings were low-passed (4 kHz cutoff and 12 dB roll-off) prior to the feature extraction, hence the 'LP' suffix. 'Zylia Recordings': approximately 10 minute field recordings of the 6 study sites captured with the 19-microphone ZYLIA 'ZM-1' recorder. These recordings have been converted to third order ambisonic 'b-format' (16 channels) using Furse-Malham channel ordering and SN3D normalisation. This was achieved with the 'Zylia Ambisonics Converter' software. These third-order ambisonic recordings were used to replicate the six sites' soundscapes in the VSE. The accompanying code for this dataset has been submitted via ScholarOne with the manuscript for peer-review.
创建时间:
2024-05-10
用户留言
有没有相关的论文或文献参考?
这个数据集是基于什么背景创建的?
数据集的作者是谁?
能帮我联系到这个数据集的作者吗?
这个数据集如何下载?
点击留言
数据主题
具身智能
数据集  4099个
机构  8个
大模型
数据集  439个
机构  10个
无人机
数据集  37个
机构  6个
指令微调
数据集  36个
机构  6个
蛋白质结构
数据集  50个
机构  8个
空间智能
数据集  21个
机构  5个
5,000+
优质数据集
54 个
任务类型
进入经典数据集
热门数据集

Figshare

Figshare是一个在线数据共享平台,允许研究人员上传和共享各种类型的研究成果,包括数据集、论文、图像、视频等。它旨在促进科学研究的开放性和可重复性。

figshare.com 收录

中国劳动力动态调查

“中国劳动力动态调查” (China Labor-force Dynamics Survey,简称 CLDS)是“985”三期“中山大学社会科学特色数据库建设”专项内容,CLDS的目的是通过对中国城乡以村/居为追踪范围的家庭、劳动力个体开展每两年一次的动态追踪调查,系统地监测村/居社区的社会结构和家庭、劳动力个体的变化与相互影响,建立劳动力、家庭和社区三个层次上的追踪数据库,从而为进行实证导向的高质量的理论研究和政策研究提供基础数据。

中国学术调查数据资料库 收录

URPC系列数据集, S-URPC2019, UDD

URPC系列数据集包括URPC2017至URPC2020DL,主要用于水下目标的检测和分类。S-URPC2019专注于水下环境的特定检测任务。UDD数据集信息未在README中详细描述。

github 收录

中国区域交通网络数据集

该数据集包含中国各区域的交通网络信息,包括道路、铁路、航空和水路等多种交通方式的网络结构和连接关系。数据集详细记录了各交通节点的位置、交通线路的类型、长度、容量以及相关的交通流量信息。

data.stats.gov.cn 收录

Google Scholar

Google Scholar是一个学术搜索引擎,旨在检索学术文献、论文、书籍、摘要和文章等。它涵盖了广泛的学科领域,包括自然科学、社会科学、艺术和人文学科。用户可以通过关键词搜索、作者姓名、出版物名称等方式查找相关学术资源。

scholar.google.com 收录