Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 86
Filter
1.
Sensors (Basel) ; 24(8)2024 Apr 11.
Article in English | MEDLINE | ID: mdl-38676063

ABSTRACT

In the process of the intelligent inspection of belt conveyor systems, due to problems such as its long duration, the large number of rollers, and the complex working environment, fault diagnosis by acoustic signals is easily affected by signal coupling interference, which poses a great challenge to selecting denoising methods of signal preprocessing. This paper proposes a novel wavelet threshold denoising algorithm by integrating a new biparameter and trisegment threshold function. Firstly, we elaborate on the mutual influence and optimization process of two adjustment parameters and three wavelet coefficient processing intervals in the BT-WTD (the biparameter and trisegment of wavelet threshold denoising, BT-WTD) denoising model. Subsequently, the advantages of the proposed threshold function are theoretically demonstrated. Finally, the BT-WTD algorithm is applied to denoise the simulation signals and the vibration and acoustic signals collected from the belt conveyor experimental platform. The experimental results indicate that this method's denoising effectiveness surpasses that of traditional threshold function denoising algorithms, effectively addressing the denoising preprocessing of idler roller fault signals under strong noise backgrounds while preserving useful signal features and avoiding signal distortion problems. This research lays the theoretical foundation for the non-contact intelligent fault diagnosis of future inspection robots based on acoustic signals.

2.
Sensors (Basel) ; 24(6)2024 Mar 21.
Article in English | MEDLINE | ID: mdl-38544280

ABSTRACT

The increasing focus on the development of positioning techniques reflects the growing interest in applications and services based on indoor positioning. Many applications necessitate precise indoor positioning or tracking of individuals and assets, leading to rapid growth in products based on these technologies in certain market sectors. Ultrasonic systems have already proven effective in achieving the desired positioning accuracy and refresh rates. The typical signal used in ultrasonic positioning systems for estimating the range between the target and reference points is the linear chirp. Unfortunately, it can undergo shape aberration due to the effects of acoustic diffraction when the aperture exceeds a certain limit. The extent of the aberration is influenced by the shape and size of the transducer, as well as the angle at which the transducer is observed by the receiver. This aberration also affects the shape of the cross-correlation, causing it to lose its easily detectable characteristic of a single global peak, which typically corresponds to the correct lag associated with the signal's time of arrival. In such instances, cross-correlation techniques yield results with a significantly higher error than anticipated. In fact, the correct lag no longer corresponds to the peak of the cross-correlation. In this study, an alternative technique to global peak detection is proposed, leveraging the inherent symmetry observed in the shape of the aberrated cross-correlation. The numerical simulations, performed using the academic acoustic simulation software Field II, conducted using a typical ultrasonic chirp and ultrasonic emitter, compare the classical and the proposed range techniques in a standard office room. The analysis includes the effects of acoustical reflection in the room and of the acoustic noise at different levels of power. The results demonstrate that the proposed technique enables accurate range estimation even in the presence of severe cross-correlation shape aberrations and for signal-to-noise ratio levels common in office and room environments, even in presence of typical reflections. This allows the use of emitting transducers with a much larger aperture than that allowed by the classical cross-correlation technique. Consequently, it becomes possible to have greater acoustic power available, leading to improved signal-to-noise ratio (SNR).

3.
Annu Rev Entomol ; 69: 21-40, 2024 Jan 25.
Article in English | MEDLINE | ID: mdl-37562048

ABSTRACT

The evolution of sexual communication is critically important in the diversity of arthropods, which are declining at a fast pace worldwide. Their environments are rapidly changing, with increasing chemical, acoustic, and light pollution. To predict how arthropod species will respond to changing climates, habitats, and communities, we need to understand how sexual communication systems can evolve. In the past decades, intraspecific variation in sexual signals and responses across different modalities has been identified, but never in a comparative way. In this review, we identify and compare the level and extent of intraspecific variation in sexual signals and responses across three different modalities, chemical, acoustic, and visual, focusing mostly on insects. By comparing causes and possible consequences of intraspecific variation in sexual communication among these modalities, we identify shared and unique patterns, as well as knowledge needed to predict the evolution of sexual communication systems in arthropods in a changing world.


Subject(s)
Arthropods , Animals , Communication
4.
Curr Biol ; 34(2): 403-409.e3, 2024 01 22.
Article in English | MEDLINE | ID: mdl-38141618

ABSTRACT

The initial process by which novel sexual signals evolve remains unclear, because rare new variants are susceptible to loss by drift or counterselection imposed by prevailing female preferences.1,2,3,4 We describe the diversification of an acoustic male courtship signal in Hawaiian populations of the field cricket Teleogryllus oceanicus, which was brought about by the evolution of a brachypterous wing morph ("small-wing") only 6 years ago.5 Small-wing has a genetic basis and causes silence or reduced-amplitude signaling by miniaturizing male forewings, conferring protection against an eavesdropping parasitoid, Ormia ochracea.5 We found that wing reduction notably increases the fundamental frequency of courtship song from an average of 5.1 kHz to 6.4 kHz. It also de-canalizes male song, broadening the range of peak signal frequencies well outside normal song character space. As courtship song prompts female mounting and is sexually selected,6,7,8,9 we evaluated two scenarios to test the fate of these new signal values. Females might show reduced acceptance of small-wing males, imposing counterselection via prevailing preferences. Alternatively, females might accept small-wing males as readily as long-wing males if their window of preference is sufficiently wide. Our results support the latter. Females preferred males who produced some signal over none, but they mounted sound-producing small-wing males as often as sound-producing long-wing males. Indiscriminate mating can facilitate the persistence of rare, novel signal values. If female permissiveness is a general characteristic of the earliest stages of sexual signal evolution, then taxa with low female mate acceptance thresholds should be more prone to diversification via sexual selection.


Subject(s)
Gryllidae , Sexual Behavior, Animal , Animals , Male , Female , Wings, Animal , Hawaii , Sound , Acoustics
5.
Sensors (Basel) ; 23(21)2023 Oct 25.
Article in English | MEDLINE | ID: mdl-37960402

ABSTRACT

The method of acoustic radiation signal detection not only enables contactless measurement but also provides comprehensive state information during equipment operation. This paper proposes an enhanced feature extraction network (EFEN) for fault diagnosis of rolling bearings based on acoustic signal feature learning. The EFEN network comprises four main components: the data preprocessing module, the information feature selection module (IFSM), the channel attention mechanism module (CAMM), and the convolutional neural network module (CNNM). Firstly, the one-dimensional acoustic signal is transformed into a two-dimensional grayscale image. Then, IFSM utilizes three different-sized convolution filters to process input image data and fuse and assign weights to feature information that can attenuate noise while highlighting effective fault information. Next, a channel attention mechanism module is introduced to assign weights to each channel. Finally, the convolutional neural network (CNN) fault diagnosis module is employed for accurate classification of rolling bearing faults. Experimental results demonstrate that the EFEN network achieves high accuracy in fault diagnosis and effectively detects rolling bearing faults based on acoustic signals. The proposed method achieves an accuracy of 98.52%, surpassing other methods in terms of performance. In comparative analysis of antinoise experiments, the average accuracy remains remarkably high at 96.62%, accompanied by a significantly reduced average iteration time of only 0.25 s. Furthermore, comparative analysis confirms that the proposed algorithm exhibits excellent accuracy and resistance against noise.

6.
Phys Med Biol ; 68(23)2023 Nov 29.
Article in English | MEDLINE | ID: mdl-37820684

ABSTRACT

Radiation-induced acoustic (RA) imaging is a promising technique for visualizing the invisible radiation energy deposition in tissues, enabling new imaging modalities and real-time therapy monitoring. However, RA imaging signal often suffers from poor signal-to-noise ratios (SNRs), thus requiring measuring hundreds or even thousands of frames for averaging to achieve satisfactory quality. This repetitive measurement increases ionizing radiation dose and degrades the temporal resolution of RA imaging, limiting its clinical utility. In this study, we developed a general deep inception convolutional neural network (GDI-CNN) to denoise RA signals to substantially reduce the number of frames needed for averaging. The network employs convolutions with multiple dilations in each inception block, allowing it to encode and decode signal features with varying temporal characteristics. This design generalizes GDI-CNN to denoise acoustic signals resulting from different radiation sources. The performance of the proposed method was evaluated using experimental data of x-ray-induced acoustic, protoacoustic, and electroacoustic signals both qualitatively and quantitatively. Results demonstrated the effectiveness of GDI-CNN: it achieved x-ray-induced acoustic image quality comparable to 750-frame-averaged results using only 10-frame-averaged measurements, reducing the imaging dose of x-ray-acoustic computed tomography (XACT) by 98.7%; it realized proton range accuracy parallel to 1500-frame-averaged results using only 20-frame-averaged measurements, improving the range verification frequency in proton therapy from 0.5 to 37.5 Hz; it reached electroacoustic image quality comparable to 750-frame-averaged results using only a single frame signal, increasing the electric field monitoring frequency from 1 fps to 1k fps. Compared to lowpass filter-based denoising, the proposed method demonstrated considerably lower mean-squared-errors, higher peak-SNR, and higher structural similarities with respect to the corresponding high-frame-averaged measurements. The proposed deep learning-based denoising framework is a generalized method for few-frame-averaged acoustic signal denoising, which significantly improves the RA imaging's clinical utilities for low-dose imaging and real-time therapy monitoring.


Subject(s)
Deep Learning , Neural Networks, Computer , Tomography, X-Ray Computed/methods , Signal-To-Noise Ratio , Acoustics , Image Processing, Computer-Assisted/methods
7.
Sensors (Basel) ; 23(16)2023 Aug 20.
Article in English | MEDLINE | ID: mdl-37631819

ABSTRACT

In recent years, deep learning-based speech synthesis has attracted a lot of attention from the machine learning and speech communities. In this paper, we propose Mixture-TTS, a non-autoregressive speech synthesis model based on mixture alignment mechanism. Mixture-TTS aims to optimize the alignment information between text sequences and mel-spectrogram. Mixture-TTS uses a linguistic encoder based on soft phoneme-level alignment and hard word-level alignment approaches, which explicitly extract word-level semantic information, and introduce pitch and energy predictors to optimally predict the rhythmic information of the audio. Specifically, Mixture-TTS introduces a post-net based on a five-layer 1D convolution network to optimize the reconfiguration capability of the mel-spectrogram. We connect the output of the decoder to the post-net through the residual network. The mel-spectrogram is converted into the final audio by the HiFi-GAN vocoder. We evaluate the performance of the Mixture-TTS on the AISHELL3 and LJSpeech datasets. Experimental results show that Mixture-TTS is somewhat better in alignment information between the text sequences and mel-spectrogram, and is able to achieve high-quality audio. The ablation studies demonstrate that the structure of Mixture-TTS is effective.


Subject(s)
Linguistics , Speech , Machine Learning , Semantics
8.
Sensors (Basel) ; 23(13)2023 Jun 30.
Article in English | MEDLINE | ID: mdl-37447901

ABSTRACT

Using a novel mathematical tool called the Te-gram, researchers analyzed the energy distribution of frequency components in the scale-frequency plane. Through this analysis, a frequency band of approximately 12 Hz is identified, which can be isolated without distorting its constituent frequencies. This band, along with others, remained inseparable through conventional time-frequency analysis methods. The Te-gram successfully addresses this knowledge gap, providing multi-sensitivity in the frequency domain and effectively attenuating cross-term energy. The Daubechies 45 wavelet function was employed due to its exceptional 150 dB attenuation in the rejection band. The validation process encompassed three stages: pre-, during-, and post-seismic activity. The utilized signal corresponds to the 19 September 2017 earthquake, occurring between the states of Morelos and Puebla, Mexico. The results showcased the impressive ability of the Te-gram to surpass expectations in terms of sensitivity and energy distribution within the frequency domain. The Te-gram outperformed the procedures documented in the existing literature. On the other hand, the results show a frequency band between 0.7 Hz and 1.75 Hz, which is named the planet Earth noise.


Subject(s)
Acoustics , Noise , Environment , Mexico
9.
Sensors (Basel) ; 23(11)2023 May 23.
Article in English | MEDLINE | ID: mdl-37299714

ABSTRACT

Monitoring and maintaining the health of wind turbine blades has long been one of the challenges facing the global wind energy industry. Detecting damage to a wind turbine blade is important for planning blade repair, avoiding aggravated blade damage, and extending the sustainability of blade operation. This paper firstly introduces the existing wind turbine blade detection methods and reviews the research progress and trends of monitoring of wind turbine composite blades based on acoustic signals. Compared with other blade damage detection technologies, acoustic emission (AE) signal detection technology has the advantage of time lead. It presents the potential to detect leaf damage by detecting the presence of cracks and growth failures and can also be used to determine the location of leaf damage sources. The detection technology based on the blade aerodynamic noise signal has the potential of blade damage detection, as well as the advantages of convenient sensor installation and real-time and remote signal acquisition. Therefore, this paper focuses on the review and analysis of wind power blade structural integrity detection and damage source location technology based on acoustic signals, as well as the automatic detection and classification method of wind power blade failure mechanisms combined with machine learning algorithm. In addition to providing a reference for understanding wind power health detection methods based on AE signals and aerodynamic noise signals, this paper also points out the development trend and prospects of blade damage detection technology. It has important reference value for the practical application of non-destructive, remote, and real-time monitoring of wind power blades.


Subject(s)
Acoustics , Noise , Industry , Technology
10.
Article in English | MEDLINE | ID: mdl-37124321

ABSTRACT

In the development of acoustic signal processing algorithms, their evaluation in various acoustic environments is of utmost importance. In order to advance evaluation in realistic and reproducible scenarios, several high-quality acoustic databases have been developed over the years. In this paper, we present another complementary database of acoustic recordings, referred to as the Multi-arraY Room Acoustic Database (MYRiAD). The MYRiAD database is unique in its diversity of microphone configurations suiting a wide range of enhancement and reproduction applications (such as assistive hearing, teleconferencing, or sound zoning), the acoustics of the two recording spaces, and the variety of contained signals including 1214 room impulse responses (RIRs), reproduced speech, music, and stationary noise, as well as recordings of live cocktail parties held in both rooms. The microphone configurations comprise a dummy head (DH) with in-ear omnidirectional microphones, two behind-the-ear (BTE) pieces equipped with 2 omnidirectional microphones each, 5 external omnidirectional microphones (XMs), and two concentric circular microphone arrays (CMAs) consisting of 12 omnidirectional microphones in total. The two recording spaces, namely the SONORA Audio Laboratory (SAL) and the Alamire Interactive Laboratory (AIL), have reverberation times of 2.1 s and 0.5 s, respectively. Audio signals were reproduced using 10 movable loudspeakers in the SAL and a built-in array of 24 loudspeakers in the AIL. MATLAB and Python scripts are included for accessing the signals as well as microphone and loudspeaker coordinates. The database is publicly available (https://zenodo.org/record/7389996).

11.
Diagnostics (Basel) ; 13(10)2023 May 16.
Article in English | MEDLINE | ID: mdl-37238233

ABSTRACT

Lung auscultation has long been used as a valuable medical tool to assess respiratory health and has gotten a lot of attention in recent years, notably following the coronavirus epidemic. Lung auscultation is used to assess a patient's respiratory role. Modern technological progress has guided the growth of computer-based respiratory speech investigation, a valuable tool for detecting lung abnormalities and diseases. Several recent studies have reviewed this important area, but none are specific to lung sound-based analysis with deep-learning architectures from one side and the provided information was not sufficient for a good understanding of these techniques. This paper gives a complete review of prior deep-learning-based architecture lung sound analysis. Deep-learning-based respiratory sound analysis articles are found in different databases including the Plos, ACM Digital Libraries, Elsevier, PubMed, MDPI, Springer, and IEEE. More than 160 publications were extracted and submitted for assessment. This paper discusses different trends in pathology/lung sound, the common features for classifying lung sounds, several considered datasets, classification methods, signal processing techniques, and some statistical information based on previous study findings. Finally, the assessment concludes with a discussion of potential future improvements and recommendations.

12.
Sensors (Basel) ; 23(7)2023 Mar 31.
Article in English | MEDLINE | ID: mdl-37050691

ABSTRACT

Wireless acoustic sensor networks (WASNs) and intelligent microsystems are crucial components of the Internet of Things (IoT) ecosystem. In various IoT applications, small, lightweight, and low-power microsystems are essential to enable autonomous edge computing and networked cooperative work. This study presents an innovative intelligent microsystem with wireless networking capabilities, sound sensing, and sound event recognition. The microsystem is designed with optimized sensing, energy supply, processing, and transceiver modules to achieve small size and low power consumption. Additionally, a low-computational sound event recognition algorithm based on a Convolutional Neural Network has been designed and integrated into the microsystem. Multiple microsystems are connected using low-power Bluetooth Mesh wireless networking technology to form a meshed WASN, which is easily accessible, flexible to expand, and straightforward to manage with smartphones. The microsystem is 7.36 cm3 in size and weighs 8 g without housing. The microsystem can accurately recognize sound events in both trained and untrained data tests, achieving an average accuracy of over 92.50% for alarm sounds above 70 dB and water flow sounds above 55 dB. The microsystems can communicate wirelessly with a direct range of 5 m. It can be applied in the field of home IoT and border security.

13.
Sensors (Basel) ; 23(5)2023 Feb 28.
Article in English | MEDLINE | ID: mdl-36904847

ABSTRACT

Weld site inspection is a research area of interest in the manufacturing industry. In this study, a digital twin system for welding robots to examine various weld flaws that might happen during welding using the acoustics of the weld site is presented. Additionally, a wavelet filtering technique is implemented to remove the acoustic signal originating from machine noise. Then, an SeCNN-LSTM model is applied to recognize and categorize weld acoustic signals according to the traits of strong acoustic signal time sequences. The model verification accuracy was found to be 91%. In addition, using numerous indicators, the model was compared with seven other models, namely, CNN-SVM, CNN-LSTM, CNN-GRU, BiLSTM, GRU, CNN-BiLSTM, and LSTM. A deep learning model, and acoustic signal filtering and preprocessing techniques are integrated into the proposed digital twin system. The goal of this work was to propose a systematic on-site weld flaw detection approach encompassing data processing, system modeling, and identification methods. In addition, our proposed method could serve as a resource for pertinent research.

14.
Sensors (Basel) ; 23(6)2023 Mar 12.
Article in English | MEDLINE | ID: mdl-36991761

ABSTRACT

This study proposes a high-efficiency method using a co-prime circular microphone array (CPCMA) for the bearing fault diagnosis, and discusses the acoustic characteristics of three fault-type signals at different rotation speeds. Due to the close positions of various bearing components, radiation sounds are seriously mixed, and it is challenging to separate the fault features. Direction-of-arrival (DOA) estimation can be used to suppress noise and directionally enhance sound sources of interest; however, classical array configurations usually require a large number of microphones to achieve high accuracy. To address this, a CPCMA is introduced to raise the array's degrees of freedom in order to reduce the dependence on the microphone numbers and computation complexity. The estimation of signal parameters via rotational invariance techniques (ESPRIT) applied to a CPCMA can quickly figure out the DOA estimation without any prior knowledge. By using the techniques above, a sound source motion-tracking diagnosis method is proposed according to the movement characteristics of impact sound sources for each fault type. Additionally, more precise frequency spectra are obtained, which are used in combination to determine the fault types and locations.

15.
Ecol Evol ; 13(3): e9909, 2023 Mar.
Article in English | MEDLINE | ID: mdl-36969923

ABSTRACT

Contact calling is a ubiquitous behavior of group-living animals. Yet in birds, beyond a general connection with group cohesion, its precise function is not well-understood, nor is it clear what stimulates changes in contact call rate. In an aviary experiment, we asked whether Swinhoe's White-eyes, Zosterops simplex, would regulate their own production of contact calls to maintain a specific rate at the group level. Specifically, we hypothesized that the sudden cessation of the group-level call rate could indicate an immediate predation threat, and we predicted that birds in smaller groups would call more to maintain a high call rate. We also investigated the effects of environmental characteristics, such as vegetation density, and social stimuli, such as the presence of certain individuals, on the rate of three different contact call types. To calculate mean individual-level rates, we measured the group-level rate and divided it by the number of birds in the aviary. We found that the individual-level rate of the most common call types increased with a greater group size, the opposite pattern to what would be expected if birds were maintaining a specific group-level rate. Vegetation density did not affect any call rate. However, individual-level rates of all call types decreased when birds were in subgroups with individuals of differing dominance status, and the rate of some call types increased when birds were with affiliated individuals. Our results do not support the hypothesis that contact calls are related to habitat structure or immediate predation risk. Rather, they appear to have a social function, used for communication within or between groups depending on the call type. Increases in call rates could recruit affiliated individuals, whereas subordinates could withhold calls so that dominants are unable to locate them, leading to fluctuations in contact calling in different social contexts.

16.
IEEE Open J Eng Med Biol ; 3: 134-141, 2022.
Article in English | MEDLINE | ID: mdl-36578775

ABSTRACT

Goal: The evaluation of respiratory events using audio sensing in an at-home setting can be indicative of worsening health conditions. This paper investigates the use of image-based transfer learning applied to five audio visualizations to evaluate three classification tasks (C1: wet vs. dry vs. whooping cough vs. restricted breathing; C2: wet vs. dry cough; C3: cough vs. restricted breathing). Methods: The five visualizations (linear spectrogram, logarithmic spectrogram, Mel-spectrogram, wavelet scalograms, and aggregate images) are applied to a pre-trained AlexNet image classifier for all tasks. Results: The aggregate image-based classifier achieved the highest overall performance across all tasks with C1, C2, and C3 having testing accuracies of 0.88, 0.88, and 0.91 respectively. However, the Mel-spectrogram method had the highest testing accuracy (0.94) for C2. Conclusions: The classification of respiratory events using aggregate image inputs to transfer learning approaches may help healthcare professionals by providing information that would otherwise be unavailable to them.

17.
Sensors (Basel) ; 22(19)2022 Sep 27.
Article in English | MEDLINE | ID: mdl-36236443

ABSTRACT

With the emergence of COVID-19, social distancing detection is a crucial technique for epidemic prevention and control. However, the current mainstream detection technology cannot obtain accurate social distance in real-time. To address this problem, this paper presents a first study on smartphone-based social distance detection technology based on near-ultrasonic signals. Firstly, according to auditory characteristics of the human ear and smartphone frequency response characteristics, a group of 18 kHz-23 kHz inaudible Chirp signals accompanied with single frequency signals are designed to complete ranging and ID identification in a short time. Secondly, an improved mutual ranging algorithm is proposed by combining the cubic spline interpolation and a two-stage search to obtain robust mutual ranging performance against multipath and NLoS affect. Thirdly, a hybrid channel access protocol is proposed consisting of Chirp BOK, FDMA, and CSMA/CA to increase the number of concurrencies and reduce the probability of collision. The results show that in our ranging algorithm, 95% of the mutual ranging error within 5 m is less than 10 cm and gets the best performance compared to the other traditional methods in both LoS and NLoS. The protocol can efficiently utilize the limited near-ultrasonic channel resources and achieve a high refresh rate ranging under the premise of reducing the collision probability. Our study can realize high-precision, high-refresh-rate social distance detection on smartphones and has significant application value during an epidemic.


Subject(s)
COVID-19 , Smartphone , Humans , Physical Distancing , Technology , Ultrasonics
18.
Smart Health (Amst) ; 26: 100329, 2022 Dec.
Article in English | MEDLINE | ID: mdl-36275046

ABSTRACT

With the emergence of the COVID-19 pandemic, early diagnosis of lung diseases has attracted growing attention. Generally, monitoring the breathing sound is the traditional means for assessing the status of a patient's respiratory health through auscultation; for that a stethoscope is one of the clinical tools used by physicians for diagnosis of lung disease and anomalies. On the other hand, recent technological advances have made telehealth systems a practical and effective option for health status assessment and remote patient monitoring. The interest in telehealth solutions have further grown with the COVID-19 pandemic. These telehealth systems aim to provide increased safety and help to cope with the massive growth in healthcare demand. Particularly, employing acoustic sensors to collect breathing sound would enable real-time assessment and instantaneous detection of anomalies. However, existing work focuses on autonomous determination of respiratory rate which is not suitable for anomaly detection due to inability to deal with noisy data recording. This paper presents a novel approach for effective breathing sound analysis. We promote a new segmentation mechanism of the captured acoustic signals to identify breathing cycles in recorded sound signals. A scoring scheme is applied to qualify the segment based on the targeted respiratory illness by the overall breathing sound analysis. We demonstrate the effectiveness of our approach via experiments using published COPD datasets.

19.
Front Psychol ; 13: 964209, 2022.
Article in English | MEDLINE | ID: mdl-36312201

ABSTRACT

Taxonomies and ontologies for the characterization of everyday sounds have been developed in several research fields, including auditory cognition, soundscape research, artificial hearing, sound design, and medicine. Here, we surveyed 36 of such knowledge organization systems, which we identified through a systematic literature search. To evaluate the semantic domains covered by these systems within a homogeneous framework, we introduced a comprehensive set of verbal sound descriptors (sound source properties; attributes of sensation; sound signal descriptors; onomatopoeias; music genres), which we used to manually label the surveyed descriptor classes. We reveal that most taxonomies and ontologies were developed to characterize higher-level semantic relations between sound sources in terms of the sound-generating objects and actions involved (what/how), or in terms of the environmental context (where). This indicates the current lack of a comprehensive ontology of everyday sounds that covers simultaneously all semantic aspects of the relation between sounds. Such an ontology may have a wide range of applications and purposes, ranging from extending our scientific knowledge of auditory processes in the real world, to developing artificial hearing systems.

20.
Sensors (Basel) ; 22(18)2022 Sep 17.
Article in English | MEDLINE | ID: mdl-36146382

ABSTRACT

This work presents the design of a wireless acoustic sensor network (WASN) that monitors indoor spaces. The proposed network would enable the acquisition of valuable information on the behavior of the inhabitants of the space. This WASN has been conceived to work in any type of indoor environment, including houses, hospitals, universities or even libraries, where the tracking of people can give relevant insight, with a focus on ambient assisted living environments. The proposed WASN has several priorities and differences compared to the literature: (i) presenting a low-cost flexible sensor able to monitor wide indoor areas; (ii) balance between acoustic quality and microphone cost; and (iii) good communication between nodes to increase the connectivity coverage. A potential application of the proposed network could be the generation of a sound map of a certain location (house, university, offices, etc.) or, in the future, the acoustic detection of events, giving information about the behavior of the inhabitants of the place under study. Each node of the network comprises an omnidirectional microphone and a computation unit, which processes acoustic information locally following the edge-computing paradigm to avoid sending raw data to a cloud server, mainly for privacy and connectivity purposes. Moreover, this work explores the placement of acoustic sensors in a real scenario, following acoustic coverage criteria. The proposed network aims to encourage the use of real-time non-invasive devices to obtain behavioral and environmental information, in order to take decisions in real-time with the minimum intrusiveness in the location under study.


Subject(s)
Acoustics , Humans , Monitoring, Physiologic
SELECTION OF CITATIONS
SEARCH DETAIL
...