Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 7.291
Filtrar
1.
Front Neurorobot ; 18: 1430155, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39220587

RESUMEN

Introduction: Unmanned aerial vehicles (UAVs) are widely used in various computer vision applications, especially in intelligent traffic monitoring, as they are agile and simplify operations while boosting efficiency. However, automating these procedures is still a significant challenge due to the difficulty of extracting foreground (vehicle) information from complex traffic scenes. Methods: This paper presents a unique method for autonomous vehicle surveillance that uses FCM to segment aerial images. YOLOv8, which is known for its ability to detect tiny objects, is then used to detect vehicles. Additionally, a system that utilizes ORB features is employed to support vehicle recognition, assignment, and recovery across picture frames. Vehicle tracking is accomplished using DeepSORT, which elegantly combines Kalman filtering with deep learning to achieve precise results. Results: Our proposed model demonstrates remarkable performance in vehicle identification and tracking with precision of 0.86 and 0.84 on the VEDAI and SRTID datasets, respectively, for vehicle detection. Discussion: For vehicle tracking, the model achieves accuracies of 0.89 and 0.85 on the VEDAI and SRTID datasets, respectively.

2.
Heliyon ; 10(16): e35929, 2024 Aug 30.
Artículo en Inglés | MEDLINE | ID: mdl-39224340

RESUMEN

A considerable number of vehicular accidents occur in low-millage zones like school streets, neighborhoods, and parking lots, among others. Therefore, the proposed work aims to provide a novel ADAS system to warn about dangerous scenarios by analyzing the driver's attention and the corresponding distances between the vehicle and the detected object on the road. This approach is made possible by concurrent Head Pose Estimation (HPE) and Object/Pedestrian Detection. Both approaches have shown independently their viable application in the automotive industry to decrease the number of vehicle collisions. The proposed system takes advantage of stereo vision characteristics for HPE by enabling the computation of the Euler Angles with a low average error for classifying the driver's attention on the road using neural networks. For Object Detection, stereo vision is used to detect the distance between the vehicle and the approaching object; this is made with a state-of-the-art algorithm known as YOLO-R and a fast template matching technique known as SoRA that provides lower processing times. The result is an ADAS system designed to ensure adequate braking time, considering the driver's attention on the road and the distances to objects.

3.
Artículo en Inglés | MEDLINE | ID: mdl-39209101

RESUMEN

Chromogranin A (CgA), a âˆ¼ 49 kDa acidic secretory protein, is ubiquitously distributed in endocrine and neuroendocrine cells and neurons. As a propeptide, CgA is proteolytically cleaved to generate several peptides of biological importance, including pancreastatin (PST: hCgA250-301), Vasostatin 1 (VS1: hCgA1-76), and catestatin (CST: CgA 352-372). VS1 represents the most conserved fragment of CgA. A 20 amino acid domain within VS1 (CgA 47-66) exhibits potent antimicrobial and anti-inflammatory activities. Autism is known to be associated with inflammation. Therefore, we seek to test the hypothesis that VS1 modulates autism behaviors by reducing inflammation in the hippocampus. Treatment of C57BL/6 (B6) and BTBR (a mouse model of idiopathic autism) mice with VS1 revealed the following: BTBR mice showed a significant decrease in chamber time in the presence of a stranger or a novel object. Treatment with VS1 significantly increased chamber time in both cases, underscoring a crucial role for VS1 in improving behavioral deficits in BTBR mice. In contrast to chamber time, sniffing time in BTBR mice in the presence of a stranger was less compared to B6 control mice. VS1 did not improve this latter parameter. Surprisingly, sniffing time in BTBR mice in the presence of a novel object was comparable with B6 mice. Proinflammatory cytokines such as IL-6 and IL-1b, as well as other inflammatory markers, were elevated in BTBR mice, which were dramatically reduced after supplementation with VS1. Interestingly, even Beclin-1/p62, pAKT/AKT, and p-p70-S6K/p70-S6K ratios were notably reduced by VS1. We conclude that VS1 plays a crucial role in restoring autistic spectrum disorders (ASD) plausibly by attenuating neuroinflammation.

4.
Sci Rep ; 14(1): 20086, 2024 08 29.
Artículo en Inglés | MEDLINE | ID: mdl-39209919

RESUMEN

This study compared the multiple object tracking (MOT) performance of athletes vs. non-athletes and expert athletes vs. novice athletes by systematically reviewing and meta-analyzing the literature. A systematic literature search was conducted using five databases for articles published until July 2024. Healthy people were included, specifically classified as athletes and non-athletes, or experts and novices. Potential sources of heterogeneity were selected using a random-effects model. Moderator analyses were also performed. A total of 23 studies were included in this review. Regarding the overall effect, athletes were significantly better at MOT tasks than non-athletes, and experts performed better than novices. Subgroup analyses showed that expert athletes had a significantly larger effect than novices, and that the type of sport significantly moderated the difference in MOT performance between the two groups. Meta-regression revealed that the number of targets and duration of tracking moderated the differences in performance between experts and novices, but did not affect the differences between athletes and non-athletes. This meta-analysis provides evidence of performance advantages for athletes compared with nonathletes, and experts compared with novices in MOT tasks. Moreover, the two effects were moderated by different factors; therefore, future studies should classify participants more specifically according to sports levels.


Asunto(s)
Atletas , Rendimiento Atlético , Humanos , Rendimiento Atlético/fisiología , Desempeño Psicomotor/fisiología , Masculino , Deportes/fisiología
5.
Sci Rep ; 14(1): 20090, 2024 Aug 29.
Artículo en Inglés | MEDLINE | ID: mdl-39209928

RESUMEN

Remote Sensing Image Object Detection (RSIOD) faces the challenges of multi-scale objects, dense overlap of objects and uneven data distribution in practical applications. In order to solve these problems, this paper proposes a YOLO-ACPHD RSIOD algorithm. The algorithm adopts Adaptive Condition Awareness Technology (ACAT), which can dynamically adjust the parameters of the convolution kernel, so as to adapt to the objects of different scales and positions. Compared with the traditional fixed convolution kernel, this dynamic adjustment can better adapt to the diversity of scale, direction and shape of the object, thus improving the accuracy and robustness of Object Detection (OD). In addition, a High-Dimensional Decoupling Technology (HDDT) is used to reduce the amount of calculation to 1/N by performing deep convolution on the input data and then performing spatial convolution on each channel. When dealing with large-scale Remote Sensing Image (RSI) data, this reduction in computation can significantly improve the efficiency of the algorithm and accelerate the speed of OD, so as to better adapt to the needs of practical application scenarios. Through the experimental verification of the RSOD RSI data set, the YOLO-ACPHD model in this paper shows very satisfactory performance. The F1 value reaches 0.99, the Precision value reaches 1, the Precision-Recall value reaches 0.994, the Recall value reaches 1, and the mAP value reaches 99.36 % , which indicates that the model shows the highest level in the accuracy and comprehensiveness of OD.

6.
Sci Rep ; 14(1): 19945, 2024 Aug 27.
Artículo en Inglés | MEDLINE | ID: mdl-39198568

RESUMEN

Soil texture is one of the most important elements to consider before planting and tillage. These features affect the product selection and regulate its water permeability. Discrimination of soils by determining soil texture features requires an intense workload and is time-consuming. Therefore, having a powerful tool and knowledge for texture-based soil discrimination could enable rapid and accurate discrimination of soils. This study focuses on presenting new models for 6 different soil sample groups (Soil_1 to Soil_6) based on 12 different machine learning algorithms that can be utilized for various problems. As a result, overall accuracy values were determined as greater than 99.2% (Trilayered Neural Network). The greatest accuracy value was found in Bayes Net (99.83%) and followed by Subspace Discriminant (99.80%). In the Bayes Net algorithm, MCC (Matthews Correlation Coefficient) and F-measure values were obtained as 0.994 and 0.995 for Soil_4 and Soil_6 sample groups while these values were 1.000 for other soil groups. Soil types can visually vary based on their texture, mineral composition, and moisture levels. The variability of this can be influenced by fertilization, precipitation levels, and soil cultivation. It is important to capture the images in soil conditions that are more stable. In conclusion, the present study has proven the feasibility of rapid, non-destructive, and accurate discrimination of soils by image processing-based machine learning.

7.
BioData Min ; 17(1): 27, 2024 Aug 28.
Artículo en Inglés | MEDLINE | ID: mdl-39198921

RESUMEN

Cardiovascular diseases are the main cause of death in the world and cardiovascular imaging techniques are the mainstay of noninvasive diagnosis. Aortic stenosis is a lethal cardiac disease preceded by aortic valve calcification for several years. Data-driven tools developed with Deep Learning (DL) algorithms can process and categorize medical images data, providing fast diagnoses with considered reliability, to improve healthcare effectiveness. A systematic review of DL applications on medical images for pathologic calcium detection concluded that there are established techniques in this field, using primarily CT scans, at the expense of radiation exposure. Echocardiography is an unexplored alternative to detect calcium, but still needs technological developments. In this article, a fully automated method based on Convolutional Neural Networks (CNNs) was developed to detect Aortic Calcification in Echocardiography images, consisting of two essential processes: (1) an object detector to locate aortic valve - achieving 95% of precision and 100% of recall; and (2) a classifier to identify calcium structures in the valve - which achieved 92% of precision and 100% of recall. The outcome of this work is the possibility of automation of the detection with Echocardiography of Aortic Valve Calcification, a lethal and prevalent disease.

8.
Sensors (Basel) ; 24(16)2024 Aug 08.
Artículo en Inglés | MEDLINE | ID: mdl-39204832

RESUMEN

Camera-based object detection is integral to advanced driver assistance systems (ADAS) and autonomous vehicle research, and RGB cameras remain indispensable for their spatial resolution and color information. This study investigates exposure time optimization for such cameras, considering image quality in dynamic ADAS scenarios. Exposure time, the period during which the camera sensor is exposed to light, directly influences the amount of information captured. In dynamic scenarios, such as those encountered in typical driving scenarios, optimizing exposure time becomes challenging due to the inherent trade-off between Signal-to-Noise Ratio (SNR) and motion blur, i.e., extending exposure time to maximize information capture increases SNR, but also increases the risk of motion blur and overexposure, particularly in low-light conditions where objects may not be fully illuminated. The study introduces a comprehensive methodology for exposure time optimization under various lighting conditions, examining its impact on image quality and computer vision performance. Traditional image quality metrics show a poor correlation with computer vision performance, highlighting the need for newer metrics that demonstrate improved correlation. The research presented in this paper offers guidance into the enhancement of single-exposure camera-based systems for automotive applications. By addressing the balance between exposure time, image quality, and computer vision performance, the findings provide a road map for optimizing camera settings for ADAS and autonomous driving technologies, contributing to safety and performance advancements in the automotive landscape.

9.
Sensors (Basel) ; 24(16)2024 Aug 11.
Artículo en Inglés | MEDLINE | ID: mdl-39204886

RESUMEN

To achieve Level 4 and above autonomous driving, a robust and stable autonomous driving system is essential to adapt to various environmental changes. This paper aims to perform vehicle pose estimation, a crucial element in forming autonomous driving systems, more universally and robustly. The prevalent method for vehicle pose estimation in autonomous driving systems relies on Real-Time Kinematic (RTK) sensor data, ensuring accurate location acquisition. However, due to the characteristics of RTK sensors, precise positioning is challenging or impossible in indoor spaces or areas with signal interference, leading to inaccurate pose estimation and hindering autonomous driving in such scenarios. This paper proposes a method to overcome these challenges by leveraging objects registered in a high-precision map. The proposed approach involves creating a semantic high-definition (HD) map with added objects, forming object-centric features, recognizing locations using these features, and accurately estimating the vehicle's pose from the recognized location. This proposed method enhances the precision of vehicle pose estimation in environments where acquiring RTK sensor data is challenging, enabling more robust and stable autonomous driving. The paper demonstrates the proposed method's effectiveness through simulation and real-world experiments, showcasing its capability for more precise pose estimation.

10.
Sensors (Basel) ; 24(16)2024 Aug 11.
Artículo en Inglés | MEDLINE | ID: mdl-39204895

RESUMEN

Poplar (Populus) trees play a vital role in various industries and in environmental sustainability. They are widely used for paper production, timber, and as windbreaks, in addition to their significant contributions to carbon sequestration. Given their economic and ecological importance, effective disease management is essential. Convolutional Neural Networks (CNNs), particularly adept at processing visual information, are crucial for the accurate detection and classification of plant diseases. This study introduces a novel dataset of manually collected images of diseased poplar leaves from Uzbekistan and South Korea, enhancing the geographic diversity and application of the dataset. The disease classes consist of "Parsha (Scab)", "Brown-spotting", "White-Gray spotting", and "Rust", reflecting common afflictions in these regions. This dataset will be made publicly available to support ongoing research efforts. Employing the advanced YOLOv8 model, a state-of-the-art CNN architecture, we applied a Contrast Stretching technique prior to model training in order to enhance disease detection accuracy. This approach not only improves the model's diagnostic capabilities but also offers a scalable tool for monitoring and treating poplar diseases, thereby supporting the health and sustainability of these critical resources. This dataset, to our knowledge, will be the first of its kind to be publicly available, offering a valuable resource for researchers and practitioners worldwide.


Asunto(s)
Redes Neurales de la Computación , Enfermedades de las Plantas , Hojas de la Planta , Populus , Procesamiento de Imagen Asistido por Computador/métodos , República de Corea
11.
Sensors (Basel) ; 24(16)2024 Aug 14.
Artículo en Inglés | MEDLINE | ID: mdl-39204965

RESUMEN

Winter is the season of main concern for beekeepers since the temperature, humidity, and potential infection from mites and other diseases may lead the colony to death. As a consequence, beekeepers perform invasive checks on the colonies, exposing them to further harm. This paper proposes a novel design of an instrumented beehive involving color cameras placed inside the beehive and at the bottom of it, paving the way for new frontiers in beehive monitoring. The overall acquisition system is described focusing on design choices towards an effective solution for internal, contactless, and stress-free beehive monitoring. To validate our approach, we conducted an experimental campaign in 2023 and analyzed the collected images with YOLOv8 to understand if the proposed solution can be useful for beekeepers and what kind of information can be derived from this kind of monitoring, including the presence of Varroa destructor mites inside the beehive. We experimentally found that the observation point inside the beehive is the most challenging due to the frequent movements of the bees and the difficulties related to obtaining in-focus images. However, from these images, it is possible to find Varroa destructor mites. On the other hand, the observation point at the bottom of the beehive showed great potential for understanding the overall activity of the colony.


Asunto(s)
Varroidae , Abejas/fisiología , Abejas/parasitología , Animales , Varroidae/fisiología , Varroidae/patogenicidad , Apicultura/métodos
12.
Sensors (Basel) ; 24(16)2024 Aug 16.
Artículo en Inglés | MEDLINE | ID: mdl-39205011

RESUMEN

When it comes to road environment perception, millimeter-wave radar with a camera facilitates more reliable detection than a single sensor. However, the limited utilization of radar features and insufficient extraction of important features remain pertinent issues, especially with regard to the detection of small and occluded objects. To address these concerns, we propose a camera-radar fusion with radar channel extension and a dual-CBAM-FPN (CRFRD), which incorporates a radar channel extension (RCE) module and a dual-CBAM-FPN (DCF) module into the camera-radar fusion net (CRF-Net). In the RCE module, we design an azimuth-weighted RCS parameter and extend three radar channels, which leverage the secondary redundant information to achieve richer feature representation. In the DCF module, we present the dual-CBAM-FPN, which enables the model to focus on important features by inserting CBAM at the input and the fusion process of FPN simultaneously. Comparative experiments conducted on the NuScenes dataset and real data demonstrate the superior performance of the CRFRD compared to CRF-Net, as its weighted mean average precision (wmAP) increases from 43.89% to 45.03%. Furthermore, ablation studies verify the indispensability of the RCE and DCF modules and the effectiveness of azimuth-weighted RCS.

13.
Sensors (Basel) ; 24(16)2024 Aug 18.
Artículo en Inglés | MEDLINE | ID: mdl-39205035

RESUMEN

In optical remote sensing image object detection, discontinuous boundaries often limit detection accuracy, particularly at high Intersection over Union (IoU) thresholds. This paper addresses this issue by proposing the Spatial Adaptive Angle-Aware (SA3) Network. The SA3 Network employs a hierarchical refinement approach, consisting of coarse regression, fine regression, and precise tuning, to optimize the angle parameters of rotated bounding boxes. It adapts to specific task scenarios using either class-aware or class-agnostic strategies. Experimental results demonstrate its effectiveness in significantly improving detection accuracy at high IoU thresholds. Additionally, we introduce a Gaussian transform-based IoU factor during angle regression loss calculation, leading to the development of Edge-aware Skewed Bounding Box Loss (EAS Loss). The EAS loss enhances the loss gradient at the final stage of angle regression for bounding boxes, addressing the challenge of further learning when the predicted box angle closely aligns with the real target box angle. This results in increased training efficiency and better alignment between training and evaluation metrics. Experimental results show that the proposed method substantially enhances the detection accuracy of ReDet and ReBiDet models. The SA3 Network and EAS loss not only elevate the mAP of the ReBiDet model on DOTA-v1.5 to 78.85% but also effectively improve the model's mAP under high IoU threshold conditions.

14.
Sensors (Basel) ; 24(16)2024 Aug 19.
Artículo en Inglés | MEDLINE | ID: mdl-39205049

RESUMEN

Robots need to sense information about the external environment before moving, which helps them to recognize and understand their surroundings so that they can plan safe and effective paths and avoid obstacles. Conventional algorithms using a single sensor cannot obtain enough information and lack real-time capabilities. To solve these problems, we propose an information perception algorithm with vision as the core and the fusion of LiDAR. Regarding vision, we propose the YOLO-SCG model, which is able to detect objects faster and more accurately. When processing point clouds, we integrate the detection results of vision for local clustering, improving both the processing speed of the point cloud and the detection effectiveness. Experiments verify that our proposed YOLO-SCG algorithm improves accuracy by 4.06% and detection speed by 7.81% compared to YOLOv9, and our algorithm excels in distinguishing different objects in the clustering of point clouds.

15.
Sensors (Basel) ; 24(16)2024 Aug 20.
Artículo en Inglés | MEDLINE | ID: mdl-39205068

RESUMEN

Referring video object segmentation (R-VOS) is a fundamental vision-language task which aims to segment the target referred by language expression in all video frames. Existing query-based R-VOS methods have conducted in-depth exploration of the interaction and alignment between visual and linguistic features but fail to transfer the information of the two modalities to the query vector with balanced intensities. Furthermore, most of the traditional approaches suffer from severe information loss in the process of multi-scale feature fusion, resulting in inaccurate segmentation. In this paper, we propose DCT, an end-to-end decoupled cross-modal transformer for referring video object segmentation, to better utilize multi-modal and multi-scale information. Specifically, we first design a Language-Guided Visual Enhancement Module (LGVE) to transmit discriminative linguistic information to visual features of all levels, performing an initial filtering of irrelevant background regions. Then, we propose a decoupled transformer decoder, using a set of object queries to gather entity-related information from both visual and linguistic features independently, mitigating the attention bias caused by feature size differences. Finally, the Cross-layer Feature Pyramid Network (CFPN) is introduced to preserve more visual details by establishing direct cross-layer communication. Extensive experiments have been carried out on A2D-Sentences, JHMDB-Sentences and Ref-Youtube-VOS. The results show that DCT achieves competitive segmentation accuracy compared with the state-of-the-art methods.

16.
Sensors (Basel) ; 24(16)2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39205117

RESUMEN

3D object-detection based on LiDAR point clouds can help driverless vehicles detect obstacles. However, the existing point-cloud-based object-detection methods are generally ineffective in detecting small objects such as pedestrians and cyclists. Therefore, a small-object-detection algorithm based on clustering is proposed. Firstly, a new segmented ground-point clouds segmentation algorithm is proposed, which filters out the object point clouds according to the heuristic rules and realizes the ground segmentation by multi-region plane-fitting. Then, the small-object point cloud is clustered using an improved DBSCAN clustering algorithm. The K-means++ algorithm for pre-clustering is used, the neighborhood radius is adaptively adjusted according to the distance, and the core point search method of the original algorithm is improved. Finally, the detection of small objects is completed using the directional wraparound box model. After extensive experiments, it was shown that the precision and recall of our proposed ground-segmentation algorithm reached 91.86% and 92.70%, respectively, and the improved DBSCAN clustering algorithm improved the recall of pedestrians and cyclists by 15.89% and 9.50%, respectively. In addition, visualization experiments confirmed that our proposed small-object-detection algorithm based on the point-cloud clustering method can realize the accurate detection of small objects.

17.
Sensors (Basel) ; 24(16)2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39205134

RESUMEN

LiDAR sensors have been shown to generate data with various common corruptions, which seriously affect their applications in 3D vision tasks, particularly object detection. At the same time, it has been demonstrated that traditional defense strategies, including adversarial training, are prone to suffering from gradient confusion during training. Moreover, they can only improve their robustness against specific types of data corruption. In this work, we propose LiDARPure, which leverages the powerful generation ability of diffusion models to purify corruption in the LiDAR scene data. By dividing the entire scene into voxels to facilitate the processes of diffusion and reverse diffusion, LiDARPure overcomes challenges induced from adversarial training, such as sparse point clouds in large-scale LiDAR data and gradient confusion. In addition, we utilize the latent geometric features of a scene as a condition to assist the generation of diffusion models. Detailed experiments show that LiDARPure can effectively purify 19 common types of LiDAR data corruption. Further evaluation results demonstrate that it can improve the average precision of 3D object detectors to an extent of 20% in the face of data corruption, much higher than existing defence strategies.

18.
Neurobiol Learn Mem ; 214: 107971, 2024 Aug 11.
Artículo en Inglés | MEDLINE | ID: mdl-39137861

RESUMEN

Exercise provides a range of cognitive benefits, including improved memory performance. Previously, we demonstrated that 14 days of continuous voluntary wheel-running exercise enables learning in a hippocampus-dependent Object Location Memory (OLM) task under insufficient, subthreshold training conditions in adult mice. Whether similar exercise benefits can be obtained from consistent intermittent exercise as continuous exercise is unknown. Here, we examine whether intermittent exercise (the weekend warrior effect: 2 days of exercise a week for 7 weeks) displays similar or distinct cognitive benefits as previously examined with 14 days of continuous exercise. We find that both continuous and intermittent exercise parameters similarly enable hippocampus-dependent OLM compared to the 2-day exercise control group. Mice receiving intermittent exercise maintained cognitive benefits following a 7-day sedentary delay, whereas mice that underwent 14 continuous days of exercise showed diminished cognitive benefits as previously reported. Further, compared to continuous exercise, intermittent exercise mice exhibited persistently elevated levels of the genes Acvr1c and Bdnf which we know to be critically involved in hippocampus-dependent long-term memory in the dorsal hippocampus. Together findings suggest that consistent intermittent exercise persistently enables hippocampal-dependent long-term memory. Understanding the optimal parameters for persistent cognitive function and the mechanisms mediating persistent effects will aid in therapeutic pursuits investigating the mitigation of cognitive ailments.

19.
Plant Methods ; 20(1): 127, 2024 Aug 16.
Artículo en Inglés | MEDLINE | ID: mdl-39152496

RESUMEN

BACKGROUND: Ripeness is a phenotype that significantly impacts the quality of fruits, constituting a crucial factor in the cultivation and harvesting processes. Manual detection methods and experimental analysis, however, are inefficient and costly. RESULTS: In this study, we propose a lightweight and efficient melon ripeness detection method, MRD-YOLO, based on an improved object detection algorithm. The method combines a lightweight backbone network, MobileNetV3, a design paradigm Slim-neck, and a Coordinate Attention mechanism. Additionally, we have created a large-scale melon dataset sourced from a greenhouse based on ripeness. This dataset contains common complexities encountered in the field environment, such as occlusions, overlapping, and varying light intensities. MRD-YOLO achieves a mean Average Precision of 97.4% on this dataset, achieving accurate and reliable melon ripeness detection. Moreover, the method demands only 4.8 G FLOPs and 2.06 M parameters, representing 58.5% and 68.4% of the baseline YOLOv8n model, respectively. It comprehensively outperforms existing methods in terms of balanced accuracy and computational efficiency. Furthermore, it maintains real-time inference capability in GPU environments and demonstrates exceptional inference speed in CPU environments. The lightweight design of MRD-YOLO is anticipated to be deployed in various resource constrained mobile and edge devices, such as picking robots. Particularly noteworthy is its performance when tested on two melon datasets obtained from the Roboflow platform, achieving a mean Average Precision of 85.9%. This underscores its excellent generalization ability on untrained data. CONCLUSIONS: This study presents an efficient method for melon ripeness detection, and the dataset utilized in this study, alongside the detection method, will provide a valuable reference for ripeness detection across various types of fruits.

20.
Neural Netw ; 179: 106623, 2024 Aug 10.
Artículo en Inglés | MEDLINE | ID: mdl-39154419

RESUMEN

LiDAR point clouds can effectively depict the motion and posture of objects in three-dimensional space. Many studies accomplish the 3D object detection by voxelizing point clouds. However, in autonomous driving scenarios, the sparsity and hollowness of point clouds create some difficulties for voxel-based methods. The sparsity of point clouds makes it challenging to describe the geometric features of objects. The hollowness of point clouds poses difficulties for the aggregation of 3D features. We propose a two-stage 3D object detection framework, called MS23D. (1) We propose a method using voxel feature points from multi-branch to construct the 3D feature layer. Using voxel feature points from different branches, we construct a relatively compact 3D feature layer with rich semantic features. Additionally, we propose a distance-weighted sampling method, reducing the loss of foreground points caused by downsampling and allowing the 3D feature layer to retain more foreground points. (2) In response to the hollowness of point clouds, we predict the offsets between deep-level feature points and the object's centroid, making them as close as possible to the object's centroid. This enables the aggregation of these feature points with abundant semantic features. For feature points from shallow-level, we retain them on the object's surface to describe the geometric features of the object. To validate our approach, we evaluated its effectiveness on both the KITTI and ONCE datasets.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA