Applied Intelligence, Год журнала: 2024, Номер 55(2)
Опубликована: Дек. 9, 2024
Язык: Английский
Applied Intelligence, Год журнала: 2024, Номер 55(2)
Опубликована: Дек. 9, 2024
Язык: Английский
Proceedings of the IEEE, Год журнала: 2024, Номер 112(6), С. 544 - 584
Опубликована: Июнь 1, 2024
Язык: Английский
Процитировано
132022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Год журнала: 2024, Номер 34, С. 5610 - 5619
Опубликована: Июнь 16, 2024
Язык: Английский
Процитировано
6Lecture notes in computer science, Год журнала: 2024, Номер unknown, С. 253 - 272
Опубликована: Ноя. 22, 2024
Язык: Английский
Процитировано
4Biomimetics, Год журнала: 2025, Номер 10(1), С. 48 - 48
Опубликована: Янв. 14, 2025
A future unmanned system needs the ability to perceive, decide and control in an open dynamic environment. In order fulfill this requirement, it construct a method with universal environmental perception ability. Moreover, perceptual process be interpretable understandable, so that interactions between systems humans can unimpeded. However, current mainstream DNN (deep learning neural network)-based AI (artificial intelligence) is 'black box'. We cannot interpret or understand how decision made by these AIs. An SNN (spiking network), which more similar biological brain than DNN, has potential implement understandable AI. work, we propose neuron group-based structural for better capture spatial temporal information from external environment, time-slicing scheme of responses generated SNN. Results show our indeed helps enhance environment SNN, possesses certain degree robustness, enhancing build future.
Язык: Английский
Процитировано
0Computer Science and Application, Год журнала: 2025, Номер 15(01), С. 187 - 198
Опубликована: Янв. 1, 2025
Язык: Английский
Процитировано
0Electronics, Год журнала: 2025, Номер 14(6), С. 1105 - 1105
Опубликована: Март 11, 2025
Robust object detection in challenging scenarios remains a critical challenge for autonomous driving systems. Inspired by human visual perception, integrating the complementary modalities of RGB frames and event streams presents promising approach to achieving robust detection. However, existing multimodal detectors achieve superior performance at cost significant computational power consumption. To address this challenge, we propose novel spiking RGB–event fusion-based network (SFDNet), fully detector capable both low-power high-performance Specifically, first introduce Leaky Integrate-and-Multi-Fire (LIMF) neuron model, which combines soft hard reset mechanisms enhance feature representation SNNs. We then develop multi-scale hierarchical residual attention lightweight aggregation module efficient dual-modality extraction fusion. Experimental results on two public datasets demonstrate that our SFDNet achieves state-of-the-art with remarkably low The scenarios, such as motion blur low-light conditions, highlights robustness effectiveness SFDNet, significantly advancing applicability SNNs real-world tasks.
Язык: Английский
Процитировано
0Scientific Reports, Год журнала: 2025, Номер 15(1)
Опубликована: Апрель 21, 2025
Abstract In response to the increasing demands for improved model performance and reduced energy consumption in object detection tasks relevant autonomous driving, this research presents an advanced YOLO model, designated as ECSLIF-YOLO, which is based on Leaky Integrate-and-Fire with Extracellular Space (ECS-LIF) framework. The primary aim of tackle issues associated high traditional artificial neural networks (ANNs) suboptimal existing spiking (SNNs). Empirical findings demonstrate that ECSLIF-YOLO achieves a peak mean Average Precision (mAP) 0.917 BDD100K KITTI datasets, thereby aligning accuracy levels conventional ANNs while exceeding current direct-training SNN approaches without incurring additional costs. These suggest particularly well-suited assist development efficient reliable systems driving.
Язык: Английский
Процитировано
0Neurocomputing, Год журнала: 2025, Номер unknown, С. 130253 - 130253
Опубликована: Апрель 1, 2025
Язык: Английский
Процитировано
0IEEE Transactions on Cognitive and Developmental Systems, Год журнала: 2024, Номер 16(4), С. 1521 - 1532
Опубликована: Март 22, 2024
Data augmentation is an effective way to overcome the over-fitting problem of deep learning models. However, most existing studies on data work frame-like (e.g., images), and few tackles with event-based data. Event-based are different from data, rendering techniques designed for unsuitable This deals object detection classification, which important self-driving, robot manipulation. Specifically, we introduce EventAugment, a new method augment asynchronous by automatically policies. We first identify 13 types operations augmenting Next, formulate finding optimal policies as hyperparameter optimization problem. To tackle this problem, propose random search-based framework. Finally, evaluate proposed six public datasets including N-Caltech101, N-Cars, ST-MNIST, N-MNIST, DVSGesture DDD17. Experimental results demonstrate that EventAugment exhibits substantial performance improvements both neural network-based spiking models, gains up approximately 4%. Notably, outperform state-of-the-art methods in terms overall performance.
Язык: Английский
Процитировано
3Remote Sensing, Год журнала: 2024, Номер 16(9), С. 1641 - 1641
Опубликована: Май 4, 2024
Object detection in remote sensing plays a crucial role various ground identification tasks. However, due to the limited feature information contained within small targets, which are more susceptible being buried by complex backgrounds, especially extreme environments (e.g., low-light, motion-blur scenes). Meanwhile, event cameras offer unique paradigm with high temporal resolution and wide dynamic range for object detection. These advantages enable without intensity of light, perform better challenging conditions compared traditional cameras. In this work, we introduce Multi-Vision Transformer (MVT), comprises three efficiently designed components: downsampling module, Channel Spatial Attention (CSA) Global (GSA) module. This architecture simultaneously considers short-term long-term dependencies semantic information, resulting improved performance Additionally, propose Cross Deformable (CDA), progressively fuses high-level low-level features instead considering all scales at each layer, thereby reducing computational complexity multi-scale features. Nevertheless, scarcity camera datasets, provide Event Detection (EOD) dataset, is first dataset that includes scenarios specifically introduced using Moreover, conducted experiments on EOD two typical unmanned aerial vehicle datasets (VisDrone2019 UAVDT Dataset). The comprehensive results demonstrate proposed MVT-Net achieves promising competitive performance.
Язык: Английский
Процитировано
3