A Comparative Analysis of Compression and Transfer Learning Techniques in DeepFake Detection Models DOI Creative Commons

A. D. Karathanasis,

John Violos, Ioannis Kompatsiaris

и другие.

Mathematics, Год журнала: 2025, Номер 13(5), С. 887 - 887

Опубликована: Март 6, 2025

DeepFake detection models play a crucial role in ambient intelligence and smart environments, where systems rely on authentic information for accurate decisions. These integrating interconnected IoT devices AI-driven systems, face significant threats from DeepFakes, potentially leading to compromised trust, erroneous decisions, security breaches. To mitigate these risks, neural-network-based have been developed. However, their substantial computational requirements long training times hinder deployment resource-constrained edge devices. This paper investigates compression transfer learning techniques reduce the demands of deploying models, while preserving performance. Pruning, knowledge distillation, quantization, adapter modules are explored enable efficient real-time detection. An evaluation was conducted four benchmark datasets: “SynthBuster”, “140k Real Fake Faces”, “DeepFake Images”, “ForenSynths”. It compared compressed with uncompressed baselines using widely recognized metrics such as accuracy, precision, recall, F1-score, model size, time. The results showed that at 10% original size retained only 56% baseline but fine-tuning similar scenarios increased this nearly 98%. In some cases, accuracy even surpassed original’s performance by up 12%. findings highlight feasibility computing scenarios.

Язык: Английский

Lightweight Deep Learning for Resource-Constrained Environments: A Survey DOI Open Access
Hou-I Liu, Marco Antonio Gutiérrez Galindo, Hongxia Xie

и другие.

ACM Computing Surveys, Год журнала: 2024, Номер 56(10), С. 1 - 42

Опубликована: Май 11, 2024

Over the past decade, dominance of deep learning has prevailed across various domains artificial intelligence, including natural language processing, computer vision, and biomedical signal processing. While there have been remarkable improvements in model accuracy, deploying these models on lightweight devices, such as mobile phones microcontrollers, is constrained by limited resources. In this survey, we provide comprehensive design guidance tailored for detailing meticulous models, compression methods, hardware acceleration strategies. The principal goal work to explore methods concepts getting around constraints without compromising model’s accuracy. Additionally, two notable paths future: deployment techniques TinyML Large Language Models. Although undoubtedly potential, they also present significant challenges, encouraging research into unexplored areas.

Язык: Английский

Процитировано

26

Neuromorphic intermediate representation: A unified instruction set for interoperable brain-inspired computing DOI Creative Commons
Jens Egholm Pedersen, Steven Abreu, Matthias Jobst

и другие.

Nature Communications, Год журнала: 2024, Номер 15(1)

Опубликована: Сен. 16, 2024

Spiking neural networks and neuromorphic hardware platforms that simulate neuronal dynamics are getting wide attention being applied to many relevant problems using Machine Learning. Despite a well-established mathematical foundation for dynamics, there exists numerous software solutions stacks whose variability makes it difficult reproduce findings. Here, we establish common reference frame computations in digital systems, titled Neuromorphic Intermediate Representation (NIR). NIR defines set of computational composable model primitives as hybrid systems combining continuous-time discrete events. By abstracting away assumptions around discretization constraints, faithfully captures the model, while bridging differences between evaluated implementation underlying formalism. supports an unprecedented number which demonstrate by reproducing three spiking network models different complexity across 7 simulators 4 platforms. decouples development software, enabling interoperability improving accessibility multiple technologies. We believe is key next step brain-inspired hardware-software co-evolution, research towards energy efficient principles nervous systems. available at neuroir.org.

Язык: Английский

Процитировано

14

CBin-NN: An Inference Engine for Binarized Neural Networks DOI Open Access
Fouad Sakr, Riccardo Berta, Joseph Doyle

и другие.

Electronics, Год журнала: 2024, Номер 13(9), С. 1624 - 1624

Опубликована: Апрель 24, 2024

Binarization is an extreme quantization technique that attracting research in the Internet of Things (IoT) field, as it radically reduces memory footprint deep neural networks without a correspondingly significant accuracy drop. To support effective deployment Binarized Neural Networks (BNNs), we propose CBin-NN, library layer operators allows building simple yet flexible convolutional (CNNs) with binary weights and activations. CBin-NN platform-independent thus portable to virtually any software-programmable device. Experimental analysis on CIFAR-10 dataset shows our library, compared set state-of-the-art inference engines, speeds up by 3.6 times required store model activations 7.5 28 times, respectively, at cost slightly lower (2.5%). An ablation study stresses importance Quantized Input Kernel Convolution improve reduce latency slight increase size.

Язык: Английский

Процитировано

13

A Comprehensive Survey of Deep Learning Approaches in Image Processing DOI Creative Commons
Μαρία Τρίγκα, Ηλίας Δρίτσας

Sensors, Год журнала: 2025, Номер 25(2), С. 531 - 531

Опубликована: Янв. 17, 2025

The integration of deep learning (DL) into image processing has driven transformative advancements, enabling capabilities far beyond the reach traditional methodologies. This survey offers an in-depth exploration DL approaches that have redefined processing, tracing their evolution from early innovations to latest state-of-the-art developments. It also analyzes progression architectural designs and paradigms significantly enhanced ability process interpret complex visual data. Key such as techniques improving model efficiency, generalization, robustness, are examined, showcasing DL's address increasingly sophisticated image-processing tasks across diverse domains. Metrics used for rigorous evaluation discussed, underscoring importance performance assessment in varied application contexts. impact is highlighted through its tackle challenges generate actionable insights. Finally, this identifies potential future directions, including emerging technologies like quantum computing neuromorphic architectures efficiency federated privacy-preserving training. Additionally, it highlights combining with edge explainable artificial intelligence (AI) scalability interpretability challenges. These advancements positioned further extend applications DL, driving innovation processing.

Язык: Английский

Процитировано

1

Hybrid Solution Through Systematic Electrical Impedance Tomography Data Reduction and CNN Compression for Efficient Hand Gesture Recognition on Resource-Constrained IoT Devices DOI Creative Commons

Salwa Sahnoun,

Mahdi Mnif,

Bilel Ghoul

и другие.

Future Internet, Год журнала: 2025, Номер 17(2), С. 89 - 89

Опубликована: Фев. 14, 2025

The rapid advancement of edge computing and Tiny Machine Learning (TinyML) has created new opportunities for deploying intelligence in resource-constrained environments. With the growing demand intelligent Internet Things (IoT) devices that can efficiently process complex data real-time, there is an urgent need innovative optimisation techniques overcome limitations IoT enable accurate efficient computations. This study investigates a novel approach to optimising Convolutional Neural Network (CNN) models Hand Gesture Recognition (HGR) based on Electrical Impedance Tomography (EIT), which requires signal processing, energy efficiency, real-time by simultaneously reducing input complexity using advanced model compression techniques. By systematically halving 1D CNN from 40 20 Boundary Voltages (BVs) applying method, we achieved remarkable size reductions 91.75% 97.49% BVs EIT inputs, respectively. Additionally, Floating-Point operations (FLOPs) are significantly reduced, more than 99% both cases. These have been with minimal loss accuracy, maintaining performance 97.22% 94.44% most significant result compressed model. In fact, at only 8.73 kB our demonstrates potential design strategies creating ultra-lightweight, high-performance CNN-based solutions near-full capabilities specifically case HGR inputs.

Язык: Английский

Процитировано

1

Advanced Deep Learning Models for 6G: Overview, Opportunities and Challenges DOI Creative Commons
Licheng Jiao, Y Shao, Long Sun

и другие.

IEEE Access, Год журнала: 2024, Номер 12, С. 133245 - 133314

Опубликована: Янв. 1, 2024

The advent of the sixth generation mobile communications (6G) ushers in an era heightened demand for advanced network intelligence to tackle challenges expanding landscape and increasing service demands. Deep Learning (DL), as a crucial technique instilling into 6G, has demonstrated powerful promising development. This paper provides comprehensive overview pivotal role DL exploring myriad opportunities that arise. Firstly, we present detailed vision emphasizing areas such adaptive resource allocation, intelligent management, robust signal processing, ubiquitous edge intelligence, endogenous security. Secondly, this reviews how models leverage their unique learning capabilities solve complex demands 6G. discussed include Convolutional Neural Networks (CNN), Generative Adversarial (GAN), Graph (GNN), Reinforcement (DRL), Transformer, Federated (FL), Meta Learning. Additionally, examine specific each model faces within 6G context. Moreover, delve rapidly evolving field Artificial Intelligence Generated Content (AIGC), examining its development impact framework. Finally, culminates discussion ten critical open problems integrating with setting stage future research field.

Язык: Английский

Процитировано

7

StripeRust-Pocket: A Mobile-Based Deep Learning Application for Efficient Disease Severity Assessment of Wheat Stripe Rust DOI Creative Commons
Weizhen Liu, Yuxi Chen,

Zhaoxin Lu

и другие.

Plant Phenomics, Год журнала: 2024, Номер 6

Опубликована: Янв. 1, 2024

Wheat stripe rust poses a marked threat to global wheat production. Accurate and effective disease severity assessments are crucial for resistance breeding timely management of field diseases. In this study, we propose practical solution using mobile-based deep learning model-assisted labeling. StripeRust-Pocket, user-friendly mobile application developed based on models, accurately quantifies in leaf images, even under complex backgrounds. Additionally, StripeRust-Pocket facilitates image acquisition, result storage, organization, sharing. The underlying model employed by called StripeRustNet, is balanced lightweight 2-stage model. first stage utilizes MobileNetV2-DeepLabV3+ segmentation, followed ResNet50-DeepLabV3+ the second lesion segmentation. Disease estimated calculating ratio pixel area area. StripeRustNet achieves 98.65% mean intersection over union (MIoU) segmentation 86.08% MIoU Validation an additional 100 images demonstrated correlation 0.964 with 3 expert visual scores. To address challenges manual labeling, introduce labeling pipeline that combines correction, spatial complementarity. We apply our self-collected dataset, reducing annotation time from 20 min per image. Our method provides efficient assessments, empowering breeders pathologists implement management. It also demonstrates how “last mile” challenge applying computer vision technology plant phenomics.

Язык: Английский

Процитировано

5

Efficiently Gluing Pre-trained Language and Vision Models for Image Captioning DOI Open Access
Peipei Song, Yuanen Zhou, Xun Yang

и другие.

ACM Transactions on Intelligent Systems and Technology, Год журнала: 2024, Номер 15(6), С. 1 - 16

Опубликована: Июль 29, 2024

Vision-and-language pre-training models have achieved impressive performance for image captioning. But most of them are trained with millions paired image-text data and require huge memory computing overhead. To alleviate this, we try to stand on the shoulders large-scale pre-trained language (PLM) vision (PVM) efficiently connect There two major challenges: one is that modalities different semantic granularity (e.g., a noun may cover many pixels), other gap still exists between models. this end, design lightweight efficient connector glue PVM PLM, which holds criterion selection-then-transformation . Specifically, in selection phase, treat each as set patches instead pixels. We select salient cluster into visual regions align text. Then, effectively reduce gap, propose map selected text space through spatial channel transformations. With training captioning datasets, learns bridge via backpropagation, preparing PLM generate descriptions. Experimental results MSCOCO Flickr30k datasets demonstrate our method yields comparable existing works. By solely small connector, achieve CIDEr 132.2% Karpathy test split. Moreover, findings reveal fine-tuning can further enhance potential, resulting score 140.6%. Code available at https://github.com/YuanEZhou/PrefixCap

Язык: Английский

Процитировано

5

EMG-YOLO: an efficient fire detection model for embedded devices DOI
Linsong Xiao, Wenzao Li, Xiaoqiang Zhang

и другие.

Digital Signal Processing, Год журнала: 2024, Номер 156, С. 104824 - 104824

Опубликована: Окт. 18, 2024

Язык: Английский

Процитировано

4

Automatic channel pruning by neural network based on improved poplar optimisation algorithm DOI
Yijie Hu, Debao Chen, Feng Zou

и другие.

Knowledge-Based Systems, Год журнала: 2025, Номер 310, С. 113002 - 113002

Опубликована: Янв. 22, 2025

Язык: Английский

Процитировано

0