Articles

Access the latest knowledge in applied science, electrical engineering, computer science and information technology, education, and health.

Filter Icon

Filters article

Years

FAQ Arrow
0
0

Source Title

FAQ Arrow

Authors

FAQ Arrow

28,428 Article Results

Multilevel and multisource data fusion approach for network intrusion detection system using machine learning techniques

10.11591/ijece.v15i4.pp3938-3948
Harshitha Somashekar , Pramod Halebidu Basavaraju
To enhance the performance of network intrusion detection systems (NIDS), this paper proposes a novel multilevel and multisource data fusion approach, applied to NSL-KDD and UNSW-NB15 datasets. The proposed approach includes three various levels of operations, which are feature level fusion, dimensionality reduction, and prediction level fusion. In the first stage features of NSL-KDD and UNSW-NB15 both datasets are fused by applying the inner join joint operation by selecting common features like protocol, service and label. Once the data sets are fused in the first level, linear discriminant analysis is applied for 12 feature columns which is reduced to a single feature column leading to dimensionality reduction at the second level. Finally, in the third level, the prediction level fusion technique is applied to two neural network models, where one neural network model has a single input node, two hidden nodes, and two output nodes, and another model having a single input node, three hidden nodes, and two output nodes. The outputs obtained from these two models are then fused using a prediction fusion technique. The proposed approach achieves a classification accuracy of 97.5%.
Volume: 15
Issue: 4
Page: 3938-3948
Publish at: 2025-08-01

Adaptive multi-radio quality of service model using neural network approach for robust wireless sensor network transmission in multipath fading environment

10.11591/ijece.v15i4.pp3795-3802
Galang Persada Nurani Hakim , Dian Widi Astuti , Ahmad Firdausi , Huda A. Majid
Wireless sensor network loss in wireless data transmission is one of the problems that needs attention. Interference, fading, congestion, and delay are some factors that cause loss in wireless data transmission. This paper used an adaptive multi-radio model to enhance the wireless data transmission to be more robust to disturbance in a multipath fading environment. A neural network approach was used to generate the adaptive model. If we use 433 MHz as our carrier frequency with 250 kHz bandwidth and 12 spreading factors, we can get signal noise ratio (SNR) for 20 meters at about -9.8 dB. Thus, we can use the adaptive model to enhance the WSN wireless data transmission's SNR to 9 dB, automatically changing the radio configuration to 797.1 MHz frequency, with 378.1 bandwidth and 7.111 for spreading factor. Based on the result, the wireless data transmission link has been successfully enhanced using the proposed adaptive model for wireless sensor networks (WSN) in a multipath fading environment.
Volume: 15
Issue: 4
Page: 3795-3802
Publish at: 2025-08-01

Blockchain and internet of things synergy: transforming smart grids for the future

10.11591/ijece.v15i4.pp4239-4248
Mouad Bensalah , Abdellatif Hair , Reda Rabie
Conventional smart grid systems face challenges in security, transparency, and efficiency. This study addresses these limitations by integrating blockchain and internet of things (IoT) technologies, presenting proof-of-concept implemented on an Orange Pi 4 single-board computer. The realized prototype demonstrated secure and transparent energy transaction management with consistent throughput between 7.45 and 7.81 transactions per second, and efficient resource utilization across varying transaction volumes. However, scalability challenges, including a linear increase in processing time with larger block sizes, emphasize the need for optimized consensus mechanisms. The findings underscore the feasibility of blockchain-based smart grids in resource-constrained settings, paving the way for advancements in peer-to-peer energy trading, decentralized energy storage, and integration with artificial intelligence for dynamic energy optimization. This work contributes to developing secure, efficient, and sustainable energy systems.
Volume: 15
Issue: 4
Page: 4239-4248
Publish at: 2025-08-01

A comparative study of deep learning-based network intrusion detection system with explainable artificial intelligence

10.11591/ijece.v15i4.pp4109-4119
Tan Juan Kai , Lee-Yeng Ong , Meng-Chew Leow
In the rapidly evolving landscape of cybersecurity, robust network intrusion detection systems (NIDS) are crucial to countering increasingly sophisticated cyber threats, including zero-day attacks. Deep learning approaches in NIDS offer promising improvements in intrusion detection rates and reduction of false positives. However, the inherent opacity of deep learning models presents significant challenges, hindering the understanding and trust in their decision-making processes. This study explores the efficacy of explainable artificial intelligence (XAI) techniques, specifically Shapley additive explanations (SHAP) and local interpretable model-agnostic explanations (LIME), in enhancing the transparency and trustworthiness of NIDS systems. With the implementation of TabNet architecture on the AWID3 dataset, it is able to achieve a remarkable accuracy of 99.99%. Despite this high performance, concerns regarding the interpretability of the TabNet model's decisions persist. By employing SHAP and LIME, this study aims to elucidate the intricacies of model interpretability, focusing on both global and local aspects of the TabNet model's decision-making processes. Ultimately, this study underscores the pivotal role of XAI in improving understanding and fostering trust in deep learning -based NIDS systems. The robustness of the model is also being tested by adding the signal-to-noise ratio (SNR) to the datasets.
Volume: 15
Issue: 4
Page: 4109-4119
Publish at: 2025-08-01

Breast cancer identification using a hybrid machine learning system

10.11591/ijece.v15i4.pp3928-3937
Toni Arifin , Ignatius Wiseto Prasetyo Agung , Erfian Junianto , Dari Dianata Agustin , Ilham Rachmat Wibowo , Rizal Rachman
Breast cancer remains one of the most prevalent malignancies among women and is frequently diagnosed at an advanced stage. Early detection is critical to improving patient prognosis and survival rates. Messenger ribonucleic acid (mRNA) gene expression data, which captures the molecular alterations in cancer cells, offers a promising avenue for enhancing diagnostic accuracy. The objective of this study is to develop a machine learning-based model for breast cancer detection using mRNA gene expression profiles. To achieve this, we implemented a hybrid machine learning system (HMLS) that integrates classification algorithms with feature selection and extraction techniques. This approach enables the effective handling of heterogeneous and high-dimensional genomic data, such as mRNA expression datasets, while simultaneously reducing dimensionality without sacrificing critical information. The classification algorithms applied in this study include support vector machine (SVM), random forest (RF), naïve Bayes (NB), k-nearest neighbors (KNN), extra trees classifier (ETC), and logistic regression (LR). Feature selection was conducted using analysis of variance (ANOVA), mutual information (MI), ETC, LR, whereas principal component analysis (PCA) was employed for feature extraction. The performance of the proposed model was evaluated using standard metrics, including recall, F1-score, and accuracy. Experimental results demonstrate that the combination of the SVM classifier with MI feature selection outperformed other configurations and conventional machine learning approaches, achieving a classification accuracy of 99.4%.
Volume: 15
Issue: 4
Page: 3928-3937
Publish at: 2025-08-01

Renewable energy impact integration in Moroccan grid-load flow analysis

10.11591/ijece.v15i4.pp3632-3648
Safaa Essaid , Loubna Lazrak , Mouhsine Ghazaoui
This paper analyzes the behavior of a Moroccan electric transportation system in the presence of an integration of renewable energy sources, which represents a significant challenge due to their intermittent nature. The aim is to evaluate the performance of the transportation system in various situations and possible configurations. The current study enables the calculation of power flow in the network using the Newton-Raphson method under the MATLAB/Simulink software. To achieve this, a series of power flow simulations were conducted on a 5-bus Moroccan electrical network, examining four distinct scenarios. In addition, this article offers an evaluation of the power flow performance of the same electric transportation system with varying percentages of renewable energy penetration. In order to provide a complete critical analysis, many simulations were conducted to obtain the voltage and active power profile generated at different bus locations, as well as an evaluation of the losses in the studied network.
Volume: 15
Issue: 4
Page: 3632-3648
Publish at: 2025-08-01

Enhancing voltage stability of transmission network using proportional integral controlled high voltage direct current system

10.11591/ijece.v15i4.pp3593-3602
Chibuike Peter Ohanu , Uche C. Ogbuefi , Emenike Ejiogu , Tole Sutikno
The contingencies experienced in transmission power networks often lead to unstable voltage profiles, challenging grid reliability and stability. This research aim is to enhance voltage stability using a proportional-integral (PI) controlled high voltage direct current (HVDC) system on a real life 330 kV network. The Newton-Raphson (NR) method is used for power flow analysis of the test network, and stability analysis identified Makurdi bus as the candidate bus for improvement due to its low eigenvalue and damping ratio. Application of a balanced three-phase fault at this bus resulted in a minimum voltage of 0.70 per unit (p.u.), falling outside the statutory voltage limit requirements of 0.95 to 1.05 p.u. The PI-based HVDC system was then applied along the Makurdi to Jos transmission line, which has a low loading capacity. The application of this model optimized the system response to disturbances, significantly improve voltage stability and raised the minimum voltage profile on the network to 0.80 p.u. This demonstrates 10% voltage profile improvement from the base case and reaffirms the effectiveness of the PI-based HVDC system in enhancing voltage stability during major disturbances. This research highlights the potential of integrating control systems into power networks to improve voltage stability and ensure reliable operation, even during large disturbances.
Volume: 15
Issue: 4
Page: 3593-3602
Publish at: 2025-08-01

Ensemble of convolutional neural network and DeepResNet for multimodal biometric authentication system

10.11591/ijece.v15i4.pp4279-4295
Ashwini Kailas , Madhusudan Girimallaih , Mallegowda Madigahalli , Vasantha Kumara Mahadevachar , Pranothi Kadirehally Somashekarappa
Multimodal biometrics technology has garnered attention recently for its ability to address inherent limitations found in single biometric modalities and to enhance overall recognition rates. A typical biometric recognition system comprises sensing, feature extraction, and matching modules. The system’s robustness heavily relies on its capability to effectively extract pertinent information from individual biometric traits. This study introduces a novel feature extraction technique tailored for a multimodal biometric system utilizing electrocardiogram (ECG) and iris traits. The ECG helps to incorporate the liveliness related information and Iris helps to produce the unique pattern for each individual. Therefore, this work presents a multimodal authentication system where data pre-processing is performed on image and ECG data where noise removal and quality enhancement tasks are performed. Later, feature extraction is carried out for ECG signals by estimating the Heart rate variability feature analysis in time and frequency domain. Finally, the ensemble of convolution neural network (CNN) and DeepResNet models are used to perform the classification. The overall accuracy is reported as 0.8900, 0.8400, 0.7900, 0.8932, 0.87, and 0.97 by using convolutional neural network-long short-term memory (CNN-LSTM), support vector machine (SVM), random forest (RF), CNN, decision tree (DT), and proposed MBANet approach respectively.
Volume: 15
Issue: 4
Page: 4279-4295
Publish at: 2025-08-01

Real-time machine learning-based posture correction for enhanced exercise performance

10.11591/ijece.v15i4.pp3843-3850
Anish Khadtare , Vasistha Ved , Himanshu Kotak , Akhil Jain , Pinki Vishwakarma
Poor posture and associated physical health problems have grown more common as technology use increases, especially during workout sessions. Maintaining proper posture is essential to increasing the efficacy of your workouts and avoiding injuries. The research paper presents the development of a machine-learning model designed to provide real-time posture correction and feedback for exercises such as squats and planks. The model uses MediaPipe for precise real-time posture estimation and OpenCV for analyzing video frames. It detects poor posture and provides users with instant corrective feedback on their posture by examining the angles between important body parts, such as the arms, knees, back, and hips. This innovative method enables a thorough evaluation of form without requiring face-to-face supervision, opening it up to a wider audience. The model is trained on real-world workout datasets of people performing exercises in different positions and postures to ensure that posture detection is reliable under various user circumstances. The system utilizes cutting-edge machine-learning algorithms to demonstrate scalability and adaptability for future training types beyond squats and planks. The main goal is to provide users with a model that increases the efficacy of workouts, lowers the risk of injury, and encourages better exercise habits. The model's emphasis on usability and accessibility makes it potentially a vital tool for anyone looking to enhance their posture and general fitness levels.
Volume: 15
Issue: 4
Page: 3843-3850
Publish at: 2025-08-01

Instance segmentation for PCB defect detection with Detectron2

10.11591/ijece.v15i4.pp4172-4180
Aravalli Sainath Chaithanya , Lavadya Nirmala Devi , Putty Srividya
Printed circuit boards (PCBs) are essential in modern electronics, where even minor defects can lead to failures. Traditional inspection methods struggle with complex PCB designs, necessitating automated deep learning techniques. Object detection models like Faster R-CNN and YOLO rely on bounding boxes for defect localization but face overlap issues, limiting precise defect isolation. This paper presents a segmentation-based PCB defect detection model using Detectron2’s Mask R-CNN. By leveraging instance segmentation, the model enables pixel-level defect localization and classification, addressing challenges such as shape variations, complex structures, and occlusions. Trained on a dataset of 690 COCO-annotated images, the model underwent rigorous experimentation and parameter tuning. Evaluation metrics, including loss functions and mean average precision (mAP), assessed performance. Results showed a steady decline in loss values and high precision for defects like mouse bites and missing holes. However, performance was lower for complex defects like spurs and spurious copper. This study highlights the effectiveness of instance segmentation in PCB defect detection, contributing to improved quality control and manufacturing automation.
Volume: 15
Issue: 4
Page: 4172-4180
Publish at: 2025-08-01

Enhancing multi-class text classification in biomedical literature by integrating sequential and contextual learning with BERT and LSTM

10.11591/ijece.v15i4.pp4202-4212
Oussama Ndama , Ismail Bensassi , Safae Ndama , El Mokhtar En-Naimi
Classification of sentences in biomedical abstracts into predefined categories is essential for enhancing readability and facilitating information retrieval in scientific literature. We propose a novel hybrid model that integrates bidirectional encoder representations from transformers (BERT) for contextual learning, long short-term memory (LSTM) for sequential processing, and sentence order information to classify sentences from biomedical abstracts. Utilizing the PubMed 200k randomized controlled trial (RCT) dataset, our model achieved an overall accuracy of 88.42%, demonstrating strong performance in identifying methods and results sections while maintaining balanced precision, recall, and F1-scores across all categories. This hybrid approach effectively captures both contextual and sequential patterns of biomedical text, offering a robust solution for improving the segmentation of scientific abstracts. The model's design promotes stability and generalization, making it an effective tool for automatic text classification and information retrieval in biomedical research. These results underscore the model's efficacy in handling overlapping categories and its significant contribution to advancing biomedical text analysis.
Volume: 15
Issue: 4
Page: 4202-4212
Publish at: 2025-08-01

Multi-layer convolutional autoencoder for recognizing three-dimensional patterns in attention deficit hyperactivity disorder using resting-state functional magnetic resonance imaging

10.11591/ijece.v15i4.pp3965-3976
Zarina Begum , Kareemulla Shaik
Attention deficit hyperactivity disorder (ADHD) is a neurological disorder that develops over time and is typified by impulsivity, hyperactivity, and attention deficiency. There have been noticeable changes in the patterns of brain activity in recent studies using functional magnetic resonance imaging (fMRI). Particularly in the prefrontal cortex. Machine learning algorithms show promise in distinguishing ADHD subtypes based on these neurobiological signatures. However, the inherent heterogeneity of ADHD complicates consistent classification, while small sample sizes limit the generalizability of findings. Additionally, methodological variability across studies contributes to inconsistent results, and the opaque nature of machine learning models hinders the understanding of underlying mechanisms. We suggest a novel deep learning architecture to overcome these issues by combining spatio-temporal feature extraction and classification through a hierarchical residual convolutional noise reduction autoencoder (HRCNRAE) and a 3D convolutional gated memory unit (GMU). This framework effectively reduces spatial dimensions, captures key temporal and spatial features, and utilizes a sigmoid classifier for robust binary classification. Our methodology was rigorously validated on the ADHD-200 dataset across five sites, demonstrating enhancements in diagnostic accuracy ranging from 1.26% to 9.6% compared to existing models. Importantly, this research represents the first application of a 3D Convolutional GMU for diagnosing ADHD with fMRI data. The improvements highlight the efficacy of our architecture in capturing complex spatio-temporal features, paving the way for more accurate and reliable ADHD diagnoses.
Volume: 15
Issue: 4
Page: 3965-3976
Publish at: 2025-08-01

Energy-efficient secure software-defined networking with reinforcement learning and Weierstrass cryptography

10.11591/ijece.v15i4.pp4227-4238
Nagaraju Tumakuru Andanaiah , Malode Vishwanatha Panduranga Rao
In the age of rapidly advancing 5G connectivity, artificial intelligence (AI), and the internet of things (IoT), network data has grown enormously, demanding more efficient and secure management solutions. Traditional networking systems, limited by manual controls and static environments, are unable to fulfill the dynamic demands of modern internet services. This paper proposes an innovative software-defined networking (SDN) framework that utilizes exponential spline regression reinforcement learning (ESR-RL) with genus Weierstrass curve cryptography (GWCC) to boost energy efficiency and data security. The ESR-RL algorithm reliably anticipates network traffic patterns, optimizing path selection to enhance routing efficiency while minimizing consumption of energy. GWCC also enables strong encryption and decryption, considerably increasing data security without impacting system performance. To further improve network reliability, the Skellam distributed Siberian TIGER optimization algorithm (SDSTOA) is used to dynamically acquire features and balance loads, resulting in optimal network performance. Extensive simulations show that the proposed framework performs better than existing models in terms of accuracy, precision, recall, F-measure, sensitivity, and specificity. Improvements in latency, turnaround time, and network throughput demonstrate the framework's success. This scalable and adaptive technology establishes a new standard for SDN systems by providing a safe, energy-efficient, and performance-optimized strategy for future network infrastructures.
Volume: 15
Issue: 4
Page: 4227-4238
Publish at: 2025-08-01

Two-step majority voting of convolutional neural networks for brain tumor classification

10.11591/ijece.v15i4.pp4087-4098
Irwan Budi Santoso , Shoffin Nahwa Utama , Supriyono Supriyono
Brain tumor type classification is essential for determining further examinations. Convolutional neural network (CNN) model with magnetic resonance imaging (MRI) image input can improve brain tumor classification performance. However, due to the highly variable shape, size, and location of brain tumors, increasing the performance of tumor classification requires consideration of the results of several different CNN models. Therefore, we proposed a two-step majority voting (MV) on the results of several CNN models for tumor classification. The CNN models included InceptionV3, Xception, DensNet201, EfficientNetB3, and ResNet50; each was customized at the classification layer. The initial step of the method is transfer-learning for each CNN model. The next step is to carry out two steps of MV, namely MV on the three CNN model classification results at different training epochs and MV on the results of the first step. The performance evaluation of the proposed method used the Nickparvar dataset, which included MRI images of glioma, pituitary, no tumor, and meningioma. The test results showed that the proposed method obtained an accuracy of 99.69% with a precision and sensitivity average of 99.67% and a specificity of 99.90%. With these results, the proposed method is better than several other methods.
Volume: 15
Issue: 4
Page: 4087-4098
Publish at: 2025-08-01

High-speed field-programmable gate array implementation for mmWave orthogonal frequency-division multiplexing transmitters: design and evaluation

10.11591/ijece.v15i4.pp3813-3823
Kidsanapong Puntsri , Bussakorn Bunsri , Puripong Suthisopapan
This paper presents a field-programmable gate array (FPGA)-based implementation of an orthogonal frequency-division multiplexing (OFDM) transmitter signal processing chain optimized for high-speed millimeter wave (mmWave) communication systems. The design prioritizes real-time processing efficiency and flexibility. A high-throughput 2048-point inverse fast Fourier transform (IFFT) module, realized using a Radix-2 algorithm, forms the core of the design, showcasing efficient hardware resource utilization. The implementation further includes cyclic prefix (CP) insertion and configurable support for various quadrature amplitude modulation (QAM) modulation orders and pilot arrangements. The design is implemented in VHSIC Hardware Description Language (VHDL) using Vivado 2020 and evaluated on the Zynq UltraScale+ RFSoC ZCU111 evaluation kit. The processing pipeline employs eight parallel lanes for concurrent data computation. Experimental results demonstrate a mean squared error (MSE) of only 0.00013 between the FPGA-generated waveform and its MATLAB-simulated counterpart. Additionally, post-implementation resource utilization analysis shows efficient usage of FPGA resources. These findings validate the efficacy and real-time capability of the proposed FPGA-based OFDM transmitter leverages parallelism and high-speed architecture to efficiently process massive data streams, making it suitable for a wide range of mmWave OFDM applications. In contrast to recent works that focus on lower-order IFFT modules, this paper employs a high-throughput IFFT computation, showcasing efficient hardware resource utilization for high-speed mmWave applications.
Volume: 15
Issue: 4
Page: 3813-3823
Publish at: 2025-08-01
Show 55 of 1896

Discover Our Library

Embark on a journey through our expansive collection of articles and let curiosity lead your path to innovation.

Explore Now
Library 3D Ilustration