Articles

Access the latest knowledge in applied science, electrical engineering, computer science and information technology, education, and health.

Filter Icon

Filters article

Years

FAQ Arrow
0
0

Source Title

FAQ Arrow

Authors

FAQ Arrow

29,922 Article Results

An intelligent and explainable IoT-Edge-Cloud architecture for real-time water quality monitoring

10.11591/ijai.v15.i2.pp1109-1120
Sara Bouziane , Badraddine Aghoutane , Aniss Moumen , Anas El Ouali , Ali Essahlaoui , Abdellah El Hmaidi
Continuous and reliable monitoring of water quality is critical for early detection of environmental deterioration, yet conventional monitoring approaches are often slow and lack timely data availability. This study proposes an intelligent and explainable internet of things (IoT)–Edge–Cloud architecture to monitor water quality in real time, using IoT sensing, edge based artificial intelligence (Edge AI), cloud-stream processing, and explainable artificial intelligence (XAI). The system calculates the water quality index (WQI) directly at the edge and predicts its evolution using a stacking ensemble model trained on physicochemical measurements taken from the Moulouya River Basin in Morocco. An explainability module based on Shapley additive explanations (SHAP) values gives a clearer image of the contribution of various parameters to WQI predictions, providing transparency of the features, which builds trust in the model’s output. The proposed architecture was implemented as an end-to-end prototype and validated using a simulation-based experimental that mimicked realistic sensor dynamics and connectivity interruptions. The experimental results show strong predictive performance (R² =0.945), stable system operations, and reliable interpretability highlighting the potential of the proposed approach for scalable, intelligent, and transparent environmental monitoring.
Volume: 15
Issue: 2
Page: 1109-1120
Publish at: 2026-04-01

Adaptive control of ball and beam system using SNA-PID combined with recurrent fuzzy neural network identifier

10.11591/ijai.v15.i2.pp1202-1210
Minh-Thanh Le , Chi-Ngon Nguyen
The ball and beam system is a nonlinear and inherently unstable single input, multiple-output (SIMO) system, which poses significant challenges for control design. Intelligent control algorithms are often applied to autonomously control complex systems when there are changes in parameters or the control environment. Therefore, in this paper, we research and develop two methods: proportional integral derivative (PID) and single neuron adaptive (SNA)-PID-recurrent fuzzy neural network identifier (RFNNI) to control the ball and beam system. Simulation results on MATLAB/Simulink show that the SNA-PID-RFNNI controller provides a more stable output signal than the traditional PID controller, with minimal overshoot and a settling time of about 15 seconds. Next, we will conduct real-time experiments on the object using the proposed algorithm through the MEGA2560 control board with an ultrasonic positioning mechanism.
Volume: 15
Issue: 2
Page: 1202-1210
Publish at: 2026-04-01

A hybrid model for enhanced aspect-based sentiment analysis using large language models

10.11591/ijai.v15.i2.pp1825-1838
Mohammed Ziaulla , Arun Biradar
Aspect-based sentiment analysis (ABSA) is a crucial task within natural language processing (NLP), enabling fine-grained opinion mining by identifying sentiments associated with specific aspects of a product or service. While transformer-based models like bidirectional encoder representations from transformers (BERT) have improved sentiment classification, they still struggle with limited contextual adaptability, especially in customer reviews containing complex expressions. Most existing approaches rely heavily on benchmark datasets such as semantic evaluation (SemEval) and multi-aspect multi-sentiment (MAMS), which do not fully capture the diversity of real-world review scenarios. Hence, this research addresses these limitations by proposing a novel hybrid model, called as hybrid-BERT (H-BERT), that integrates span-aware BERT (SpanBERT) with bidirectional long short-term memory (BiLSTM), conditional random field (CRF), and large language models (LLMs). The objective is to enhance aspect extraction and sentiment classification performance using both annotated and synthetic data. The methodology includes preprocessing, hybrid model training, and evaluation using the SemEval 2014 dataset. Experimental results show that H-BERT achieved 90.58% accuracy and 90.56% F-score in the laptop domain and 91.21% accuracy with a 92.03% F-score in the restaurant domain. These results outperform existing models, confirming H-BERT’s robustness and effectiveness. In conclusion, H-BERT improves sentiment understanding in customer reviews.
Volume: 15
Issue: 2
Page: 1825-1838
Publish at: 2026-04-01

Fetal organ detection using feature enhancement with attention and residual block

10.11591/ijai.v15.i2.pp1593-1604
Nuswil Bernolian , Siti Nurmaini , Ade Iriani Sapitri , Annisa Darmawahyuni , Muhammad Naufal Rachmatullah , Bambang Tutuko , Firdaus Firdaus
The rapid advancements in fetal ultrasonography have significantly enhanced prenatal diagnosis in recent years. Deep learning (DL) architectures have further streamlined the process of organ detection, improved diagnostic accuracy, and reduced observer dependency. This study proposes a computer-aided DL approach for fetal organ segmentation using the you only look once (YOLO) algorithm, a state-of-the-art method for object detection and image segmentation. This study identified and classified 15 fetal organs, including the umbilical vein, stomach, abdomen, brain (trans-cerebellum, trans-thalamic, and trans-ventricular regions), femur, head, thorax (chest cavity), heart (circumference, left atrium, left ventricle, right atrium, right ventricle), and aorta. We compared the performance of YOLOv7, YOLOv8, YOLOv9, and YOLOv11 architectures. The results showed that YOLOv9 outperformed YOLOv7, YOLOv8, and YOLOv11 achieving mAP50 and mAP95 scores of 91.90% and 94.50%, respectively. This performance surpasses previous studies that focused on classifying only a limited number of fetal organs.
Volume: 15
Issue: 2
Page: 1593-1604
Publish at: 2026-04-01

Automated bacteria and fungi classification using convolutional neural network on embedded system

10.11591/ijai.v15.i2.pp1132-1142
Tarik Bouganssa , Maryem Ait Moulay , Samar Aarabi , Abedelali Lasfar , Abdelatif EL Afia
In this study, we created and applied novel concepts for hardware-based image identification and categorization. For artificial intelligence (AI) and image recognition applications, this includes putting algorithms for recognizing colors, textures, and shapes into practice. Our contribution uses an embedded device with a camera and a microcomputer (Raspberry-Pi4 type) to replace the optical assessment of Petri dishes. Our object recognition system processes images efficiently by using a state-of-the-art kernel function and a new neighborhood architecture. Using the well-known convolutional neural network (CNN) architecture, YOLOv8, as a pre-trained model, we evaluated the proposed CNN-based method for object recognition in a number of demanding scenarios. Several Petri plates, uncontrolled settings, and different backgrounds and illumination were used to evaluate the technology. Our dynamic mode integrates a CNN network with an attention mask to highlight the traits of bacteria and fungi, ensuring robust recognition. We implemented our algorithm on a Raspberry Pi 400, connected to a CMOS 3.0 camera sensor and a human-machine interface (HMI) for instant display of results.
Volume: 15
Issue: 2
Page: 1132-1142
Publish at: 2026-04-01

RBC_Frame_Net: a hybrid deep learning framework for detection of red blood cells in malaria diagnostic smear

10.11591/ijai.v15.i2.pp1486-1496
Muhammad Shameem P. , Mathiarasi Balakrishnan
Malaria continues to pose a major global health threat, especially in areas where timely and accurate diagnosis is essential for effective treatment. Conventional diagnostic techniques, such as manually examining Giemsa stained blood smears, are often time-intensive, laborious, and susceptible to human error. To overcome these challenges, this study presents red blood cell frame network (RBC_Frame_Net), a novel deep-learning framework that combines convolutional neural networks (CNNs) with transformer based architectures, augmented by attention mechanisms, for the automated identification of RBCs in malaria smear images. The framework leverages the convolutional block attention modules (CBAM)-UNet model for segmentation, enhancing both spatial and channel features through CBAM and integrates the detection transformer (DETR) to accurately detect and classify RBCs within the diagnostic images. The model achieved outstanding performance with a segmentation intersection over union (IoU) of 0.97, a Dice coefficient of 0.98, and near-perfect detection results (precision: 0.999, recall: 0.998, and mean average precision (mAP): 0.995). When compared to leading models such as YOLOv8, faster region-based convolutional neural network (Faster R-CNN), and EfficientDet-D3, and RBC_Frame_Net demonstrated superior accuracy and robustness. The inclusion of attention mechanisms and a hybrid architecture enhance its adaptability, making it well-suited for deployment in real-world, resource limited environments and positioning it as a valuable asset in automated malaria diagnostics.
Volume: 15
Issue: 2
Page: 1486-1496
Publish at: 2026-04-01

The effects of data imbalance on fraud detection model accuracy

10.11591/ijai.v15.i2.pp1402-1408
Rusma Anieza Ruslan , Nureize Arbaiy , Pei-Chun Lin
Machine learning (ML) model performance is often assessed by accuracy, but the quality and balance of data also play crucial roles. Imbalanced datasets, where the minority class has fewer samples than the majority class, can lead to biased predictions favoring the majority class. This study addresses the issue of class imbalance through resampling techniques, including random undersampling (RUS) and random oversampling (ROS), specifically applied to a fraud detection dataset. We classify the resampled datasets using random forest (RF) and gradient boosting (GB) models. Our findings indicate that the RF model, when combined with ROS, achieves an accuracy of 97.4%, surpassing the 96.1% accuracy of the GB model with RUS. This approach demonstrates the importance of addressing class imbalance to improve prediction accuracy in ML.
Volume: 15
Issue: 2
Page: 1402-1408
Publish at: 2026-04-01

Improvised mask faster recurrent convolutional neural network for breast cancer classification using histopathology images

10.11591/ijai.v15.i2.pp1999-2008
Pattan M. D. Ali Khan , Xavier Arputha Rathina
Despite the prevalence of this disease, the existing method for obtaining an exact breast cancer diagnosis would need a lot of time and labor. It needs a qualified pathologist to manually process and review histopathological images to distinguish the characteristics that characterize different cancer severity levels. Building a model for automatically detecting, segmenting, and classifying breast lesions using histopathological images seems to be the goal of this work. Various deep learning methods have been used in computational pathology for the diagnosis of cancer. Improved faster recurrent convolutional neural network (IMFRCNN) is a supervised learning system with proposed for recognizing small items like mitotic and non mitotic nuclei. To protect small items from vanishing in the deep layers, this system uses expanded layers in the spine. To close image and the things gap size includes, this approach uses expanded layers. The region proposal network has been created for precise tiny object identification. Researchers examined time for training and testing time for various techniques for identifying objects. The total accuracy of benign/malignant categorization in proposed system reaches 96.5%. The proposed technique offers a thorough and non-invasive method for identifying and categorizes an area of abnormal breast tissue.
Volume: 15
Issue: 2
Page: 1999-2008
Publish at: 2026-04-01

Energy-efficient and secure WSN clustering for IoT using particle swarm optimization and advanced encryption standard

10.11591/ijai.v15.i2.pp1275-1285
S. Swapna Kumar , Kalli Satyanarayan Reddy
Wireless sensor networks (WSNs) are made up of distributed sensor nodes that work together under energy and communication constraints. They support diverse internet of things (IoT) applications such as smart agriculture and environmental monitoring. This paper proposes a technique to optimize the WSN framework for secure and energy-efficient data transmission. To improve cluster formation and network energy consumption, the suggested model combines k-means clustering with particle swarm optimization (PSO). Inter-cluster data is encrypted by the cluster head (CH) using the advanced encryption standard (AES)-128. To protect data and save energy, the low-energy adaptive clustering hierarchy (LEACH) protocol uses a number of techniques. Energy efficiency, model accuracy, likelihood of privacy breaches, and network longevity are examples of performance metrics. The system is tested by Python simulations on the Intel Berkeley Research Lab (IBRL) real-world dataset, which includes 54 sensor nodes measuring temperature and humidity. The results demonstrate significant energy savings and a model accuracy of 96.50%, thereby reducing privacy breaches and extending network lifetime. The framework offers scalability, effective privacy monitoring, and adaptability to changing topologies.
Volume: 15
Issue: 2
Page: 1275-1285
Publish at: 2026-04-01

The role of prompt engineering in enhancing LLMs: a systematic review of applications and ethical implications

10.11591/ijai.v15.i2.pp1071-1086
Izzul Fatawi , Muhammad Roil Bilad , Muhammad Asy'ari
Large language models (LLMs) have transformed natural language processing (NLP), demonstrating exceptional proficiency in tasks such as text generation, translation, and summarization. However, LLMs are prone to generating biased, inaccurate, or contextually irrelevant outputs, posing significant risks in high-stakes domains such as healthcare, legal reasoning, and engineering. This paper systematically investigates the role of prompt engineering as a solution to these challenges. By strategically designing inputs, prompt engineering enhances LLM performance, yielding more accurate, contextually relevant, and ethically aligned outputs. Advanced techniques, including chain-of-thought (CoT) prompting and retrieval augmented generation (RAG), are examined for their ability to improve reasoning capabilities, reduce errors, and mitigate bias. CoT prompting facilitates structured, stepwise reasoning, while RAG incorporates real-time data, ensuring output accuracy in rapidly evolving fields. In addition, we present a novel comparative perspective on these techniques, highlighting their distinct strengths and limitations across specialized applications such as healthcare diagnostics and scientific data extraction. The findings demonstrate that sophisticated prompt engineering significantly elevates the reliability and precision of LLM outputs, while addressing critical ethical concerns such as data privacy, bias, and hallucination. These insights underscore the necessity of advanced prompt design in optimizing LLMs for high-impact applications, ensuring both performance and ethical integrity.
Volume: 15
Issue: 2
Page: 1071-1086
Publish at: 2026-04-01

Drone-assisted deep learning weed detection for sustainable agriculture and environmental resilience

10.11591/ijai.v15.i2.pp1428-1440
Agustan Latif , Handaru Jati , Herman Dwi Surjono , Mani Yusuf
Effective weed detection plays a crucial role in sustainable agriculture, boosting crop productivity and supporting environmental conservation. This study compares three deep learning models—YOLOv5, YOLO-NAS, and mask region-based convolutional neural network (Mask R-CNN)-against traditional methods in terms of accuracy, processing speed, and adaptability in tropical agricultural conditions, with Merauke, Indonesia, as the case study. The results show that YOLO-NAS delivers the highest accuracy at 96% with a processing time of 25 ms per image, making it suitable for high precision applications. YOLOv5 balances strong accuracy (94%) with faster processing at 12 ms per image, establishing it as the most effective for real time scenarios. Mask R-CNN also achieves 94% accuracy and provides advanced segmentation capabilities, but its slower processing speed of 31 ms limits large-scale implementation. Traditional methods perform poorly in comparison, with only 85% accuracy and processing time above 50 ms per image. These findings highlight the transformative potential of artificial intelligence (AI)-based weed detection for precision agriculture, particularly in tropical regions like Merauke. Adoption of models such as YOLOv5 reduces manual labor dependence while advancing efficient, eco-friendly weed management. Future research should expand datasets and explore newer models like YOLOv8, YOLO-NAS, vision transformers (ViTs), and hybrid approaches.
Volume: 15
Issue: 2
Page: 1428-1440
Publish at: 2026-04-01

Enhanced VGG-19 model for rice plant disease detection and classification

10.11591/ijai.v15.i2.pp1691-1700
Aye Thida Win , Khin Mar Soe , Myint Myint Lwin
Rice is the main staple food and rice farming plays a crucial role in the agriculture sector of Myanmar. It is also an essential pillar in generating foreign income. However, rice diseases seriously reduced the rice production and quality. Early detection of rice diseases is one of the effective ways to reduce the disease spreading and increase yields. Most Myanmar farmers detect rice diseases based on visual judgment and their experience, which leads to delay in taking efficient action. To overcome this challenge, we intend to propose an enhanced rice plant disease classification model that contributes as artificial intelligence (AI) in Myanmar agriculture sector. The proposed model enhances original visual geometry group 19 (VGG-19) by integrating the algorithms: mixture of Gaussians 2 (MOG2), GrabCut, and relevance estimation with linear feature (RELIEF) for classification. It was trained on 6,326 rice plant images of Kaggle and Eastern Shan State and validated using 5-fold nested cross-validation. The training and testing of proposed model are followed as 80:20. The proposed model experimental result is (98.3%) and lowest standard deviation (0.004) across seven classes than the original VGG-19, MobileNet, Efficient Net, and RestNet50 respectively. Future work will expand dataset diversity, enhance early-stage disease prediction, and support mobile diagnostics for real-world agricultural application.
Volume: 15
Issue: 2
Page: 1691-1700
Publish at: 2026-04-01

Unified voting-based ensemble learning for rice leaf disease detection using improved pretrained models

10.11591/ijai.v15.i2.pp1646-1663
Govindarajan Subburaman , Mary Vennila Selvadurai
As a staple food for a large portion of the global population, rice is particularly susceptible to leaf diseases that adversely affect its yield and overall quality. This study utilizes four pretrained convolutional neural network (CNN) models to construct a unified voting-based ensemble approach for rice leaf disease classification. The models include VGG16, DenseNet121, InceptionV3, and Xception. The dataset used in this study was collected from Kaggle and further enriched with images obtained from Google sources. It comprises a total of 4,000 images categorized into six classes: bacterial leaf blight, brown spot, leaf blast, leaf scald, narrow brown spot, and healthy leaves. It was split into training (327 images/class), validation (140 images/class), and testing (200 images/class). Images were normalized to [0,1] and augmented through rotation, flipping, shifting, shear, zoom, brightness, and channel adjustments to improve generalization. Individually, the fine-tuned models achieved accuracies of 91.3% (VGG16), 95.6% (DenseNet121), 92.1% (InceptionV3), and 89.8% (Xception). The ensemble leveraged majority voting (93.6%), weighted voting (96.5%), and soft voting (97%), yielding an absolute gain of 1.4% over the best individual model and 4.8% over the average of all models. To our knowledge, this is the first ensemble combining these four architectures with unified voting for identifying diseases in rice leaves, delivering a scalable and computationally efficient solution suitable in advance diagnosis and timely execution in agricultural settings with limited resources.
Volume: 15
Issue: 2
Page: 1646-1663
Publish at: 2026-04-01

2D-CNN-GACL-ECGNet graph attention: a robust framework for electrocardiogram-based stress detection

10.11591/ijai.v15.i2.pp1529-1538
P. Kavitha , L. Shakkeera
Early detection of cardiovascular diseases (CVDs) via electrocardiogram (ECG) classification during physiological stress is critical and remains challenging due to stress-induced morphological variability, noise from ambulatory settings, and inter-class ambiguities. Existing models, such as 1D signal-based models with convolutional neural networks (CNNs) and graph convolutional networks (GCNs), struggle to adapt to dynamic stress conditions and generate interpretable insights. In response, we propose 2D CNN and graph attention network (GAT) for optimizer. The model 2D-CNN and GACL-ECG-Net, an innovative framework integrating GATs with adaptive contrastive learning (ACL) and morpho-temporal graph construction. Key innovations include 2D-CNN denoising, 2D transformation, dynamic morpho-temporal graphs modeling ECG beats as nodes with hybrid edges (70% morphological similarity, 30% temporal proximity), and stress-adaptive contrastive loss with learnable margins on stress-conditioned labels, reducing class ambiguity by 18%. Multi-head attention mechanisms provide interpretable heatmaps aligned with cardiologist annotations (κ =0.82) and are evaluated using Massachusetts Institute of Technology-Beth Israel Hospital (MIT-BIH) arrhythmia database, wearable stress and affect detection (WESAD) dataset for emotional stress, and stress at work, knowledge work (SWELL-KW) dataset for cognitive stress. 2D-CNN-GACL-ECG-Net achieves state-of-the-art performance with 98.7% F1-score (MIT-BIH), 94.2% (WESAD), and 92.8% (SWELL-KW), outperforming CNN-bidirectional long short-term memory (BiLSTM) and GCN baselines by 95%. The framework is computationally efficient and clinically validated for wearable health monitoring.
Volume: 15
Issue: 2
Page: 1529-1538
Publish at: 2026-04-01

Time encoded signal processing and recognition with vector quantization: applied to Arabic numerals

10.12928/telkomnika.v24i2.27443
Abdelmajid; Sidi Mohamed Ben Abdellah University Lamkadam , Mohammed; Sidi Mohamed Ben Abdellah University Karim
This article presents our contribution to speaker recognition using Arabic numerals. This recognition is based on hybridization between the time encoded signal processing and recognition (TESPAR) technique and vector quantization (VQ), in order to consolidate the classification step thanks to this combination. To set up an effective and efficient recognition system, we used a corpus recorded under ideal conditions, minimizing the differences between the reference corpus and the test corpus. We also applied the linear discriminant analysis (LDA) technique in order to discriminate the acoustic vectors and minimize the representative space. This hybridization indicated a quantifiable increase in the speaker recognition rate with the ten Arabic numerals (0–9).
Volume: 24
Issue: 2
Page: 481-489
Publish at: 2026-04-01
Show 16 of 1995

Discover Our Library

Embark on a journey through our expansive collection of articles and let curiosity lead your path to innovation.

Explore Now
Library 3D Ilustration