Articles

Access the latest knowledge in applied science, electrical engineering, computer science and information technology, education, and health.

Filter Icon

Filters article

Years

FAQ Arrow
0
0

Source Title

FAQ Arrow

Authors

FAQ Arrow

29,393 Article Results

The bootstrap procedure for selecting the number of principal components in PCA

10.11591/ijict.v14i3.pp1136-1145
Borislava Toleva
The initial step in determining the number of principal components for both classification and regression involves evaluating how much each component contributes to the total variance in the data. Based on this analysis, a subset of components that explains the highest percentage of variance is typically selected. However, multiple valid combinations may exist, and the final choice is often made manually by the researcher. This study introduces a novel yet straightforward algorithm for the automatic selection of the number of principal components. By integrating ANOVA and bootstrapping with principal component analysis (PCA), the proposed method enables automatic component selection in classification tasks. The algorithm is evaluated using three publicly available datasets and applied with both decision tree and support vector machine (SVM) classifiers. Results indicate that this automated procedure not only eliminates researcher bias in selecting components but also improves classification accuracy. Unlike traditional methods, it selects a single optimal combination of principal components without manual intervention, offering a new and efficient approach to PCAbased model development.
Volume: 14
Issue: 3
Page: 1136-1145
Publish at: 2025-12-01

An artificial intelligent system for cotton leaf disease detection

10.11591/ijict.v14i3.pp950-959
Priyanka Nilesh Jadhav , Pragati Prashant Patil , Nitesh Sureja , Nandini Chaudhari , Heli Sureja
This study aims to develop a deep learning-based system for the detection and classification of diseases in cotton leaves, with the goal of aiding in early diagnosis and disease management, thereby enhancing agricultural productivity in India. The study utilizes a dataset of cotton leaf images, classified into four categories: Fusarium wilt, Curl virus, Bacterial blight, and Healthy leaves. The dataset is used to train and evaluate various CNN models such as basic CNN, VGG19, Xception, InceptionV3, and ResNet50. These models were evaluated on their accuracy in identifying the presence of diseases and classifying cotton leaf images into the respective categories. The models were trained using standard deep learning frameworks and optimized for high performance. The results indicated that ResNet50 achieved the highest accuracy of 100%, followed by InceptionV3 with 98.75%, and VGG19 and Xception both with 97.50%. The basic CNN model showed an accuracy of 96.25%. These models demonstrated strong potential for accurate multi-class classification of cotton leaf diseases. This study emphasizes the potential of deep learning in agricultural diagnostics. Future research can focus on improving model robustness, incorporating larger datasets, and deploying the system for real-time field use to assist farmers in disease management and improving cotton production.
Volume: 14
Issue: 3
Page: 950-959
Publish at: 2025-12-01

Optimizing brain tumor MRI classification using advanced preprocessing techniques and ensemble learning methods

10.11591/ijai.v14.i6.pp5106-5119
Akim Manaor Hara Pardede , Ahmad Zamsuri , Indi Nuroini , Putrama Alkhairi
Brain tumor classification is a critical task in medical imaging that directly impacts the accuracy of diagnosis and treatment planning. However, the complexity and variability of magnetic resonance imaging (MRI) images pose significant challenges, often resulting in reduced model reliability and generalization. This study addresses these limitations by proposing a novel ResNet+Bagging model, leveraging the strengths of residual networks and ensemble learning to enhance classification performance. Using publicly available brain tumor MRI datasets, including images labeled as benign, malignant, and normal, the study employs advanced preprocessing techniques such as normalization, data augmentation, and noise reduction to ensure high-quality inputs. The proposed model demonstrated significant improvements, achieving the highest testing accuracy of 72%, outperforming other tested models such as LeNet, standard ResNet, GoogleNet, and VGGNet. Precision (0.6010), recall (0.6000), and F1-score (0.5990) metrics further highlight its superior balance in detecting positive and negative classes. The novelty of this research lies in the application of Bagging to ResNet, which effectively mitigates overfitting and enhances predictive stability in complex medical datasets. These findings underscore the proposed model's potential as a robust solution for brain tumor classification, contributing to more accurate and reliable diagnostics.
Volume: 14
Issue: 6
Page: 5106-5119
Publish at: 2025-12-01

A web-based learning platform to assess student performance using online session activity engagement

10.11591/ijai.v14.i6.pp5240-5250
Shashirekha Hanumanthappa , Chetana Prakash
Predicting students' performance and engagement is crucial for academic eLearning partners in colleges and universities as well as students themselves considering post-COVID-19 pandemic and university grant commission (UGC) dual degree regulation era. An educational system's data on students’ engagement in taking courses that are a significant component of an institution of higher learning with a cogent vertical syllabus can be used to make predictions. By examining how closely a student's course-taking actions correspond with the requirements of the syllabus, one can utilize the student's conduct in the classroom and online eLearning web tool as a predictor of future achievement. This paper presents a study that uses an eLearning web-based dataset to predict students' success throughout a series of online interactive sessions. The dataset records how students engage with each other during online lab work, including how many keystrokes they make, how long they spend on each task, and how well they perform on exams overall. The current methods lacks accuracy to assess student performance and engagement with high precision. In addressing this paper introduces novel multi-label ensemble learning (MLEL) using XGBoost (XGB) and K-fold cross validation. Experiment outcome shows the proposed (MLEL-XGB) achieves much improved outcome than other existing models.
Volume: 14
Issue: 6
Page: 5240-5250
Publish at: 2025-12-01

Chatbot for virtual medical assistance

10.11591/ijict.v14i3.pp914-922
Aravalli Sainath Chaithanya , Sampangi Lahari Vishista , Adepu MadhuSri
A healthy population is vital for societal prosperity and happiness. Amidst busy lifestyles and the challenges posed by the COVID-19 pandemic, individuals often neglect their health needs. To address this, we introduce a novel approach utilizing a chatbot for virtual medical assistance. Tailored for individuals confined indoors or hesitant to visit hospitals for minor ailments, our chatbot offers personalized medical support by diagnosing ailments based on user-reported symptoms and engaging in interactive conversations. Leveraging a robust dataset containing 132 symptoms, 41 diseases, and corresponding medications, our chatbot employs a systematic approach for symptom refinement, enhancing diagnostic precision. Upon identifying a disease, the chatbot promptly suggests basic medications tailored to the specific ailment. Furthermore, our system integrates user demographics to evaluate medication history and current state, allowing for personalized medication recommendations based on individual needs. Through extensive testing and validation, we demonstrate the effectiveness of our chatbot in accurately predicting ailments and providing timely treatment advice. Our study introduces a novel paradigm for medicine recommendation and disease prediction, with the potential to enhance healthcare accessibility and effectiveness.
Volume: 14
Issue: 3
Page: 914-922
Publish at: 2025-12-01

Soil moisture prototype soil moisture sensor YL-69 for Gaharu (Aquilaria malaccensis) tree planting media

10.11591/ijict.v14i3.pp1163-1171
Rikie Kartadie , Muhammad Agung Nugroho , Adiyuda Prayitna , Adi Kusjani , Ardeana Galih Mardika
Soil moisture, defined as the amount of water present in the spaces between soil particles, plays a critical role in plant growth. Excessive soil moisture can lead to issues such as root rot, deviating from the ideal conditions required for root absorption. To address this, we developed a prototype tool using the YL-69 soil moisture sensor to monitor and control the soil moisture levels in Agarwood/Gaharu tree planting media. The prototype was designed to activate a water pump when soil moisture exceeded 80%, ensuring optimal humidity for plant growth. Once the moisture level dropped below 80%, the pump was deactivated to prevent overwatering. The YL-69 sensor demonstrated an accuracy of 88.76% under controlled conditions. This study highlights the potential of using low-cost sensors for automated soil moisture management in small-scale Gaharu cultivation.
Volume: 14
Issue: 3
Page: 1163-1171
Publish at: 2025-12-01

Prediction of flood-affected areas based on geographic information system data using machine learning

10.11591/ijai.v14.i6.pp4675-4683
Amrul Faruq , Lailis Syafaah , Muhammad Irfan , Shahrum Shah Abdullah , Shamsul Faisal Mohd Hussein , Fitri Yakub
Flood disasters have become more frequent and severe due to climate variability, posing significant threats to human lives, agriculture, and infrastructure. Effective disaster management and mitigation require accurate identification of flood-prone areas. This study develops an intelligent flood prediction system by integrating machine learning algorithms with geographic information systems (GIS) data to enhance flood risk assessment. The proposed system utilizes two machine learning models, including random forest (RF) and support vector machine (SVM), to predict flood-susceptible areas. The models are trained on historical flood data and GIS-derived features, including elevation, slope, topographic wetness index (TWI), aspect, and curvature. The dataset undergoes preprocessing, including normalization and feature selection, before being divided into training, validation, and test sets. The models are then trained and evaluated based on their predictive performance. Evaluation metrics, particularly the area under the curve (AUC), demonstrate that RF outperforms SVM in predicting flood-prone areas. RF achieves an accuracy of 82%, while SVM records a lower accuracy of 68%. The superior performance of RF is attributed to its ability to handle complex, nonlinear relationships in flood prediction. These results highlight the effectiveness of machine learning algorithms in flood susceptibility modeling and support the integration of data-driven techniques into flood and disaster risk reduction management strategies.
Volume: 14
Issue: 6
Page: 4675-4683
Publish at: 2025-12-01

Impact of smoothing techniques for text classification: implementation in hidden Markov model

10.11591/ijai.v14.i6.pp5183-5192
Norsyela Muhammad Noor Mathivanan , Roziah Mohd Janor , Shukor Abd Razak , Nor Azura Md. Ghani
A hidden Markov model (HMM) is widely used for sequence modeling in various text classification tasks. This study investigates the impact of different smoothing techniques, such as Laplace, absolute discounting, and Gibbs sampling on HMM performance across three distinct domains: e-commerce products, spam filtering, and occupational data mining. Through the comparative analysis, Laplace smoothing consistently outperforms other techniques in handling zero-probability issues, demonstrating superior performance in the e-commerce and SMS spam datasets. The HMM without any smoothing technique achieved the best results for job title classification. This divergence underscores the dataset-specific nature of smoothing requirements, where the simplicity of parameter estimation proves effective in contexts characterized by a limited and repetitive vocabulary. Hence, the findings suggest that tailored smoothing strategies are crucial for optimizing HMM performance in different textual analysis applications.
Volume: 14
Issue: 6
Page: 5183-5192
Publish at: 2025-12-01

A bibliometric analysis of feature selection techniques: trends, innovations, and future directions

10.11591/ijai.v14.i6.pp4403-4414
Oumaima Semmar , Wissal El Habti , Donalson Wilson , Abdellah Azmani
Feature selection techniques have become increasingly important in addressing the challenges of high dimensionality in machine learning and other artificial intelligence domains. In this study, we present a comprehensive bibliometric analysis of research on feature selection techniques over the past decade, focusing on mapping the intellectual structure, identifying emerging trends, and highlighting productive collaborations in the field. Using merged data from Scopus and Web of Science databases, we collected and analyzed 2,079 relevant documents published between 2014 and 2024, applying citation analysis, co-authorship networks, and keyword co-occurrence mapping. Our findings reveal that feature selection methodologies, including supervised, unsupervised, and hybrid approaches across filter, wrapper, and embedded techniques, have been widely applied across various domains. The authors who have most contributed to the development of these methods are primarily affiliated with institutions in China, India, and the USA. The insights provided by this analysis offer researchers and practitioners a valuable foundation for guiding future research directions in feature selection.
Volume: 14
Issue: 6
Page: 4403-4414
Publish at: 2025-12-01

Scaling of Facebook architecture and technology stack with heavy workload: past, present and future

10.11591/ijict.v14i3.pp772-782
Tole Sutikno , Laksana Talenta Ahmad
Leading social media Facebook has improved its architecture to meet user needs. Facebook has improved its systems to handle millions of users with heavy workloads and large datasets using innovative architectural solutions and adaptive strategies. The study examines Facebook’s architectural and technological advances in heavy workload and big data. To understand how Facebook scaled with a growing user base and data volume, history and system architecture will be examined. It will also examine how cloud storage and high-performance computing optimize resource utilization and maintain performance during peak user activity. Facebook is managing big data and heavy workloads with new technologies like the hybrid communication model that uses PULL and PUSH strategies for real-time messaging. Facebook switched from HBase to MyRocks for message storage to improve performance as data grew. Architectural scaling and technology stack research must prioritize data storage innovations and optimized communication protocols to handle heavy workloads and big data. The messenger Sync protocol reduces network congestion and improves synchronous communication, reducing resource consumption and maintaining performance under high load. High-performance computing (HPC) and cloud storage should be studied together to support complex compute workflows. This convergence may improve large-scale application infrastructures and encourage interdisciplinary collaboration for scalable and resilient systems.
Volume: 14
Issue: 3
Page: 772-782
Publish at: 2025-12-01

Catalysing precision in bone x-ray analysis for image detection and classification: the triple context attention model advancement

10.11591/ijai.v14.i6.pp4957-4970
Tabassum N. Sultana , Nagaratna P. Hegde , Asma Parveen
Accurate detection and classification of fractures in bone x-ray images are crucial for effective medical diagnosis and treatment. In this study, we propose the triple context attention model (TCAN) as a novel approach to address the challenges in this domain. TCAN offers several key contributions that significantly enhance the accuracy and efficiency of bone x-ray image recognition and classification. Firstly, TCAN introduces the coordination attention mechanism, which considers both horizontal and vertical positional data during the recognition process. Secondly, TCAN mitigates the common issue of mislabelling fractures in bone x-ray images, particularly in the you only look once (YOLO) model, due to the absence of positional data during training. Thirdly, TCAN efficiently enhances positional data by focusing on weights, and increasing feature dimension while maintaining a manageable model size. This allows for effective utilization of positional data without computational overhead. Lastly, TCAN combines the visual attention network (VAN) with its capabilities, resulting in a comprehensive system that can handle diverse image dimensions and accurately classify various types of fractures across different body regions. Overall, TCAN presents a promising advancement in medical image analysis, improving fracture detection accuracy and classification efficiency in bone x-ray images, thus aiding in more effective clinical decision-making.
Volume: 14
Issue: 6
Page: 4957-4970
Publish at: 2025-12-01

Hybrid N-gram-based framework for payload distributed denial of service detection and classification

10.11591/ijai.v14.i6.pp4763-4774
Andi Maslan , Cik Feresa Mohd Foozy , Kamaruddin Malik Bin Mohamad , Abdul Hamid , Dedy Fitriawan , Joni Hasugian
There are three primary approaches to DDoS detection: anomaly-based, pattern-based, and heuristic-based. The heuristic-based method integrates both anomaly- and pattern-based techniques. However, existing DDoS detection systems face challenges in performing HTTP payload-level analysis, mainly due to high false positive rates and insufficient granularity in current datasets. To address this, the study introduces a novel heuristic approach based on a hybrid N-Gram model. This hybrid combines two components: CSDPayload+N-Gram and CSPayload+N-Gram. CSDPayload represents the gap (measured via Chi-Square Distance) between a given payload and normal traffic payloads, while CSPayload reflects the similarity (measured via Cosine Similarity) between them. These metrics form a new feature set evaluated using three datasets: CIC2019, MIB2016, and H2N-Payload. The methodology begins with packet extraction and conversion of TCP/IP traffic—specifically HTTP traffic—into hexadecimal payloads. N-Gram analysis (from 1-Gram to 6-Gram) is then applied to these payloads. For each N-Gram, frequency counts are computed, followed by calculations of Chi-Square Distance (CSD), Cosine Similarity (CS), and Pearson’s Chi-Square test to classify payloads as either benign or malicious. Subsequently, feature selection is performed using weight correlation, and the resulting features are fed into three machine learning classifiers: Support Vector Machine (SVM), K-Nearest Neighbors (KNN), and Neural Network. Experimental results demonstrate high detection accuracy, particularly in the 4-Gram feature category: Neural Network achieves 99.65%, KNN 95.14%, and SVM 99.73% accuracy on average.
Volume: 14
Issue: 6
Page: 4763-4774
Publish at: 2025-12-01

Comparative evaluation of machine learning models for intrusion detection in WSNs using the IDSAI dataset

10.11591/ijai.v14.i6.pp4913-4922
Mansour Lmkaiti , Houda Moudni , Hicham Mouncif
This paper provides comparative assessment of three lightweight machine learning (ML) models (logistic regression (LR), random forest (RF), and gradient boosting (GB)), which are employed to detect intrusions in wireless sensor networks (WSNs) using the IDSAI dataset. The goal is to determine the most effective and deployable classifier within the constraints of WSN resources. In order to prevent data leakage and report accuracy, precision, recall, F1-score, and receiver operating characteristic-area under the curve (ROC-AUC) with mean±SD, we implement stratified 5-fold cross validation with in fold preprocessing. The results indicate that RF provides the most optimal generalization and overall performance (accuracy 0.9994 ± 0.0001, precision 0.9995±0.0001, recall 0.9994±0.0001, F1-score 0.9994±0.0001, ROC–AUC 0.9998 ± 0.0000). RF is closely followed by GB (accuracy 0.9990±0.0001, precision 0.9995±0.0001, recall 0.9985±0.0001, F1-score 0.9990 ± 0.0001, ROC-AUC ≈ 1.0000). LR demonstrates limitations in linearly overlapping classes, as evidenced by its high precision but reduced recall (accuracy 0.9167±0.0010, precision 0.9829±0.0002, recall 0.8481±0.0018, F1-score 0.9105 ± 0.0011, ROC–AUC 0.9707 ± 0.0001). In order to evaluate deployability, we characterize the inference throughput on a modest PC: LR ∼ 6.5 × 105 samples/s, GB ∼ 2.2 × 105 samples/s, and RF ∼ 1.3 × 105 samples/s, indicating a tiered intrusion detection system (IDS) (LR at sensors, RF at cluster-heads, and GB at the gateway). We also address the potential dangers of overfitting that may arise from the cleanliness of the dataset and provide a roadmap for future validation on a more diverse set of traffic. The research establishes a baseline for lightweight IDS in actual WSNs that is deployable and reproducible.
Volume: 14
Issue: 6
Page: 4913-4922
Publish at: 2025-12-01

Adaptive intelligent PSO-Based MPPT technique for PV systems under dynamic irradiance and partial shading conditions

10.11591/ijpeds.v16.i4.pp2841-2859
Muhammad Gul E. Islam , Mohammad Faridun Naim Tajuddin , Azralmukmin Azmi , Rini Nur Hasanah , Shahrin Md. Ayob , Tole Sutikno
This research introduces an adaptive improved particle swarm optimization (AIPSO) approach for maximum power point tracking (MPPT) approach designed to enhance energy harvesting from photovoltaic (PV) systems under dynamic irradiance conditions. The proposed AIPSO algorithm addresses the challenges associated with traditional MPPT methods, particularly in scenarios characterized by fluctuating solar irradiance, such as step changes and partial shading. By incorporating a robust reinitialization strategy along with updated velocity and position equations, the algorithm demonstrates superior performance in terms of convergence accuracy, tracking speed, and tracking efficiency. This modification enables the algorithm to effectively escape local maxima and explore a wider search space, leading to improved convergence and optimal power point tracking. Furthermore, the adaptive nature of the PSO enhances the algorithm’s ability to respond to real-time changes in environmental conditions, making it particularly suitable for large- scale PV systems subjected to varying atmospheric factors. Here, “adaptive” denotes coefficient scheduling (C3) and a re-initialization trigger that responds to irradiance regime changes; “intelligent” denotes robust regime shift detection and safe duty ratio clamping. Across uniform, step change, and partial shading conditions, the proposed AIPSO achieves fast reconvergence and high tracking efficiency with negligible steady state oscillations, as summarized in the results. Building on this contribution, future research will focus on evaluating its scalability across different PV architectures and large-scale grid integration with real hardware setup.
Volume: 16
Issue: 4
Page: 2841-2859
Publish at: 2025-12-01

Optimizing sparse ternary compression with thresholds for communication-efficient federated learning

10.11591/ijai.v14.i6.pp4902-4912
Nithyanianjan Murthy Chittaiah , Manjula Sunkadakatte Haladappa
Federated learning (FL) enables decentralized model training while preserving client data privacy, yet suffers from significant communication overhead due to frequent parameter exchanges. This study investigates how varying sparse ternary compression (STC) thresholds impact communication efficiency and model accuracy across the CIFAR-10 and MedMNIST datasets. Experiments tested thresholds ranging from 1.0 to 1.9 and batch sizes of 10, 15, and 20. Results demonstrated that selecting thresholds between 1.2 and 1.5 reduced total communication costs by approximately 10–15%, while maintaining acceptable accuracy levels. These findings suggest that careful threshold tuning can achieve substantial communication savings with minimal compromise in model performance, offering practical guidance for improving the efficiency and scalability of FL systems.
Volume: 14
Issue: 6
Page: 4902-4912
Publish at: 2025-12-01
Show 25 of 1960

Discover Our Library

Embark on a journey through our expansive collection of articles and let curiosity lead your path to innovation.

Explore Now
Library 3D Ilustration