A Quantum Circuit Learning-based Investigation: A Case Study in Iris Benchmark Dataset Binary Classification
(Muhamad Akrom, Wise Herowati, De Rosal Ignatius Moses Setiadi)
DOI : 10.62411/jcta.11779
- Volume: 2,
Issue: 3,
Sitasi : 0 05-Jan-2025
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
This study presents a Quantum Machine Learning (QML) architecture for perfectly classifying the Iris flower dataset. The research addresses improving classification accuracy using quantum models in machine-learning tasks. The objective is to demonstrate the effectiveness of QML approaches, specifically the Variational Quantum Circuit (VQC), Quantum Neural Network (QNN), and Quantum Support Vector Machine (QSVM), in achieving high performance on the Iris dataset. The proposed methods result in perfect classification, with all models attaining accuracy, precision, recall, and an F1-score of 1.00. The main finding is that the QML architecture successfully achieves flawless classification, contributing significantly to the field. These results underscore the potential of QML in solving complex classification problems and highlight its promise for future applications across various domains. The study concludes that QML techniques can offer transformative solutions in machine learning tasks, particularly those leveraging VQC, QNN, and QSVM.
|
0 |
2025 |
Hypertension Detection via Tree-Based Stack Ensemble with SMOTE-Tomek Data Balance and XGBoost Meta-Learner
(Christopher Chukwufunaya Odiakaose, Fidelis Obukohwo Aghware, Margaret Dumebi Okpor, Andrew Okonji Eboka, Amaka Patience Binitie, Arnold Adimabua Ojugo, De Rosal Ignatius Moses Setiadi, Ayei Egu Ibor, Rita Erhovwo Ako, Victor Ochuko Geteloma, Eferhire Valentine Ugbotu, Tabitha Chukwudi Aghaunor)
DOI : 10.62411/faith.3048-3719-43
- Volume: 1,
Issue: 3,
Sitasi : 114 01-Dec-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
High blood pressure (or hypertension) is a causative disorder to a plethora of other ailments – as it succinctly masks other ailments, making them difficult to diagnose and manage with a targeted treatment plan effectively. While some patients living with elevated high blood pressure can effectively manage their condition via adjusted lifestyle and monitoring with follow-up treatments, Others in self-denial leads to unreported instances, mishandled cases, and in now rampant cases – result in death. Even with the usage of machine learning schemes in medicine, two (2) significant issues abound, namely: (a) utilization of dataset in the construction of the model, which often yields non-perfect scores, and (b) the exploration of complex deep learning models have yielded improved accuracy, which often requires large dataset. To curb these issues, our study explores the tree-based stacking ensemble with Decision tree, Adaptive Boosting, and Random Forest (base learners) while we explore the XGBoost as a meta-learner. With the Kaggle dataset as retrieved, our stacking ensemble yields a prediction accuracy of 1.00 and an F1-score of 1.00 that effectively correctly classified all instances of the test dataset.
|
114 |
2024 |
Outlier Detection Using Gaussian Mixture Model Clustering to Optimize XGBoost for Credit Approval Prediction
(De Rosal Ignatius Moses Setiadi, Ahmad Rofiqul Muslikh, Syahroni Wahyu Iriananda, Warto Warto, Jutono Gondohanindijo, Arnold Adimabua Ojugo)
DOI : 10.62411/jcta.11638
- Volume: 2,
Issue: 2,
Sitasi : 0 01-Nov-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
Credit approval prediction is one of the critical challenges in the financial industry, where the accuracy and efficiency of credit decision-making can significantly affect business risk. This study proposes an outlier detection method using the Gaussian Mixture Model (GMM) combined with Extreme Gradient Boosting (XGBoost) to improve prediction accuracy. GMM is used to detect outliers with a probabilistic approach, allowing for finer-grained anomaly identification compared to distance- or density-based methods. Furthermore, the data cleaned through GMM is processed using XGBoost, a decision tree-based boosting algorithm that efficiently handles complex datasets. This study compares the performance of XGBoost with various outlier detection methods, such as LOF, CBLOF, DBSCAN, IF, and K-Means, as well as various other classification algorithms based on machine learning and deep learning. Experimental results show that the combination of GMM and XGBoost provides the best performance with an accuracy of 95.493%, a recall of 91.650%, and an AUC of 95.145%, outperforming other models in the context of credit approval prediction on an imbalanced dataset. The proposed method has been proven to reduce prediction errors and improve the model's reliability in detecting eligible credit applications.
|
0 |
2024 |
Exploring Deep Q-Network for Autonomous Driving Simulation Across Different Driving Modes
(Marcell Adi Setiawan, De Rosal Ignatius Moses Setiadi, Erna Zuni Astuti, T. Sutojo, Noor Ageng Setiyanto)
DOI : 10.62411/faith.3048-3719-31
- Volume: 1,
Issue: 3,
Sitasi : 29 18-Oct-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
The rapid growth in vehicle ownership has led to increased traffic congestion, making the need for autonomous driving solutions more urgent. Autonomous Vehicles (AVs) offer a promising solution to improve road safety and reduce traffic accidents by adapting to various driving conditions without human intervention. This research focuses on implementing Deep Q-Network (DQN) to enhance AV performance in different driving modes: safe, normal, and aggressive. DQN was selected for its ability to handle complex, dynamic environments through experience replay, asynchronous training, and epsilon-greedy exploration. We designed a simulation environment using the Highway-env platform and evaluated the DQN model under varying traffic densities. The performance of the AV was assessed based on two key metrics: success rate and total reward. Our findings show that the DQN model achieved a success rate of 90.75%, 94.625%, and 95.875% in safe, normal, and aggressive modes, respectively. Although the success rate increased with traffic intensity, the total reward remained lower in aggressive driving scenarios, indicating room for optimization in decision-making processes under highly dynamic conditions. This study demonstrates that DQN can adapt effectively to different driving needs, but further optimization is needed to enhance performance in more challenging environments. Future work will focus on improving the DQN algorithm to maximize both success rate and reward in high-traffic scenarios and testing the model in more diverse and complex environments.
|
29 |
2024 |
Exploring Machine Learning and Deep Learning Techniques for Occluded Face Recognition: A Comprehensive Survey and Comparative Analysis
(Keny Muhamada, De Rosal Ignatius Moses Setiadi, Usman Sudibyo, Budi Widjajanto, Arnold Adimabua Ojugo)
DOI : 10.62411/faith.2024-30
- Volume: 1,
Issue: 2,
Sitasi : 54 26-Sep-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
Face recognition occluded by occlusions, such as glasses or shadows, remains a challenge in many security and surveillance applications. This study aims to analyze the performance of various machine learning and deep learning techniques in face recognition scenarios with occlusions. We evaluate KNN (standard and FisherFace), CNN, DenseNet, Inception, and FaceNet methods combined with a pre-trained DeepFace model using three public datasets: YALE, Essex Grimace, and Georgia Tech. The results show that KNN maintains the highest accuracy, reaching 100% on two datasets (Essex Grimace and YALE), even in the presence of occlusions. Meanwhile, CNN shows strong performance, with accuracy remaining 100% on YALE, both with and without occlusions, although its performance drops slightly on Essex Grimace (94% with occlusion). DenseNet and Inception show a more significant drop in accuracy when faced with occlusion, with DenseNet dropping from 81% to 72% on Essex Grimace and Inception dropping from 100% to 92% on the same dataset. FaceNet + DeepFace excels on more large dataset (Georgia Tech) with 98% accuracy, but its performance drops dramatically to 53% and 70% on Essex Grimace and YALE with occlusion. These findings indicate that while deep learning methods show high accuracy under ideal conditions, machine learning methods such as KNN are more flexible and robust to occlusion in face recognition.
|
54 |
2024 |
Phishing Website Detection Using Bidirectional Gated Recurrent Unit Model and Feature Selection
(De Rosal Ignatius Moses Setiadi, Suyud Widiono, Achmad Nuruddin Safriandono, Setyo Budi)
DOI : 10.62411/faith.2024-15
- Volume: 1,
Issue: 2,
Sitasi : 47 12-Jul-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
Phishing attacks continue to be a significant threat to internet users, necessitating the development of advanced detection systems. This study explores the efficacy of a Bidirectional Gated Recurrent Unit (BiGRU) model combined with feature selection techniques for detecting phishing websites. The dataset used for this research is sourced from the UCI Machine Learning Repository, specifically the Phishing Websites dataset. This approach involves cleaning and preprocessing the data, then normalizing features and employing feature selection to identify the most relevant attributes for classification. The BiGRU model, known for its ability to capture temporal dependencies in data, is then applied. To ensure robust evaluation, we utilized cross-validation, dividing the data into five folds. The experimental results are highly promising, demonstrating a Mean Accuracy, Mean Precision, Mean Recall, Mean F1 Score, and Mean AUC of 1.0. These results indicate the model's exceptional performance distinguishing between phishing and legitimate websites. This study highlights the potential of combining BiGRU models with feature selection and cross-validation to create highly accurate phishing detection systems, providing a reliable solution to enhance cybersecurity measures.
|
47 |
2024 |
Effects of Data Resampling on Predicting Customer Churn via a Comparative Tree-based Random Forest and XGBoost
(Rita Erhovwo Ako, Fidelis Obukohwo Aghware, Margaret Dumebi Okpor, Maureen Ifeanyi Akazue, Rume Elizabeth Yoro, Arnold Adimabua Ojugo, De Rosal Ignatius Moses Setiadi, Chris Chukwufunaya Odiakaose, Reuben Akporube Abere, Frances Uche Emordi, Victor Ochuko Geteloma, Patrick Ogholuwarami Ejeh)
DOI : 10.62411/jcta.10562
- Volume: 2,
Issue: 1,
Sitasi : 0 27-Jun-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
Customer attrition has become the focus of many businesses today – since the online market space has continued to proffer customers, various choices and alternatives to goods, services, and products for their monies. Businesses must seek to improve value, meet customers' teething demands/needs, enhance their strategies toward customer retention, and better monetize. The study compares the effects of data resampling schemes on predicting customer churn for both Random Forest (RF) and XGBoost ensembles. Data resampling schemes used include: (a) default mode, (b) random-under-sampling RUS, (c) synthetic minority oversampling technique (SMOTE), and (d) SMOTE-edited nearest neighbor (SMOTEEN). Both tree-based ensembles were constructed and trained to assess how well they performed with the chi-square feature selection mode. The result shows that RF achieved F1 0.9898, Accuracy 0.9973, Precision 0.9457, and Recall 0.9698 for the default, RUS, SMOTE, and SMOTEEN resampling, respectively. Xgboost outperformed Random Forest with F1 0.9945, Accuracy 0.9984, Precision 0.9616, and Recall 0.9890 for the default, RUS, SMOTE, and SMOTEEN, respectively. Studies support that the use of SMOTEEN resampling outperforms other schemes; while, it attributed XGBoost enhanced performance to hyper-parameter tuning of its decision trees. Retention strategies of recency-frequency-monetization were used and have been found to curb churn and improve monetization policies that will place business managers ahead of the curve of churning by customers.
|
0 |
2024 |
Analyzing Preprocessing Impact on Machine Learning Classifiers for Cryotherapy and Immunotherapy Dataset
(De Rosal Ignatius Moses Setiadi, Hussain Md Mehedul Islam, Gustina Alfa Trisnapradika, Wise Herowati)
DOI : 10.62411/faith.2024-2
- Volume: 1,
Issue: 1,
Sitasi : 36 01-Jun-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
In the clinical treatment of skin diseases and cancer, cryotherapy and immunotherapy offer effective and minimally invasive alternatives. However, the complexity of patient response demands more sophisticated analytical strategies for accurate outcome prediction. This research focuses on analyzing the effect of preprocessing in various machine learning models on the prediction performance of cryotherapy and immunotherapy. The preprocessing techniques analyzed are advanced feature engineering and Synthetic Minority Over-sampling Technique (SMOTE) and Tomek links as resampling techniques and their combination. Various classifiers, including support vector machine (SVM), Naive Bayes (NB), Decision Tree (DT), Random Forest (RF), XGBoost, and Bidirectional Gated Recurrent Unit (BiGRU), were tested. The findings of this study show that preprocessing methods can significantly improve model performance, especially in the XGBoost model. Random Forest also gets the same results as XGBoost, but it can also work better without significant preprocessing. The best results were 0.8889, 0.8889, 0.6000, 0.9037, and 0.8790, respectively, for accuracy, recall, specificity, precision, and f1 on the Immunotherapy dataset, while on the Cryotherapy dataset, respectively, they were 0.8889, 0.8889, 0.6000, 0.9037, and 0.8790. This study confirms the potential of customized preprocessing and machine learning models to provide deep insights into treatment dynamics, ultimately improving the quality of diagnosis.
|
36 |
2024 |
Analyzing Quantum Feature Engineering and Balancing Strategies Effect on Liver Disease Classification
(Achmad Nuruddin Safriandono, De Rosal Ignatius Moses Setiadi, Akhmad Dahlan, Farah Zakiyah Rahmanti, Iwan Setiawan Wibisono, Arnold Adimabua Ojugo)
DOI : 10.62411/faith.2024-12
- Volume: 1,
Issue: 1,
Sitasi : 38 01-Jun-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
This research aims to improve the accuracy of liver disease classification using Quantum Feature Engineering (QFE) and the Synthetic Minority Over-sampling Tech-nique and Tomek Links (SMOTE-Tomek) data balancing technique. Four machine learning models were compared in this research, namely eXtreme Gradient Boosting (XGB), Random Forest (RF), Support Vector Machine (SVM), and Logistic Regression (LR) on the Indian Liver Patient Dataset (ILPD) dataset. QFE is applied to capture correlations and complex patterns in the data, while SMOTE-Tomek is used to address data imbalances. The results showed that QFE significantly improved LR performance in terms of recall and specificity up to 99%, which is very important in medical diagnosis. The combination of QFE and SMOTE-Tomek gives the best results for the XGB method with an accuracy of 81%, recall of 90%, and f1-score of 83%. This study concludes that the use of QFE and data balancing techniques can improve liver disease classification performance in general.
|
38 |
2024 |
Integrating SMOTE-Tomek and Fusion Learning with XGBoost Meta-Learner for Robust Diabetes Recognition
(De Rosal Ignatius Moses Setiadi, Kristiawan Nugroho, Ahmad Rofiqul Muslikh, Syahroni Wahyu Iriananda, Arnold Adimabua Ojugo)
DOI : 10.62411/faith.2024-11
- Volume: 1,
Issue: 1,
Sitasi : 40 23-May-2024
| Abstrak
| PDF File
| Resource
| Last.29-Jan-2026
Abstrak:
This research aims to develop a robust diabetes classification method by integrating the Synthetic Minority Over-sampling Technique (SMOTE)-Tomek technique for data balancing and using a machine learning ensemble led by eXtreme Gradient Boosting (XGB) as a meta-learner. We propose an ensemble model that combines deep learning techniques such as Bidirectional Long Short-Term Memory (BiLSTM) and Bidirectional Gated Recurrent Units (BiGRU) with XGB classifier as the base learner. The data used included the Pima Indians Diabetes and Iraqi Society Diabetes datasets, which were processed by missing value handling, duplication, normalization, and the application of SMOTE-Tomek to resolve data imbalances. XGB, as a meta-learner, successfully improves the model's predictive ability by reducing bias and variance, resulting in more accurate and robust classification. The proposed ensemble model achieves perfect accuracy, precision, recall, specificity, and F1 score of 100% on all tested datasets. This method shows that combining ensemble learning techniques with a rigorous preprocessing approach can significantly improve diabetes classification performance.
|
40 |
2024 |