(Support Vector Machines) model and the
Perceptron Multilayer (MLP) model follow
with the same performance.
5 Conclusion
In this paper, we proposed machine learning
models for the detection of fraud in customs
declarations in Senegal. These models were
built using data from customs declarations and
using the following supervised learning
methods:: Multilayer Perceptron, Support
Vector Machines, Random Forest, and Extreme
Gradient Boosting. The model obtained with
Random Forest was found to perform the best
according to the performance measures we
used, namely precision, recall, F1-score, and
accuracy. Then follow, in order, the model
obtained with Extreme Gradient Boosting and
the models obtained with Multilayer Perceptron
and Support Vector Machines.
In perspective, it would be interesting to
combine these models to form an ensemble
model that would be very efficient in fraud
detection.
These models could be integrated into the
Senegalese Customs' fraud risk management
system to improve the efficiency of controls,
and facilitate the work of customs officers.
References:
[1] T. Mitchell, “Machine learning”, McGraw
Hill, 1997.
[2] D. A. N. Seck and F. B. R. Diakité,
"Supervised Machine Learning Models for the
Prediction of Renal Failure in Senegal," 2023
International Conference on Control,
Artificial Intelligence, Robotics &
Optimization (ICCAIRO), Crete, Greece,
2023, pp. 94-98, DOI:
10.1109/ICCAIRO58903.2023.00022.
[3] N. Paranoan, S. Y. Sabandar, A. Paranoan, E.
Pali, I. Pasulu, "The Effect of Prevention
Measures, Fraud Detection, and Investigative
Audits on Efforts to Minimize Fraud in The
Financial Statements of Companies, Makassar
City Indonesia," WSEAS Transactions on
Information Science and Applications, vol. 19,
pp. 54-62, 2022,
https://doi.org/10.37394/23209.2022.19.6.
[4] P. J. Werbos, “Beyond Regression: New Tools
for Prediction and Analysis in the Behavioral
Sciences”, Doctoral Dissertation, Harvard
University, Cambridge, 1974.
[5] D. E. Rumelhart, G. E. Hinton, R. J. Williams
(1986) , “Learning representations by back-
propagating errors”, Nature, Vol 323, 533-
536.
[6] B. E. Boser, I. M. Guyon, and V. N. Vapnik,
“A training algorithm for optimal margin
classifiers”, In Proceedings of the fifth annual
workshop on Computational learning theory,
pages 144–152, 1992.
[7] C. Cortes and V. Vapnik, “Support-vector
networks”, Machine learning, 20(3):273–297,
1995.
[8] N. Cristianini and J. Shawe-Taylor, “An
introduction to support vector machines and
other kernel-based learning methods”,
Cambridge University Press, 2000, DOI:
10.1017/CBO9780511801389.
[9] L. Breiman, “Random forests”, Machine
learning, 45(1):5–32, 2001.
[10] L. Breiman, “Bagging predictors”, Machine
Learning 24(2), 123-140, 1996.
[11] J. Quinlan, “C4.5: Programs for Machine
Learning”, Morgan Kaufman, San Mateo,
California, 1993.
[12] L. Breiman, J. Friedman, R. Olshen, C. Stone,
“Classification and Regression Trees”,
Wadsworth, Belmont, California, 1984.
[13] T. Chen and C. Guestrin, “Xgboost: A
scalable tree boosting system”, In
Proceedings of the 22nd Acm sigkdd
international conference on knowledge
discovery and data mining, pages 785–794,
2016.
[14] R. Schapire, “The strength of weak
learnability”, Machine Learning, 5(2):197–
227, 1990.
[15] F. Pedregosa, G. Varoquaux, A. Gramfort, V.
Michel, B. Thirion, O. Grisel, M. Blondel, P.
Prettenhofer, R. Weiss, V. Dubourg, J.
Vanderplas, A. Passos, D. Cournapeau.,
“Scikit-learn: Machine Learning in Python”,
JMLR 12, pp. 2825-2830, 2011.
[16] T. Chen, C. Guestrin, “XGBoost: A Scalable
Tree Boosting System”, In: Proceedings of
the 22nd ACM SIGKDD International
Conference on Knowledge Discovery and
Data Mining, New York, NY, USA: ACM;
2016, p. 785–94, (KDD '16), Available from:
http://doi.acm.org/10.1145/2939672.2939785.
WSEAS TRANSACTIONS on INFORMATION SCIENCE and APPLICATIONS
DOI: 10.37394/23209.2024.21.20
Djamal Abdoul Nasser Seck