American Accounting Association DOI: 10.2308/ajpt-50009
Financial Statement Fraud Detection: An Analysis of Statistical and Machine Learning Algorithms
Johan Perols
SUMMARY: This study compares the performance of six popular statistical and machine learning models in detecting financial statement fraud under different assumptions of misclassification costs and ratios of fraud firms to nonfraud firms. The results show, somewhat surprisingly, that logistic regression and support vector machines perform well relative to an artificial neural network, bagging, C4.5, and stacking. The results also reveal some diversity in predictors used across the classification algorithms. Out of 42 predictors examined, only six are consistently selected and used by different classification algorithms: auditor turnover, total discretionary accruals, Big 4 auditor, accounts receivable, meeting or beating analyst forecasts, and unexpected employee productivity. These findings extend financial statement fraud research and can be used by practitioners and regulators to improve fraud risk models. Keywords: analytical auditing; financial statement fraud; fraud detection; fraud predictors; classification algorithms. Data Availability: A list of fraud companies used in this study is available from the author upon request. All other data sources are described in the text.
INTRODUCTION
T
he cost of financial statement fraud is estimated at $572 billion1 per year in the U.S. (Association of Certified Fraud Examiners [ACFE] 2008). In addition to direct costs, financial statement fraud negatively affects employees and investors and undermines the
Johan Perols is an Assistant Professor at the University of San Diego.
This study is based on one of my three dissertation papers completed at the University of South Florida. I thank my dissertation co-chairs, Jacqueline Reck and Kaushal Chari, and committee
References: American Institute of Certified Public Accountants (AICPA). 1988. The Auditor’s Responsibility to Detect and Report Errors and Irregularities. Statement on Auditing Standards (SAS) No. 53. New York, NY: AICPA. American Institute of Certified Public Accountants (AICPA). 1997. Consideration of Fraud in a Financial Statement Audit. Statement on Auditing Standards (SAS) No. 82. New York, NY: AICPA. Association of Certified Fraud Examiners (ACFE). 2008. Report to the Nation on Occupational Fraud and Abuse. Austin, TX: ACFE. Bayley, L., and S. Taylor. 2007. Identifying earnings management: A financial statement analysis (red flag) approach. Working paper ABN AMRO and University of New South Wales. Beasley, M. 1996. An empirical analysis of the relation between the board of director composition and financial statement fraud. The Accounting Review 71 (4): 443–465. Bell, T., and J. Carcello. 2000. A decision aid for assessing the likelihood of fraudulent financial reporting. Auditing: A Journal of Practice & Theory 19 (1): 169–184. Beneish, M. 1997. Detecting GAAP violation: Implications for assessing earnings management among firms with extreme financial performance. Journal of Accounting and Public Policy 16: 271–309. Beneish, M. 1999. Incentives and penalties related to earnings overstatements that violate GAAP. The Accounting Review 74 (4): 425–457. Breiman, L. 1996. Bagging predictors. Machine Learning 24 (2): 123–140. Auditing: A Journal of Practice & Theory May 2011 Financial Statement Fraud Detection: An Analysis of Statistical and Machine Learning Algorithms 49 Breiman, L., J. Friedman, R. Olshen, and C. Stone. 1984. Classification and Regression Trees. Boca Raton, FL: Chapman and Hall/CRC Press. Cecchini, M., H. Aytug, G. Koehler, and P. Pathak. 2010. Detecting management fraud in public companies. Management Science 56 (7): 1146–1160. Chan, P. K., W. Fan, A. L. Prodromidis, and S. J. Stolfo. 1999. Distributed data mining in credit card fraud detection. IEEE Intelligent Systems and Their Applications 14 (6): 67–74. Chawla, N. V. 2005. Data mining for imbalanced datasets: An overview. In The Data Mining and Knowledge Discovery Handbook, edited by Maimon, O., and L. Rokach, 853–867. Secaucus, NJ: Springer-Verlag New York, Inc. Chawla, N. V., K. W. Bowyer, L. O. Hall, and W. P. Kegelmeyer. 2002. SMOTE: Synthetic minority oversampling technique. Journal of Artificial Intelligence Research 16: 321–357. Chen, C., and J. Sennetti. 2005. Fraudulent financial reporting characteristics of the computer industry under a strategic-systems lens. Journal of Forensic Accounting 6 (1): 23–54. Dechow, P., R. Sloan, and A. Sweeney. 1996. Causes and consequences of earnings manipulations: An analysis of firms subject to enforcement actions by the SEC. Contemporary Accounting Research 13 (1): 1–36. Dopuch, N., R. Holthausen, and R. Leftwich. 1987. Predicting audit qualifications with financial and market variables. The Accounting Review 62 (3): 431–454. Drummond, C., and R. C. Holte. 2003. C4.5, class imbalance, and cost sensitivity: Why undersampling beats over-sampling. In The Proceedings of the Workshop on Learning from Imbalanced Datasets II, International Conference on Machine Learning, Washington, D.C. Fan, A., and M. Palaniswami. 2000. Selecting bankruptcy predictors using a support vector machine approach. Neural Networks 6: 354–359. Fanning, K., and K. Cogger. 1998. Neural network detection of management fraud using published financial data. International Journal of Intelligent Systems in Accounting, Finance and Management 7 (1): 21–41. Feroz, E., T. Kwon, V. Pastena, and K. Park. 2000. The efficacy of red flags in predicting the SEC’s targets: An artificial neural networks approach. International Journal of Intelligent Systems in Accounting, Finance and Management 9 (3): 145–157. Fries, T., N. Cristianini, and C. Campbell. 1998. The Kernel-Adatron algorithm: A fast and simple learning procedure for support vector machines. In The Proceedings of the 15th International Conference on Machine Learning, Madison, WI. Green, B. P., and J. H. Choi. 1997. Assessing the risk of management fraud through neural network technology. Auditing: A Journal of Practice & Theory 16 (1): 14–28. Hall, M., and G. Holmes. 2003. Benchmarking attribute selection techniques for discrete class data mining. IEEE Transactions on Knowledge and Data Engineering 15 (3): 1–16. Kaminski, K., S. Wetzel, and L. Guan. 2004. Can financial ratios detect fraudulent financial reporting? Managerial Auditing Journal 19 (1): 15–28. Kirkos, E., C. Spathis, and Y. Manolopoulos. 2007. Data mining techniques for the detection of fraudulent financial statements. Expert Systems with Applications 32 (4): 995–1003. Kotsiantis, S., E. Koumanakos, D. Tzelepis, and V. Tampakas. 2006. Forecasting fraudulent financial statements using data mining. International Journal of Computational Intelligence 3 (2): 104–110. Lee, T. A., R. W. Ingram, and T. P. Howard. 1999. The difference between earnings and operating cash flow as an indicator of financial reporting fraud. Contemporary Accounting Research 16 (4): 749– 786. Lin, J., M. Hwang, and J. Becker. 2003. A fuzzy neural network for assessing the risk of fraudulent financial reporting. Managerial Auditing Journal 18 (8): 657–665. Perlich, C., F. Provost, and J. Simonoff. 2003. Tree induction vs. logistic regression: A learning-curve analysis. Journal of Machine Learning Research 4: 211–255. Perols, J., and B. Lougee. 2009. Prior earnings management, forecast attainment, unexpected revenue per employee, and fraud. In The Proceedings of the American Accounting Association Western Region Annual Meeting, San Diego, CA. Auditing: A Journal of Practice & Theory May 2011 50 Perols Phua, C., D. Alahakoon, and V. Lee. 2004. Minority report in fraud detection: Classification of skewed data. SIGKDD Explorations 6 (1): 50–59. Platt, J. 1999. Fast training of support vector machines using sequential minimal optimization. In Advances in Kernel Methods: Support Vector Learning, edited by Scholkopf, B., C. J. C. Burges, and A. J. Smola, 185–208. Cambridge, MA: MIT. Prodromidis, A., P. Chan, and S. Stolfo. 2000. Meta-learning in distributed data mining systems: Issues and approaches. In Advances in Distributed and Parallel Knowledge Discovery, edited by Kargupta, H., and P. Chan, 81–114. Menlo Park, CA: AAAI/MIT. Provost, F., and T. Fawcett. 1997. Analysis and visualization of classifier performance: comparison under imprecise class and cost distributions. In Proceedings of the Third International Conference on Knowledge Discovery and Data Mining, Menlo Park, CA. Provost, F., T. Fawcett, and R. Kohavi. 1998. The case against accuracy estimation for comparing induction algorithms. In The Proceedings of the Fifteenth International Conference on Machine Learning, Madison, WI. Quinlan, J. R. 1993. C4.5: Programs for Machine Learning. San Francisco, CA: Morgan Kaufmann Publishers. Shin, K. S., T. Lee, and H. J. Kim. 2005. An application of support vector machines in bankruptcy prediction model. Expert Systems with Application 28: 127–135. Summers, S. L., and J. T. Sweeney. 1998. Fraudulently misstated financial statements and insider trading: An empirical analysis. The Accounting Review 73 (1): 131–146. Uzun, H., S. H. Szewczyk, and R. Varma. 2004. Board composition and corporate fraud. Financial Analysts Journal 60 (3): 33–43. Weiss, G. M. 2004. Mining with rarity: A unifying framework. ACM SIGKDD Explorations Newsletter 6 (1): 7–19. West, D., S. Dellana, and J. Qian. 2005. Neural network ensemble strategies for decision applications. Computer and Operations Research 32 (10): 2543–2559. Witten, I. H., and E. Frank. 2005. Data Mining: Practical Machine Learning Tools and Techniques. San Francisco, CA: Morgan Kaufmann Publishers. Wolpert, D. 1992. Stacked generalization. Neural Networks 5 (2): 241–259. Auditing: A Journal of Practice & Theory May 2011 Copyright of Auditing is the property of American Accounting Association and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder 's express written permission. However, users may print, download, or email articles for individual use.