Back to Search Start Over

Automation for Interpretable Machine Learning Through a Comparison of Loss Functions to Regularisers

Authors :
Parkes, A. I.
Camilleri, J.
Hudson, D. A.
Sobey, A. J.
Publication Year :
2021

Abstract

To increase the ubiquity of machine learning it needs to be automated. Automation is cost-effective as it allows experts to spend less time tuning the approach, which leads to shorter development times. However, while this automation produces highly accurate architectures, they can be uninterpretable, acting as `black-boxes' which produce low conventional errors but fail to model the underlying input-output relationships -- the ground truth. This paper explores the use of the Fit to Median Error measure in machine learning regression automation, using evolutionary computation in order to improve the approximation of the ground truth. When used alongside conventional error measures it improves interpretability by regularising learnt input-output relationships to the conditional median. It is compared to traditional regularisers to illustrate that the use of the Fit to Median Error produces regression neural networks which model more consistent input-output relationships. The problem considered is ship power prediction using a fuel-saving air lubrication system, which is highly stochastic in nature. The networks optimised for their Fit to Median Error are shown to approximate the ground truth more consistently, without sacrificing conventional Minkowski-r error values.<br />Comment: 11 pages, 5 figures, under review

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2106.03428
Document Type :
Working Paper