Opening the 'Black Box' of Neural Networks

Lead Research Organisation: University of Cambridge
Department Name: Engineering

Abstract

In the field of Artificial Intelligence, the neural network (NN) has become an incredibly popular tool of late, achieving impressive accuracy on previously very challenging applications, including object recognition, image classification and prediction in high-dimensional spaces. Systems have been developed which independently learn intuitively reasonable data features, and the methodology has been found to generally excel at processing features in this way. However, very little intuition exists for the workings of NNs, especially in very deep multi-layer applications (Deep Learning); frequently, they are deployed by practitioners as 'black box' function approximators, with little fundamental understanding of their abilities and limitations. This is problematic for two reasons.
Firstly, the performance of an NN depends heavily on the selection of a suitable architecture, with the number of layers, number of perceptron 'units' per layer, unit interconnection structure, use of convolutional filters (and their associated sizes) and use of generalisation aids such as batch normalisation and dropout all impacting the network's test accuracy. Currently, appropriate architectures can only be identified by referencing designs used in other successful systems and evolving these by trial and error. The field would massively benefit from additional intuition for this task, which would permit more rigorous architecture optimisation and possibly the development of more accurate systems.
Secondly, by our non-understanding of how NNs work at a fundamental level, we are similarly blind to why they might break in certain situations. Recent research has considered adversarial examples - carefully perturbed sample inputs which are almost indistinguishable from the original by humans, but cause a classifier NN to confidently assign a completely incorrect classification - and found that many networks are highly susceptible to failing under this kind of attack. There have also been cases of networks exhibiting racial bias in their decisions, presumably learned from training data subject to ever-present human biases. NNs are beginning to be investigated in socially and economically important contexts, where there is potential for substantial harm to be done by an improperly functioning system; it is thus very dangerous to proceed down this route without thoroughly understanding networks' underlying behaviour. Consequently, the development of this kind of knowledge is a strong prerequisite for continuing research on this path.
I propose to investigate these issues by carefully analysing NNs of various scales, deployed in various applications. This could partly comprise a naïve grid search over possible architectures to investigate the resulting accuracy of each, or alternatively the exploration of mathematically provable aspects of the network's computations. One potentially fruitful avenue for immediate exploration is the somewhat paradoxical behaviour of NN weights - particularly in large-scale networks, the trained weight collection is curiously sparse, with relatively few weights differing non-trivially from zero, suggesting large swathes of the network contribute very little to the overall objective. Observing that additional weights increase the dimensionality of the weight-training search space, it seems intuitively reasonable to slim down large networks; indeed, existing work has noted minimal accuracy loss from substantial post-training weight pruning. However, further studies have found these seemingly superfluous weights to be surprisingly essential to the optimiser's ability to find a performant collection of weights, with training markedly inhibited in slimmed-down networks, even where we know a good-quality solution to exist. Understanding the mechanisms behind this phenomenon could provide valuable insight into this aspect of NNs' operation, thus advancing progress towards fuller general understanding of NNs.

Publications

10 25 50
 
Description 1) We have extended existing techniques used to automatically configure machine learning algorithms, making them applicable to a broader, more useful class of configuration settings. This will benefit a wide variety of applications, improving the final performance obtained from machine learning systems and reducing the time required to tune parameters to achieve that performance. Our work thus far will be published in a paper at the 2022 International Conference on Learning Representations, which proposes several avenues for advancing our results; we hope to investigate several of these in the remainder of the funded period.

2) In follow-up work, we investigated theoretical failings of a broad class of optimisation methods, arguing for an alternative approach which avoided these concerns. Surprisingly, our proposed alternative did not outperform established methods to the degree expected, which has incited further work into the mechanisms underlying existing techniques.
Exploitation Route 1) Our technique is applicable to a wide range of machine learning applications, so will be relevant to practitioners across a range of sectors as ML-driven 'intelligence' becomes more commonly used. We have detailed our approach in a submission to an open-access conference, and have made our research code publicly available, so all interested parties may deploy the logic in their own applications, and other researchers may pick up the ideas to advance further.

2) The unexpected results of our work will form part of my PhD thesis, and we hope to publish them in an open-access venue for further consideration by the research community.
Sectors Digital/Communication/Information Technologies (including Software)

 
Description Contributed Presentation to 19th EUROPT Workshop on Advances in Continuous Optimization 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Postgraduate students
Results and Impact Contributed a presentation to the "Multiobjective optimization and machine learning" track of the conference, engaging with other researchers with more theoretical and mathematical backgrounds in optimisation.
Year(s) Of Engagement Activity 2022
URL https://sites.fct.unl.pt/europt2022/home
 
Description Invited Talk at Vector Institute 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Postgraduate students
Results and Impact Invited to give a talk to the Vector Institute at the University of Toronto, Canada, by a PhD student there whose work was a key foundation in developing our contribution. Attendance included current PhD students and principal investigators. In follow-up question-and-answer, we discussed the outstanding issues in this area of research.
Year(s) Of Engagement Activity 2022