Loading [MathJax]/extensions/MathMenu.js
Pseudoinvertible Neural Networks | IEEE Journals & Magazine | IEEE Xplore
Impact Statement:While neural networks have proven highly capable in a myriad of tasks, a lack of labeled data often hinders progress. In complex domains, such as medical imaging, the cos...Show More

Abstract:

This work introduces novel architecture components and training procedures to create augmented neural networks with the ability to process data bidirectionally via an end...Show More
Impact Statement:
While neural networks have proven highly capable in a myriad of tasks, a lack of labeled data often hinders progress. In complex domains, such as medical imaging, the cost in human expert time to label a large dataset may be too high. Unsupervised and semi-supervised learning are highly advantageous in such scenarios, as they allow models to learn useful transformations with minimal or no information from an expert. Previously proposed solutions require that a system learn a task and its inverse, allowing it to self-supervise by reconstructing its own input. Our method provides an improved framework for such tasks by allowing a single neural network to learn both the forward and backward directions of a task, rather than the usual approach using two opposite models. This improves not only efficiency by reducing system size, but may also improve results by teaching the model a more comprehensive view of its task.

Abstract:

This work introduces novel architecture components and training procedures to create augmented neural networks with the ability to process data bidirectionally via an end-to-end approximate inverse. We develop pseudoinvertible neural network (PsI-NN) layers which function as drop-in replacements for corresponding convolutional and fully connected layers; by using these, existing architectures gain a pseudoinverse function which, with training, approximately reverses the forward function. For cases where learning both a task and its inverse are necessary or desirable, we show that PsI-NN enabled models match or exceed the quality of results generated by systems that use two separate models while drastically reducing system parameter count. We demonstrate this on two tasks: unpaired image translation and semisupervised classification. In both, PsI-NN greatly reduces parameter count without any loss of output quality; in semisupervised image classification, PsI-NN improves classification ...
Published in: IEEE Transactions on Artificial Intelligence ( Volume: 5, Issue: 2, February 2024)
Page(s): 602 - 612
Date of Publication: 18 July 2023
Electronic ISSN: 2691-4581

I. Introduction

In TRAINING neural network systems, obtaining the inverse of a multilayer operation is often either desirable or necessary, as evidenced by the widespread use of approximate inverse networks whose objective includes the reconstruction of the input from the output such as in autoencoders [6] and unpaired image translation [3]. These solutions require a second model with mirrored tensor dimensions to be trained on the opposite task of the first network. While this approach has proven functional in practice, it usually doubles the number of trainable parameters in the system and requires the mirrored model to be trained to approximately invert the function represented by the first model, which itself is already an approximation of the true mapping function. This approach has been necessary to obtain an approximation to a bijective mapping between two data distributions because multilayer neural networks are not invertible functions. In this article, we introduce pseudoinvertible neural networks (PsI-NNs), a class of neural networks with the capability to process data both forwards and backwards, which through training enables them to approximate a bijective mapping function with a single model.

Contact IEEE to Subscribe

References

References is not available for this document.