Autoencoders For Anomaly Detection Are Unreliable
Autoencoders For Anomaly Detection Are Unreliable
RELIABLE
A BSTRACT
Autoencoders are frequently used for anomaly detection, both in the unsupervised
and semi-supervised settings. They rely on the assumption that when trained using
the reconstruction loss, they will be able to reconstruct normal data more accu-
rately than anomalous data. Some recent works have posited that this assumption
may not always hold, but little has been done to study the validity of the assump-
tion in theory. In this work we show that this assumption indeed does not hold,
and illustrate that anomalies, lying far away from normal data, can be perfectly re-
constructed in practice. We revisit the theory of failure of linear autoencoders for
anomaly detection by showing how they can perfectly reconstruct out of bounds,
or extrapolate undesirably, and note how this can be dangerous in safety critical
applications. We connect this to non-linear autoencoders through experiments on
both tabular data and real-world image data, the two primary application areas of
autoencoders for anomaly detection.
1 I NTRODUCTION
Autoencoders are one of the most popular architectures within anomaly detection, either directly, or
as a scaffold or integral part in larger pipelines or architectures. They are commonly used across a
variety of domains, such as predictive maintenance (Kamat & Sugandhi, 2020), network anomaly
detection (Said Elsayed et al., 2020), and intrusion detection (Farahnakian & Heikkonen, 2018),
but find much contemporary use in computer vision anomaly detection, with applications such as
industrial inspection (Tsai & Jen, 2021), medical imaging (Wei et al., 2018), structural health mon-
itoring (Chow et al., 2020) and video surveillance (Zhao et al., 2017; Cruz-Esquivel & Guzman-
Zavaleta, 2022). Many of these applications are safety critical, meaning that the reliability of these
algorithms is of utmost importance in order to prevent catastrophic failure and associated dangers
and consequences.
Anomaly detection using autoencoders typically relies on using the reconstruction loss, often the
mean squared error (MSE), as a proxy for “anomalousness”. The underlying assumption is that
anomalies are harder to reconstruct, and will therefore have a higher reconstruction loss. How-
ever, the validity of this assumption has been questioned in recent years. Merrill & Eskandarian
(2020) and Beggel et al. (2020) for example state that anomalies in the training data might lead to
reconstruction of anomalies, leading to unreliable detectors. Some researchers have noted that re-
construction of unseen anomalies might also occur in the semi-supervised setting, where the training
data is assumed to contain no anomalies (Astrid et al., 2021; 2024; Gong et al., 2019; Zong et al.,
2018). Yet, little work has been done on the nature of failure and reliability of autoencoders.
In this work we provide valuable insights into the reliability of autoencoders for anomaly detection.
Following the seminal works of Bourlard & Kamp (1988) and Baldi & Hornik (1989) we mathemat-
ically and experimentally study autoencoder failure modes, whilst briefly examining how various
activation functions influence these failures. We show that failure is not just a rarely occurring edge
case, but also show failure cases on tabular data and on real-world image data commonly used in
anomaly detection benchmarks. By doing this we provide a foundation for future research in solving
the demonstrated unreliability of autoencoders for anomaly detection and furthering our understand-
1
ing of autoencoders. We show that for different architectures and activations functions, even with
sufficiently low-dimensional latent spaces, these problems persist. By developing our understanding
of autoencoders for anomaly detection, we hope to function as a warning regarding the reliability of
autoencoders in practice, and as a scaffold for future developments striving for more robust anomaly
detection algorithms.
2 R ELATED W ORK
This work is not the first to recognize that autoencoders have several issues as anomaly detectors.
The most discussed type of failure is the unwanted reconstruction of anomalies, which is also the
focus of this work. Several causes of this unwanted behavior have been proposed.
Many works focus on the unsupervised setting, and observe that contrary to prior belief, autoen-
coders will fairly easily reconstruct any anomalies present in the training data, leading to an unusable
detector (Merrill & Eskandarian, 2020; Beggel et al., 2020; Cheng et al., 2021; Tong et al., 2022).
Several works cover the anomaly reconstruction problem within the semi-supervised setting. Most
commonly, it is only experimentally observed that anomalies are well reconstructed (Gong et al.,
2019; Zong et al., 2018; Cheng et al., 2021; Astrid et al., 2021; 2024; Salehi et al., 2021; Nalisnick
et al., 2019; Lyudchik, 2016). Based on these experimental results, some solutions have been pro-
posed. Gong et al. (2019) mention that out-of-bounds reconstruction can occur and propose adding a
memory module to the autoencoder to alleviate the issue. While the addition of the memory module
can aid in limiting out-of-bounds reconstruction, it also leads to a severely decreased reconstruc-
tion ability and substantial added complexity in training and optimizing the network. Zong et al.
(2018) note that while some anomalies have a high reconstruction loss, some occupy the region of
normal reconstruction loss, and add a Gaussian mixture model to the latent space to aid in detection
of anomalies under the assumption that anomalies occupy a low-density region in the latent space.
Similarly, Cheng et al. (2021) aim to detect anomalies in the latent space by looking at the distance
to the centroid. From our experiments we can glean that relying on distances or densities in the
latent space does not always work in practice. Astrid et al. (2021; 2024) make use of constructed
pseudo anomalies in training the autoencoder. They add adaptive noise to normal samples to gen-
erate pseudo anomalies. In the reconstruction, they then optimize the reconstruction loss between
the pseudo anomaly and the normal sample used to generate it. While they show promising results
and greater discriminative power on benchmark datasets, they do not quantify to which degree their
performance gains can be attributed to the reduction of the out-of-bounds reconstruction. Salehi
et al. (2021) aim to limit the reconstruction of anomalies by generating adversarial samples. The ad-
versarial examples are generated by perturbing the normal samples, and minimizing the effect those
perturbations have in the latent space. This is similar to the concept of denoising autoencoders.
Based on our experiments, we do not think this results in a reliable autoencoder, as often adversarial
anomalies can occupy the latent space close to normal data.
Some authors have moved beyond the experimental, and propose explanations for the anomaly re-
construction problem. For example You et al. (2022), Lu et al. (2023) and Bercea et al. (2023)
propose that anomaly reconstruction can happen because an autoencoder can learn an “identical
shortcut”, where both normal data and anomalous data is effectively reconstructed using an identity
mapping. This point has however been countered by Cai et al. (2024) who show that by constraining
the latent space to be sufficiently low dimensional, this problem can be avoided.
The second line of thought follows from VAE anomaly detection, where Havtorn et al. (2021) theo-
rize that in out-of-distribution detection, unwanted reconstruction can happen due to a high correla-
tion between learned low-level features for in- and out-of-distribution data.
A third line of thought is proposed by Zhou (2022) who propose that reconstruction of out-of-
distribution samples can happen due to out-of-distribution data having smaller neural activations
than in-distribution data.
Finally, some works theorize that autoencoders can perfectly reconstruct data due to the anomalies
occupying the reconstruction hyperplane, or latent space manifold. Denouden et al. (2018) for
example note this phenomenon, and aim to solve it by adding the Mahalanobis distance in the latent
space to the loss. The most detailed work is that of Yoon et al. (2021) who provide an example of
the hyperplane interpolating between clusters of data. They solve this by introducing a normalized
2
autoencoder which reinterprets the autoencoder as a likelihood-based energy model. We specifically
follow up on this line of reasoning, and revisit mathematical proofs and provide new experimental
evidence of anomaly reconstruction due to both unwanted extrapolation and inter-class interpolation.
3 BACKGROUND
3.1 A NOMALY D ETECTION
In practical anomaly detection, we attribute a score si = fanomaly score (xi ) for each sample xi ∈ X =
Rn , i.e. the i-th row of dataset X with size m-by-n. The score should then be higher for anomalous
samples than for normal data. When applied to some dataset consisting of both normal data and
anomalies, i.e. X = {X normal , X anomalous }, a perfect anomaly detector will assign higher scores to
anomalies than to normal data: mini (fanomaly score (xanomalous
i )) > maxi (fanomaly score (xnormal
i )).
The two most common anomaly detection settings are unsupervised and semi-supervised. Un-
supervised anomaly detection is characterized by having no discernible “train” and “test” splits.
Instead, we only consider a single dataset X = {X normal , X anomalous }, where we are uncer-
tain which samples are anomalous and which are not. In the semi-supervised setting we instead
have a conventional “train” and “test” set. The train set consists out of only normal samples:
X train = {X train, normal }, while the test set is unknown, and can contain both normal and anoma-
lous samples: X test = {X test, normal , X test, anomalous }. In this paper we will only consider the semi-
supervised case, and simplify the notation with X = X train and xi referring to an individual training
data point, which in the semi-supervised case by definition is not an anomaly. We then consider a
new data point a to determine whether this is an anomaly or not. In older literature, semi-supervised
anomaly detection is often called one-class classification.
Both PCA and autoencoders are dimensionality reduction techniques that can be used to detect
anomalies using their reconstruction loss, commonly known as the mean squared error, or MSE.
We can calculate the reconstruction loss by comparing a sample xi to its reconstruction x̂i :
Pn 2
LR (xi , x̂i ) = n1 j=1 (xi,j − x̂i,j ) , for each sample vector xi . This reconstruction loss often
serves as a proxy for detecting anomalies, with the underlying assumption that a higher reconstruc-
tion loss indicates a higher likelihood of the sample being an anomaly.
For both PCA and autoencoders we want to find a lower-dimensional encoding Y , e.g. d < n, in the
encoding space Y = Rd by applying the function g : X → Y. We then decode Y by transforming
it back into the space X through the decoder h : Y → X , yielding the reconstructed data X̂.
g h
Summarizing, we learn the concrete transformations X −→Y − → X̂.
We can then formulate the anomaly scoring function in terms of the reconstruction loss, encoder,
and decoder: fanomaly score (xi ) = LR (xi , h(g(xi ))). The worst case can then be formulated as: there
exists an a far from all training data such that LR (a, â) ≤ mini (LR (xi , x̂i )).
3
4.2 PCA
Now we will prove that there exists some adversarial example a ∈ Rn that is far from all normal
data, i.e. mini (dist(xi , a)) > δ, for an arbitrary δ and the Euclidean distance metric, but still has a
reconstruction loss LR (a, g(h(a))) = 0.
Let us first recall that any a in the column space of Vd will have zero reconstruction loss.
If we then define a = xi Vd VdT + cVdT , a will still have zero reconstruction loss.
Then for the Euclidean distance it follows that:
dist(xi , a)2 = ∥xi − xi Vd VdT ∥2 + ∥a − xi Vd VdT ∥2 ,
or the squared Euclidean distance is equal to the distance from xi to its projection onto the hyper-
plane xi Vd VdT plus the distance from that projection xi Vd VdT to a.
It then follows that:
dist(xi , a)2 ≥ ∥a − xi Vd VdT ∥2 = ∥xi Vd VdT + cVdT − xi Vd VdT ∥2 = ∥cVdT ∥2 ,
which we can increase to arbitrary length. This can be intuited as moving the projection of xi along
the hyperplane.
To ensure that we increase the distance to all points xi rather than just a single one, we need to move
outward starting from a sample on the convex hull enclosing XVd . Any point in this convex set,
that is the set of points occupying the convex hull, can be moved along the hyperplane to increase
the distance to all points xi Vd , and therefore to all points xi as long as c lies in the direction from
xi Vd to the exterior of the convex hull.
We can thus always find a a = xi Vd VdT + cVdT , for some xi Vd in the convex set of XVd and
choose c so that it points from xi Vd to the exterior of the convex hull and is of sufficient length such
that mini (dist(xi , a)) > δ, for an arbitrary δ.
4
Hence, all vectors a ∈ Rn found in this way are constructed anomalies, or adversarial examples,
that are far from all normal data, but still have zero reconstruction loss.
We posit that the same principle applies to other distance metrics, and the intuition is that this follows
the same line of reasoning as presented here for the Euclidean distance.
Linear neural networks, like PCA, can also exhibit out-of-bounds reconstruction for certain anoma-
lous data points.
Linear autoencoders consist of a single linear encoding layer and a single linear decoding layer.
Given a mean-centered dataset X, the encoding and decoding transformations can be represented
as follows:
Y = g(X) = XWenc ,
T T
X̂ = h(Y ) = Y Wdec = XWenc Wdec
T
where Wenc is the n-by-d weight matrix of the encoder, and Wdec is the d-by-n weight matrix of
the decoder. We assume the autoencoder to have converged to the global optimum. Note that we
T
define Wdec in its transposed form to illustrate its relation to VdT . Due to the full linearity of the
model, even multiple layer networks can be simplified to a single multiplication. It is known that
a well-converged linear autoencoder finds an encoding in the space spanned by the first d principal
components (Baldi & Hornik, 1989). In other words, the encoding weight matrix can be expressed
in terms of the first d principal components and some invertible matrix C:
Wenc = Vd C.
T
At the global optimum Wdec can be expressed as the inverse of Wenc :
T
Wdec −1
= Wenc = (Vd C)−1 = C −1 Vd−1 = C −1 VdT .
To show that linear autoencoders can exhibit perfect out-of-bounds reconstruction, we follow the
same lines as for PCA.
Let us now look for some a ∈ Rn . We aim to prove that if a lies in the column space of Vd , then
the reconstruction loss LR (a, h(g(a))) = 0.
We need to show that there exists some a such that h(g(a)) = a. For linear autoencoders, this
condition can be written as:
T
aWenc Wdec = a.
T
Assume a is in the row space of Vd . Then a can be expressed as a linear combination of the rows
in Vd . Let c ∈ Rd be such that:
a = cVdT .
Then it follows that:
T
aWenc Wdec = cVdT Wenc Wdec
T
= cVdT Vd CC −1 VdT = cVdT Vd VdT = cVdT = a,
indicating that a satisfies the condition h(g(a)) = a, implying that the reconstruction loss
LR (a, g(h(a))) = 0.
By proving this, the case of the linear autoencoder reduces to that of PCA, with the same
proof that adversarial examples satisfying mini (dist(xi , a)) > δ, with a reconstruction loss
LR (a, g(h(a))) = 0, exist.
An extension of this proof to the case of linear networks with bias terms applied on non-centered
data can be found in Appendix A.1.
In this section we show that datasets exist for which we can prove that non-linear neural networks
perform the same unwanted out-of-bounds reconstruction. Then we experimentally show that this
behavior indeed occurs in more complex real-world examples. In this way, we illustrate how au-
toencoders demonstrate unreliability even in real-world scenarios.
5
4.4.1 FAILURE OF A N ON -L INEAR N ETWORK WITH R E LU ACTIVATIONS
We can show on a simple dataset that unwanted reconstruction behavior can occur in non-linear
autoencoders. We consider a two-dimensional dataset X consisting out of normal samples xi =
αi (1, 1), where αi is some scalar. Simply put, every normal sample xi occupies the diagonal. This
dataset can be perfectly reconstructed by a linear autoencoder with Wenc = β(1, 1)T , where β is
some scalar. The simplest non-linear autoencoder with ReLU activation will find the same weights,
but with a bias such that xi Wenc + benc > 0 for all xi ∈ X, i.e. benc ≥ mini (xi Wenc ). This will
then lead to a perfect reconstruction for all xi . Adversarial anomalies a can also be easily found as
a = c(1, 1), where c ≫ max i (xi )
(1,1) is some sufficiently large scalar such that mini (dist(xi , a)) > δ
is satisfied. We theorize that even beyond this simple case, similar linear behavior can occur beyond
the convex hull that the normal data occupies. We experimentally show this anomaly reconstruction
behavior in later sections.
On more complex datasets, we observe similar behavior. We have synthesized several two-
dimensional datasets to show how non-linear autoencoders behave when used for anomaly detec-
tion. These datasets, as well as contours of the MSE of autoencoders trained on these datasets,
are visualized in Figure 1. In each of these cases, we have synthesized a dataset by sampling 100
points per distribution, either from a single or from two Gaussians as in 1a, 1b, 1e, and 1f, or from
x2 = x21 with added Gaussian noise in 1c and 1d. In all cases we use an autoencoder with layer
sizes [2,5,1,5,2], except for 1d, where we use layer sizes of [2,100,20,1,20,100,2] to better model
the non-linearity of the data. All layers have ReLU (Subfigures 1a, 1b, 1c, and 1d ), or sigmoid
(Subfigures 1e and 1f) activations, except for the last layer, which has a linear activation. In these
figures the color red is used to highlight those areas where the autoencoder is able to nearly perfectly
reconstruct the data, i.e. MSE < ϵ = 0.1.
ReLU Activation Failure on Tabular Data We can readily observe some of the problematic
behavior of autoencoders as anomaly detectors. Firstly, in Figure 1a we observe that well outside
the bounds of the training data there is an area with a near-perfect reconstruction. Worryingly, the
reconstruction loss is lower than for a large part of the area which the normal data occupies. If we
move in the (−1, −1) direction, the encoder and decoder will no longer match perfectly. Even so,
problematically low reconstruction losses can be found in this direction. In Figures 1c and 1d we
see the same linear out-of-bounds behavior. In each of these cases, the mismatch between encoder
and decoder in the linear domain is less noticeable, leading to even larger areas of near-perfect
reconstruction. Lastly, in Figure 1b that there is an area between the two clusters with a good
reconstruction. Likely the units responsible for this area are still close to their initialization, and due
to the simple shallow network architecture can not meaningfully contribute to the reconstruction of
any samples.
Our intuition of this behavior directly relates to the proof of out-of-bounds reconstruction we have
provided for linear autoencoders. At the edge of the data space, only a few of the ReLU neurons
activate. Beyond this edge, no new neurons will activate, nor will any of the activated ones deacti-
vate. This can lead to linear behavior on some edge of the data space, i.e., in this area the network
reduces to a linear transformation Wenc . If we now synthesize some a such that it lies in the column
T
space of Wenc , we can again find some adversarial anomalies a = cWenc . Like we have observed
in Figure 1a, there may be a mismatch between the encoder and decoder, even at the global opti-
mum, so we might not be able to increase c towards infinity and still find adversarial examples with
LR (a, g(h(a))) < ϵ.
Sigmoid Activation Autoencoders Nowadays, full sigmoid networks have mostly fallen out of
favor in deeper networks due to their vanishing gradients (Hochreiter, 1991; Glorot et al., 2011).
However, sigmoids are more attractive to use in anomaly detection because they lead to networks
that do not exhibit the hyperplane issues that the ReLU suffers from. While sigmoids have the more
desirable property of tapering off at high or low input, making it hard to perfectly reconstruct data
far away from normal data, autoencoders with just sigmoid activation can still behave unexpectedly,
albeit less so than those with ReLU activation.
6
626.9 632.0 397.1
X2
X2
1.8 1.6
0.7
0.3 0.1 0.2
X2
X2
2.8 1.9
0.8
0.6 0.5
0.1 0.1 0.1
X1 0.0 X1 0.0 X1 0.0
Figure 1: Plots of the contours of the reconstruction loss of non-linear autoencoders when applied
to 3 distinct datasets. The datasets consist out of a 100 samples from a 2D Gaussian (a, e), 100
samples for each of 2 different 2D Gaussians (b, f), and a 100 samples from a classic banana shaped
function with Gaussian noise (c, d). In (a,b,c,e,f) a [2,5,1,5,2] autoencoder is used, while for (d)
a deeper [2,100,20,1,20,100,2] autoencoder is used. The contourplot is colored red whenever the
MSE is below a set threshold ϵ < 0.1 to indicate a near-perfect reconstruction. Note that the color
scaling is exponential to better visualize the MSE loss.
We can see in Figure 1e that the data is nicely approximated by a sigmoid autoencoder. It extends
nicely to the first and last samples on the direction of the first principal component, and does not
extend beyond that. When we extend this example to multimodal data, as in Figure 1f, we can see
different undesirable behavior arising. There exists an area where the sigmoids reconstructing both
clusters intersect. Due to the two distinct sigmoids mixing, we can find a hyperplane orthogonal to
the first principal component where the reconstruction loss is much lower than would be expected.
While in this case there are no points on the hyperplane which would have a lower reconstruction
loss than all normal data, there is still a substantial area where the reconstruction loss is lower than
for many of the normal data points.
Other Activation Functions While we have explicitly discussed the ReLU and sigmoid activation
functions, the behavior shown is similar for other activation functions. Effectively, we can categorize
most activation functions as those having an order of continuity of C 0 like the ReLU, or C ∞ like the
sigmoid. In summary, activation functions with an order of continuity of C 0 suffer most from out-
of-bounds reconstruction, but allow for more easily trainable deep networks. In contrast, activation
functions with an order of continuity of C ∞ generally have more desirable properties for anomaly
detection, but are harder to use in deep networks due to the vanishing gradient.
All the previous examples clearly illustrate autoencoders’ possible failure and unreliability when
used for anomaly detection on tabular data. Yet, many applications of anomaly detection are in
computer vision, where anomaly detection can be used to detect foreign objects. Typical examples
of computer vision anomaly detection are surveillance, where videos are analyzed to find possible
security threats (Nayak et al., 2021; Sultani et al., 2018), structural health monitoring (Bao et al.,
2019), and industrial inspection (Bergmann et al., 2021).
7
For most applications of autoencoders on image data, the architecture is fairly straightforward.
ReLU activation functions are most commonly used throughout the network, with a sigmoid ac-
tivation at the final layer. Connections to and from the bottleneck layer are often chosen to be just
linear, to allow for a richer internal representation. As most layers have ReLU activation functions,
these networks do not suffer from the vanishing gradient. Yet, due to using a sigmoid at the last
layer, these networks suffer less from the issues encountered in full ReLU/linear networks as dis-
cussed in Section 4.4.2. Nonetheless, we will show that even on more complex real-world problems,
autoencoders remain unreliable and are often able to reconstruct out-of-bounds.
8
Normal 4 130.0 Normal 0 69.06
Normal 5 Normal 1
Y2
2.3
0.82
0.6
0.2 0.19
Y1 0.0 Y1 0.00
(a) (b)
Figure 2: Plots of the contours of the reconstruction loss in the 2D latent space of a convolutional
autoencoder when applied on subset of MNIST (a, b), plots of constructed adversarial anomalies
(c, d), and a plot of non-problematic out-of-bounds reconstruction (e). Subplots (a, c) show the
results for an autoencoder trained on digits 4, 5, and 7, and Subplots (b, d, e) show the results for an
autoencoder trained on digits 0, and 1. The visualized samples, i.e. the points in (a, b) are the latent
representations of the training data. The shown digits are constructed by sampling from the ϵ < 0.1
zone within the marked area, and correspond to these from left to right. The contourplot is colored
red whenever the MSE is below a set threshold ϵ < 0.1 to indicate a near-perfect reconstruction.
Note that the color scaling is exponential to better visualize the MSE loss.
contrast to our previous experiment, this area corresponds to a more uncommon diagonally drawn
1, as shown in Figure 2e, which is still within the bounds of what we can consider normal data.
From this we can conclude that although out-of-bounds reconstruction can be unwanted, in some
cases it aligns with the expectations of an anomaly detector. More generally speaking we observe
that some generalization of the autoencoder can align with the expectation of a user. In some cases,
generalization can lead to unwanted inter- or extrapolation. This unwanted generalization can cause
anomalous data to stay fully undetected. This is similar to the phenomena observed by Nalisnick
et al. (2019) for variational autoencoders, who observe some out-of-distribution samples to remain
fully undetected.
While conducting the experiments on the MNIST data, we found that the problems shown above do
not seem to arise in every case. We observe that depending on the random seed used for initialization
and the digits selected as normal data, the out-of-bounds reconstruction may or may not be easily
detected. In some cases, the out-of-bounds behavior seems very non-monotonous, meaning that the
regions of reconstruction disappear one epoch, and reappear the next. This solidifies our belief that
autoencoders may fail, but in many cases outwardly seem to work well. The crux lies in the fact
that in a semi-supervised or unsupervised setting, it is not possible to accurately judge whether a
network will fail on future data. This is further complicated by the fact that due to the heterogeneity
of anomalies, some may be detected while others go unnoticed.
9
5 C ONCLUSION
In this work we provide an analysis of the unwanted reconstruction of anomalies that autoencoders
can exhibit when used for anomaly detection. We move beyond existing theories of unwanted re-
construction happening in interpolation and show how unwanted out-of-bounds reconstruction can
occur when extrapolating as well, and how this can lead to anomalies staying fully undetected. We
show through several experiments that these issues can arise in real-world data and not just in theory.
This leads us to some safety concerns, where autoencoders can catastrophically fail to detect obvi-
ous anomalies. This unreliability can have major consequences when trust is put into the anomaly
detector in safety-critical applications.
In general, we solidify the growing belief that the reconstruction loss is not a reliable proxy for
anomaly detection, especially when the network is explicitly trained to lower the reconstruction
loss for normal data without constraining the reconstruction capability beyond the bounds of the
normal training data such as has been done by Yoon et al. (2021). We find that this issue is most
prevalent for (conditionally) linear units such as the ReLU, but similar issues exist for sigmoid
networks, albeit to a lesser degree. The reconstruction issue is mostly caused by the fact that a point
in the lower-dimensional latent space corresponds to a hyperplane in the original space that the data
occupies. Next to interpolation and out-of-bounds reconstruction, we find that anomalies can remain
undetected when they occupy the latent space where normal classes border.
Users of autoencoders for anomaly detection should be aware of these issues. Good practice would
be to at least check whether a trained non-linear autoencoder exhibits the undesirable out-of-bounds
reconstruction. In this paper’s illustrative examples, we checked for this by searching for adversar-
ial anomalies. This was relatively easy, as it could be done either visually in the latent space, or
through a simple 2D grid search. For more complex datasets, requiring larger latent spaces, a feasi-
ble strategy might be to again synthesize samples from the latent space and formulate the search for
adversarial anomalies as an optimization in terms of projected gradient descent (Madry et al., 2017).
By describing exactly how autoencoders are unreliable anomaly detectors by describing anomaly
reconstruction, we hope to provide a scaffold for future research on fixing and avoiding the identified
issues in a targeted manner.
AUTHOR C ONTRIBUTIONS
Roel Bouman: Conceptualization, Methodology, Software, Validation, Formal Analysis, Investiga-
tion, Writing - Original Draft, Writing - Review & Editing, and Visualization. Tom Heskes: Writing
- Review & Editing, Supervision, Project Administration, and Funding Acquisition.
ACKNOWLEDGMENTS
The research reported in this paper has been partly funded by the NWO grant NWA.1160.18.238
(PrimaVera); as well as BMK, BMDW, and the State of Upper Austria in the frame of the SCCH
competence center INTEGRATE [(FFG grant no. 892418)] part of the FFG COMET Competence
Centers for Excellent Technologies Programme. We also want to thank Alex Kolmus and Marco
Loog for the excellent input and discussions on this topic.
10
R EFERENCES
Marcella Astrid, Muhammad Zaigham Zaheer, Jae-Yeong Lee, and Seung-Ik Lee. Learning not to
reconstruct anomalies. arXiv preprint arXiv:2110.09742, 2021.
Marcella Astrid, Muhammad Zaigham Zaheer, Djamila Aouada, and Seung-Ik Lee. Exploiting
autoencoder’s weakness to generate pseudo anomalies. Neural Computing and Applications, pp.
1–17, 2024.
Pierre Baldi and Kurt Hornik. Neural networks and principal component analysis: Learning from
examples without local minima. Neural networks, 2(1):53–58, 1989.
Yuequan Bao, Zhiyi Tang, Hui Li, and Yufeng Zhang. Computer vision and deep learning–based
data anomaly detection method for structural health monitoring. Structural Health Monitoring,
18(2):401–421, 2019.
Laura Beggel, Michael Pfeiffer, and Bernd Bischl. Robust anomaly detection in images using ad-
versarial autoencoders. In Machine Learning and Knowledge Discovery in Databases: European
Conference, ECML PKDD 2019, Würzburg, Germany, September 16–20, 2019, Proceedings, Part
I, pp. 206–222. Springer, 2020.
Cosmin I Bercea, Daniel Rueckert, and Julia A Schnabel. What do aes learn? challenging common
assumptions in unsupervised anomaly detection. In International Conference on Medical Image
Computing and Computer-Assisted Intervention, pp. 304–314. Springer, 2023.
Paul Bergmann, Kilian Batzner, Michael Fauser, David Sattlegger, and Carsten Steger. The MVTec
anomaly detection dataset: a comprehensive real-world dataset for unsupervised anomaly detec-
tion. International Journal of Computer Vision, 129(4):1038–1059, 2021.
Roel Bouman, Zaharah Bukhsh, and Tom Heskes. Unsupervised anomaly detection algorithms on
real-world data: How many do we need? Journal of Machine Learning Research, 25(105):1–34,
2024. URL [Link]
Hervé Bourlard and Yves Kamp. Auto-association by multilayer perceptrons and singular value
decomposition. Biological cybernetics, 59(4):291–294, 1988.
Yu Cai, Hao Chen, and Kwang-Ting Cheng. Rethinking autoencoders for medical anomaly detection
from a theoretical perspective. In International Conference on Medical Image Computing and
Computer-Assisted Intervention, pp. 544–554. Springer, 2024.
Zhen Cheng, Siwei Wang, Pei Zhang, Siqi Wang, Xinwang Liu, and En Zhu. Improved autoencoder
for unsupervised anomaly detection. International Journal of Intelligent Systems, 36(12):7103–
7125, 2021.
Jun Kang Chow, Zhaoyu Su, Jimmy Wu, Pin Siang Tan, Xin Mao, and Yu-Hsing Wang. Anomaly
detection of defects on concrete structures with the convolutional autoencoder. Advanced Engi-
neering Informatics, 45:101105, 2020.
Taylor Denouden, Rick Salay, Krzysztof Czarnecki, Vahdat Abdelzad, Buu Phan, and Sachin
Vernekar. Improving reconstruction autoencoder out-of-distribution detection with mahalanobis
distance. arXiv preprint arXiv:1812.02765, 2018.
Fahimeh Farahnakian and Jukka Heikkonen. A deep auto-encoder based approach for intrusion de-
tection system. In 2018 20th International Conference on Advanced Communication Technology
(ICACT), pp. 178–183. IEEE, 2018.
Xavier Glorot, Antoine Bordes, and Yoshua Bengio. Deep sparse rectifier neural networks. In
Proceedings of the fourteenth international conference on artificial intelligence and statistics, pp.
315–323. JMLR Workshop and Conference Proceedings, 2011.
11
Dong Gong, Lingqiao Liu, Vuong Le, Budhaditya Saha, Moussa Reda Mansour, Svetha Venkatesh,
and Anton van den Hengel. Memorizing normality to detect anomaly: Memory-augmented deep
autoencoder for unsupervised anomaly detection. In Proceedings of the IEEE/CVF international
conference on computer vision, pp. 1705–1714, 2019.
Jakob D Havtorn, Jes Frellsen, Søren Hauberg, and Lars Maaløe. Hierarchical vaes know what they
don’t know. In International Conference on Machine Learning, pp. 4117–4128. PMLR, 2021.
Pooja Kamat and Rekha Sugandhi. Anomaly detection for predictive maintenance in industry 4.0-a
survey. In E3S web of conferences, volume 170, pp. 02007. EDP Sciences, 2020.
Yann LeCun. The mnist database of handwritten digits. [Link] lecun. com/exdb/mnist/, 1998.
Ruiying Lu, YuJie Wu, Long Tian, Dongsheng Wang, Bo Chen, Xiyang Liu, and Ruimin Hu. Hier-
archical vector quantized transformer for multi-class unsupervised anomaly detection. Advances
in Neural Information Processing Systems, 36:8487–8500, 2023.
Olga Lyudchik. Outlier detection using autoencoders. Technical report, CERN, 2016.
Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu.
Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083,
2017.
Nicholas Merrill and Azim Eskandarian. Modified autoencoder training and scoring for robust
unsupervised anomaly detection in deep learning. IEEE Access, 8:101824–101833, 2020.
Eric Nalisnick, Akihiro Matsukawa, Yee Whye Teh, Dilan Gorur, and Balaji Lakshminarayanan. Do
deep generative models know what they don’t know? In International Conference on Learning
Representations, 2019.
Rashmiranjan Nayak, Umesh Chandra Pati, and Santos Kumar Das. A comprehensive review on
deep learning-based methods for video anomaly detection. Image and Vision Computing, 106:
104078, 2021.
Mahmoud Said Elsayed, Nhien-An Le-Khac, Soumyabrata Dev, and Anca Delia Jurcut. Network
anomaly detection using lstm based autoencoder. In Proceedings of the 16th ACM Symposium on
QoS and Security for Wireless and Mobile Networks, pp. 37–45, 2020.
Mohammadreza Salehi, Atrin Arya, Barbod Pajoum, Mohammad Otoofi, Amirreza Shaeiri, Moham-
mad Hossein Rohban, and Hamid R Rabiee. Arae: Adversarially robust training of autoencoders
improves novelty detection. Neural Networks, 144:726–736, 2021.
Waqas Sultani, Chen Chen, and Mubarak Shah. Real-world anomaly detection in surveillance
videos. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.
6479–6488, 2018.
Alexander Tong, Guy Wolf, and Smita Krishnaswamy. Fixing bias in reconstruction-based anomaly
detection with lipschitz discriminators. Journal of Signal Processing Systems, 94(2):229–243,
2022.
Du-Ming Tsai and Po-Hao Jen. Autoencoder-based anomaly detection for surface defect inspection.
Advanced Engineering Informatics, 48:101272, 2021.
Qi Wei, Yinhao Ren, Rui Hou, Bibo Shi, Joseph Y Lo, and Lawrence Carin. Anomaly detection for
medical images based on a one-class classification. In Medical Imaging 2018: Computer-Aided
Diagnosis, volume 10575, pp. 375–380. SPIE, 2018.
Sangwoong Yoon, Yung-Kyun Noh, and Frank Park. Autoencoding under normalization constraints.
In International Conference on Machine Learning, pp. 12087–12097. PMLR, 2021.
12
Zhiyuan You, Lei Cui, Yujun Shen, Kai Yang, Xin Lu, Yu Zheng, and Xinyi Le. A unified model for
multi-class anomaly detection. Advances in Neural Information Processing Systems, 35:4571–
4584, 2022.
Yiru Zhao, Bing Deng, Chen Shen, Yao Liu, Hongtao Lu, and Xian-Sheng Hua. Spatio-temporal au-
toencoder for video anomaly detection. In Proceedings of the 25th ACM international conference
on Multimedia, pp. 1933–1941, 2017.
Bo Zong, Qi Song, Martin Renqiang Min, Wei Cheng, Cristian Lumezanu, Daeki Cho, and Haifeng
Chen. Deep autoencoding gaussian mixture model for unsupervised anomaly detection. In Inter-
national conference on learning representations, 2018.
A A PPENDIX
Linear neural networks with a bias term, similar to those without a bias term, still exhibit out-of-
bounds reconstruction that leads to zero reconstruction loss for certain anomalous data points.
Linear autoencoders with bias terms consist of a single linear encoding layer and a single linear
decoding layer, each with an added bias term. Like for linear networks without a bias, all multi-
layer networks can be reduced to a single layer autoencoder. At the global optimum the bias terms
will recover the process of mean-centering. Note that a simplified version of this proof was presented
by Bourlard & Kamp (1988).
1
Let us now consider x̄ = m 1m X, so the vector of length n where each element contains the corre-
sponding column-wise mean of X. We will prove that the reconstruction loss LR (benc , bdec ; X, X̂)
T
for fixed Wenc , Wdec is minimized by benc = −x̄Wenc , and bdec = x̄.
First let us acknowledge that
m
1 X
x̄ = xi ,
m i=1
and thus
m
X
(xi − x̄) = 0.
i=1
We can then express the average reconstruction loss over the entire dataset as:
13
m
1 X
LR (benc , bdec ; X, X̂) = |xi − x̂i |2
mn i=1
m
1 X
= |xi − h(g(xi ))|2
mn i=1
m
1 X T
= |xi − ((xi Wenc + benc )Wdec + bdec )|2
mn i=1
m
1 X T T
= |xi − xi Wenc Wdec − benc Wdec − bdec |2
mn i=1
m
1 X T T
= |xi (1 − Wenc Wdec ) − benc Wdec − bdec |2
mn i=1
m
1 X T T T
= |(xi − x̄)(1 − Wenc Wdec ) + (x̄ − x̄Wenc Wdec ) − benc Wdec − bdec |2
mn i=1
m
1 X T
= |(xi − x̄)(1 − Wenc Wdec )|2
mn i=1
m
1 X T T
+ |(x̄ − x̄Wenc Wdec ) − benc Wdec − bdec |2
mn i=1
m
1 X T T T
+ ((xi − x̄)(1 − Wenc Wdec ))((x̄ − x̄Wenc Wdec ) − benc Wdec − bdec )
mn i=1
m
1 X T T
+ ((x̄ − x̄Wenc Wdec ) − benc Wdec − bdec )((xi − x̄)(1 − Wenc Wdec ))
mn i=1
m
1 X T
= |(xi − x̄)(1 − Wenc Wdec )|2
mn i=1
m
1 X T T
+ |(x̄ − x̄Wenc Wdec ) − benc Wdec − bdec |2 .
mn i=1
Notice thatPthe left term is constant with respect to bdec and benc , and the right term is minimized
1 m T T 2
when mn i=1 |(x̄ − x̄Wenc Wdec ) − benc Wdec − bdec | = 0. If we now substitute benc = −x̄Wenc ,
and bdec = x̄:
m
1 X T T
|(x̄ − x̄Wenc Wdec ) − benc Wdec − bdec |2 =
mn i=1
m
1 X T T
|(x̄ − x̄Wenc Wdec ) + x̄Wenc Wdec − x̄|2 = 0
mn i=1
thereby showing that the optimal solution for the biases indeed recovers the process of mean center-
ing.
Also note that the other term now mimics the reconstruction loss on the mean-centered data. This
means that we find Vd by performing PCA not on X, but on (X − X̄). Again, we can use the same
strategy a = cVdT to find adversarial anomalies.
14