Over the past years, literature has shown that attacks exploiting the microarchitecture of modern processors pose a serious threat to the privacy of mobile phone users. This is because applications leave distinct footprints in the processor, which can be used by malware to infer user activities. In this work, we show that these inference attacks are considerably more practical when combined with advanced AI techniques. In particular, we focus on profiling the activity in the last-level cache (LLC) of ARM processors. We employ a simple Prime+Probe based monitoring technique to obtain cache traces, which we classify with Deep Learning methods including Convolutional Neural Networks. We demonstrate our approach on an off-the-shelf Android phone by launching a successful attack from an unprivileged, zeropermission App in well under a minute. The App thereby detects running applications with an accuracy of 98% and reveals opened websites and streaming videos by monitoring the LLC for at most 6 seconds. This is possible, since Deep Learning compensates measurement disturbances stemming from the inherently noisy LLC monitoring and unfavorable cache characteristics such as random line replacement policies. In summary, our results show that thanks to advanced AI techniques, inference attacks are becoming alarmingly easy to implement and execute in practice. This once more calls for countermeasures that confine microarchitectural leakage and protect mobile phone applications, especially those valuing the privacy of their users.CNNs are a so-called Deep Learning technique. They have received broad interest in the field of supervised learning, especially for complex tasks such as image and language classification [8,28]. A typical CNN is made up of neurons, which are interconnected and grouped into layers. Each neuron computes a weighted sum of its inputs using a (non-) linear activation function. Those inputs stem from the actual inputs to the network or from previous layers. A typical CNN comprises multiple layers of neurons, as shown in Figure 1:Convolutional Layer. This layer is the core block of the CNN. It consists of learnable filters that are slid over the width and height of the input to learn any two-dimensional patterns. The activation functions in the neurons thereby extract important features from each input. For efficiency, the neurons are connected only to a local region of the input.