I. Introduction
In recent years, the number of applications adopting mi-croservice architectures in cloud datacenters has proliferated rapidly [1]–[3]. Compared to monolithic architectures where all components of an application are tightly coupled, microser-vice architectures decompose applications into loosely coupled services. This enables advantages like cross-team development and friendly deployment. Microservices interconnect via re-mote procedure calls (RPCs), HTTP, or other methods. The time that a microservice takes to execute an RPC or HTTP request, noted as request latency, directly impacts application performance, and thus becomes the most critical performance metric for datacenter operators. When micro service request latency spikes, operators must leverage application and system performance data to pinpoint anomalies.