The CMS experiment at the CERN LHC developed the workflow management archive system to persistently store unstructured framework job report documents produced by distributed workflow management agents. In this paper, we present its architecture, implementation, deployment, and integration with the CMS and CERN computing infrastructures, such as central HDFS and Hadoop Spark cluster. The system leverages modern technologies such as a document-oriented database and the Hadoop ecosystem to provide the necessary flexibility to reliably process, store, and aggregate O(1M) documents on a daily basis. We describe the data transformation, the short-and long-term storage layers, and the query language, along with the aggregation pipeline developed to visualize various performance metrics, to assist CMS data operators in assessing the performance of the CMS computing system.