We establish sharp concentration inequalities for sums of dependent random matrices. Our results concern two models. First, a model where summands are generated by a $\psi$-mixing Markov chain. Second, a model where summands are expressed as deterministic matrices multiplied by scalar random variables. In both models, the leading-order term is provided by free probability theory. This leading-order term is often asymptotically sharp and, in particular, does not suffer from the logarithmic dimensional dependence which is present in previous results such as the matrix Khintchine inequality. A key challenge in the proof is that techniques based on classical cumulants, which can be used in a setting with independent summands, fail to produce efficient estimates in the Markovian model. Our approach is instead based on Boolean cumulants and a change-of-measure argument. We discuss applications concerning community detection in Markov chains, random matrices with heavy-tailed entries, and the analysis of random graphs with dependent edges.
Comment: 69 pages, 4 figures