Trust-region methods have received massive attention in a variety of continuous optimization. They aim to obtain a trial step by minimizing a quadratic model in a region of a certain trust-region radius around the current iterate. This paper proposes an adaptive Riemannian trust-region algorithm for optimization on manifolds, in which the trust-region radius depends linearly on the norm of the Riemannian gradient at each iteration. Under mild assumptions, we establish the liminf-type convergence, lim-type convergence, and global convergence results of the proposed algorithm. In addition, the proposed algorithm is shown to reach the conclusion that the norm of the Riemannian gradient is smaller than ϵO(1ϵ2) within ϵO(1ϵ2) iterations. Some numerical examples of tensor approximations are carried out to reveal the performances of the proposed algorithm compared to the classical Riemannian trust-region algorithm.