The intentional spread of misinformation can have serious consequences in our society. This motivates us to address the task of identifying sock puppet accounts (i.e. fabricated online personas) created by individuals or organizations with the intention of deceiving their target audience. By approaching the problem as authorship attribution, we develop a sock puppet detection framework that relies solely on the texts posted by sock puppets without using any meta-information. We employed a large pre-trained language model and used interpretability methods to extract linguistic cues to answer our research question. We curated a high-quality sock puppet dataset to enable a comprehensive study for the research communities where existing datasets may include false positives due to their approximate approaches in identifying sock puppets. This dataset enables us to study the research question of whether authorship attribution works on sock puppets’ written texts. Our experiment shows that our method remarkably outperformed human annotators by 31.8% points. Furthermore, we employed an interpretability method, Integrated Gradient, to extract linguistic cues from our model for explanations of why an account is a sock puppet or not.