The abuse of clickbait headlines by online news media has kept increasing, causing bad experiences for the reader and reducing online news reading engagement. Since the advances in self-attention model, model such as BERT and other variant has been considered to be the state-of-the-art method in many NLP (natural language processing) related tasks. In this experiment, we'll be focusing on one of BERT variants called RoBERTa which is an improved model over BERT. The goal of this experiment is to compare how well the currently available RoBERTa model perform in detecting clickbait In-donesia news headline. Using a total of 6632 annotated news headlines, sampled from the CLICK-ID dataset, we experimented with some of huggingface community's top voted RoBERTa and BERT Indonesia language models, and compare each of their performances in classifying Indonesia clickbait headlines. We evaluate the accuracy, precission, recall, and F1 score of each models and found that cahya/XLM-RoBERTa-large and indobenchmarkIndoBERT-p1 as our top-performing model in this experiment with a 92% accuracy. Resource and performance wise we recommend indobenchmark/IndoBERT-p1 as a more suitable model, however, XLM-RoBERTa-large also comes with its merit in terms of having a more consistent output across validation and unseen set. In this paper, we propose each of the model configurations along with the exploratory data analysis, preprocessing method on the data, and the training performance of the configuration of each model.