Canonical deep learning-based remaining useful life prediction relies on supervised learning methods, which in turn requires large data sets of run-to-failure data to ensure model performance. In a considerable class of cases, run-to-failure data is difficult to collect in practice as it may be expensive and unsafe to operate assets until failure. As such, there is a need to leverage data that are not run-to-failure but may still contain some measurable, and thus learnable, degradation signal. In this paper, we propose utilizing self-supervised learning as a pretraining step to learn representations of data which will enable efficient training on the downstream task of remaining useful life prediction. The self-supervised learning task chosen is time series sequence ordering, a task that involves constructing tuples each consisting of n sequences sampled from the time series and reordered with some probability p. Subsequently, a classifier is trained on the resulting binary classification task; distinguishing between correctly ordered and shuffled tuples. The classifier’s weights are then transferred to the remaining useful life prediction model and fine-tuned using run-to-failure data. To conduct our experiments, we use a data set of simulated run-to-failure turbofan jet engines. We show that the proposed self-supervised learning scheme can retain performance when training on a fraction of the full data set. In addition, we show indications that self-supervised learning as a pretraining step can enhance the performance of the model even when training on the full run-to-failure data set.