Transformer models have grown in popularity in recent years.
This is primarily due to the self-attention mechanism in their architecture and the ability for parallelization.
This paper proposes experiments for text summarisation using the WikiHow dataset.