Text Summarization
using BERT and T5


Transformer models have grown in popularity in recent years. This is primarily due to the self-attention mechanism in their architecture and the ability for parallelization. This paper proposes experiments for text summarisation using the WikiHow dataset.

PDF of Paper