@inproceedings{gao-etal-2018-hierarchical,
    title = "Hierarchical Convolutional Attention Networks for Text Classification",
    author = "Gao, Shang  and
      Ramanathan, Arvind  and
      Tourassi, Georgia",
    editor = "Augenstein, Isabelle  and
      Cao, Kris  and
      He, He  and
      Hill, Felix  and
      Gella, Spandana  and
      Kiros, Jamie  and
      Mei, Hongyuan  and
      Misra, Dipendra",
    booktitle = "Proceedings of the Third Workshop on Representation Learning for {NLP}",
    month = jul,
    year = "2018",
    address = "Melbourne, Australia",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/iwcs-25-ingestion/W18-3002/",
    doi = "10.18653/v1/W18-3002",
    pages = "11--23",
    abstract = "Recent work in machine translation has demonstrated that self-attention mechanisms can be used in place of recurrent neural networks to increase training speed without sacrificing model accuracy. We propose combining this approach with the benefits of convolutional filters and a hierarchical structure to create a document classification model that is both highly accurate and fast to train {--} we name our method Hierarchical Convolutional Attention Networks. We demonstrate the effectiveness of this architecture by surpassing the accuracy of the current state-of-the-art on several classification tasks while being twice as fast to train."
}Markdown (Informal)
[Hierarchical Convolutional Attention Networks for Text Classification](https://preview.aclanthology.org/iwcs-25-ingestion/W18-3002/) (Gao et al., RepL4NLP 2018)
ACL