Impact of Grammar on Language Model Comprehension

  • Kimia Ameri
  • , Michael Hempel
  • , Hamid Sharif
  • , Juan Lopez
  • , Kalyan Perumalla

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Machine Learning and Natural Language Processing are playing an increasingly vital role in many different areas, including cybersecurity in Information Technology and Operational Technology networking, with many associated research challenges. In this paper, we introduce a new language model based on transformers with the addition of syntactical information into the embedding process. We show that our proposed Structurally Enriched Transformer (SET) language model outperforms baseline datasets on a number of downstream tasks from the GLUE benchmark. Our model improved CoLA classification by 11 points over the BERT-Base model. The performance of attention-based models has been demonstrated to be significantly better than that of traditional algorithms in several NLP tasks. Transformers are comprised of multi attention heads stacked on top of each others. A Transformer is capable of generating abstract representations of tokens input to an encoder based on their relationship to all tokens in a sequence. Despite the fact that such models can learn syntactic features based on examples alone, researchers have found that explicitly feeding this information to deep learning models can significantly boost their performance. A complex model like transformers may benefit from leveraging syntactic information such as part of speech (POS).

Original languageEnglish
Title of host publication2023 International Conference on Computing, Networking and Communications, ICNC 2023
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages19-24
Number of pages6
ISBN (Electronic)9781665457194
DOIs
StatePublished - 2023
Event2023 International Conference on Computing, Networking and Communications, ICNC 2023 - Honolulu, United States
Duration: Feb 20 2023Feb 22 2023

Publication series

Name2023 International Conference on Computing, Networking and Communications, ICNC 2023

Conference

Conference2023 International Conference on Computing, Networking and Communications, ICNC 2023
Country/TerritoryUnited States
CityHonolulu
Period02/20/2302/22/23

Funding

This manuscript has been authored by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The publisher acknowledges the US government license to provide public access under the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).

Keywords

  • BERT
  • Grammar Enriched
  • Natural Language Processing
  • Part of Speech
  • Transfer Learning
  • Transformers

Fingerprint

Dive into the research topics of 'Impact of Grammar on Language Model Comprehension'. Together they form a unique fingerprint.

Cite this