Abstract
Machine Learning and Natural Language Processing are playing an increasingly vital role in many different areas, including cybersecurity in Information Technology and Operational Technology networking, with many associated research challenges. In this paper, we introduce a new language model based on transformers with the addition of syntactical information into the embedding process. We show that our proposed Structurally Enriched Transformer (SET) language model outperforms baseline datasets on a number of downstream tasks from the GLUE benchmark. Our model improved CoLA classification by 11 points over the BERT-Base model. The performance of attention-based models has been demonstrated to be significantly better than that of traditional algorithms in several NLP tasks. Transformers are comprised of multi attention heads stacked on top of each others. A Transformer is capable of generating abstract representations of tokens input to an encoder based on their relationship to all tokens in a sequence. Despite the fact that such models can learn syntactic features based on examples alone, researchers have found that explicitly feeding this information to deep learning models can significantly boost their performance. A complex model like transformers may benefit from leveraging syntactic information such as part of speech (POS).
| Original language | English |
|---|---|
| Title of host publication | 2023 International Conference on Computing, Networking and Communications, ICNC 2023 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 19-24 |
| Number of pages | 6 |
| ISBN (Electronic) | 9781665457194 |
| DOIs | |
| State | Published - 2023 |
| Event | 2023 International Conference on Computing, Networking and Communications, ICNC 2023 - Honolulu, United States Duration: Feb 20 2023 → Feb 22 2023 |
Publication series
| Name | 2023 International Conference on Computing, Networking and Communications, ICNC 2023 |
|---|
Conference
| Conference | 2023 International Conference on Computing, Networking and Communications, ICNC 2023 |
|---|---|
| Country/Territory | United States |
| City | Honolulu |
| Period | 02/20/23 → 02/22/23 |
Funding
This manuscript has been authored by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The publisher acknowledges the US government license to provide public access under the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).
Keywords
- BERT
- Grammar Enriched
- Natural Language Processing
- Part of Speech
- Transfer Learning
- Transformers