Submitted:
20 September 2023
Posted:
21 September 2023
You are already at the latest version
Abstract
Keywords:
1. Introduction
- We construct a small-scale Chinese dataset specifically focused on Tibet’s railway traffic texts, which has been manually annotated to ensure its suitability for training relation extraction models.
- In this paper, we propose a relation extraction model (SFRARE) that enhances subject features and relational attention. The model relies on Long Short Term Memory (LSTM) to enhance subject features and mitigate the effects of error propagation. Furthermore, we incorporate a multi-head attention mechanism to enable the fusion of relation-focused attention with the text vector. This allows the model to prioritize relations that are closely related to the subject.
- We evaluated SFRARE on the Tibet Railway Traffic text dataset and achieved a F1-score of 93.3%, which is 0.8% higher than the baseline model CasRel. To further assess the generalization ability of our model, we evaluated SFRARE on two widely-used English public datasets, namely NYT and WebNLG. The experimental results demonstrated that SFRARE outperforms mainstream models in terms of extraction performance.
2. Related Work
3. The SFRARE Model
3.1. Relational triplet representation
3.2. Overall framework of SFRARE model
3.3. Text Embedding Layer
3.4. Subject Recognizer
3.5. Subject Features Enhancement Module
3.6. Relational Attention Enhancement Module
3.7. Relation-object Recognizer
3.8. Loss Function
4. Experiments
4.1. Tibet Railway Traffic text Dataset
| Dataset | Train | Valid | Test | ALL |
|---|---|---|---|---|
| Tibet Railway Traffic text Dataset | 1645 | 205 | 206 | 2056 |
| Relation Type | The number of triplets | Relation Type | The number of triplets |
|---|---|---|---|
| railway station-be located in-location | 205 | railway station-cover an area of-area | 163 |
| railway-alternative name-name | 23 | railway-construction date-date | 35 |
| railway-finish date-date | 63 | railway-opening date-date | 32 |
| railway-length of railway-length | 22 | railway-width of railway-width | 17 |
| railway-height of railway-elevation | 26 | railway-belong to-railway | 46 |
| train-departure station-railway station | 54 | train-terminal station-railway station | 54 |
| train-departure time-time | 54 | train-journey time-time | 54 |
| constructors-construct-railway | 468 | passenger-take the train-train | 762 |
| The number of all triplets | 2078 | ||
4.2. Public Datasets
| Category | NYT | WebNLG | ||||
|---|---|---|---|---|---|---|
| Train | Valid | Test | Train | Valid | Test | |
| All | 56195 | 4999 | 5000 | 5019 | 500 | 703 |
4.3. Experimental setup
| Items | Specific setting |
|---|---|
| Word embedding layer | bert_base&bert_chinese |
| Recognition threshold | 0.5 |
| Hidden layer dimension of LSTM | 768 |
| Number of heads of attention mechanism | num-rels |
| Learning rate | 1e-5 |
| Batch-size | 16 |
| Dropout | 0.2 |
| Experimental device | NVIDIA GeForce RTX 3090(24G) |
4.4. Experimental result
4.4.1. Experimental results on Tibet Railway Traffic text Dataset
- CasRel : A novel cascade binary tagging framework for relational triplet extraction proposed by Wei et al.
| Model | Tibet Railway Traffic text Dataset | ||
|---|---|---|---|
| Pre | Rec | F1 | |
| CasRel | 93.1% | 91.9% | 92.5% |
| SFRARE | 93.7% | 93.0% | 93.3% |
4.4.2. Experimental results on public datasets
- CopyR[20] : An end-to-end relation extraction model with replication mechanism proposed by Zeng et al., a single decoder is used at the decoding layer.
- CopyR : An end-to-end relational extraction model with replication mechanism proposed by Zeng et al., multiple decoders are used at the decoding layer.
- GraphRel[21] : The first phase of relational extraction model based on relational graph structure proposed by Tsu-Jui Fu et al.
- GraphRel : The second phase of relational extraction model based on relational graph structure proposed by Tsu-Jui Fu et al.
- CopyR[22] : Zeng et al. proposed an end-to-end relation extraction model based on CopyR that applies reinforcement learning to the generation of relational triplets.
- CasRel : A novel cascade binary tagging framework for relational triplet extraction proposed by Wei et al.
| Model | NYT | WebNLG | ||||
|---|---|---|---|---|---|---|
| Pre | Rec | F1 | Pre | Rec | F1 | |
| CopyR | 59.4% | 53.1% | 56.0% | 32.2% | 28.9% | 30.5% |
| CopyR | 61.0% | 56.6% | 58.7% | 37.7% | 36.4% | 37.1% |
| GraphRel | 62.9% | 57.3% | 60.0% | 42.3% | 39.2% | 40.7% |
| GraphRel | 63.9% | 60.0% | 61.9% | 44.7% | 41.1% | 42.9% |
| CopyR | 77.9% | 67.2% | 72.1% | 63.3% | 59.9% | 61.6% |
| CasRel | 89.7% | 89.5% | 89.6% | 93.4% | 90.1% | 91.8% |
| SFRARE | 90.5% | 91.7% | 91.1% | 92.6% | 92.7% | 92.6% |
4.4.3. Ablation experiment
| Methods | NYT | WebNLG | ||||
|---|---|---|---|---|---|---|
| Pre | Rec | F1 | Pre | Rec | F1 | |
| CasRel | 89.7% | 89.5% | 89.6% | 93.4% | 90.1% | 91.8% |
| SFRARE | 90.1% | 89.8% | 89.9% | 92.2% | 91.8% | 92.0% |
| SFRARE | 89.9% | 91.0% | 90.5% | 93.7% | 90.7% | 92.2% |
| SFRARE | 90.5% | 91.7% | 91.1% | 92.6% | 92.7% | 92.6% |
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Miller, S.; Fox, H.; Ramshaw, L. ; Weischedel, R A Novel Use of Statistical Parsing to Extract Information from Text. In Proceedings of the 6th Applied Natural Language Processing Conference, Seattle, Washington, USA, 29 April–4 May 2000; pp. 226–233. [Google Scholar]
- Chen, Y.; Zheng, D.; Zhao, T. Chinese relation extraction based on Deep Belief Nets. Journal of Software 2012, 23, 2572–2585. [Google Scholar] [CrossRef]
- E, H.; Zhang, W.; Xiao, S.; Cheng, R.; Hu, Y.; Zhou, X.; Niu, P. Joint entity relationship extraction based on deep learning. Journal of Software 2019, 30, 1793–1818. [Google Scholar]
- Zhang, Y. A historical survey of the construction of Xizang Railway. China’s Borderland History and Geography Studies 2015, 3, 32–43. [Google Scholar]
- Rong, Z. The role of Qinghai-Xizang Railway in the evolution of temporal and spatial structure of traffic in Xizang. China Tibetology 2016, 2, 62–71. [Google Scholar]
- Zhang, S.; Wang, X.; Chen, Z.; Wang, L.; Xu, D.; Jia, Y. Survey of Supervised Joint Entity Relation Extraction Methods. Journal of Frontiers of Computer Science and Technology 2022, 16, 713–733. [Google Scholar]
- Miwa, M.; Bansal, M. End-to-end relation extraction using LSTMs on sequences and tree structures. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; pp. 1105–1116. [Google Scholar]
- Tan, Z.; Zhao, X.; Wang, W.; Xiao, W. Jointly extracting multiple triplets with multilayer translation constraints. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence, Honolulu, Hawaii, USA, 27 January–1 February 2019; pp. 7080–7087. [Google Scholar]
- Liu, J.; Chen, S.; Wang, B.; Zhang, J.; Xu, T. Attention as relation:learning supervised multi-head self-attention for relation extraction. In Proceedings of the 29th International Joint Conference on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 3787–3793. [Google Scholar]
- Sun, K.; Zhang, R.; Mensah, S.; Mao, Y.; Liu, X. Recurrent interaction network for jointly extracting entities and classifying relations. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, Online, 16–20 November 2020; pp. 3722–3732. [Google Scholar]
- Wei, Z.; Su, j.; Wang, Y.; Tian, Y.; Chang, Y. A Novel Cascade Binary Tagging Framework for Relational Triple Extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Seattle, Washington, USA, 5–10 July 2020; pp. 1476–1488. [Google Scholar]
- Wang, Y.; Yu, B.; Zhang, Y.; Liu, T.; Zhu, H.; Sun, L. TPLinker: Single-stage Joint Extraction of Entities and Relations Through Token Pair Linking. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain (Online), 8–13 December 2020; pp. 1572–1582. [Google Scholar]
- Wang, J.; Lu, W. Two are better than one: joint entity and relation extraction with table-sequence encoders. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, Online, 16–20 November 2020; pp. 1706–1721. [Google Scholar]
- Zheng, S.; Wang, F.; Bao, H.; Hao, Y.; Zhou, P.; Xu, B. Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, Canada, 30 July-4 August 2017; pp. 1227–1236. [Google Scholar]
- Dai, D.; Xiao, X.; Lyu, Y.; Dou, S.; Wang, H. Joint extraction of entities and overlapping relations using position-attentive sequence labeling. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence, Honolulu, Hawaii, USA, 27 January-1 February 2019; pp. 6300–6308. [Google Scholar]
- Sui, D.; Zeng, X.; Chen, Y.; Liu, K.; Zhao, J. Joint Entity and Relation Extraction With Set Prediction Networks. IEEE Transactions on Neural Networks and Learning Systems 2020, PP. [Google Scholar] [CrossRef] [PubMed]
- Vaswani, A.; Shazeer, N.; Parmer, N.; Uszkoreit, J.; Jones, L.; Gomez, N.A.; Kaiser, L.; Polosukhin, l. Attention Is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, California, USA, 4-9 December 2017; pp. 5998–6008. [Google Scholar]
- Jacob, D.; Chang, M.W.; Kenton, L.; Kristina, T. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies Volume 1 (Long and Short Papers), Minneapolis, USA, 2-7 June 2019; pp. 4171–4186. [Google Scholar]
- Ren, X.; Wu, Z.; He, W.; Meng, Q.; Voss, R.C.; Ji, H.; Abdelzaher, F.T.; Han, J. CoType: joint extraction of typed entities and relations with knowledge bases. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3-7 April 2017; pp. 1015–1024. [Google Scholar]
- Zeng, X.; Zeng, D.; He, S.; Liu, K.; Zhao, J. Extracting relational facts by an end-to-end neural model with copy mechanism. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume1: Long Papers), Melbourne, Australia, 15-20 July 2018; pp. 506–514. [Google Scholar]
- Fu, T.J.; Li, P.H.; Ma, W.Y. Modeling text as relational graphs for joint entity and relation extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 30 July-4 August; pp. 1409–1418.
- Zeng, X.; He, S.; Zeng, D.; Liu, K.; Liu, S.; Zhao, J. Learning the extraction order of multiple relational facts in a sentence with reinforcement learning. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3-7 November 2019; pp. 367–377. [Google Scholar]



Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).