APPLYING DEEP LEARNING TO EXTRACT CAUSALITY FROM TEXT USING SYNTHETIC DATA
Abstract
This article addresses the problem of developing a causal full-tuples extraction model from unstructured texts to represent decision-making situations in complex social and humanitarian environments. We present a causal full-tuples extraction model using a pre-trained BERT with additional feature-based special fine-tuning. To refine the causal classification, the model uses two types of features (verb causality and cause-and-effect quality metrics) to recognize a causal tuple, automatically extracts semantic features from sentences, increasing the accuracy of extraction. Text preprocessing is performed using the open source SpaCy library. The extracted cause-and-effect tuples in the format <cause phrase, verb phrase, effect phrase, polarity> are easily transformed into the corresponding elements of the graph <outgoing graph node, graph arc direction, incoming graph node, graph connection weight sign> and can then be used to construct a directed weighted signed graph with deterministic causality on arcs. In order to reduce dependence on external knowledge, synthetic generated annotated datasets are used to fine-tune and test the BERT model. Experimental results show that the accuracy of extracting cause-and-effect relationships on synthetic data reaches 94%, and the F1 value is 95%. The advantages of the presented technological solution are that the model does not require high operating costs, is implemented on a computer with standard characteristics, uses free software, which makes it accessible to a wide variety of users. It is expected that the proposed model can be used to automate text analysis and support decision-making in conditions of high uncertainty, which is especially important for social and humanitarian environments.
References
1. Li Z. et al. Causality extraction based on self-attentive BiLSTM-CRF with transferred embeddings,
Neurocomputing, 2021, Vol. 423, pp. 207-219.
2. Banko M., Etzioni O. The Tradeoffs Between Open and Traditional Relation Extraction, Annual
Meeting of the Association for Computational Linguistics, 2008.
3. Shao Y. et al. Extraction of causal relations based on SBEL and BERT model, Database, 2021, Vol. 2021.
4. Zhao X. et al. A Comprehensive Survey on Relation Extraction: Recent Advances and New Frontiers, 2023.
5. Bojduj B.N. Extraction of Causal-Association Networks from Unstructured Text Data. San Luis
Obispo. California: California Polytechnic State University, 2009, 61 p.
6. Alibage A. Achieving High Reliability Organizations Using Fuzzy Cognitive Maps - the Case of
Offshore Oil and Gas. Portland, OR: Portland State University, 2020, 342 p.
7. An N. et al. Extracting causal relations from the literature with word vector mapping, Comput. Biol.
Med., 2019, Vol. 115, pp. 103524.
8. Devlin J. et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,
2018.
9. Paulheim H. Knowledge graph refinement: A survey of approaches and evaluation methods, Semant.
Web, ed. Cimiano P., 2016, Vol. 8, No. 3, pp. 489-508.
10. Kulikowski C.A., Weiss S.M. Representation of Expert Knowledge for Consultation: The CASNET and
EXPERT Projects, Artificial Intelligence in Medicine. Routledge, 2019, pp. 21-55.
11. Kosko B. Hidden patterns in combined and adaptive knowledge networks, Int. J. Approx. Reason,
1988, Vol. 2, No. 4, pp. 377-393.
12. Yang J., Han S.C., Poon J. A survey on extraction of causal relations from natural language text,
Knowl. Inf. Syst., 2022, Vol. 64, No. 5, pp. 1161-1186.
13. Park J., Cardie C. Identifying Appropriate Support for Propositions in Online User Comments,
Proceedings of the First Workshop on Argumentation Mining. Stroudsburg, PA, USA: Association for
Computational Linguistics, 2014, pp. 29-38.
14. Job S. et al. Exploring Causal Learning through Graph Neural Networks: An In-depth Review, 2023.
15. Li Z. et al. CausalBERT: Injecting Causal Knowledge Into Pre-trained Models with Minimal
Supervision, ArXiv. 2021, Vol. abs/2107.0.
16. Felgueira T. et al. The Impact of Feature Causality on Normal Behaviour Models for SCADA-based
Wind Turbine Fault Detection, 2019.
17. Tselykh A., Vasilev V., Tselykh L. A Method for Modeling the Control Impact Strategy Based on the
Mental Frame of References of the Decision-Maker, 2023, pp. 315-324.
18. Tselykh A. et al. Influence control method on directed weighted signed graphs with deterministic
causality, Ann. Oper. Res., 2022, Vol. 311, No. 2, pp. 1281-1305.
19. Tselykh A., Vasilev V., Tselykh L. Assessment of influence productivity in cognitive models, Artif.
Intell. Rev., 2020.
20. Fellbaum C. WordNet / ed. Fellbaum C. The MIT Press, 1998.
21. Vasiliev Y. Natural language processing with Python and spaCy: A practical introduction. No Starch
Press, 2020.
22. Schmitt X. et al. A Replicable Comparison Study of NER Software: StanfordNLP, NLTK, OpenNLP,
SpaCy, Gate, 2019 Sixth International Conference on Social Networks Analysis, Management and
Security (SNAMS). IEEE, 2019, pp. 338-343.