Please use this identifier to cite or link to this item: https://doi.org/10.21256/zhaw-30386
Publication type: Conference paper
Type of review: Peer review (publication)
Title: Divide et impera : multi-transformer architectures for complex NLP-tasks
Authors: Helland, Solveig
Gavagnin, Elena
de Spindler, Alexandre
et. al: No
DOI: 10.21256/zhaw-30386
Proceedings: Proceedings of the 8th edition of the Swiss Text Analytics Conference
Page(s): 70
Pages to: 75
Conference details: 8th Swiss Text Analytics Conference – SwissText 2023, Neuchâtel, Switzerland, 12-14 June 2023
Issue Date: 2023
Publisher / Ed. Institution: Association for Computational Linguistics
Language: English
Subject (DDC): 410.285: Computational linguistics
Abstract: The growing capabilities of transformer models pave the way for solving increasingly complex NLP tasks. A key to supporting applicationspecific requirements is the ability to fine-tune. However, compiling a fine-tuning dataset tailored to complex tasks is tedious and results in large datasets, limiting the ability to control transformer output. We present an approach in which complex tasks are divided into simpler subtasks. Multiple transformer models are fine-tuned to one subtask each, and lined up to accomplish the complex task. This simplifies the compilation of fine-tuning datasets and increases overall controllability. Using the example of reducing gender bias as a complex task, we demonstrate our approach and show that it performs better than using a single model.
URI: https://aclanthology.org/2023.swisstext-1.7
https://digitalcollection.zhaw.ch/handle/11475/30386
Fulltext version: Published version
License (according to publishing contract): CC BY 4.0: Attribution 4.0 International
Departement: School of Management and Law
Organisational Unit: Institute of Business Information Technology (IWI)
Appears in collections:Publikationen School of Management and Law

Files in This Item:
File Description SizeFormat 
2023_Helland-etal_Multi-transformer-architectures-for-complex-NLP-tasks.pdf157.83 kBAdobe PDFThumbnail
View/Open
Show full item record
Helland, S., Gavagnin, E., & de Spindler, A. (2023). Divide et impera : multi-transformer architectures for complex NLP-tasks [Conference paper]. Proceedings of the 8th Edition of the Swiss Text Analytics Conference, 70–75. https://doi.org/10.21256/zhaw-30386
Helland, S., Gavagnin, E. and de Spindler, A. (2023) ‘Divide et impera : multi-transformer architectures for complex NLP-tasks’, in Proceedings of the 8th edition of the Swiss Text Analytics Conference. Association for Computational Linguistics, pp. 70–75. Available at: https://doi.org/10.21256/zhaw-30386.
S. Helland, E. Gavagnin, and A. de Spindler, “Divide et impera : multi-transformer architectures for complex NLP-tasks,” in Proceedings of the 8th edition of the Swiss Text Analytics Conference, 2023, pp. 70–75. doi: 10.21256/zhaw-30386.
HELLAND, Solveig, Elena GAVAGNIN und Alexandre DE SPINDLER, 2023. Divide et impera : multi-transformer architectures for complex NLP-tasks. In: Proceedings of the 8th edition of the Swiss Text Analytics Conference [online]. Conference paper. Association for Computational Linguistics. 2023. S. 70–75. Verfügbar unter: https://aclanthology.org/2023.swisstext-1.7
Helland, Solveig, Elena Gavagnin, and Alexandre de Spindler. 2023. “Divide et Impera : Multi-Transformer Architectures for Complex NLP-Tasks.” Conference paper. In Proceedings of the 8th Edition of the Swiss Text Analytics Conference, 70–75. Association for Computational Linguistics. https://doi.org/10.21256/zhaw-30386.
Helland, Solveig, et al. “Divide et Impera : Multi-Transformer Architectures for Complex NLP-Tasks.” Proceedings of the 8th Edition of the Swiss Text Analytics Conference, Association for Computational Linguistics, 2023, pp. 70–75, https://doi.org/10.21256/zhaw-30386.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.