@inProceedings{zhou-etal-2023-finer-325541,
title = {The Finer They Get: Combining Fine-Tuned Models For Better Semantic Change Detection},
abstract = {In this work we investigate the hypothesis that enriching contextualized models using fine-tuning tasks can improve their
capacity to detect lexical semantic change (LSC). We include tasks aimed to capture both low-level linguistic information like part-of-speech tagging, as well as higher level (semantic) information.
Through a series of analyses we demonstrate that certain combinations of fine-tuning tasks, like sentiment, syntactic information, and logical inference, bring large improvements to standard LSC models that are based only on standard language modeling. We test on the binary classification and ranking tasks of SemEval-2020 Task 1 and evaluate using both permutation tests and under transfer-learning scenarios.},
booktitle = {24th Nordic Conference on Computational Linguistics (NoDaLiDa)},
author = {Zhou, Wei and Tahmasebi, Nina and Dubossarsky, Haim},
year = {2023},
publisher = {Linköping University Electronic Press},
ISBN = {978-99-1621-999-7},
}
@incollection{tahmasebi-dubossarsky-2023-computational-325543,
title = {Computational modeling of semantic change},
abstract = {In this chapter we provide an overview of computational modeling for semantic change using large and semi-large textual corpora. We aim to provide a key for the interpretation of relevant methods and evaluation techniques, and also provide insights into important aspects of the computational study of semantic change. We discuss the pros and cons of different classes of models with respect to the properties of the data from which one wishes to model semantic change, and which avenues are available to evaluate the results. This chapter is forthcoming as the book has not yet been published. },
booktitle = {Routledge Handbook of Historical Linguistics, 2nd edition},
author = {Tahmasebi, Nina and Dubossarsky, Haim},
year = {2023},
publisher = {Routledge},
}