Open Access. Powered by Scholars. Published by Universities.®

Engineering Commons

Open Access. Powered by Scholars. Published by Universities.®

Electrical and Computer Engineering

TÜBİTAK

Agglutinative languages

Publication Year

Articles 1 - 2 of 2

Full-Text Articles in Engineering

Deep Learning-Based Turkish Spelling Error Detection With A Multi-Class False Positive Reduction Model, Burak Aytan, Cemal Okan Şakar May 2023

Deep Learning-Based Turkish Spelling Error Detection With A Multi-Class False Positive Reduction Model, Burak Aytan, Cemal Okan Şakar

Turkish Journal of Electrical Engineering and Computer Sciences

Spell checking and correction is an important step in the text normalization process. These tasks are more challenging in agglutinative languages such as Turkish since many words can be derived from the root word by combining many suffixes. In this study, we propose a two-step deep learning-based model for misspelled word detection in the Turkish language. A false positive reduction model is integrated into the system to reduce the false positive predictions originating from the use of foreign words and abbreviations that are commonly used in Internet sharing platforms. For this purpose, we create a multi-class dataset by developing a …


Transmorph: A Transformer Based Morphological Disambiguator For Turkish, Hi̇lal Özer, Emi̇n Erkan Korkmaz Jul 2022

Transmorph: A Transformer Based Morphological Disambiguator For Turkish, Hi̇lal Özer, Emi̇n Erkan Korkmaz

Turkish Journal of Electrical Engineering and Computer Sciences

The agglutinative nature of the Turkish language has a complex morphological structure, and there are generally more than one parse for a given word. Before further processing, morphological disambiguation is required to determine the correct morphological analysis of a word. Morphological disambiguation is one of the first and crucial steps in natural language processing since its success determines later analyses. In our proposed morphological disambiguation method, we used a transformer-based sequence-to-sequence neural network architecture. Transformers are commonly used in various NLP tasks, and they produce state-of-the-art results in machine translation. However, to the best of our knowledge, transformer-based encoder-decoders have …