The first step in nearly all natural language processing (NLP) applications is applying preprocessing operations [1] to the text. Preprocessing operations include tokenization (segmenting the text into tokens), sentence splitting (dividing the text into sentences), normalization (converting the text into a canonical form), and the like. In this project, you will develop and implement algorithms for preprocessing of Turkish text using deep learning approaches. First, a literature review will be conducted and similar systems for English will be analyzed (e.g. UDPipe [2], Stanza [3]).