Morphological skip-gram : replacing fast text characters n-gram with morphological knowledge
<?xml version="1.0" encoding="UTF-8"?><modsCollection xmlns="http://www.loc.gov/mods/v3" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.loc.gov/mods/v3 http://www.loc.gov/standards/mods/v3/mods-3-8.xsd">
<mods version="3.8">
<titleInfo>
<title>Morphological skip-gram</title>
<subTitle>: replacing fast text characters n-gram with morphological knowledge</subTitle>
</titleInfo>
<name type="personal" usage="primary" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20210005589">
<namePart>Santos, Flavio Arthur O.</namePart>
<nameIdentifier>MAPA20210005589</nameIdentifier>
</name>
<typeOfResource>text</typeOfResource>
<genre authority="marcgt">periodical</genre>
<originInfo>
<place>
<placeTerm type="code" authority="marccountry">esp</placeTerm>
</place>
<dateIssued encoding="marc">2021</dateIssued>
<issuance>serial</issuance>
</originInfo>
<language>
<languageTerm type="code" authority="iso639-2b">spa</languageTerm>
</language>
<physicalDescription>
<form authority="marcform">print</form>
<internetMediaType>application/pdf</internetMediaType>
</physicalDescription>
<abstract displayLabel="Summary">Natural language processing systems have attracted much interest of the industry. This branch of study is composed of some applications such as machine translation, sentiment analysis, named entity recognition, question and answer, and others. Word embeddings (i.e., continuous word representations) are an essential module for those applications generally used as word representation to machine learning models. Some popular methods to train word embeddings are GloVe and Word2Vec. They achieve good word representations, despite limitations: both ignore morphological information of the words and consider only one representation vector for each word. This approach implies the word embeddings does not consider dierent word contexts properly and are unaware of its inner structure. To mitigate this problem, the other word embeddings method FastText represents each word as a bag of characters n-grams. Hence, a continuous vector describes each n-gram, and the nal word representation is the sum of its characters n-grams vectors. Nevertheless, the use of all n-grams character of a word is a poor approach since some n-grams have no semantic relation with their words and increase the amount of potentially useless information. This approach also increase the training phase time. In this work, we propose a new method for training word embeddings, and its goal is to replace the FastText bag of character n-grams for a bag of word morphemes through the morphological analysis of the word. Thus, words with similar context and morphemes are represented by vectors close to each other. To evaluate our new approach, we performed intrinsic evaluations considering 15 dierent tasks, and the results show a competitive performance compared to FastText. Moreover, the proposed model is 40% faster than FastText in the training phase. We also outperform the baseline approaches in extrinsic evaluations through Hate speech detection and NER tasks using dierent scenarios </abstract>
<note type="statement of responsibility">Flavio Arthur O. Santos</note>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080611200">
<topic>Inteligencia artificial</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080617479">
<topic>Lenguajes de programación</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080561871">
<topic>Conocimiento</topic>
</subject>
<classification authority="">922.134</classification>
<relatedItem type="host">
<titleInfo>
<title>Revista Iberoamericana de Inteligencia Artificial</title>
</titleInfo>
<originInfo>
<publisher>IBERAMIA, Sociedad Iberoamericana de Inteligencia Artificial , 2018-</publisher>
</originInfo>
<identifier type="issn">1988-3064</identifier>
<identifier type="local">MAP20200034445</identifier>
<part>
<text>15/02/2021 Volumen 24 Número 67 - febrero 2021 , p. 1-17</text>
</part>
</relatedItem>
<recordInfo>
<recordContentSource authority="marcorg">MAP</recordContentSource>
<recordCreationDate encoding="marc">210413</recordCreationDate>
<recordChangeDate encoding="iso8601">20220911190026.0</recordChangeDate>
<recordIdentifier source="MAP">MAP20210011719</recordIdentifier>
<languageOfCataloging>
<languageTerm type="code" authority="iso639-2b">spa</languageTerm>
</languageOfCataloging>
</recordInfo>
</mods>
</modsCollection>