Cross-lingual Embeddings Reveal Universal and Lineage ...

11
Proceedings of the 24th Conference on Computational Natural Language Learning, pages 265–275 Online, November 19-20, 2020. c 2020 Association for Computational Linguistics https://doi.org/10.18653/v1/P17 265 Cross-lingual Embeddings Reveal Universal and Lineage-Specific Patterns in Grammatical Gender Assignment Hartger Veeman Uppsala University Department of linguistics and philology Box 635, 75126 Uppsala [email protected] Aleksandrs Berdicevskis University of Gothenburg Spr˚ akbanken Box 200, 40530 Gothenburg [email protected] Marc Allassonni` ere-Tang University Lyon 2 Lab Dynamics of Language 14 avenue Berthelot 69363 Lyon [email protected] Ali Basirat Uppsala University Department of linguistics and philology Box 635, 75126 Uppsala [email protected] Abstract Grammatical gender is assigned to nouns dif- ferently in different languages. Are all fac- tors that influence gender assignment idiosyn- cratic to languages or are there any that are uni- versal? Using cross-lingual aligned word em- beddings, we perform two experiments to ad- dress these questions about language typology and human cognition. In both experiments, we predict the gender of nouns in language X using a classifier trained on the nouns of lan- guage Y, and take the classifier’s accuracy as a measure of transferability of gender systems. First, we show that for 22 Indo-European lan- guages the transferability decreases as the phy- logenetic distance increases. This correlation supports the claim that some gender assign- ment factors are idiosyncratic, and as the lan- guages diverge, the proportion of shared in- herited idiosyncrasies diminishes. Second, we show that when the classifier is trained on two Afro-Asiatic languages and tested on the same 22 Indo-European languages (or vice versa), its performance is still significantly above the chance baseline, thus showing that universal factors exist and, moreover, can be captured by word embeddings. When the classifier is tested across families and on inanimate nouns only, the performance is still above baseline, indicating that the universal factors are not lim- ited to biological sex. 1 Grammatical gender assignment Grammatical gender is one of the nominal classifi- cation systems found in natural languages (Seifart, 2010). In languages with grammatical gender, cer- tain words agree in a specific form with the noun they modify depending on the gender of the mod- ified noun (Corbett, 1991, 2001). For instance, Swedish has a binary gender system with the com- mon/neuter values, in which the articles and ad- jectives must have grammatical gender agreement with the noun they are modifying, c.f., ett stor-t ¨ apple (SG. NEUT big-SG. NEUT apple. SG. NEUT) ‘a big apple’ and en stor-h ¨ ast (a. SG. UTER big- SG. UTER horse. SG. UTER) ‘a big horse’. The most common gender distinctions are mas- culine/feminine (e.g., in French and Italian), mascu- line/feminine/neuter (e.g., in Russian and German), and common/neuter (e.g., in Swedish and Danish) (Corbett, 2013a). 1 Within these distinctions, gram- matical gender does not necessarily fully agree with the biological sex. By way of illustration, the German word for ‘girl’, M ¨ adchen, is a neuter noun. Moreover, nouns with the same meaning may belong to different grammatical gender in dif- ferent languages. For example, the German noun for ‘sun’, Sonne, is feminine, but its French equiva- lent, soleil, is masculine. Based on these observations, several questions have been developed in the literature. First of all, what are the main factors that influence gen- der assignment in individual languages? Sec- ond, are there any principles that are shared cross- linguistically or are they all language- and/or culture-specific? With regard to the first question, two main factors have been identified in the liter- ature: formal features of the noun and its mean- ing (Corbett and Fraser, 2000; Rice, 2006; Corbett, 2013b; Fedden and Corbett, 2019). With regard to the second question, it is generally believed that 1 More complex distinctions are also found. As an example, Swahili has a more complex system with 18 classes. These systems are generally referred to as ‘noun classes’ in the liter- ature and are not covered by the term ‘grammatical gender’ in the current paper.

Transcript of Cross-lingual Embeddings Reveal Universal and Lineage ...

Page 1: Cross-lingual Embeddings Reveal Universal and Lineage ...

Proceedings of the 24th Conference on Computational Natural Language Learning, pages 265–275Online, November 19-20, 2020. c©2020 Association for Computational Linguistics

https://doi.org/10.18653/v1/P17

265

Cross-lingual Embeddings Reveal Universal and Lineage-SpecificPatterns in Grammatical Gender AssignmentHartger Veeman

Uppsala UniversityDepartment of linguistics and philology

Box 635, 75126 [email protected]

Aleksandrs BerdicevskisUniversity of Gothenburg

SprakbankenBox 200, 40530 Gothenburg

[email protected]

Marc Allassonniere-TangUniversity Lyon 2

Lab Dynamics of Language14 avenue Berthelot 69363 [email protected]

Ali BasiratUppsala University

Department of linguistics and philologyBox 635, 75126 Uppsala

[email protected]

Abstract

Grammatical gender is assigned to nouns dif-ferently in different languages. Are all fac-tors that influence gender assignment idiosyn-cratic to languages or are there any that are uni-versal? Using cross-lingual aligned word em-beddings, we perform two experiments to ad-dress these questions about language typologyand human cognition. In both experiments,we predict the gender of nouns in language Xusing a classifier trained on the nouns of lan-guage Y, and take the classifier’s accuracy asa measure of transferability of gender systems.First, we show that for 22 Indo-European lan-guages the transferability decreases as the phy-logenetic distance increases. This correlationsupports the claim that some gender assign-ment factors are idiosyncratic, and as the lan-guages diverge, the proportion of shared in-herited idiosyncrasies diminishes. Second, weshow that when the classifier is trained on twoAfro-Asiatic languages and tested on the same22 Indo-European languages (or vice versa),its performance is still significantly above thechance baseline, thus showing that universalfactors exist and, moreover, can be capturedby word embeddings. When the classifier istested across families and on inanimate nounsonly, the performance is still above baseline,indicating that the universal factors are not lim-ited to biological sex.

1 Grammatical gender assignment

Grammatical gender is one of the nominal classifi-cation systems found in natural languages (Seifart,2010). In languages with grammatical gender, cer-tain words agree in a specific form with the nounthey modify depending on the gender of the mod-ified noun (Corbett, 1991, 2001). For instance,

Swedish has a binary gender system with the com-mon/neuter values, in which the articles and ad-jectives must have grammatical gender agreementwith the noun they are modifying, c.f., ett stor-tapple (SG.NEUT big-SG.NEUT apple.SG.NEUT) ‘abig apple’ and en stor-∅ hast (a.SG.UTER big-SG.UTER horse.SG.UTER) ‘a big horse’.

The most common gender distinctions are mas-culine/feminine (e.g., in French and Italian), mascu-line/feminine/neuter (e.g., in Russian and German),and common/neuter (e.g., in Swedish and Danish)(Corbett, 2013a).1 Within these distinctions, gram-matical gender does not necessarily fully agreewith the biological sex. By way of illustration,the German word for ‘girl’, Madchen, is a neuternoun. Moreover, nouns with the same meaningmay belong to different grammatical gender in dif-ferent languages. For example, the German nounfor ‘sun’, Sonne, is feminine, but its French equiva-lent, soleil, is masculine.

Based on these observations, several questionshave been developed in the literature. First ofall, what are the main factors that influence gen-der assignment in individual languages? Sec-ond, are there any principles that are shared cross-linguistically or are they all language- and/orculture-specific? With regard to the first question,two main factors have been identified in the liter-ature: formal features of the noun and its mean-ing (Corbett and Fraser, 2000; Rice, 2006; Corbett,2013b; Fedden and Corbett, 2019). With regard tothe second question, it is generally believed that

1More complex distinctions are also found. As an example,Swahili has a more complex system with 18 classes. Thesesystems are generally referred to as ‘noun classes’ in the liter-ature and are not covered by the term ‘grammatical gender’ inthe current paper.

Page 2: Cross-lingual Embeddings Reveal Universal and Lineage ...

266

gender assignment is based on a mix of sharedcognitive principles (Kemmerer, 2017) and linguis-tic/cultural idiosyncrasies (Takamura et al., 2016;Di Garbo et al., 2019).

One of the most common way to answer the sec-ond question is to measure the transferability ofgender across languages. If the gender assignmentrules of a language can be easily used to predictthe gender of nouns in another language, it showsthat the principles of gender assignment are partlyshared and transferable between the two languages.If such a transfer is possible within most languages,one can then assume that gender assignment is toa large extent based on universal patterns. Mostempirical studies that adopted this approach fol-lowed the perspective of language acquisition andanalyzed how native speakers of language X couldpredict the gender of a selected amount of nounsfrom language Y (Sabourin et al., 2006; Jarvis andPavlenko, 2010, p.132-136). No studies known tothe authors investigated the transferability of gram-matical gender for a large amount of nouns froma large sample of languages by using natural lan-guage processing methods, which is the gap weaim at filling.

We use a transfer learning setting to measurethe transferability of grammatical gender acrosslanguages. In this setting, a neural classificationmodel is trained to predict the grammatical genderof nouns in a source language. This model is thenapplied to a set of test nouns in a target language topredict their grammatical gender. The classifier’sability to classify the test nouns is interpreted as anindication of the transferability of grammatical gen-der system from the source language to the targetlanguage (i.e., the higher the accuracy is, the moretransferable the gender systems are). The entiresetting is founded on the cross-lingual represen-tation of words, providing for knowledge transferbetween the gender classification models acrosslanguages. The embeddings are used to representnouns in both source and target languages. The useof word embeddings for the study of grammaticalgender is based on the premise that they can capturelinguistically-motivated information about words(Nastase and Popescu, 2009; Andreas and Klein,2014; Artetxe et al., 2018; Basirat and Tang, 2019;Williams et al., 2019), including information aboutthe gender of nouns within a language (Basiratand Tang, 2019; Williams et al., 2019; Nastase andPopescu, 2009; Basirat et al., in press).

We ask the following research questions. Issuccessful gender transfer possible between non-related languages (if yes, it means that there existuniversal factors in gender assignment)? Whenthe classifier is applied to related languages, doesits success depend on how related they are (if yes,this in an indication that some factors are not uni-versal)? Does gender transfer work in the sameway for all nouns or are there differences betweencertain noun classes?

2 Experimental materials and settings

In this section, we present the languages involved inthis study along with the source of our data. Then,we provide an overview of the cross-lingual wordembedding method and the settings of the classifierused for gender transfer.

2.1 Materials

Two sources of data are selected for each language.First, a noun-gender dictionary is constructed fromthe morphological annotations of the Universal De-pendencies 2.6 (Zeman et al., 2020). For each lan-guage, a dictionary is created by iterating throughall available UD treebanks for the given language.For every unique downcased noun form in thesetreebanks, the grammatical gender is extractedfrom the treebank and labeled to the noun as one offour classes: neuter, feminine, masculine and com-mon (underspecified values such as “Fem,Masc”were ignored). The same four-class label structurewas used for all languages, to ensure compatibilityof the models. Second, word embeddings are se-lected from pre-trained cross-lingual embeddingspublished on the fastText website (Joulin et al.,2018).2 Further details about the embeddings areprovided in the following subsection.

We selected all languages that have grammaticalgender and are present in both data sources, withthe exception of Albanian due to its small treebanksize and Norwegian because in pilot experiments,our classifier showed unexpectedly poor perfor-mance for reasons we were not able to establish.This results in the selection of 24 languages thatare shown in Table 1. Three types of gender sys-tems are found: masculine/feminine (42%, 10/24),masculine/feminine/neuter (46%, 11/24), and com-mon/neuter (12%, 3/24). Only a few languagesbelong to the third type, which is actually the resultof a merge between the masculine and the feminine

2https://fasttext.cc/docs/en/aligned-vectors.html

Page 3: Cross-lingual Embeddings Reveal Universal and Lineage ...

267

categories existing originally in those languages(Enger, 2017, p.1439). The Indo-European lan-guage family is over-represented, which is due topractical limitations from the available resources.Nevertheless, our sample has its advantages. First,the Indo-European language family is consideredto be one of the ‘typical’ grammatical gender lan-guage families (Audring, 2016, p.2), which repre-sents an ideal starting point for a quantitative anal-ysis. Second, comparing languages mostly fromthe same family allows us to address our secondresearch question about the correlation between re-latedness of the languages and the transferabilityof gender.

Language m f n c SizeArabic* ar 33 67 - - 3Bulgarian bg 24 33 43 - 9Catalan ca 49 51 - - 9Czech cs 17 41 43 - 44Danish da - - 72 28 7German de 24 40 36 - 56Greek el 25 52 23 - 4Spanish es 44 56 - - 1French fr 43 57 - - 13Hebrew* he 44 57 - - 6Hindi hi 33 67 - - 8Croatian hr 17 39 45 - 12Italian it 45 55 - - 13Lithuanian lt 38 62 - - 7Latvian lv 49 51 - - 13Dutch nl - - 72 28 9Polish pl 21 35 45 - 26Portuguese pt 45 55 - - 8Romanian3 ro 64 37 - - 17Russian ru 17 34 49 - 44Slovak sk 18 39 43 - 9Slovenian sl 16 41 43 - 12Swedish sv - - 75 25 11Ukrainian uk 14 38 48 - 11

Table 1: Languages included in the data. Asterisk(*) denotes Afro-Asiatic languages, the rest are Indo-European. The gender distribution (in %) is shown incolumns m = masculine, f = feminine, c = common(uter), n = neuter. The ”Size” column indicates the num-ber of nouns in thousand tokens (K).

3Traditionally, Romanian is considered to have three gen-ders: masculine, feminine, and neuter, but an alternative two-gender analysis has also been proposed (Bateman and Polin-sky, 2010). UD follows the two-gender annotation.

2.2 Cross-lingual Word Embeddings

Cross-lingual word embeddings aim at representingwords of multiple languages in a joint embeddingspace such that similar words (in each language andacross all languages) are clustered together. Theseresources provide a foundation for the cross-lingualstudy of words and the development of transferlearning models between languages.

The cross-lingual word embeddings can betrained in different ways (Ruder et al., 2019). Oneof the main approaches is to find a mapping be-tween monolingual word embedding spaces using aseed dictionary that contains words and their trans-lations in different languages. This approach isbased on the observation made by Mikolov et al.(2013) that word embeddings exhibit similar struc-tures across languages. Mikolov et al. (2013) for-mulate the mapping as a least-square linear regres-sion between the monolingual embeddings of theseed lexicon to minimize the mean square error ofthe word translations. The mapping model is thengeneralized to all words in the languages. Thisapproach is improved by Xing et al. (2015); Smithet al. (2017), imposing an orthogonal constraint onthe transformation weights. Later attempts weremade to reduce the need for the seed dictionary(Smith et al., 2017; Artetxe et al., 2017). Conneauet al. (2018); Zhang et al. (2017) leverage adversar-ial training to automatically produce the dictionaryduring training and completely eliminate its neces-sity as a supervision source. Joulin et al. (2018)further enhance the loss function of the regressionmodel using the retrieval model of Conneau et al.(2018), providing for the representation of unseenwords.

In this study, we use fastText cross-lingual wordembeddings trained on the monolingual word em-beddings of Bojanowski et al. (2017) using themapping approach of Joulin et al. (2018). Themonolingual embeddings are trained on Wikipediadata for words that appear at least five times. Theembeddings encode information about the formand semantics of words from sub-word units andword co-occurrences, respectively. The informa-tion about the form and semantics plays a criticalrole in the assignment of grammatical gender tonouns (Corbett, 1991; Rice, 2006). This motivatesus to use fastText embeddings for the study of cross-lingual grammatical gender transfer. The originalembeddings are distributed in a 300-dimensionalspace and cover 44 languages belonging to differ-

Page 4: Cross-lingual Embeddings Reveal Universal and Lineage ...

268

ent language families. In the current study, weretrieved the embeddings for the 24 languages thathave gender systems and a sufficiently large datasize.

2.3 Settings

A multi-layer perceptron is used to predict thegrammatical gender of nouns from their cross-lingual embeddings. The choice of a multi-layerperceptron instead of a recurrent model is moti-vated by 1) the fact that the grammatical gender isan inherent static property of a noun that does notchange in different contexts, and 2) the proven abil-ity of a multi-layer perceptron for the task (Basiratand Tang, 2019). The network has three layers, aninput layer that reads the 300-dimensional wordembeddings, a single hidden layer twice the sizeof the input layer with ReLu activation, and anoutput layer with softmax activation consisting offour neurons related to the four genders mascu-line, feminine, neuter, and common. This pro-vides for modeling the three gender systems mas-culine/feminine, masculine/feminine/neuter, andcommon/neuter and analysing the extent to whichthese systems are transferable.

The classifier is trained on pairs of the noun em-beddings and genders collected from the dictionary.The data is split into 80%, 10% and 10% for train-ing, validation and testing. The data is randomlysplit in folds, of which the designations of training,test and validation are rotated between runs. The fi-nal results are the average of multiple runs coveringa full rotation of the folds. We label the languagethe classifier is trained on source and the languageto which it attempts predicting gender target. Thetrain and the validation data is used for training aclassification model on the source language and thetest data is used for testing the model on the targetlanguage. We go through all possible source-targetcombinations, 576 (24× 24) language pairs.

PyTorch (Paszke et al., 2019) is used to imple-ment the classifier using the stochastic gradientdescent optimizer with a learning rate of 0.1 andthe cross-entropy loss function. Early stopping wasemployed if the model stopped improving over 20epochs, with a minimum of 2200 epochs and amaximum of 25000 epochs. We ran the classifierten times with different random seeds to measurethe variability between training runs. For everylanguage pair, we calculate Fleiss’ kappa acrossthe ten runs. The kappas vary from 0.73 (substan-

tial agreement) to 0.99, the average value is 0.91(almost perfect agreement), and the standard devia-tion is 0.04. We conclude that the results are robustwith respect to random seed.

3 Gender transfer at the language level

In this section, we analyse the results of the ex-periment from two different perspectives. First,we consider the broad transferability of genderacross languages by measuring the accuracy acrossall possible pairs of languages in the dataset (sec-tion 3.1, figures 1 and 3). While this step providesan overview of the accuracy of gender transfer, it isalso extremely influenced by the different gendersystems across languages. For instance, asking alanguage that has the masculine/feminine systemto predict the categories on a language that has amasculine/feminine/neuter system is by definitiongoing to result in a low accuracy since the sourcelanguage does not have information about neuternouns. To overcome this issue, we perform an ad-ditional analysis with narrower scope, where wecompare pairs of only those languages that haveisomorphic systems (section 3.2). As an example,we use languages that have masculine/feminine topredict the gender in languages that also have mas-culine/feminine. The rationale behind narrowingthe scope is that it enables us to focus on the ques-tion of how similar the distributions of nouns acrossgender classes are, abstracting away from possibledifferences between the number of classes and theirtypes.

We define a random guessing baseline for thetransfer between each pair of languages. The base-line is the accuracy that would have been achievedby a classifier that makes a random guess basedsolely on gender probabilities in the source lan-guage. The accuracy it would achieve is∑

g∈{m,f,c,n}

p(gs)p(gt)

where p(gs) is the probability of the given genderin the source language and p(gt) is the probabilityof the given gender in the target language. In all ourexperiments, we report the absolute improvement(or degradation) of the transfer learning accuracyfrom the random baseline accuracy. In this way, anegative value indicates that the transfer accuracyis below the baseline, a positive value indicates thatthe transfer accuracy is higher than the baseline,

Page 5: Cross-lingual Embeddings Reveal Universal and Lineage ...

269

and a zero value indicates that the transfer accuracyis only as good as the random baseline.4

3.1 Gender transfer between all systemsThe mean accuracy from the ten different-seed runsof each transfer is compared with the random base-line and plotted in Figure 1. Each entry is thedifference between the result and the baseline, i.e.an improvement (or degradation) from the baselineresults.

We run three two-sided paired t-tests to checkwhether the accuracy is significantly different fromthe baseline: one for language pairs within theIndo-European family (t(483) = 27.013, p-value <0.001), one for language pairs where the sourcelanguage is from an Afro-Asiatic family and thetarget language is Indo-European (t(43) = 10.454,p-value < 0.001), one for language pairs wherethe source is Indo-European and the target is Afro-Asiatic (t(43) = 9.5251, p-value < 0.001), in allcases the average classifier accuracy is higher thanthe baseline.

Three main observations are worth noting. First,word embeddings do provide sufficient informa-tion to generate an accuracy significantly above thebaseline for the majority of languages, as shownby the diagonal line in the plot and the output ofthe t-tests. Second, Danish, Dutch, and Swedishdo not transfer well to other languages. Thesethree languages are the only languages that havea common/neuter system, which explains the lowaccuracy of gender transfer. Third, even thoughArabic and Hebrew are not related to the Indo-European language family, both as source and tar-get languages they yield accuracy that is compara-ble to that yielded by Indo-European languages andis significantly higher than the baseline (see Section4). These points imply that while the relatedness oflanguages affect the transferability of gender, weare also likely to find some shared principles ofgender assignment across non-related languages.

Then, we compare the accuracy of the trans-fer with phylogenetic distance within the Indo-European language family. To do so, we extractthe phylogenetic distance from the broad Indo-European tree published by Chang et al. (2015, tree

4Note that our measure is not of course a perfect quan-tification of gender-system similarity, since it does not yieldthe accuracy of 1 for all the cases when source and targetlanguages are the same (the diagonal in Figure 1). It can,however, be viewed as an approximation (in principle, theaccuracy of transfer X → Y can be normalized by dividing itby the accuracy of X → X).

A3), as shown in Figure 2. The branch lengths areannotated in terms of years, which allows a directcomparison of phylogenetic distance defined as thetime depth of the first common ancestor shared bya pair of compared languages. The larger the dis-tance, the less related is a pair of languages. Theoutput of the comparison is shown in Figure 3.

A linear regression shows a significant relation-ship between accuracy and phylogenetic distance(t(4838) = −60.12, p < 0.001). The slope coeffi-cient for phylogenetic distance is−0.00005, whichmeans that the accuracy decreases by 5% for each1000 years of phylogenetic distance. The R2 valueshows that 43% of the variation in accuracy canbe explained by phylogenetic distance. A closeranalysis indicates that the transfer accuracy of apair of languages sharing the same system is gen-erally higher than the transfer accuracy of pair oflanguages having different systems. For instance,the lower values between 1000 and 2000 years ofphylogenetic distance are gender transfers betweencommon/neuter and masculine/feminine/neuter lan-guages. Further details are explained in subsec-tion 3.2.

Finally, we performed a correlation test to seewhether gender transferability in a language pairis affected by how different gender distributions inthe two languages are. By gender distribution wemean a distribution of the marginal probabilities ofseeing each gender over all nouns in the vocabularyset, and we measure the difference between twodistributions as the KL-divergence. We find thatthe transferability correlates negatively with theKL-divergence both globally over all languages(Spearman ρ = −0.7, p < 0.001) and locallywithin each branch (Slavic: ρ = −0.4, p = 0.002,Germanic: ρ = −0.9, p < 0.001, and Romance:ρ = −0.8, p < 0.001), indicating that gender trans-fer becomes weaker as the marginal distributionsof gender become more different.

3.2 Gender transfer between isomorphicsystems

Three types of comparisons are made. First, theaccuracy of transfer between languages with mascu-line/feminine systems is measured. Then, the sameprocess is conducted for languages with mascu-line/feminine/neuter systems and common/neutersystems.

To estimate the combined effect of phylogeneticdistance and system type (masculine/feminine,

Page 6: Cross-lingual Embeddings Reveal Universal and Lineage ...

270

Figure 1: Average difference between accuracy and the random baseline. A positive value represents an accuracyabove the baseline while a negative value indicates an accuracy below the baseline.

Figure 2: The phylogenetic tree of the Indo-Europeanlanguages included in the analysis.

masculine/feminine/neuter, common/neuter), wefit a linear regression model with the values of theaccuracy improvement (or degradation) from thebaseline as the dependent variable, phylogeneticdistance a continuous predictor, and system typea categorical predictor (common/neuter is the ref-erence level). The two-way interaction betweenthe predictors is also included. The summary ofthe model is presented in Table 2. The R2 valueshows that 74% of the variation of accuracy in thesample can be explained by phylogenetic distanceand gender system types.

The results again show a negative relationship be-tween accuracy and phylogenetic distance. With re-gard to gender systems, having masculine/feminineor masculine/feminine/neuter has a positive effect

Figure 3: A comparison of the accuracy improvement(or degradation) of gender transfer (Y-axis) and the phy-logenetic distance between each language pair (X-axis).The dashed line refers to the random baseline. The phy-logenetic distance refers to the years separating eachpair of languages in the Indo-European tree. Each pointrepresents the average of the transfer accuracy over 10runs.

on the accuracy, when considering common/neutersystems as the reference level. Within all threesystems, masculine/feminine has the highest coeffi-cient (0.12), which implies that transfers betweenlanguages having masculine/feminine gender sys-tems generally result in higher accuracy than mas-culine/feminine/neuter and common/neuter. Theinteractions show that the negative effect of phylo-genetic distance on accuracy is attenuated if bothlanguages in the pair have masculine/feminine or

Page 7: Cross-lingual Embeddings Reveal Universal and Lineage ...

271

Predictor Estimate t(1894) P valuePhyDis -0.00009 -10.926 < 0.001m/f 0.12251 12.053 < 0.001m/f/n 0.06577 6.589 < 0.001PhyDis:m/f 0.00003 4.104 < 0.001PhyDis:m/f/n 0.00004 4.679 < 0.001

Table 2: Summary of the regression model: Accuracyas predicted by phylogenetic distance and gender sys-tem type (m = masculine, f = feminine, n = neuter).

masculine/feminine/neuter gender systems.

4 Gender transfer at the word level

In this section, we perform a finer-grained analysis:focus not on languages, but on individual nouns.Our main question is if there are any patterns in thedistribution of errors. Is it random or are certainclasses of nouns systematically more difficult topredict than others?

To obtain a single prediction for every noun fromthe 10 random-seed runs, we pick the gender whichhas the largest sum of confidence scores (softmaxactivation values). This is virtually equivalent totaking the gender that gets most votes across theruns, but has an advantage of avoiding ties.

We test whether the following factors play a role:how frequent a noun is, whether it is animate ornot and whether its form is equivalent to lemma(citation form, baseform) or not.

It is reasonable to expect that embeddings of fre-quent nouns will capture more useful informationand thus yield better accuracy. Note, however, thatvery infrequent nouns (frequency <5) have alreadybeen excluded from consideration, since for themthe embeddings are not available. We calculate fre-quency of every noun form using the UD corpora.

Nouns denoting living beings, especially humanbeings, can be expected to yield higher accuracy,since for them the semantic motivation behind gen-der assignment is often more transparent (basedon biological sex). That is not always the case(cf. the already-mentioned German Madchen ‘girl’,which is neuter), and the proportion of nouns wheregrammatical gender is predicted by biological sexis likely to vary across languages. Furthermore,it is unknown to what extent the embeddings canactually capture the relevant semantics. Nonethe-less, at least in some cases sex can predict gender(cf. French garcon ‘boy’ and fille ‘girl’, or Russiankot ‘tomcat’ and koska ‘female cat’ that are resp.

masculine and feminine). For nouns that do notdenote living things no such predictor is known.

As a proxy for “denoting a living thing” we usethe animacy category available in some UD tree-banks for Slavic languages (Czech, Slovak, Polish,Russian, Ukrainian, Slovenian and Croatian). InSlavic, animacy is manifested on the grammaticallevel, primarily through differential object marking(Janda, forthcoming). Animacy annotation is alsoavailable in the Hindi-PUD treebank, but that tree-bank lacks lemmas, which makes it unsuitable forour analysis; see below.

We extract animacy information in the follow-ing way: we go through all treebanks available forevery language and calculate how often a form isannotated as “animate” or “inanimate”. Polish hasmore detailed annotation: animate human and ani-mate non-human, we collapse these two categoriesinto “animate”. Note that Slavic animacy is a for-mal feature that is not exactly isomorphic to livingvs. non-living distinction.

Finally, at least in some languages it is easierto infer gender from the citation form (that is, theform which is equivalent to lemma) than from in-flected forms (see (Berdicevskis, forthcoming) foran overview of Slavic languages). This presum-ably happens because cues are more transparentin the citation form. It can also be easier to learnthe cues for the citation form, which is often themost frequent one. We test whether this tendencyis observed in our data. Alternatively, we couldhave tested whether certain morphological features(number: singular vs plural, case: nominative vsoblique etc.) play a role, but the analysis we chooseis more universal: it can be applied to any languagewithout adjustments.

Other formal, semantic and historical propertiescan potentially affect how easy it is to infer thegender of a noun (e.g. whether a noun is a recentborrowing or not), but there is no straightforwardway to reliably extract this information for all thenouns in our datasets.

We run two logistic regression models. Bothinclude data only from those language pairs wherethe target language is one of the seven Slavic lan-guages with detailed animacy information: Czech,Slovak, Polish, Russian, Ukrainian, Slovenian,Croatian. In both models, the dependent variable iswhether a noun has its gender correctly predictedby the classifier. In Model 1, the independent vari-ables are frequency, animacy, citation form and

Page 8: Cross-lingual Embeddings Reveal Universal and Lineage ...

272

phylogenetic distance between source and targetlanguages. It is applied to all language pairs exceptthose where source language is Arabic or Hebrew(since for them the distance cannot be estimated).To the remaining pairs we apply Model 2, whichhas only three independent variables: frequency,animacy, citation form. The results are reported inTable 3 and Table 4.

Predictor Estimate z P valueFreq -0.004 -0.95 0.340Inan -0.999 -13.4 < 0.001Non-cit -0.789 -10.4 < 0.001PhyDis -0.0004 -24.4 < 0.001Freq:Inan 0.006 1.4 0.168Freq:Non-cit 0.009 2.0 0.042Inan:Non-cit -0.311 -3.9 < 0.001Freq:PhyDis 1e-06 1.0 < 0.302Inan:PhyDis 5.7e-05 3.0 < 0.003Non-cit:PhyDis 6.9e-05 3.6 < 0.001Freq:Inan:Non-cit -0.009 -2.1 0.034Freq:Inan:PhyDis -1.7e-06 -1.6 0.106Freq:Non-cit:PhyDis -2.1e-06 -1.9 0.059Inan:Non-cit:PhyDis 6.5e-05 3.1 0.002Freq:Inan:Non-cit:PhyDis 2.4e-06 2.2 0.030

Table 3: Summary of Model 1: Correctness of theguess as predicted by noun frequency, animacy (Animvs Inan), citation form (Cit vs Non-Cit) and phyloge-netic distance (PhyDis).

Coefficient Estimate z P valueFreq -0.003 -0.4 0.707Inan -1.871 -14.0 < 0.001Non-cit -1.455 -10.7 < 0.001Freq:Inan 0.002 0.3 0.800Freq:Non-cit 0.002 0.3 0.7744Inan:Non-cit 0.852 6.0 < 0.001Freq:Inan:Non-cit -0.001 -0.1 0.897

Table 4: Summary of Model 2 (Arabic and Hebrewas source languages): Correctness of the guess as pre-dicted by noun frequency, animacy (Anim vs Inan) andcitation form (Cit vs Non-Cit).

Model 1 (Indo-European languages only) showsthat the prediction accuracy is significantly lowerfor inanimate nouns than for animate nouns andfor inflected forms than for citation forms. The fol-lowing predictors also have significant negative ef-fects: the interaction of inanimacy and non-citationform; the interaction of frequency, inanimacy andnon-citation form; phylogenetic distance betweensource and target languages (the coefficient is small,but it shows change per year, and the distance in

our dataset vary from approx. 300 to 5000 years).Interestingly, all other significant predictors (mostof which all are interactions of distance with otherpredictors) are positive. It means that the negativeeffects (for instance, those of inanimacy and non-citation form) described above are smaller for lessrelated languages. The negative effect of inflectedforms is also smaller for more frequent nouns.

Model 2 (Afro-Asiatic languages as source)shows similar results: inanimacy and non-citationforms have significant and strong negative effect.This similarity with Model 1 provides further ev-idence in favor of the universal factors in genderassignment. What is different, however, is that theinteraction of these two factors has a strong pos-itive effect (that is, inflected forms of inanimatenouns have higher accuracy than can be expected).

Finally, we perform one more test. Our resultsindicate that gender systems are partly transfer-able across non-related languages (see Section 3.1),which suggests there are certain universalities ingender assignment. We want to test whether theseuniversalities are limited to the aforementioned factthat grammatical gender for living creatures closely(even though not perfectly) matches biological sex.To investigate that, we focus on those languagepairs where source language is Afro-Asiatic andtarget language is one of those for which the ani-macy information is available. From these pairs,we exclude those treebanks where animacy is anno-tated only for a small proportion of nouns (Slove-nian and Croatian), and that leaves us with 10 pairs(Arabic and Hebrew as source, Russian, Czech, Pol-ish, Slovak and Ukrainian as target). We focus oninanimate nouns only (thus eliminating any possi-ble contribution of biological sex) and test whetherthe classifier still performs above the chance base-line. With only 10 datapoints, we cannot run areliable t-test. Instead, we perform 10 simulationtests, running a naive classifier that guesses thegender relying solely on the source-language prob-abilities 10000 times for every language pair andtaking the proportion of cases when it achieves thesame accuracy as our classifier (or higher) as thep-value. The p-value is 0 in all pairs apart fromfour: Arabic → Czech: 0.009, Arabic→ Polish:0.718, Arabic→ Slovak: 0.923, Hebrew→ Polish:0.003. In other words, in eight cases out of 10,the classifier, applied only to inanimate nouns, stillperforms significantly better than chance.

Page 9: Cross-lingual Embeddings Reveal Universal and Lineage ...

273

5 Conclusion

This study investigates how grammatical gender istransferable across languages from a transfer learn-ing point of view. The cross-lingual word embed-dings are considered as the source of knowledgeshared between languages from which the gram-matical gender of nouns are predicted using a multi-layer perceptron. The empirical results reveals thatthere exist some universal and lineage-specific pat-terns in the grammatical gender assignment.

First, our analysis of gender transfer betweenAfro-Asiatic and Indo-European languages indi-cated that partly successful gender transfer is pos-sible between non-related languages. This observa-tion supports the existence of universal factors ingender assignment. The accuracy of the classifieris higher than the random baseline even when it istested on inanimate nouns only, which means thatthe universal factors are not limited to biologicalsex.

Second, our analysis of gender transfer betweenIndo-European languages demonstrates that thephylogenetic distance between languages has a neg-ative effect on the success of the transfer, whichsuggests that some factors of gender assignmentare not universal. These results match with theliterature by showing that gender assignment is amixture of universal and idiosyncratic factors.

Third, we also found that gender transfer doesnot work in the same way for all nouns. The predic-tion accuracy is significantly lower for inanimatenouns than for animate nouns and for inflectedforms than for citation forms. This effect is foundwhen considering both family-internal and family-external transfers, which provides further evidencein favor of the universal factors in gender assign-ment.

We would like to make a few caveats and sug-gestions about the future development of the cur-rent study. While we address the universality ofgender assignment cross-linguistically, our data isrestricted to languages from two families and ourword embeddings are trained on data from specificdomains. Additional data from a more diverse sam-ple is needed to further confirm our observations.Furthermore, we cannot fully exclude that the ob-served similarities are an areal effect caused bycontact.

It should also be noted that we cannot identifywhich universal factors enable the classifier to per-form above the baseline. A more fine-grained word-

level analysis would be required to find the possiblecontributors to this. Linguistically, grammaticalgender is strongly tied to the semantic and for-mal properties of nouns. Since the cross-lingualword embeddings used in this study encode boththe formal and semantic information, we cannotdisentangle the relative contributions of form andsemantics to gender transfer.

Finally, it should be mentioned that an importantline of research in modern NLP focuses on genderbias present in naturally occurring texts (Caliskanet al., 2017; Gonen et al., 2019). The combina-tion of these questions and approaches with ourperspective might become an interesting researchdirection.

Supplementary materials, including raw data andscripts for analysis are openly available.5

Acknowledgments

The authors are thankful for the constructive com-ments from the anonymous referees and editors,which helped to significantly improve the qualityof the paper. The second author is thankful forthe support of the IDEXLYON (16-IDEX-0005)Fellowship grant.

ReferencesJacob Andreas and Dan Klein. 2014. How much do

word embeddings encode about syntax? In Proceed-ings of the 52nd Annual Meeting of the Associationfor Computational Linguistics (Volume 2: Short Pa-pers), pages 822–827, Baltimore, Maryland. Associ-ation for Computational Linguistics.

Mikel Artetxe, Gorka Labaka, and Eneko Agirre. 2017.Learning bilingual word embeddings with (almost)no bilingual data. In Proceedings of the 55th AnnualMeeting of the Association for Computational Lin-guistics (Volume 1: Long Papers), pages 451–462,Vancouver, Canada. Association for ComputationalLinguistics.

Mikel Artetxe, Gorka Labaka, Inigo Lopez-Gazpio,and Eneko Agirre. 2018. Uncovering divergentlinguistic information in word embeddings withlessons for intrinsic and extrinsic evaluation. In Pro-ceedings of the 22nd Conference on ComputationalNatural Language Learning, pages 282–291, Brus-sels, Belgium. Association for Computational Lin-guistics.

Jenny Audring. 2016. Gender. In Mark Aronoff, editor,Oxford research encyclopedia of linguistics. OxfordUniversity Press, Oxford.5https://github.com/marctang/Cross-lingual-embeddings-

Grammatical-gender

Page 10: Cross-lingual Embeddings Reveal Universal and Lineage ...

274

Ali Basirat, Marc Allassonniere-Tang, and Aleksan-drs Berdicevskis. in press. An empirical study onthe contribution of formal and semantic features tothe grammatical gender of nouns. Linguistics Van-guard.

Ali Basirat and Marc Tang. 2019. Linguistic informa-tion in word embeddings. In Agents and ArtificialIntelligence, pages 492–513, Cham. Springer Inter-national Publishing.

Nicoleta Bateman and Maria Polinsky. 2010. Ro-manian as a two-gender language. In HypothesisA/Hypothesis B, pages 41–77. MIT Press.

Aleksandrs Berdicevskis. forthcoming. Gender anddeclension. In Neil Bermel and Jan Fellerer, edi-tors, Oxford Guides to the World’s Languages: TheSlavonic Languages. Oxford University Press.

Piotr Bojanowski, Edouard Grave, Armand Joulin, andTomas Mikolov. 2017. Enriching word vectors withsubword information. Transactions of the Associa-tion for Computational Linguistics, 5:135–146.

Aylin Caliskan, Joanna J Bryson, and ArvindNarayanan. 2017. Semantics derived automaticallyfrom language corpora contain human-like biases.Science, 356(6334):183–186.

Will Chang, Chundra Cathcart, David Hall, and An-drew Garrett. 2015. Ancestry-constrained phyloge-netic analysis supports the Indo-European steppe hy-pothesis. Language, 91(1):194–244.

Alexis Conneau, Guillaume Lample, Marc’AurelioRanzato, Ludovic Denoyer, and Herve Jegou. 2018.Word translation without parallel data.

G. G. Corbett. 2001. Grammatical gender. In Inter-national Encyclopedia of the Social Sciences, pages6335–6340.

Greville G Corbett. 1991. Gender. Cambridge Univer-sity Press, Cambridge.

Greville G Corbett. 2013a. Number of Genders. InMatthew S Dryer and Martin Haspelmath, editors,The World Atlas of Language Structures Online.Max Planck Institute for Evolutionary Anthropol-ogy, Leipzig.

Greville G Corbett. 2013b. Sex-based and non-sex-based gender systems. In Matthew S Dryer and Mar-tin Haspelmath, editors, The world atlas of languagestructures online. Max Planck Institute for Evolu-tionary Anthropology, Leipzig.

Greville G Corbett and Norman Fraser. 2000. Gen-der assignment: A typology and a model. InGunter Senft, editor, Systems of nominal classifica-tion, pages 293–325. Cambridge University Press,Cambridge.

Francesca Di Garbo, Bruno Olsson, and BernhardWalchli. 2019. Grammatical gender and linguisticcomplexity : Volume ii: World-wide comparativestudies.

Hans-Olav Enger. 2017. The Nordic languages in the19th century II: Morphology. In Oskar Bandle, KurtBraunmuller, Ernst Hakon Jahr, Allan Karker, Hans-Peter Naumann, Ulf Teleman, Lennart Elmevik, andGun Widmark, editors, The Nordic Languages, Part2, pages 1437–1442. De Gruyter, Berlin.

Sebastian Fedden and Greville G Corbett. 2019. Thecontinuing challenge of the German gender system.Paper presented at the International Symposium ofMorphology.

Hila Gonen, Yova Kementchedjhieva, and Yoav Gold-berg. 2019. How does grammatical gender affectnoun representations in gender-marking languages?In Proceedings of the 23rd Conference on Computa-tional Natural Language Learning (CoNLL), pages463–471, Hong Kong, China. Association for Com-putational Linguistics.

Laura Janda. forthcoming. Gender and animacy. InNeil Bermel and Jan Fellerer, editors, Oxford Guidesto the World’s Languages: The Slavonic Languages.Oxford University Press.

Scott Jarvis and Aneta Pavlenko. 2010. Crosslinguis-tic influence in language and cognition, paperbacked edition. Routledge, New York, NY. OCLC:845735473.

Armand Joulin, Piotr Bojanowski, Tomas Mikolov,Herve Jegou, and Edouard Grave. 2018. Loss intranslation: Learning bilingual word mapping with aretrieval criterion. In Proceedings of the 2018 Con-ference on Empirical Methods in Natural LanguageProcessing.

David Kemmerer. 2017. Categories of object conceptsacross languages and brains: the relevance of nomi-nal classification systems to cognitive neuroscience.Language, Cognition and Neuroscience, 32(4):401–424.

Tomas Mikolov, Quoc V Le, and Ilya Sutskever. 2013.Exploiting similarities among languages for ma-chine translation. arXiv preprint arXiv:1309.4168.

Vivi Nastase and Marius Popescu. 2009. What’s ina name? In some languages, grammatical gender.In Proceedings of the 2009 Conference on Empiri-cal Methods in Natural Language Processing, pages1368–1377, Singapore. Association for Computa-tional Linguistics.

Adam Paszke, Sam Gross, Francisco Massa, AdamLerer, James Bradbury, Gregory Chanan, TrevorKilleen, Zeming Lin, Natalia Gimelshein, LucaAntiga, Alban Desmaison, Andreas Kopf, EdwardYang, Zachary DeVito, Martin Raison, Alykhan Te-jani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang,

Page 11: Cross-lingual Embeddings Reveal Universal and Lineage ...

275

Junjie Bai, and Soumith Chintala. 2019. Py-torch: An imperative style, high-performance deeplearning library. In H. Wallach, H. Larochelle,A. Beygelzimer, F. dAlche-Buc, E. Fox, and R. Gar-nett, editors, Advances in Neural Information Pro-cessing Systems 32, pages 8024–8035. Curran Asso-ciates, Inc.

Curt Rice. 2006. Optimizing gender. Lingua,116(9):1394–1417.

Sebastian Ruder, Ivan Vulic, and Anders Søgaard.2019. A survey of cross-lingual word embeddingmodels. J. Artif. Int. Res., 65(1):569–630.

Laura Sabourin, Laurie A Stowe, and Ger J De Haan.2006. Transfer effects in learning a second languagegrammatical gender system. Second Language Re-search, 22(1):1–29.

Frank Seifart. 2010. Nominal classification. Languageand Linguistics Compass, 4(8):719–736.

Samuel L Smith, David HP Turban, Steven Hamblin,and Nils Y Hammerla. 2017. Offline bilingual wordvectors, orthogonal transformations and the invertedsoftmax.

Hiroya Takamura, Ryo Nagata, and YoshifumiKawasaki. 2016. Discriminative analysis of linguis-tic features for typological study. In Proceedingsof the Tenth International Conference on LanguageResources and Evaluation (LREC’16), pages 69–76,Portoroz, Slovenia. European Language ResourcesAssociation (ELRA).

Adina Williams, Damian Blasi, Lawrence Wolf-Sonkin, Hanna Wallach, and Ryan Cotterell. 2019.Quantifying the semantic core of gender systems. InProceedings of the 2019 Conference on EmpiricalMethods in Natural Language Processing and the9th International Joint Conference on Natural Lan-guage Processing (EMNLP-IJCNLP), pages 5734–5739, Hong Kong, China. Association for Computa-tional Linguistics.

Chao Xing, Dong Wang, Chao Liu, and Yiye Lin. 2015.Normalized word embedding and orthogonal trans-form for bilingual word translation. In Proceedingsof the 2015 Conference of the North American Chap-ter of the Association for Computational Linguistics:Human Language Technologies, pages 1006–1011,Denver, Colorado. Association for ComputationalLinguistics.

Daniel Zeman, Joakim Nivre, et al. 2020. Universal de-pendencies 2.6. LINDAT/CLARIAH-CZ digital li-brary at the Institute of Formal and Applied Linguis-tics (UFAL), Faculty of Mathematics and Physics,Charles University.

Meng Zhang, Yang Liu, Huanbo Luan, and MaosongSun. 2017. Adversarial training for unsupervisedbilingual lexicon induction. In Proceedings of the55th Annual Meeting of the Association for Com-putational Linguistics (Volume 1: Long Papers),

pages 1959–1970, Vancouver, Canada. Associationfor Computational Linguistics.