Introduction

These pages present some basic frequencies based on the Electronic Text Corpus of Sumerian Literature. The frequencies are in the form of alphabetic and rank frequency lists, and the basic units of the lists are either word forms (types), lexemes, or sign names. There are also lists based on only a part of the corpus, which can be used to analyse subgroups of the compositions making up the ETCSL.

Since the ETCSL is constantly being expanded and amended, the frequencies will necessarily change over time. It is therefore important to register any change in the dating of the various lists.

Before making use of the many lists provided, be sure to read how they were produced and take note of the fact that certain parts of the corpus were excluded before the frequencies were put together. This means that a search in the corpus will yield a slightly different count from the frequency provided for the same word, lexeme, or sign name in any of the frequency lists.

How ...

When extracting the information to be included in the various lists, certain (arbitrary) decisions regarding format had to be made, partly due to requirements imposed by the programs used to produce the lists.

The main programs used to produce the lists are: WordSmith Tools 4.0 (http://www.lexically.net/wordsmith/) and the Ngram Statistics Package (http://ngram.sourceforge.net/). In addition, some home-made Perl programs were developed.


Exclusions

In order to compile a base-profile of the corpus the following have been excluded: additions and non-primary variants within compositions; the catalogues; and those proverbs numbered 6.2 (to a large degree these duplicate proverbs which also occur in the files numbered 6.1). The Sumerian King List (2.1.1) has also been excluded because its statistical profile is atypical. If a composition has more than one version, Me-Turan and Susa versions are excluded and the remaining longest version is regarded as primary. Thus, the program extracting the material used as basis for the frequencies excludes:

c.0.1.1.xml
c.0.1.2.xml
c.0.2.01.xml
c.0.2.02.xml
c.0.2.03.xml
c.0.2.04.xml
c.0.2.05.xml
c.0.2.06.xml
c.0.2.07.xml
c.0.2.08.xml
c.0.2.11.xml
c.0.2.12.xml
c.0.2.13.xml
c.1.2.2.xml:<text id="c122.ver1" n="2">
c.1.8.1.2.xml:<text n="2">,<text n="3">
c.1.8.1.3.xml:<text id="c1813.ver1" n="2">,<text n="3">
c.1.8.1.4.xml:<text n="2">,<text n="3">,<text n="4">,<text n="5">
c.2.1.1.xml
c.2.1.5.xml:<text n="2">
c.2.2.6.xml:<text id="c226.ver1" n="2">
c.2.4.1.1.xml:<text n="2">
c.2.4.1.4.xml:<text n="2">,<text n="3">
c.2.5.4.13.xml:<text id="c25413.ver1" n="2">
c.2.5.4.15.xml:<text n="2">
c.3.1.06.1.xml:<text n="2">
c.3.1.08.xml:<text n="1">
c.3.1.20.xml:<text n="2">
c.4.07.8.xml:<text n="2">
c.4.08.18.xml:<text n="2">,<text n="3">
c.4.16.1.xml:<text n="2">
c.4.31.1.xml:<text n="2">
c.6.2.1.xml
c.6.2.2.xml
c.6.2.3.xml
c.6.2.4.xml
c.6.2.5.xml


Comparing a subset with the whole corpus

When comparing a subset of the compositions with the whole corpus, i.e. using the whole corpus as a reference corpus, all the compositions making up the subset are excluded from the reference corpus. This seemed to yield the best result.

The comparison is presented as a list of key words and is done with the help of the WordSmith Tool. Below is an extract from the help menu of that program. To read more about WordSmith Tools, go to http://www.lexically.net/wordsmith/.

Key words (from WordSmith Tools Help (c) Mike Scott)

The term "key word", though it is in common use, is not defined in Linguistics. This program identifies key words on a mechanical basis by comparing patterns of frequency. (A human being, on the other hand, may choose a phrase or a superordinate as a key word.)

A word is said to be "key" if
a) it occurs in the text at least as many times as the user has specified as a Minimum Frequency
b) its frequency in the text when compared with its frequency in a reference corpus is such that the statistical probability as computed by an appropriate procedure is smaller than or equal to a p value specified by the user.

positive and negative keyness
A word which is positively key occurs more often than would be expected by chance in comparison with the reference corpus. A word which is negatively key occurs less often than would be expected by chance in comparison with the reference corpus.

typical key words KeyWords will usually throw up 3 kinds of words as "key".

First, there will be proper nouns. Proper nouns are often key in texts, though a text about racing could wrongly identify as key, names of horses which are quite incidental to the story. This can be avoided by specifying a higher Minimum Frequency.

Second, there are key words that human beings would recognise. The program is quite good at finding these, and they give a good indication of the text's "aboutness". (All the same, the program does not group synonyms, and a word which only occurs once in a text may sometimes be "key" for a human being. And KeyWords will not identify key phrases unless you are comparing wordlists based on word clusters.)

Third, there are high-frequency words like because or shall or already. These would not usually be identified by the reader as key. They may be key indicators more of style than of "aboutness". But the fact that KeyWords identifies such words should prompt you to go back to the text, perhaps with Concord [another WordSmith Tool, JE], to investigate why such words have cropped up with unusual frequencies.


Batch comparison

A routine has been set up to simplify the comparison of compositions. In principle, all the compositions in the corpus can now be compared with each other. The program that performs the comparisons outputs a tab-delimited list, which can easily be pasted into e.g. Excel. In the example below, all the compositions starting with c.2.4, Ur III praise poetry, were compared. The list was then pasted into Excel and sorted according to the percentage of matching lexeme bigrams. (A nice thing about Excel is that it can sort on multiple columns.) Only a fraction of the 1,892 lines of the complete list is shown in Table 1.1. A brief discussion of the table and a way of highlighting the similarities are found below the table.

Table 1.1. Comparison of individual compositions sorted according to % of common lexeme bigrams (column 11)
C1/2 = composition 1/2, W = word/token, L = lexeme, WL = word line, LL = lexeme line
Compos. (C1 > C2)W in C1Overlap with C2%L in C1Overlap with C2%Biwords%Bilexemes%WL overlap%LL overlap%
c.2.4.4.9 > c.2.4.2.2443754375266.7266.70000
c.2.4.1.5 > c.2.4.1.618111362.418115585.65533.18351.6512.5717.5
c.2.4.1.6 > c.2.4.1.524413053.324218676.95524.6833959.6713.5
c.2.4.4.9 > c.2.4.4.14125437500133.30000
c.2.4.4.9 > c.2.4.2.1742504250133.3133.30000
c.2.4.1.8 > c.2.4.1.1271348.127176314.5313.60000
c.2.4.5.5 > c.2.4.5.4935761.3936974.277.71213.20000
c.2.4.2.b > c.2.4.2.041133732.71098174.332.81312.90000
c.2.4.2.b > c.2.4.2.021133732.71098174.321.91312.90000
c.2.4.1.a > c.2.4.2.0275304073547434.1912.70000
c.2.4.2.25 > c.2.4.2.031284938.31239274.864.71512.513.413.4
c.2.4.2.12 > c.2.4.2.0252917.3493571.400612.50015.6
c.2.4.4.5 > c.2.4.5.3451226.742215024.5512.500210
c.2.4.2.17 > c.2.4.2.0421912356.220817182.283.92312.311.912
c.2.4.2.17 > c.2.4.2.242199543.420815775.573.42211.80000
c.2.4.1.a > c.2.4.2.04753749.3736082.234.1811.30000
c.2.4.1.a > c.2.4.2.24752837.373547434.1811.30000
c.2.4.1.5 > c.2.4.1.31816938.118112770.2531811.20000
c.2.4.5.5 > c.2.4.2.04935660.2937277.422.210110000
c.2.4.2.23 > c.2.4.2.05491530.6483266.712.1510.90014.3
c.2.4.2.b > c.2.4.2.011132723.91097266.132.81110.90000
c.2.4.2.05 > c.2.4.2.0293138741.690169677.2222.49210.810.531.4
c.2.4.2.17 > c.2.4.2.152199744.320814871.294.41910.20000
c.2.4.5.1 > c.2.4.5.415063421429063.496.91210.10000
c.2.4.2.25 > c.2.4.2.051284837.51238972.410.812100000
c.2.4.2.a > c.2.4.2.071112421.61116760.411.66100000
c.2.4.1.a > c.2.4.1.1753040735676.722.779.90000
c.2.4.2.b > c.2.4.2.051133833.61097770.610.9109.90000
c.2.4.2.b > c.2.4.2.151133026.51096559.621.9109.90000

From left to right the various columns contains

  1. the two compositions compared
  2. the total number of words in composition 1
  3. the number of words it shares with composition no. 2
  4. the percentage of shared words
  5. the total number of lexemes in compositions 1
  6. the number of lexemes it shares with compositions no. 2
  7. the percentage of shared lexemes
  8. the number of shared word bigrams in the two compositions
  9. the percentage of shared word bigrams
  10. the number of shared lexeme bigrams shared
  11. the percentage of shared lexeme bigrams
  12. the number of equivalent lines in the two compositions
  13. the percentage
  14. the number of similar lines when all words have been converted to lexemes AND all proper nouns have been converted to the dummy lexeme PROPER + type of proper noun, e.g. PROPER_RN
  15. the percentage.

What is potentially interesting about such a comparison is not only finding compositions that are similar in some respect, but also the ones that are (very) different within a particular grouping.

As always, there are a number of things one needs to be aware of when reading such a table. One is nicely illustrated by the first row of the table. This row yields extremely high numbers, but this is only because the composition being compared contains four words only, and hence is very "similar" to very many other compositions.

Row two in Table 1.1 contains the result of the comparison of c.2.4.1.5 with c.2.4.1.6 and may be more rewarding to follow up. These two compositions have a shared vocabulary (lexemes) of 85.6 % (seen from C1). They even have five identical lines (column 12) and seven similar lines (column 14). To illustrate the overlap between the two composition, a program called MOSS (Measure Of Software Similarity) can be used (http://www.cs.berkeley.edu/~aiken/moss.html). This program was designed to spot plagiarism in computer programming classes, but it seems to do a good job of finding similarities between Sumerian as well. Open c.2.4.1.5vsc.2.4.1.6.html and study the result. Note that we have compared the files consisting of the lexemes only. We may also compare the word files (c.2.4.1.5fvsc.2.4.1.6f.html), although the result it not equally colourful. However, MOSS does a good job of finding matches. MOSS is easy to download, install and run, provided you have Perl installed on your machine as well.

Another way of highlighting similarites between two compositions is to put in bold all the shared bigrams, either the lexeme or word bigrams. The outputs from this process can then be put side by side in a web document: sharedlexemebigrams / sharedwordbigrams.


Collocations

It is sometimes said that words can be defined by the company they keep, i.e., that the meaning of a particular word is strongly influenced or coloured by the surrounding words. With the development of large corpora and dedicated software it is very easy to study a word and its collocates. WordSmith Tools have options which let you highlight a word's left and right collocates. As an experiment, let us look at the nouns occurring immediately to the left of the following lexemes: babbar 'white', dadag 'to be bright', dalla 'to be bright', kug 'shining', and zalag 'to shine'. (Please ignore for the sake of this experiment the difference in word class marking.) As can be seen from the English labels given to these lexemes, their meaning, with perhaps the exception of babbar, seem similar, and they would very likely have been listed as synonyms in a thesaurus comparable to English words such as beaming, gleaming, glittering, glowing, shining, and luminous. We should also note that several of the Sumerian words are written with the same sign, UD (or UD.UD), i.e. babbar/dadag/zalag.

Table 1.2. Five comparable lexemes and their most frequent left collocates
Keyword (freq.)Nouns immediately to the left of the keyword
babbar (73)erin, tug2, siki, ud
dadag (79)cu-luh, a, ki, cu
dalla (60)ud, kalam, utu
kug (1,196)an (An), cag4, ki, inim, barag, ki-tuc, an (heaven), cir3
zalag (98)ud, saj-ki, cag4, igi

Before we venture some general remarks, note that the frequency with which these lexemes occur differ significantly, and that this, of course, will have a bearing on the strength of any claims made. It also vital to stress the importance of looking at the wider context of the above collocations and the forms of the five keywords' collocates. In the sequence kalam dalla, for instance, dalla is not modifying kalam, since the form is kalam-ma 'in the Land'. dalla stands out in another way as well, since the wider collocation very often is ud-/utu-gin7 dalla e3 ≅ "to appear brightly in the manner of daylight/the sun god"(?).

At first glance, the five lexemes investigated seem to modify slightly different things, which lend support to the argument that they, and especially babbar, dadag, and zalag, should be seen as different lexemes. More interesting to follow up is perhaps the instances where the lexemes observed modify the same head, e.g. cag4 kug vs. cag4 zalag. Do the two combinations convey the same meaning? And what about ki dadag vs. ki kug?

Another observation we can make is that inim (words) and cir3 (songs) are modified by kug only, perhaps indicating a sense nuance of kug not present in the other lexemes. Places to be (sit, dwell) such as barag and ki-tuc are also modified by kug, while concepts associated with liquid stuff (a, cu-luh) occur with dadag, and materials, e.g. erin and tug2, with babbar.

The reason for bringing up the English words beaming, gleaming, etc. above was not purely accidental. Looking at the nouns occurring near these English words, one gets a picture similar to the one for the five Sumerian lexemes, where some of the associated words overlap while others do not. Thus, we find that smiles and faces are beaming, eyes and faces are gleaming, eyes, prizes, and careers are glittering, eyes and faces are glowing, the sun, eyes, and light are shining, and the sun and eyes are luminous.

An exercise such as the one performed here can be done for any set of (synonymous) Sumerian words or lexemes to try and pinpoint their meanings or tease out similarities and differences between sets of semantically related words.


Sumerian scribe

© Copyright 2003, 2004, 2005, 2006 The ETCSL project, Oriental Institute, University of Oxford

University of Oxford