1 Introduction
Along with rapid technological developments in other fields over the last decades, lexicography has undergone dramatic changes. Most importantly, there has been a clear shift from print to the digital medium. Electronic dictionaries (EDs) are a relatively modern invention, defined in broad terms by Nesi (Reference Nesi2000: 839) as:
any reference material stored in electronic form that gives information about spelling, meaning, or use of words. Thus, a spell checker in a word-processing program, a device that scans and translates printed words, a glossary for on-line teaching materials, or an electronic version of a respected hard-copy dictionary are all EDs of a sort, characterised by the same system of storage and retrieval.
The best-known classification of digital dictionaries is that given in De Schryver (Reference De Schryver2003), but it pre-dates the current explosion of online dictionaries, which may be characterized as dictionaries that cannot be accessed without an active internet connection (Müller-Spitzer, Koplenig & Töpel, Reference Müller-Spitzer, Koplenig and Töpel2012: 430).
Dictionaries are an important resource for foreign language learners. With the ongoing digital revolution in lexicography, language learners have largely shifted to digital dictionaries, primarily online (L’Homme & Cormier, Reference L’Homme and Cormier2014; Lew & De Schryver, Reference Lew and De Schryver2014). According to a recent large-scale survey, digital dictionaries are the single most important resource for language learners (Levy & Steel, Reference Levy and Steel2015).
Online dictionaries come in different shapes and sizes, and many are of dubious quality. Unlike authoritative monolingual dictionaries for learners of English, for which an ad-supported business model is viable given the scale effect of millions of users worldwide, quality bilingual dictionaries for a combination of English with a lesser-used language can rarely afford to make the contents available freely online. As many users of online dictionaries are not willing to pay for their content, they are stuck with the free offerings. However, the average language learner is unable to competently assess the quality of the available lexical resources, many of which are poor lexicographically (Lew, Reference Lew2014).
Despite the enormous popularity of online dictionaries, there have been very few attempts to evaluate them systematically (Lew, Reference Lew2011; Meyer & Gurevych, Reference Meyer and Gurevych2012; Pearsons & Nichols, Reference Pearsons and Nichols2013; Yamada, Reference Yamada2010). The few studies that are available seem to be concerned with monolingual English dictionaries, completely disregarding bilingual dictionaries, even though the latter are known to be preferred by learners (Lew & Adamska-Sałaciak, Reference Lew and Adamska-Sałaciak2015). Lew (Reference Lew2011) provides an overview of selected monolingual online English dictionaries, categorized and discussed in terms of the major lexicographic issues. Meyer and Gurevych (Reference Meyer and Gurevych2012) offer a detailed description of Wiktionary as a popular collaborative dictionary. Yamada (Reference Yamada2013) is an evaluation of online dictionaries of English using the framework elaborated in Pearsons and Nichols (Reference Pearsons and Nichols2013). Yet, as far as we are aware, no comparable analysis has ever been attempted for online bilingual dictionaries.
This study presents a proposal for an evaluation framework for online bilingual dictionaries. The framework can be applied directly in the evaluation of dictionaries with English as the source language (i.e. going from English to another language); for dictionaries with another source language, a different set of test items would have to be used, specific to that language, though it can be created on analogous principles. The proposed framework is then applied to those English-Polish online dictionaries that are most popular among Polish learners of English. In order to establish a list of the dictionaries that learners use most frequently, an online survey was conducted involving Polish learners of English at different levels of language proficiency.
Any evaluation needs criteria. We take as a starting point an existing evaluation framework for online monolingual English dictionaries (Pearsons & Nichols, Reference Pearsons and Nichols2013) and construct a framework for online bilingual dictionaries with English as a source language. The framework addresses four fundamental lexicographic areas: coverage of lexical items, their treatment in the entries, access to lexical information, and presentation of lexicographic data. The first two areas are medium-independent: they apply equally to print and digital dictionaries. By contrast, access and presentation are areas that operate differently for print and digital dictionaries, and have benefited the most from the digital revolution (Lew & De Schryver, Reference Lew and De Schryver2014). The criteria for these two areas will be developed on the basis of recent findings on the effectiveness and user-friendliness of various features of digital dictionaries.
In the immediately following section, we report on an online survey employed to establish the most popular English-Polish online dictionaries. The top six of these are subjected to a systematic expert evaluation. The framework for this evaluation is given in Section 4 (evaluation framework), followed by the presentation of evaluation results and lexicographic recommendations.
2 Selection of dictionaries for evaluation
2.1 Method
In order to establish which English-Polish online dictionaries were the most popular among Polish learners of English, a questionnaire was designed on the online survey platform SurveyGizmo. An online format was selected for two reasons. First, it enabled us to use advanced features unavailable in pencil-and-paper format, such as drag-and-drop ranking, question piping and looping. These features helped improve the reliability and validity of the survey. The second reason was that an online survey was the natural choice when trying to reach an audience of online dictionary users. As the survey was intended for Polish learners of English, it was drafted in Polish. The questionnaire was piloted on a group of ten volunteers, and, after some improvements, made available to the participants (see Section 2.2). It went live between 2 and 16 March, 2014.
The questionnaire probed several aspects of online dictionary use. The two items that are most relevant in the present context concern the online dictionaries most used by participants, and their ranking of quality of these dictionaries. In addition, the questionnaire collected the following information about every single dictionary indicated by a given participant but not reported here: the purposes for which participants used the dictionary, context of use, format and type of device used for accessing them, and their level of proficiency in English.
A key item in the questionnaire inquired about online dictionaries most often used by the participants. They were presented with a clickable list of eleven dictionaries which were known to the researchers to be popular. This list had been constructed based on information gathered on a number of internet forums with substantial threads (at least ten responses) discussing recommended English-Polish online dictionaries, giving: Diki, bab.la, MEGAsłownik, DICT, Ling.pl, Getionary, Wiktionary, PONS, Angool.com, WordReference, and leksyka.pl. The above were listed explicitly within this questionnaire item.
Apart from this initial list of eleven dictionaries, the item had three extra boxes for participants to list up to three other dictionaries they used often that were not on the initial list. A participant could select as many dictionaries as they wished, combining the ones from the list with the ones supplied by the participants.
In the next step, the survey system presented a drag-and-drop ranking task with all (and only) the dictionaries ticked or added in the previous item by a given participant. Participants were here instructed to drag the dictionaries across to an empty box, placing the best at the top, and the worst at the bottom. This item allowed dictionaries to be rearranged at will until the participant was satisfied with their choice and confirmed it. There was no obligation to include every single one of the dictionaries they had selected or supplied in the ranking.
2.2 Participants
Participants were recruited online via snowball sampling. A group of initial respondents were asked to nominate, through their social networks (Facebook, email), other participants who met the eligibility criteria: participants had to be Polish learners of English, regardless of their language proficiency, aged between 18 and 25. Most of them were students of different specialisms who studied English either as a major or minor subject.
The online survey invitation included a clarification of objectives, explaining that participation was voluntary and it was possible to withdraw at any time. Participants were informed that by returning the questionnaires they were consenting to participation in the study and their responses being analysed and published, and that the survey would be anonymous. They provided written consent online by ticking a relevant box. The survey was fully anonymous and no sensitive personal data were collected (except possibly their IP addresses, HTTP Referrer, and HTTP client data, which were subsequently removed from the data file). Overall, there were 205 participants. Of these, 187 responses were used for the analysis.
The breakdown of the participants’ level of proficiency (self-declared in the online questionnaire) was as follows, based on the 180 responses returned for this item (not all participants provided a response on this optional item). Half of the participants were advanced or proficient, with declared CEFR levels C1 (48) or C2 (43). About 40% represented (upper-)intermediate proficiency (B1: 29; B2: 42). Only 10% were beginners (A2: 10; A1: 8). This should not be surprising, as most Polish learners within this age bracket now represent relatively high levels of proficiency in English.
3 Results
3.1 Dictionaries most frequently used
Table 1 presents the distribution of the participants’ choices as to which dictionaries they used most often. A dozen others (not in the table) received single nominations: Collins, Cambridge, linguee, infobot, glosbe, mtranslator, medland, translator, Oxford, Wikipedia, tłumacz.onet, thefreedictionary.
Table 1 Dictionaries most frequently used (multiple answers were possible)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab1.gif?pub-status=live)
It may be noted that not all of the frequently named resources had been supplied in the initial list of eleven dictionaries. Participants reported frequent use of machine-translation services, notably Google translate, with as many as 24 participants adding it under the other category. No doubt, the use of Google translate as a language resource is much higher; it can hardly be classified as an online dictionary, though, so it was not included in the initial list (and likely quite a few other participants did not report its use for this very reason). Other translation services also received single mentions, as did some monolingual dictionaries.
Based on these results, the seven most popular dictionaries were preselected for inclusion in the evaluation.
3.2 Dictionaries most highly ranked
Participants were also asked to rank the dictionaries they named (or a subset thereof) in terms of quality. We had expected the results to largely overlap with the dictionaries most frequently used (as it is rational to tend to use the tools you value highly). This expectation was confirmed, as the set of top seven dictionaries turned out to be the same in both cases. However, there was one notable difference: MEGAsłownik, which was by far the most popular dictionary among the participants, was not ranked as the best, but the second best. The number one dictionary according to the participants’ evaluation was Diki, the second most popular. The ranking of the remaining dictionaries, however, corresponded to their declared frequency of use, i.e. the dictionaries reported to be used more often were ranked more highly.
3.3 Final list of dictionaries selected for evaluation
MEGAsłownik unexpectedly closed down shortly after the survey was conducted. Accordingly, the list of the dictionaries for evaluation was reduced to the remaining six: Diki (http://www.diki.pl), bab.la (http://bab.la), Ling.pl (http://ling.pl),Footnote 1 PONS (http://pl.pons.com/t%C5%82umaczenie?l=enpl), Wiktionary (http://pl.wiktionary.org), and WordReference (http://www.wordreference.com/enpl/).
4 Evaluation framework
4.1 Criteria for coverage and treatment
The criteria selected for the evaluation of coverage of lexical items and lexicographic treatment were partially based on the framework proposed in Pearsons and Nichols (Reference Pearsons and Nichols2013). As in the original framework, each criterion was marked on an integer scale of 1 to 5, with verbal labels for these being as follows: 1: Unsatisfactory; 2: Partly acceptable; 3: Mostly satisfactory; 4: Satisfactory; 5: Very satisfactory. However, as most of the original criteria were somewhat vague, in an effort to make the evaluation more objective, for most of the original items, we added four additional, more detailed sub-criteria, each worth a single point. Thus, if a dictionary met none of the four sub-criteria, it received the lowest mark of 1 for a given main criterion; when a single detailed criterion was met, a score of 2 was awarded; two criteria earned a score of 3; three criteria, 4; and all criteria, 5.
For the coverage and presence of labels criteria, the proportion of features covered in a given dictionary was computed and converted to scores as follows: a mark of 5 was given to 80–100% coverage or presence of labels; 4 to 60–79%; 3 to 40–59%; 2 to 20–39%; 1 to 0–19%. The items used in evaluating coverage and labels are given in Sections 4.1.1 and 4.1.2 below.
The list of areas, main criteria, and sub-criteria used in the evaluation of coverage and treatment was as follows:
-
I Coverage
-
1. Coverage of entries (whether a certain item was present in a dictionary)
-
a) neologisms
-
b) specialized information technology (IT) vocabulary
-
c) multi-word expressions with frequent verbs
-
-
-
II Treatment
-
2. Presence of labels for
-
-
a) usage level (e.g. formal, informal, technical, literary, slang)
-
b) regional variety (British/American)
-
c) part of speech
-
3. Hyperlinked cross-references from
-
-
a) equivalents
-
b) examples
-
c) synonyms
-
d) other
-
4. Pronunciation indication (audio, transcription)
-
5. Provision of example sentences
-
6. Additional information on
-
-
a) usage
-
b) synonyms
-
c) antonyms
-
d) related words and phrases
-
7. Grammar and collocation
-
-
a) irregular forms of verbs
-
b) noun countability
-
c) comparative and superlative forms of irregular adjectives
-
d) collocations
4.1.1 Items used in the evaluation of coverage
Thirty vocabulary items were used to assess the coverage of entries. These items were selected from within three categories: neologisms, specialized vocabulary in the field of IT, and multi-word expressions containing frequent verbs.
The rationale behind using neologisms was to have a measure of how up-to-date the headword lists were. Even if some of them would not survive, their inclusion was thought to be a valid measure of an effective inclusion policy for an online dictionary. The ten items were selected randomly from among neologisms added to Oxford Dictionaries Online in 2013: babymoon, bankster, bitcoin, buzzworthy, dappy, fauxhawk, selfie, twerk, phablet and omnishambles.
The specialized IT items were selected at random from the glossaries on information technology and computer science on the website http://whatis.techtarget.com: adware, cache, firewall, spreadsheet, router, crowdsourcing, failover, scareware, freemium, and interface.
The ten test multi-word expressions were selected from Longman Dictionary of Contemporary English Online, Oxford Dictionaries Online, and Macmillan English Dictionary Online. The idea was to have less frequent expressions containing very frequent verbs. The items selected were: to take exception to, to make a beeline for, to pass the time of day with sb, to put sth behind you, to pull rank on sb, to go all out to do sth, to bring sb up short, to set the seal on sth, to go by the board and to try one’s hand at. Each of the 30 items was looked up manually in the six dictionaries and its presence or absence noted.
4.1.2 Items used in the evaluation of labelling
The ten items for assessing the inclusion of appropriate usage labelling in the dictionaries were adopted from an instructional section in Oxford Learners Dictionaries Online. In this section, the use of labels in the dictionaries was explained with representative examples and these examples were adopted for the present analysis. They included: two formal words (admonish, besmirch); two informal words (bonkers, dodgy); two technical items (accretion, adipose); two literary words (aflame, halcyon); and two slang words (dingbat, dosh).
The ten items for assessing the presence of regional variety labels were adapted from Leaney (Reference Leaney2007), and they consisted of five British English words in specific senses: flat, mark, (bank)note, pavement, petrol, and their American English counterparts: apartment, grade, bill, sidewalk, gas.
The ten items that were used to assess the presence of part of speech labels were also adapted from Leaney (Reference Leaney2007): allow, accent, commit, contract, damage, export, flower, forgive, import, level. Inflected verb forms of these items were also used in assessing the retrieval of inflected forms. Each of the thirty items was looked up manually in the six dictionaries and its presence or absence noted.
4.1.3 Items for the remaining treatment criteria
The assessment of the presence of the remaining elements in the dictionaries did not call for dedicated item lists. Therefore, for this purpose a joint list of ten words was drawn up, based on the frequency lists supplied with the Corpus of Contemporary American English (Davies, Reference Davies2008–). The most frequent words from the four major syntactic classes were selected as follows:
-
∙ verbs: be, have, do
-
∙ nouns: time, year, people
-
∙ adjectives: other, new
-
∙ adverbs: up, so
In addition, the following irregular adjectives were used to check for inclusion of their irregular comparative and superlative forms: good, bad, many.
4.2 Criteria for access and presentation
The criteria for the evaluation of access were largely based on Lew (Reference Lew2012) and they included:
-
III Access
-
8. Headword identification
-
-
a) accessing inflected forms (reducing inflected forms to the lemma)
-
b) accessing misspelled words (the “did you mean” function)
-
9. Accessing multi-word units (separate headwords for multi-word expressions)
-
10. Type-ahead search
-
11. Entry navigation devices (entry menus or signposts helping users in selecting the relevant sense)
-
The criteria for the evaluation of presentation were derived from De Schryver (Reference De Schryver2003), Dziemianko (Reference Dziemianko2011, Reference Dziemianko2012), Lew (Reference Lew2012), Lew, Grzelak and Leszkowicz (Reference Lew, Grzelak and Leszkowicz2013), Nielsen (Reference Nielsen2008), and Pearsons and Nichols (Reference Pearsons and Nichols2013), and they consisted of the following items:
-
IV Presentation
-
12. Presence of pictorial illustrations, which have been found to be helpful in dictionary consultation (Lew, Reference Lew2012)
-
13. Consistent entry form (i.e. inclusion of the same microstructural elements in consistent order and presentation), which has been claimed to make dictionary look-up process easier (Nielsen, Reference Nielsen2008: 177)
-
14. Full names of grammatical codes and symbols given in the user’s mother tongue (De Schryver, Reference De Schryver2003)
-
15. Use of bold type other than in the headword or equivalent (Lew, Grzelak & Leszkowicz, Reference Lew, Grzelak and Leszkowicz2013)
-
16. No intrusive advertisements, which were found to impede dictionary consultation (Dziemianko, Reference Dziemianko2011, Reference Dziemianko2012); this criterion also featured in Pearsons and Nichols (Reference Pearsons and Nichols2013)
-
In addition, customizability of the interface (De Schryver, Reference De Schryver2003) was also considered as a potential criterion, but, as it turned out, none of the dictionaries under evaluation offered such an option, therefore this criterion was dropped. For the same reason, we did not include another potential criterion: the use of colours for distinguishing various elements of the microstructure, as advocated in Dziemianko (Reference Dziemianko2015).
In evaluating access and presentation, to test whether specific elements were present in the dictionaries, generally the same test items were used as in the evaluation of treatment, but with a few variations. In testing the two criteria related to headword identification, in the inflected form search test, inflected forms were used rather than lemma forms; for example, for the verb do, the inflected forms did and done were searched for. In testing for the presence of the “did you mean” function (Lew & Mitton, Reference Lew and Mitton2011, Reference Lew and Mitton2013), deliberate single-letter misspellings were used, thus haeve was entered to test for have, etc. In testing whether multi-word expressions could be accessed directly, two well-known idioms were used whose presence in all six dictionaries had previously been verified: a piece of cake and kick the bucket. Finally, to test for the presence of pictures, frequent concrete nouns were used (cat, dog, chair), as such entries were most likely to feature pictures (Stein, Reference Stein1991) if they were at all offered in the dictionary.
The scoring system applied in the evaluation of access and presentation was similar to that used for coverage and treatment, with a scale ranging from 1 (unsatisfactory) to 5 (satisfactory). One score was given for access and another one for presentation, reflecting the number of features present in a given category: if all five features tested were present in a particular category, a score of 5 was awarded; four features earned a dictionary a score of 4, etc. If only one or no features were found in a dictionary, a score of 1 (unsatisfactory) was entered.
In general, when a given feature was not fully consistent across the test items, the mark reflected the situation for the majority of the items. For example, if at least six out of ten items checked included collocational information in a dictionary, a point was awarded for this feature. These points were awarded even if the quality of the lexicographic information was sometimes in doubt, but such cases are noted.
4.3 Evaluation procedure
The evaluation of the six online English-Polish dictionaries was conducted manually, independently by two experts on lexicography (the two co-authors), with all marks entered in parallel Excel workbooks. On completion, the two datasets were compared. All detailed marks by the two experts turned out to be in complete agreement (an interrater agreement of 100%). The results of the evaluation will be presented below. For data analysis, we used Microsoft Excel 2013 and R version 3.2.3 (R Core Team, 2015).
5 Evaluation results and discussion
5.1 Coverage
The coverage of entries from the three vocabulary categories was examined. The results are presented in the tables below. A +/– marking system was applied, where “–” means that a particular entry was not present in a dictionary and “+” means that an entry was present in a dictionary.
The coverage of neologisms was not comprehensive in any of the dictionaries (Table 2). Four out of ten entries were covered in Diki and Wiktionary, three in bab.la, two in PONS and none at all in Ling.pl and WordReference.
Table 2 Coverage of neologisms
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab2.gif?pub-status=live)
Coverage of IT terms was rather better (Table 3). Nine out of ten technical vocabulary items related to information technology were covered in Diki, eight in bab.la, and seven in Wiktionary. Five items were found in WordReference, four in Ling.pl, and three in PONS.
Table 3 Coverage of technical vocabulary
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab3.gif?pub-status=live)
Diki was unbeatable in its coverage of multi-word expressions (Table 4), scoring ten out of ten. Six phrases were found in Ling.pl, five in PONS, while bab.la and WordReference scored four. Wiktionary turned out to be the worst, as not a single one of the multi-word expressions tested was found there.
Table 4 Coverage of multi-word expressions
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab4.gif?pub-status=live)
5.2 Treatment
5.2.1 Labels
The inclusion of labels was checked for three types of labels: usage level, regional variety, and part of speech. We used the same scoring system as in the case of the coverage of entries (Section 5.1).
When it came to checking for usage labels (Table 5), labels in Polish were accepted (they are in fact preferable to the Polish user), as were variations in labelling, such as in marking levels of usage (informal versus slang, which is not always an easy call), or more domain-specific labels (e.g. astronomy rather than technical). Diki, bab.la, and Ling.pl got about half of the usage labels right, PONS and WordReference three out of ten, while Wiktionary did not score any marks, although the majority of the testing items were simply missing from the dictionary.
Table 5 Presence of usage labels (n/f=entry missing)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab5.gif?pub-status=live)
When it comes to regional variety labels (Table 6), Diki and PONS did very well, with 10 and 9 regional labels present, respectively. The other dictionaries scored between 3 and 5; Wiktionary did the worst.
Table 6 Presence of regional variety labels
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab6.gif?pub-status=live)
Part of speech labels were present for all ten test items in all six dictionaries. Thus, all dictionaries received perfect marks on this sub-criterion. We omit a detailed table for this.
5.2.2 Hyperlinked cross-references
Inclusion of hyperlinked cross-references in the six dictionaries is summarized in Table 7. Diki included hyperlinked cross-references for related words, idioms and phrases, but not for individual words in senses and examples. Phonetic transcription of each word was hyperlinked to instructions on the International Phonetic Alphabet. Similarly, syntactic labels were hyperlinked to their extensions with explanation. There were also links to verb conjugation patterns.
Table 7 Presence of hyperlinked cross-references
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab7.gif?pub-status=live)
In bab.la, all Polish equivalents were hyperlinked to their entries in the Polish-English section of the dictionary. Within each entry there was a list of hyperlinked synonyms and related phrases, and the dictionary also supplied links to verb conjugation patterns. In Ling.pl, no cross-references could be found. The only words that were hyperlinked served as advertisements, leading to non-lexical commercial websites such as online shops. In PONS, equivalents, related phrases, as well as all words in examples were hyperlinked. In Wiktionary, each equivalent was hyperlinked, as well as synonyms, antonyms, related phrases and individual words in example sentences. There were also links to expansions of syntactic labels. WordReference included hyperlinks for every single word as well: equivalents, synonyms, related words and phrases, individual words in example sentences, and explanations of abbreviations used in the dictionary.
5.2.3 Indication of pronunciation
A summary of the extent of pronunciation indication in the six dictionaries is given in Table 8. Diki offered both transcription and audio recordings, though only for British English. In bab.la, no transcription was provided, but audio recordings were present for both English and Polish words. There was no clear indication as to which variety of English was presented. In Ling.pl, there were audio recordings (without an indication of the variety, but the recordings sounded like American English), and transcription for both British and American English, yet the location of the transcriptions was inconsistent, given that Ling.pl aggregated content from a number of individual dictionaries. PONS offered both transcriptions and audio recordings for Polish and English, including audio recordings for phrases, with a choice of British or American versions. In Wiktionary, there were both transcriptions and audio recordings. The inclusion of varietal forms was inconsistent across entries: some provided only the American English variety, while others had British, American and even Australian English. Finally, WordReference provided both transcription and audio recordings for both American and British English.
Table 8 Indication of pronunciation
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab8.gif?pub-status=live)
5.2.4 Example sentences
Example sentences (Table 9) were present in all six dictionaries, though the details of presentation varied. In Diki, each main sense of an entry had between one and three example sentences, translated in brackets and with audio representation available at a click. In addition, most test entries had an additional substantial set of over twenty examples available on request. These extra examples seemed to be drawn automatically from a corpus, and so not all were very well-suited to the entry. A translation option was offered for these examples, though these were clearly machine translations, of low accuracy and questionable benefit to language learners.
Table 9 Example sentences
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab9.gif?pub-status=live)
In bab.la, each sense was accompanied by a single example sentence, but when this sentence was clicked on, a further four sentences appeared. All the example sentences were translated and their Internet sources were identified.
The presentation of examples in Ling.pl again suffered due to the aggregator status of this resource, being highly inconsistent. The presence and extent of exemplifying material depended on which specific dictionary happened to be drawn on at a particular point. When examples did appear, they were usually translated. In PONS, there were only a few examples under some of the entries. Most were short phrases, with very few full-sentence examples. They were all translated into Polish. Wiktionary included at least one example per sense (thus over five senses in our test polysemous entries). All were translated into Polish, but no source was identified. WordReference, like Wiktionary, typically supplied a single example sentence for each sense, with no source given. Most of the examples were accompanied by Polish translations, except when the examples illustrated multi-word expressions.
5.2.5 Additional information
Additional lexical information evaluated was as follows: usage notes, antonyms, synonyms, and related words and phrases (note that it is the presence of these information categories that is at issue here, while in Section 5.2.2 above it is the presence of the hyperlink that matters.) A summary of the results is given in Table 10.
Table 10 Provision of additional information
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab10.gif?pub-status=live)
In Diki, most entries included a list of hyperlinked related words and phrases in a sidebar. Some senses were also accompanied by hyperlinked synonyms and antonyms below the Polish equivalent, but no notes on usage were available. In bab.la, each entry was accompanied by synonyms and related phrases, but there were no antonyms or usage notes. In addition, forum posts with additional information were available. Ling.pl did not list any additional information. A minority of entries (depending on which specific dictionary they came from) included related phrases within an entry, but they were not specifically marked. In PONS, synonyms were included as sense indicators and related phrases were listed. Antonyms and usage notes were not present. Wiktionary was quite rich in additional information. Synonyms, antonyms, related phrases and usage notes were all present for numerous entries. Wiktionary was the only dictionary to score a point on each of the four sub-criteria. In WordReference, related phrases were present and some entries had accompanying usage notes. Synonyms could be called up by clicking on “Angielskie synonimy” (‘English synonyms’) below the entry word, which directed the user to an English thesaurus for a particular word. Antonyms were not available in this dictionary.
5.2.6 Grammar and collocation
A summary of the results for information on grammar and collocation is given in Table 11. Grammatical information in Diki covered irregular verb forms, noun countability, and comparative and superlative forms of irregular adjectives. Some entries supplied collocational information (in collapsed view by default but available on request), which consisted of large sets of collocations classified by part of speech. These sets appeared to be automatically extracted (so they were not always fully accurate), and supplied with machine translations into Polish. In bab.la, only irregular forms of verbs were found. No information on collocation, comparative and superlative forms of irregular adjectives, or noun countability was given. In Ling.pl, irregular forms of verbs, noun countability, and comparative and superlative forms of irregular adjectives were all present, but collocational information was not available. PONS covered all the features under evaluation, i.e. irregular forms of verbs, noun countability, comparative and superlative forms of irregular adjectives as well as collocation. In Wiktionary, grammatical information was quite exhaustive, with irregular forms, noun countability, comparative and superlative forms all given, along with collocational information.
Table 11 Provision of grammatical and collocational information
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab11.gif?pub-status=live)
5.3 Summary of the results for coverage and treatment
The areas of coverage and treatment both relate to the content of dictionaries, as opposed to what is done with that content. Therefore, we present here a summary of the evaluation in these two areas. Table 12 lists the scores for all coverage and treatment criteria (see Section 4.1 for details on how the scores were awarded). The results indicate that the best overall score was received by Diki: 4.0 out of 5. Wiktionary scored 3.5, PONS 3.4, WordReference 3.3, bab.la 3.1, and Ling.pl ended up with the lowest score of 2.6. Of course, we should not be forgetting that these are just the most used and most liked dictionaries: an elite, as it were, of the freely available English-Polish online bilingual dictionaries.
Table 12 Scores for coverage and treatment
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab12.gif?pub-status=live)
The evaluation so far has focused on the traditional lexicographic criteria of coverage and treatment. Let us now proceed to two further aspects that can take advantage of the specific possibilities afforded by the digital medium: access and presentation.
5.4 Access
Access was evaluated with the use of the five detailed criteria listed in Section 4.2; these are repeated in the first column of Table 13, which gives a summary of the results for access. In Diki, inflected forms did lead to their lemmas correctly, though the dictionary did not immediately redirect the user to the full entry. Diki featured a “did you mean” function, displaying a list of suggestions for misspelled searches. Multi-word units were accessible directly and type-ahead search was present. No entry navigation devices were available.
Table 13 Access
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab13.gif?pub-status=live)
Bab.la recognized inflected forms correctly, with automatic redirection to their parent lemmas. Misspelled words were also identified with a list of suggestions. Multi-word expressions could be accessed directly, and an incremental search function was available. Bab.la did not contain entry navigation devices.
Ling.pl did not recognize inflected forms. This dictionary resembled a print dictionary in that there were traditional cross-references to citation forms (e.g. “gone → zob. go”) that were not hyperlinked, so a user had to again type “go” in the search box. Ling.pl did not feature a “did you mean” function, so misspelled words were not found at all. No entry navigation devices were featured. However, a type-ahead search function was available, as were separate entries for multi-word expressions.
In PONS, most inflected forms were immediately displayed as their lemmas, yet, in the case of irregular verbs such as gone, there was a note that this was the past participle of the citation form (go), but without a hyperlink. Still, a hyperlink to the lemma could be found in the “see also” box, located below the grammatical information section. The dictionary was able to identify misspellings as well as direct multi-word expression searches, and featured an incremental search function. In addition, PONS included entry navigation devices in the form of synonyms or short glosses in brackets next to each sense, thus scoring a maximum of 5.
Wiktionary was highly inconsistent when it comes to identifying inflected forms. It covered mostly inflected forms of irregular verbs and forms representing lemmata of more than one syntactic class, but it usually failed on regular inflected forms. The dictionary did not accommodate misspellings, and no entry navigation devices were present. On the positive side, multi-word expressions could be accessed directly and there was a type-ahead search function.
WordReference recognized all inflected forms and automatically redirected the user to the relevant entries. It accommodated misspellings, multi-word expression searches, and had type-ahead search capability as well as sense signposts as entry navigation devices. With a score of 5, WordReference tied with PONS in respect of access features.
5.5 Presentation
A summary of the results for the presentation aspect of the evaluation is given in Table 14. Diki included pictorial illustrations. Presentation was consistent across entries. Grammatical codes and symbols were given in full and in Polish. Bold type was used for translation equivalents, and there were no distracting advertisements.
Table 14 Presentation
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab14.gif?pub-status=live)
Bab.la did not feature pictorial illustrations. Entry form was consistent. Abbreviations were given in Polish, but their expansions were not available. Bold type was used to highlight tokens of headwords within examples. In bab.la there were quite intrusive ads.
In Ling.pl, no pictures were present. The resource could not be credited with entry consistency, as each “entry” brought together original entries from a number of dictionaries, and in each case the number and order of entries presented was different. Abbreviations went unexplained and were mostly given in English. A big banner ad was displayed at the top of the screen. On a positive note, bold type was used for some microstructural elements, although this again was inconsistent due to the aggregator character of the dictionary.
PONS did not include any pictures, either. Abbreviations were in Polish, but their expansions were not given. Entry form was consistent and there were numerous bold elements in this dictionary. The page included distracting advertisements.
Wiktionary contained pictures in certain cases and each entry looked the same. There were full names of grammatical codes and symbols and they were provided in the users’ mother tongue. There were elements in bold and there were no adverts.
WordReference did not include any pictures or provide full names for abbreviations, which were in English. Entry form was consistent, but there were no bold elements other than the entry word, and there were distracting adverts.
5.6 Overall scores
Table 15 brings together evaluation scores in all four areas for the six dictionaries, with Figure 1 presenting these results visually. Overall scores reflecting all four areas evaluated are given in the bottom row of the table, assuming equal weight given to each of the four areas. The dictionary that has emerged ahead of the competition is Diki, with an overall score of 4.3. The scores of the next four dictionaries, Wiktionary, PONS, bab.la, and WordReference are quite close to each other (3.3, 3.2, 3.0, and 2.9, respectively). Ling.pl has received the lowest score of the six at 2.0. It has to be stressed that this should not be taken to mean that Ling.pl is an awful dictionary, as the dictionaries selected for evaluation are the top six from a larger set.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20170330142509-54785-mediumThumb-S0958344016000252_fig1g.jpg?pub-status=live)
Fig. 1 Summary of the scores for the four areas evaluated
Table 15 Summary of the scores for the four areas evaluated
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab15.gif?pub-status=live)
Access and presentation are two areas dependent on the digital medium. Still, the distribution of evaluation scores in these two areas turns out to be fairly similar to that in the more traditional lexicographic aspects of coverage and treatment. The two dictionaries with perfect scores for presentation are Diki and Wiktionary, whereas access received maximum possible scores in PONS and WordReference.
5.7 Evaluation-based ranks versus user survey ranks
Taking into account overall scores, the final ranking of the dictionaries is as given in the evaluation rank column of Table 16. These ranks may be compared with those emerging from the online questionnaire reports by dictionary users so as to assess the degree of overlap between expert evaluation and users’ evaluation ranks.
Table 16 Final evaluation scores, evaluation-based ranks, and user survey ranks compared
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20170330142132096-0128:S0958344016000252:S0958344016000252_tab16.gif?pub-status=live)
The correlation between evaluation ranks and survey-derived ranks turns out to be low (Spearman’s ρ=0.31, p=0.54, n.s.), suggesting that the choices made by participants tend to be motivated by factors other than lexicographic quality. Nevertheless, the top dictionary in the survey overlaps with the best dictionary in the expert evaluation. Perhaps this means that most users are able to recognize clear differences in quality, though it could well be a coincidence: we do not have enough data points to make this kind of determination with confidence.
6 Lexicographic recommendations
The majority of the dictionaries investigated in the present study appear to be quite good. The scores that prevailed were “mostly satisfactory” or “satisfactory”. Ling.pl was the exception, with scores gravitating towards “partly acceptable”. Diki was a clear winner, with the average score exceeding the “satisfactory” grade. The four remaining dictionaries received similar scores between 2.9 and 3.3, which corresponds to a label of “satisfactory”.
Diki had problems with its coverage of neologisms (“partly acceptable”), and the editors would do well to extend the coverage in this respect. Minor problems were also observed in its usage-level labelling: this aspect should be revised and the labelling applied more consistently. With regard to cross-referencing, a greater number of microstructural elements should be hyperlinked, such as individual words in example sentences so that users can get help when they run into unknown vocabulary in an example. When it comes to pronunciation, offering transcription and audio recordings in American English along with the current British English would be an improvement. Usage notes would alert Polish learners of English to specific problematic issues. Diki might also consider introducing entry navigation devices for easier navigation between the senses.
Wiktionary was ranked in second place overall, but it failed in the coverage of multi-word expressions and usage-level labelling. This should not be surprising, given that this dictionary is edited collectively by enthusiastic amateurs (Lew, Reference Lew2014; Meyer & Gurevych, Reference Meyer and Gurevych2012). Its coverage of neologisms could also be improved. Multi-word expressions should be discoverable independently of the component lexemes. Absence of words that should be labelled may indicate that the overall coverage of less central items is poor. Users of Wikipedia who tried to search for inflected forms or misspelled their target items would be at a loss. Finally, entry navigation devices ought to be introduced.
The third best dictionary, PONS, was unsuccessful in its coverage of neologisms and only partly acceptable in the coverage of technical vocabulary and multi-word expressions, presence of usage-level labels, provision of example sentences, and presentation. The scope and accuracy of usage-level labels should be improved. The dictionary should expand its exemplification material. Usage notes would be a welcome addition, as would pictorial illustrations. Expansions of grammatical codes and symbols should be provided, and advertising toned down.
Bab.la’s coverage of neologisms and multi-word expressions could be improved on, along with labelling. Individual words in examples ought to be hyperlinked and the provision of pronunciation information extended by adding transcription and the British variety (to match the British flag symbol displayed in the dictionary). Usage notes and collocational information would be welcome. Grammatical information is only rudimentary, with missing noun countability indication and comparative/superlative forms of irregular adjectives. The dictionary would benefit from pictorial illustrations and entry navigation aids. Advertisements should be less intrusive.
WordReference should improve its coverage of entries in general, as the coverage in all three categories of items examined in the study left much to be desired. As in the other dictionaries, labelling ought to be improved on. Expansions of grammatical codes should be given, along with information on noun countability and comparative and superlative forms of irregular adjectives. Advertisements could be made less intrusive, while pictorial illustrations would be a welcome addition.
Many of the problems evident in Ling.pl are a consequence of its adopted aggregator model, resulting in redundant and inconsistent presentation that is difficult to navigate and lacking in cross-references and example sentences. Even though the resource offers access to several dozen dictionaries, its coverage of vocabulary was not fully satisfactory. The dictionary fails to find inflected or misspelled forms.
To sum up, the most persistent weakness across the dictionaries evaluated was their coverage of neologisms and absence of usage-level labels. The only criterion on which all the dictionaries were very successful was part-of-speech labelling.
Going beyond the strict list of criteria, bilingual online dictionaries would do well to improve the discoverability of multi-word units, since language learners do not always immediately see such units for what they are on encountering them in a text. Doing so would discourage simplistic word-for-word translation and promote the idiom-principle approach. Further, dictionary users themselves could be engaged more fully in commercial projects, by allowing user comments and discussions to enhance entries (such as bab.la’s user forum). These, however, should remain clearly distinguished from the ‘professional’ core content.
7 Conclusion
A general aim of the present study was to propose a comprehensive framework for the evaluation of online bilingual dictionaries, building on the work of Pearsons and Nichols (Reference Pearsons and Nichols2013) for monolingual dictionaries of English, supplemented with findings from other relevant publications. Although in this study the framework has been applied in practice to English-Polish dictionaries, it is directly applicable to any online bilingual dictionary with English as the source language, and would require only minor adaptation to serve other language pairs.
A more specific aim of the study was to apply the proposed framework to evaluate selected English-Polish dictionaries freely available online, so as to offer guidance to teachers and learners on the relative quality of cost-free online dictionaries for the growing proportion of younger learners. Of course, there do exist better, professionally edited dictionaries for Polish learners of English (Lew & Adamska-Sałaciak, Reference Lew and Adamska-Sałaciak2015), but they tend not to be available online, let alone free of charge. Perhaps the growing popularity of dictionary apps for small-screen devices will improve the situation in the near future, but for the moment, given that many young learners exhibit a strong preference for online resources available at no cost, we should at least offer them guidance as to which of the freely available resources are relatively better than others.
Overall, the best dictionary according to our criteria (Diki) turned out to be quite satisfactory. This was also the dictionary that our survey participants identified most frequently and valued most (if we ignore MEGAsłownik, which had gone out of business before the study was completed).
We hope that this study will inspire metalexicographers to carry out systematic analyses of online dictionaries, particularly bilingual dictionaries, as these are most often used by language learners. Such analyses could be of use to both language learners and teachers as a reliable indicator of which reference works can be trusted, especially when it comes to non-professional dictionaries.
Before attempting a replication, it may be advisable to reconsider some aspects of the present study with a view to improving them. First, the number of items tested in the evaluation of coverage was rather small, and the selection of categories somewhat arbitrary. A larger wordlist would increase the reliability of the evaluation framework as an instrument, though it might also limit the number of dictionaries that could be evaluated. Items might be sampled from several frequency bands, and the selection of neologisms informed by checking the candidates in a reliable monitor corpus. Running automated queries to test coverage might be considered, though the degree of success would here depend on the features of the specific dictionary interfaces. A more general aspect is the choice of criteria: with a different set (or weighting), the results might differ. It also has to be noted that the quality of equivalents was not evaluated – admittedly an important aspect in gauging the quality of bilingual dictionaries. Finally, the number of criteria used in evaluating coverage and treatment was larger than that for access and presentation. A more balanced distribution across these areas, or else a different approach to weighting the partial scores, might be considered.
Acknowledgements
This work was supported in part by the Polish National Science Centre (Narodowe Centrum Nauki), under grant DEC-2013/09/B/HS2/01125 to the first author.
The authors wish to express their gratitude to all the anonymous participants.
Supplementary material
For supplementary material referred to in this article, please visit https://doi.org/10.1017/S0958344016000252