Before I started my PhD project, I worked as a part-time translator for a reading club, and my major task was to translate British detective fiction into Chinese which would be sent to the small-scale, in-group “publication”. When I was asked about my work, some annoying people believed that the job was rather easy; to put it in their own words, “if you give an English-Chinese dictionary a typewriter, it will do all the work for you”, which, I think, is a prejudiced and unfair comment, not only on the translator, but also on the task of translation. When the nature of translation is finally presented to us, we will find that it is not a simple “word-matching” game. Actually, translation involves a more complicated mechanism which may cause great frustration, because we will finally get lost in translation no matter how hard we struggle. Today, I would like to name a few issues that are notable if we analyse “translation” from the perspective of semantics and pragmatics, and, due to my knowledge and preference, I will focus on the translation between English and Chinese.
The basic form of translation, if we follow the common belief, is to find a group of words in the target language that correspond to the words in the source language, and then form the sentence using all these words. That is also the basic practice of students who are starting to learn how to translate. The fundamental problem, then, comes from “finding the corresponding words”: Is there any degree of correspondence? For an ideal translation, we need to find the words between the two languages whose meanings are strictly equivalent to each other: they will have the same, fixed, unambiguous meaning, or, if the words are ambiguous, they need to be ambiguous “in the same way”, and the ambiguous meanings should form a one-to-one match between each other.
At this stage, polysemy can only make matters worse, so we should better limit our discussion to the words with “fixed, unambiguous” literal meaning. Not a lot of words hold a stable and unambiguous meaning in English; definitely we could name a few, but the number is rather small compared with the vocabulary of an average native speaker. One simple idea for the selection of such type of words is to refer to David Kaplan’s (1989) concept of character and content. Briefly, the character of a word is the “meaning” of a word which is set by the linguistic conventions, while the content is the “meaning” of a word which is determined by the character in a particular context. Following that definition, words with a stable character are suitable for our requirement. In Kaplan’s categorisation, that class only includes indexical expressions (e.g. I, here, now) and proper names (e.g. Chris Xia, University of Cambridge). If we extend the definition of proper names, we can also include a number of abstract scientific concepts, for instance helium and Higgs Boson, but the limitation is clearly shown: Only a limited number of words can have a very definite literal meaning across different languages, and we can only expect that other words will only have “roughly the same” counterparts in another language.
Even if we can successfully identify a set of words that can have strictly the same literal meaning (namely the same character) between English and Chinese, the situation is still more complex than we might expect. Take the pronouns as an example – pronouns are usually the perfect illustration of indexical terms, but the conventional use of pronoun differs between languages. Standard Chinese clearly differentiates two second person singular pronouns, ni and nin. Nin is the honorific form, which is generally used to address strangers and seniors, while ni is more casual and can be used between family members and friends. When a piece of Chinese text with the different use of ni and nin is to be translated into English, the translator can only reduce the two forms to a single word “you”, if she does not want to bother her readers with the archaic form “thou”. Although both “you” and ni/nin constantly refer to the person who is targeted in the conversation, the appearance of honorific form conveys additional information that will be lost in the translation. I do not intend to complain that English finally lost the plain form of second person singular pronoun, but it does create obstacles during the process of translation. Finding the strictly corresponding words is far away from an ideal readable translation, and that is also the reason that a strict literal translation never really exists.
Let’s take one step back and select some words that are “roughly the same” among different languages; that is much easier for the translators, and is the common practice by both professional and amateur translators. The notion of “roughly the same” already implies that the definition of a particular word in two languages is different, which indicates that the actual concept represented by the selected words can be slightly different across two languages. In the usual case, such a difference will not cause any comprehension problem, but it will still lead to miscomprehension or difficulty of understanding in the translation of certain text, and such miscomprehension is clearly language-related, or even culture-related. Imagine an excerpt of conversation in a British story book for kids: A mother said to her son who is extremely picky about his food, “you should eat more vegetables and fewer potatoes – here, eat some sweetcorn.” That utterance is perfectly natural in British English, because British people categorise potatoes as starchy food rather than a kind of vegetable, while sweetcorn is classified as vegetable. If, however, a translator follows the literal correspondence and translates this piece of advice into Chinese as “you should eat more shucai, and fewer tudou – here, eat some yumi“, the Chinese readers will surely be confused. Chinese language regards tudou (potatoes), but not yumi (sweet corn), as a member of shucai (vegetable), and how can a child eat more vegetable by eating more sweetcorn rather than potatoes? That does not make any sense in Chinese. The sense of the text is distorted by the mismatch between concepts, although the translator makes a perfect and conventional literal translation. The variance of concept encoding is another problem that translation can hardly solve, and this problem is definitely beyond the simple translation of literal meanings.
Even if we manage to overcome the difficulties of word selection and concept matching, some more complicated situations still await us. To give you a taste of such a situation, I would like to present a case that has happened to me before. When we discussed the conversational maxims proposed by Grice in my masters programme, one of my classmates, who is born and raised British, told us that “I am washing my hair” is a poorly-formed excuse if you want to reject someone’s invitation on the telephone. Later I described the setting of the conversation in Chinese to my friends from mainland China, “if you phone your friend to ask her whether she would like to go shopping with you on a sunny Saturday afternoon, and her reply is ‘I am washing my hair’, what would you think about her reply?” To my surprise, 80% of them told me, “That’s fine! She is asking me to wait for her, and we will go after she finishes washing her hair.” The difference in that example comes from the derivation of implicatures in different languages, which is totally beyond the words and sentence structure, and cannot be eliminated even if we carefully control the selection of words and the formation of a single sentence in the process of translation. The difference will still exist if you change the wording, unless the translator chooses to spell out the implicature.
Following my analysis, you may feel that an ideally faithful translation is almost a mission impossible; and trust me, it is. Since the emergence of literal meanings of words as well as, potentially, implicatures of utterances is conventional within a particular language, it is difficult to balance these conventional meanings cross-linguistically. Any form of translation will lead to the loss of some information, and the only difference is the form and the amount – either explicit information or implicit information, either more or less. Not just a matter of study for translation theories, this is also a long-term question in the field of second language comprehension and cross-linguistic pragmatics, and I believe that some relevant research may finally help to rescue this kind of loss in translation.
Kaplan, David. 1989. ‘Demonstratives’, in Joseph Almog, John Perry and Howard Wettstein (eds.), Themes from Kaplan (Oxford: Oxford University Press), pp. 481–563
Around 15 months ago, a small group of Linguistics PhD students found themselves in that most idea-and-optimism-inducing place, the local watering hole, and hatched a plan to enter the world of web logging (or blogging, to you and me). And exactly a year ago, Cam Lang Sci was born. As well as creating more of a local linguist community feel by finding out about each others’ interests, we wanted to share with you, our dear readers, some of the many facets of Language Science that make us tick. There’s been posts on sounds – and silent sounds; on words, verbs, and relative clauses; on languages dying and languages being born; on gender and language, nations and language, and thought and language. We’ve asked: What’s in a name? Where have all the implicatures gone? And, are you a Belieber? And in doing so, we’ve dipped our toes into many of the foundational areas of Linguistics: phonology, morphology, syntax, semantics, and pragmatics. And more: sociolinguistics, bilingualism, acquisition and historical linguistics. Even the practice of being a linguist itself. But there’s much more to Linguistics than all that! So we hope you’ll join us for another year of Cam Lang Sci, continue to be part of our growing readership (now almost a thousand visits a month) and in the meantime, take the chance today to sit back and catch up on some of those posts you may have missed.
Oh, and you can now follow us on facebook, too.
A number of years ago it was observed that there is more than one type of intransitive verb (i.e. of verbs describing an action, state or change where only one noun is involved – of which many examples to follow). This is true of many languages, including English.
Sometimes, the subjects of intransitive verbs behave like the subjects of transitive verbs (verbs with two nouns involved, as in Lucy plays football, Harry loves Ancient Egyptian etc.). For example, a typical property of the subject of a transitive verb is that it can be described with the verb plus the suffix -er, e.g.:
Some intransitive subjects can also be so described: worker, talker, swimmer, runner etc. But others can’t: we don’t say arriver “one who arrives”, dier “one who dies”, beer “one who is” and so forth …
But there is evidence that the subject of many of this latter group of intransitive verbs actually functions a bit like a transitive object. The object of transitive verbs can often be described with a form of the verb called the past participle, placed before the noun, for example:
Note that often an extra adjective or other modification is required for the sentence to make sense: the loved library is a bit of an odd thing to say but the much-loved library is fine; likewise we probably wouldn’t say the eaten hamburger but we might say the recently eaten hamburger or the half-eaten hamburger.
We can do a similar thing with some intransitive verbs:
This suggests the subjects of these verbs actually behave like the objects of transitives, as previously discussed. Note as well that these intransitive verbs are amongst those which can’t take the -er suffix, and furthermore that verbs which can take this suffix don’t allow the pre-noun past participle construction: we can’t say the talked man “the man who talked” or the swum woman “the woman who swam”.
This idea that some intransitive verbs at one level really have objects, instead of subjects, is for various complex reasons referred to as the unaccusative hypothesis. Of course, at another level these “objects” do behave like subjects: e.g. in a sentence they usually come before the verb; if they take the form of pronouns they have the subject forms I, he, she rather than the object forms me, him, her; etc. Thus we say I fell not fell me, and so on. There are also various complications (which I won’t go into here, because they are, well, complicated) which suggest the unaccusative hypothesis in its simplest form may be inadequate, and that some refinement is needed.
However, the hypothesis still usefully highlights a couple of things which seem to come up again and again in modern linguistics. Firstly, the way things appear “on the surface” may mask other properties. The noun Lucy looks like a subject in both Lucy talked and Lucy arrived, but in fact in the latter it seems also to have some “underlying” object properties which don’t show up so easily. Secondly, close inspection of the details of a language – including details which we might not think obviously relate to whatever it is we’re thinking about – can help reveal these underlying properties. It is this sort of close attention to detail that has allowed many of the advances in linguistics that have been made over the course of the last few decades. And even the complications I just alluded to may be useful here, in helping us come up with an even better theory.
What’s in a colour? Since more than 80% of information comes in through vision, the way we colour the world around us is quite important. Have you ever thought that the sky can be other than blue and the grass other than green in other languages? These colour categories seem universal, but in fact they are not.
Guy Deutscher finely described the problem in his book ‘Through the Language Glass. Why the World Looks Different in Other Languages’, but I will try to make a very brief overview of the linguistic ‘colour-issue’ and to mention some interesting points for further analysis.
In fact, it was William Gladstone who first noticed colour differences (at least, brought them to public attention) and set the stage for colour debate. In 1849 he published his work on Homer where he questioned why the ancient poet described the sea as wine-dark, honey as green, and sheep and iron as violet. Why were his skies never blue but iron or copper? These oddities cannot be blamed on Homer being blind or colour-blind, since other ancient Greek writers (along with the authors of the Indian Vedas, the Bible and early Chinese texts) shared this worldview. Gladstone conjectured that there was a universal anatomical deficiency in the ancient world which gradually evolved. But evolutionary studies proved that humans must have had the same degree of colour vision for millennia, which means that our vision is hardly different from Homer’s.
But if colour distinctions are not determined by anatomy, are they formed by language? Does language really determine or, at least, influence the way we colour the world around us? (as supposed by the Sapir-Whorf theory)
We tend to think of colour names in terms of our basic 11-colour paradigm. But it’s not typical of all the languages. For example, Russian has 2 words for blue (‘goluboy’ and ‘siniy’) distinguishing between light-blue and dark-blue. At the same time, pink is not considered as ‘basic’ colour in Russian, but rather as a very light hue of red. Polish also has two words for blue — ‘niebieski’ and ‘granatowy’ — but their semantics is different from Russian’s ‘goluboy’ and ‘siniy’. Japanese initially didn’t distinguish between blue and green having only one word for these hues — ‘aoi’ (which can be determined as blue with a far broader range of shades than in English). Nowadays, under the influence of the European tradition the semantics of the word has shifted towards our ‘classical’ blue. And green is now described with another word — ‘midori’. Nevertheless, the grass is still ‘aoi’ in Japanese, as well as a green traffic light. Some New Guinea Highland languages have terms only for ‘dark’ and ‘light’. Hanuno’o language, spoken in the Philippines, has only four basic colour words: black, white, red and green. Pirahã language, spoken by an Amazonian tribe, is said to have no fixed words for colours at all. According to Dan Everett, if you show them a red cup, they’re likely to say “This looks like blood”.
The issue of whether the colour spectrum is randomly carved up into categories, or whether there are universal constraints on where these categories form has long been the centre of linguistic, anthropological, psychological, and philosophical debate. Some interesting discoveries were made, the most famous of which is the one by Berlin and Kay (1969). They considered colour cognition as an innate, physiological process and discovered that colour words emerge in all languages in a predictable order. They identified eleven possible basic colour categories (white, black, red, green, yellow, blue, brown, purple, pink, orange, and grey) and found that the colours followed a specific evolutionary pattern. Black and white come first, then red, then yellow, then green and finally blue. Researchers tried to find explanations for this phenomenon in nature. Red is probably first because it is the colour of blood and of the easiest dyes to make in the wild. Green and yellow are the colours of plants. And blue is the last one because – with the exception of the sky – few things are blue in nature.
Though the theory was much critisized later on, it revolutionized and revived colour studies, which had stayed silent for almost 100 years. In the last few decades a whole number of experiments were carried out to find out whether speakers of ‘colour-deficient’ languages can see all the colours and distinguish between them.
The fact is, that we all see more or less the same. If asked to choose a lighter or a darker colour, most of us would do it properly. But a number of tests have shown that people can remember and sort coloured objects more easily if their language has a name for that colour. For instance, bilingual children of Senegal (French-Wolof) would distinguish between red and orange colours faster than monolingual Wolof children. Wolof language has only one word for these two hues, whereas French — two. And in some tests Russian speakers were faster at distinguishing certain shades of blue than English speakers (since Russian has 2 different terms for light and dark shades of blue, as has been already mentioned).
It goes without saying, that linguists are interested in describing the semantics of colour terms not just in one language, but in a range of different languages. Yet comparisons through translations (like ‘niebieski’ = blue; ‘siniy’ = blue; ‘aoi’ = blue) are absolutely unacceptable. The semantics, collocations and connotations of every word are unique.
It seems that it would make sense to look at colour categories from a cognitive perspective. The way people of different cultures set boundaries in spectrum and distinguish between certain hues depends on how they conceptualize colours, rather than on how they perceive them. Anna Wierzbicka argues that colour concepts are bound to certain universalities of human experience, such as day and night, sun, fire, vegetation, sky and earth. The number of colour terms may depend on how important it is for a certain culture to distinguish between them. For example, yellow and red hues are more relevant for Southern cultures (because of sun, sand, etc.) than blue, green or black, which can seem equal in their significance. Why invent three different words for a phenomenon conceptualized as one? Language strives for economy, hence the differences in the number of colour categories.
A curious afterthought to all this “colour-debate” is that colours are used in very different ways in different idioms across languages. For example, in English one argues until he is blue in the face, whereas in Russian he would definetely turn red. In English hair is grey, whereas in Russian it’s rather white.
And what would you say of this short advertisement?
Green bags available in seven colours.
(Placed at Cambridge University Press bookshop)
In this context it seems to acquire far more hidden meanings.
When Charles Darwin (eventually) published On the Origin of Species by Means of Natural Selection in 1859, one species on whose origins he remained deliberately quiet or, at most, vague was Homo sapiens, i.e. us. That humans had evolved from a ‘lower animal’ was profoundly controversial in Darwin’s time (I say was, but it remains controversial for many even now, as Richard Dawkins continually reminds his readers). One of the major difficulties lay in accounting for the differences between humans’ mental capacities and those of other animals, and one of the chief differences concerned the evolution of language.
Darwin addressed the evolution of Homo sapiens in The Descent of Man published in 1871. It’s quite a hefty work and dedicates all of about 10 pages to the evolution of language, but those pages are full of insights and observations and many of the ideas and conclusions were ahead of their time. I’ll run through some of them, but it is well worth reading the original!
Darwin made the crucial distinction between articulate language, which he said was peculiar to humans, and things such as cries, gestures, facial expressions, etc, which are found in many species besides humans. This distinction is often blurred, especially when people talk about the evolution of “communication” as opposed to the evolution of language.
It is also common to blur the distinction between speech and language, but Darwin was careful to separate the two, with language being primarily a mental capacity. Drawing on several observations, Darwin concluded that the defining aspect of language was not in “the understanding of articulate sounds”, nor in “the mere articulation”, nor even in “the mere capacity of connecting definite sounds with definite ideas”, all of which are found in some other species or other. Instead, “the lower animals differ from man solely in his almost infinitely larger power of associating together the most diversified sounds and ideas; and this obviously depends on the high development of his mental powers” (the aspects of language which are uniquely human and those which are not might nowadays be referred to as Faculty of Language in the Narrow (FLN) and Broad (FLB) Senses respectively (see Hauser, Chomsky & Fitch 2002)). Exactly what is meant by this “almost infinitely larger power” (a vast lexicon, or the principle of compositionality, maybe?), the major point here is hard to miss – language is primarily a mental faculty and its evolution is tied in with the evolution of human cognition.
Darwin makes a very interesting analogy with birdsong. Songbirds show an instinctive tendency to sing and go through a ‘babbling’ stage, but ultimately they learn the particular song of their parents. In the same vein, particular languages have to be learned but there is an instinct to learn a language in the first place. Over evolutionary time, speaking and singing lead to improved modifications in the vocal organs, but Darwin pointed out that “the relation between the continued use of language and the development of the brain has no doubt been far more important.” As evidence that the brain and language are connected, Darwin observed that there are cases of “brain-disease” which specifically affect language or parts of language. In these observations, Darwin was using comparative and neurological evidence – a highly interdisciplinary undertaking!
Darwin argued that things like the human capacity for concept formation evolved from more rudimentary capacities. He showed that many animals form concepts and do so without language. Language is thus not a pre-requisite for concept formation, as was argued by some at the time. Incidentally, the claim in some of the recent Minimalist literature that the appearance of the operation Merge (the operation used to form sets) was the break-through moment in the evolution of language strikes me as implausible: concept formation presumably involves the ability to form sets and determine whether an entity is a member of a set or not, and if non-human animals can do this, then Merge presumably existed before language.
There’s a nice summary of Darwin’s arguments and the history of ideas of language evolution in Fitch’s book The Evolution of Language, including an entertaining section of the intellectual battle between Darwin and the linguist Max Müller. Darwin’s ideas on language evolution are by no means the final word on the matter (to the extent that there ever can be a final word on this subject), but Darwin’s ideas go to show that careful observation, interdisciplinary evidence, and courage and perseverance to pursue ideas despite various (apparent) challenges and problems can be both fruitful and illuminating.
Fitch, W. T. (2005). The Evolution of Language. Cambridge: Cambridge University Press.
Hauser, M., Chomsky, N., & Fitch, W. T. (2002). The language faculty: What is it, who has it, and how did it evolve? Science 298: 1569–1579.
You may remember that back in February I wrote about that little not-so-innocent word, ‘again’. It turned out to be a tricksy linguistic nut to crack because it appears to have two meanings. Our example was the following:
Frederick opened the door again.
This can be uttered in two different contexts, giving a repetitive and restitutive meaning of ‘again’:
a. Frederick opened the door, and he had opened it before.
b. Frederick opened the door, and it had been open before.
This can either taken to be a simple case of polysemy, or of a singularly repetitive meaning affecting different parts of a verb’s ‘internal’ meaning:
a. again (CAUSEFrederick (BECOME (openthe door)))
b. CAUSEFrederick (BECOME (again(openthe door)))
Today I’ll consider another puzzling aspect of this adverb.
In my last post I may have given the impression that ‘again’’s interesting properties only show themselves when they meet telic verbs like ‘open’ and ‘close’. Well, I’m sorry to admit that that’s not the whole story. What about these cases?
On the face of it, these look just like our open/close cases. There seems to be an ambiguity between the repetitive and restitutive reading – they can comfortably occur in contexts where the event is repeated, and where there is a counterdirectional movement. This is how von Stechow (1996) treats them, at any rate, suggesting an analysis of (BECOME[MORE [low]]), similar to (CAUSE[BECOME[open]]) – low and open are the end states.
But hang on, is there (always) an inherent endpoint for these verbs – are they telic? A classic, though not unproblematic, test of telicity is the ‘in X time / for X time’ test. Usually atelic verbs sit happily with a ‘for X time’ adjunct, while telic ones do not, preferring instead ‘in X time’. For example:
Ben read for an hour / *in an hour.
Belinda opened the window *for a second / in a second.
(NB ‘for a second’ sounds okay – but on a different reading, that Belinda opened the window and closed it again after a second, not that it took one second to open the window).
Note that telicity isn’t a property of verbs per se, at least not in isolation, but of predicates, because the arguments the verb takes affect it. We can quite happily, although rather improbably, say
Ben read Crime & Punishment in an hour.
And actually, we can frame the event of Ben’s reading Crime & Punishment in two ways, that seem to make it either telic or atelic, because this is also fine:
Ben read Crime & Punishment for an hour.
Julie walked around the park for 20 minutes / in 20 minutes.
The idea of telicity under our belt, lets return to our degree achievement verbs (like ‘widen’ and ‘cool’) and directed motion verbs (like ‘rise’ and ‘fall’). What happens when we run them through the telicity test?
The road widened for 2 metres.
?* The road widened in 2 metres.
(NB this sounds fine on the reading of ‘in 2 metres further down the road’, but that is not the one we’re interested in here)
The shares fell for 3 days.
?The shares fell in 3 days.
These examples seem to be happier (at least to me – what about for you?) with ‘for X time’. So they don’t have an inherent endpoint (after all, we don’t know anything about it, apart from it being wider or lower than the start point), can’t be analysed as (BECOME[MORE[x]]), and, on this reading, would only have a repetitive reading when combined with ‘again’.
But, wait, we just said a few paragraphs ago that these verbs also have a repetitive/restitutive ambiguity when they appear alongside ‘again’. How can we account for that? Well, I suggest that, just like ‘walk around the park’, predicates with ‘fall’, ‘widen’ and so on are themselves ambiguous, and may have telic or atelic readings. This is what Hay, Kennedy & Levin (1999) argued when they wrote that these verbs’ telicity “depends on the the boundedness of the difference value” – in other words, whether there is a maximum or minimum bound, or a fixed end point, provided. We can see this by tweaking our examples slightly:
The road widened to four lanes in 2 metres.
The shares fell to rock bottom in three days.
Suddenly the ‘in X time’ adjunct is quite fine! And this is where we can get both readings for ‘again’:
The road widened to four lanes again.
a. The road widened to four lanes and it had widened to four lanes before.
b. The road widened to four lanes and it had been four lanes before.
The shares fell to rock bottom again.
a. The shares fell to rock bottom and they had fallen to rock bottom before.
b. The shares fell to rock bottom and they had been at rock bottom before.
The really interesting thing is that this bound does not have to be explicit – it does not have to be stated as part of the sentence. Rather the speaker and hearer can ‘fill it in’ based on the context and their world knowledge.
The sheep fell down the cliff again.
Assuming that this is a small cliff and the sheep survives his descent, we again get two readings, presumably because the ground below implicitly provides a lower bound.
And where there is a restitutive context, one where a reversal of direction has been made explicit, there are always upper or lower bounds that allow both readings of again (although the context would might favour the restitutive one).
Yesterday the shares rose. Today the fell again.
Here, the fact that the shares fell after rising means that the shares must have stopped rising at some inferred point, and this can be taken by the speaker and hearer as the upper bound of the rising eventuality.
So there we have it, some more fun and games with ‘again’ (and the predicates it interacts with). But the real amusement lies – for the linguist at least – in looking out for real life examples that confirm – or question – the theory. How many ‘agains’ can you spot today?
Hay, J., Kennedy, C., & Levin, B. (1999). Scalar structure underlies telicity in“ Degree achievements.” In Proceedings of SALT (Vol. 9, pp. 127–144).
von Stechow, A. 1996. The different readings of wieder ‘again’: A structural account. Journal of Semantics 13: 87–138.
One of the things I’ve been looking at recently is a particular grammatical pattern in various languages including Middle English (i.e. English as spoken in the period 1066 to 1470-ish). Simplifying matters a bit, in older varieties of English some verbs employed have in the “perfect” construction, whereas other verbs took be:
(1) I am come, thou art gone, he is fallen …
(2) I have worked, thou hast made, she hath said …
In present-day English we basically only use have, so we use the following forms in place of those in (1):
(3) I have come, you have gone, he has fallen …
But when exactly did Middle English use have and when did it use be? The best way to answer this (the best I’ve been able to come up with at any rate) is to trawl through a great deal of text and see what patterns emerge. A body of texts put together for the purpose of trawling through to look for answers to particular questions in this way is known as a corpus. The corpus I’ve been using is the Helsinki corpus, a collection of texts up to the year 1710 – specifically the 609,000 words of texts from the period 1150-1500.
Obviously 609,000 words is a lot of words (The Lord of the Rings is about 480,000, for comparison, and my copy is 6.3cm thick in very small font). And the frequency of instances of what I’m looking for are pretty small: as a rough estimate, there about 6 instances of the perfect constructions in every thousand words, and only about 5% of all these constructions use be rather than have.
Thankfully advances in modern technology (specifically, in my case, the Microsoft Word search function) mean I don’t have to read through the entire length of the corpus hoping to spot the relevant constructions on the rare occasions when they do turn up. But even with the aid of the search facility, the process is still a rather drawn out one. There are two reasons for this: firstly, the irregularity of the verb to be, and secondly, the irregularity of English spelling in the period in question.
Regarding the first, observe that be in English has multiple different forms: be, am, are, is, were, was etc. For one thing, there are simply more forms than we find for any other verb: compare the following:
(4) I am, you are; I was, you were (different forms for different persons)
(5) I love, you love; I loved, you loved (same forms in each tense regardless of person)
For another, many of the forms of be are completely different from each other, with no shared material. Thus, whilst all the forms of love begin with the letters lov- (love, loves, loved, loving), there is no sequence of letters which is common to all the forms of be.
To make matters worse, in Middle English there were even more forms of be. art, as in thou art, was very common, and there were also forms like they weren (= they were), sindan (= they are), he/she bið (= he/she is). To get the full picture, these need searching for as well.
This is compounded still further by the second problem: spelling. Spelling in Middle English wasn’t standardised and there was a great deal of variation in how words were spelled. Even for a little word like is spellings found include is, iss, esse, ys, ysse, his, hys, hes, yes and so on and so forth. am is spelled am, eom, eam, æm, ham … All these various spellings need to be taken into account for a comprehensive survey.
Some corpora may allow you to get around this sort of problem through tagging. In a tagged corpus, each word is associated with a tag which tells you what sort of word it is. The tags used vary, but some corpora specifically mark forms of be and have with their own particular codes, which makes them a lot easier to track down. Obviously, though, the corpus has to be tagged in the first place, which is a lot of work. This can be mitigated to some extent by getting a computer to do it for you, although computers aren’t 100% accurate at this sort of thing so it still needs to be checked by a real person.
After all this, what have I discovered? I’m approaching my word limit, so I’ll have to be quick, but basically verbs in English which took be in the perfect seem to have been either “change of location” verbs like go, come, fall or “change of state” verbs like become. This is interesting because – whilst languages which have this construction vary in how many verbs take be rather than have – there’s been a prediction that if any verbs take be they will include the change of location verbs, and if the class of be verbs is any larger than that it will include the change of state verbs. So Middle English supports that prediction.
In fact, the class of verbs which took be in Middle English is much the same as in modern French (where you say je suis allé(e) “I am gone” and not *j’ai allé “I have gone”). Might this be due to contact between English and French? Probably not, because the French spoken at the time of Middle English allowed be with a much larger set of verbs. This suggests we need to seek out a deeper explanation for the similarities, rooted in the psychology of linguistic processing.
Ultimately, then, I’ve found something out, and so all this corpus-trawling has been worth it.
Do you pronounce your ars or your ahs?
One of the most salient differences between different varieties of English—most famously between the standard, prestige varieties of North America and those of most of the rest of the English speaking world—concerns rhoticity. Even if you’ve never heard the term, or given any particular thought to the phenomenon, I can almost guarantee that you’ll recognise it.
Rhoticity has to do with a sound change which began to affect some English varieties in the fifteenth century and then had its main period of fast expansion in the eighteenth century. This sound change deleted an /r/ (by then already pronounced in different ways in different dialects) in coda position—in layman’s terms, in all positions except before a vowel (for this reason linguists also sometimes use the term ‘nonprevocalic r’ to refer to the /r/s affected by this change). It left some traces on the pronunciation of the preceding vowel, lengthening it and for some vowels also changing their quality. This sound change has affected some varieties of English, such as British RP, but not others, such as General American—the two types are then referred to as non-rhotic and rhotic varieties respectively. The difference should be clear if you compare typical GA and RP pronunciations of words like ‘sister’, ‘car’ and ‘work’:
In each case, the GA pronunciation preserves some relatively r-like gesture in the position where historically there was an /r/: a ‘retroflex’ gesture, involving curling the tongue-tip back towards the hard palate, which is relatively similar to the GA pronunciation of /r/ as a consonant before a vowel. By contrast, the RP pronunciation has no such r-like gestures: in ‘work’, the trace of the historical /r/ is shown by the fact that vowel is lengthened and has a different quality than it once had (as shown by the spelling, this word would once have had an ‘o’-like sound, perhaps [ɔ] or [o]); in ‘car’ the only trace is the lengthening of the vowel; and in ‘sister’, no trace remains at all.
The distribution of this sound change in different dialects is actually much more complicated than just American English vs. British English varieties. In the UK, it originally affected a series of varieties in the South East of England, the Midlands, the North of England (excluding some areas in the North West), and most of Wales—that is, the English spoken in these areas became non-rhotic. This left Scottish English, Irish English, and the varieties spoken in the West Country and some of the North West of England as rhotic varieties. However, the variety that has become most prestigious, at least in England—meaning that it has become associated with financial and political success and influence—is RP, which happens to be a non-rhotic variety. Because of its prestige, RP has exerted a lot of influence on other varieties in the UK, and as a result rhoticity has been consistently retreating into smaller and smaller regions. In some of the regions it was once completely general, such as the West Country, rhoticity is now primarily only found in the speech of older speakers, as young people are switching more to eastern, non-rhotic pronunciations.
Almost the inverse picture is found in North America. Here, immigrants from different parts of the UK brought different varieties with them: some rhotic, and some non-rhotic. Non-rhotic varieties were particularly well-established in the Southern States of the US and in New York City. However, the variety which has gained the most prestige in the US and Canada happens to be a rhotic one. As a result, the areas which preserve non-rhoticity have long been shrinking, often leaving older speakers with non-rhotic pronunciations while younger generations switch to rhotic ones.
These differences in the sociolinguistic and historical status of rhoticity in North America and the UK often make themselves felt in speakers’ creative and socially marked uses. As African American Vernacular English is a non-rhotic variety, in contrast to the prestige norm, speakers can choose to spell words in a way which indicates non-rhotic pronunciations to indicate that AAVE is the variety they speak: consider book titles like The Savvy Sistahs by Brenda Jackson, track titles like Whateva Man by Redman, or the stage name of rap artist DeAndre Cortez Way, Soulja Boy. This may be completely lost on speakers of British English varieties, for whom the non-rhotic pronunciation is the prestige norm.
Nevertheless, non-rhotic spellings are also used in a creative way by speakers to communicate social information in the non-rhotic parts of the UK. However, here, the subtle difference is that such spellings do not point towards any specific variety spoken by the writer, and so don’t carry any particular ethnic or regional connotations: instead, they work simply by subverting the arbitrary, prestige orthographic norm. What connotations they carry thus have more to do with social background and attitudes to education and authority than they do with ethnicity.
There’s lots more to say about the history and sociolinguistics of rhoticity in English, so I may return to it in a future post. For the time being, though, I’ll leave you with a hint about another interesting, related phenomenon: if you’d call something that tasted like oranges ‘organgey’, what would you call something that tasted like banana?
This is essentially a follow-up to my previous post, with a more practical focus, but it shouldn’t be necessary to read the earlier post to understand this one.
The pro-gun activists in the United States have a slogan: “guns don’t kill people, people kill people” (parodied by Welsh rap act Goldie Lookin Chain in the song Guns Don’t Kill People, Rappers Do.) The basic idea, presumably, is that guns, being inanimate objects, clearly cannot take responsibility for killing: rather, the responsibility for killing lies with people who use guns to that end. (And therefore we should, the argument goes, focus our attentions on stopping people from using guns to kill people, not on getting rid of guns themselves.) Even if we disagree with the sentiments behind this, we have no trouble understanding what is meant.
This is an interesting use of language because, from a strictly literal viewpoint, it’s undeniable that guns do kill people. Not as animate, volitional “agents”, of course, but nevertheless Guns kill people is a perfectly acceptable English sentence. And indeed, it’s quite normal for inanimate, non-volitional “instruments” to be used as subjects: there’s nothing syntactically or semantically wrong with Scissors cut paper or The knife sliced easily through the soft, white cheese.
Perhaps, we might argue, kill is different from cut or slice – it requires an animate agent as its subject. (Maybe it’s a bit like eat, which can’t take an instrument as its subject: as I pointed out in my last post, we can’t usually say The fork ate the peas to mean “someone ate the peas with the fork”.) But this is clearly false: surely nobody has any problem with The avalanche killed the skier or Trains kill people who ignore red lights at level crossings.
No, Guns kill people is fine (strictly speaking, at any rate). But the aforementioned slogan does highlight something interesting about attitudes to language: although there’s nothing ungrammatical or unmeaningful about a sentence with an instrument as its subject, there is nevertheless a feeling that volitional agents make better subjects, and perhaps that it may even be in some sense incorrect to use an instrument as a subject when an agent would be available instead.
We see something similar in the arguments by cycling campaigners (e.g. this article) regarding the use of language in journalism relating to road collisions. Often, newspapers phrase things along the lines of A car collided with a cyclist or A lorry ran over a pedestrian. This, the cycling lobby claims, is undesirable because it appears to remove responsibility from the drivers of motor vehicles: cars and lorries do not generally run into things of their own accord, but because of actions taken by their drivers. In other words, given that in such incidents there is an agent (the driver), it is infelicitous to promote an instrument (the vehicle) to the status of subject.
Of course, in parallel with the gun case, an inanimate thing like a car or lorry is a perfectly acceptable subject of a verb like collide or run over as far as grammar or literal meaning is concerned. But the cyclists’ arguments nevertheless highlight, and indeed rest upon, an intuition that volitional agents are once again “better” subjects than instruments. Ordinary users of English have an impression that some types of construction are preferable to others, even when both are technically acceptable: an impression which links closely to what linguists have described as “thematic roles” like agent and instrument. This intuition may seem to support the linguistic analysis that agents are subjects by default, and instruments are only promoted to subject status when an agent is absent.
(In other cases the line between what is merely inappropriate and what is grammatically/semantically unacceptable becomes blurred. The article I linked to gives the example of [the cyclist] collided with a van, referring to an incident where the van was driven into the cyclist from behind. We would probably think of the cyclist here in terms of the thematic role of “patient”: he was not the principle cause of the action, didn’t bring it about on purpose and was the participant most affected by it. Is the use of a patient as a subject syntactically acceptable (as the journalist would appear to think), even if it is an undesirable phrasing, or is it just wrong in every way?)
So: even though things like thematic roles may seem like quite abstract linguistic concepts, it appears that they do have a role to play in the ways in which even non-linguists think about language – and in what is deemed advisable not merely semantically and syntactically, but socially as well.
Something you might find surprising if you delve into sign language literature is the familiarity of the terminology. When you see the word phonology you probably think about the study of sounds. You might even be shocked to discover that there is phonology for sign languages. This post will explain how the phonological terms of spoken languages can be applied to sign languages.
Spoken language phonology identifies the smallest contrastive sound units of language. In spoken languages phonemes differ in various ways (for example, place of articulation, voicing or aspiration). We know that phonemes are contrastive in a certain language when we find minimal pairs where only one of these features differs. For example, when we say the English words came and game, we know that the only difference between them is the voicing of the first consonant but this contrast is enough for them to be considered two different words. Place of articulation (PoA) is also contrastive. Game and dame both start with voiced stops, but one is velar and one is (usually) alveolar and this marks them as separate words. However, certain speakers pronounce dame with a dental stop (some Scottish accents, for example). As alveolar and dental stops are not contrastive in English, both would be considered acceptable variations of the same word. What this tells us is that not all contrasts are meaningful in all languages. We find the same situation when we look at contrastive units of sign languages.
PoA is contrastive in sign languages as well as in spoken languages. In sign languages PoAs are not places along the vocal tract but are various body parts where a sign takes place. These are called Locations. The same exact sign produced in two different Locations yields two different meanings. SEE and TELL in British Sign Language (BSL) are identical apart from their Location (from the eyes for SEE and from the lips for TELL) and this difference is what gives them separate meanings. Location is the first of five parameters that make up the phonology of signs.
The second parameter is Handshape. There are many possible handshapes and each sign language uses a certain sub-set of these as meaningful components of the language. Again, we can identify the handshapes used in a particular sign language by looking for minimal pairs. BSL, for example, has a contrast between a fist with the little finger raised (the [I] handshape) and a fist with the thumb raised (the [Ȧ] handshape). When we keep all other parameters the same and change just the handshape, two different signs are produced, for example PRAISE and CRITICISE. There are also handshapes that are contrastive in other languages but not contrastive in BSL. In American Sign Language there is a contrast between a fist made with the thumb over the fingers and a fist made with the fingers resting on the thumb. BSL does not have this distinction and use of either handshape for a sign such as EUROPE would not alter its meaning.
The orientation of the hand used in a sign is the third parameter. Orientation is the exact direction in which the handshape faces (upwards/downwards, leftwards/rightwards and towards/away from the signer). Even in gesture we can see how important hand orientation is, as we get a very different meaning if we turn the two-fingered peace sign around. In Britain this is offensive, yet this orientation may be seen as simply a variant of the same meaning in other cultures. Again, we can find minimal pairs in BSL where the only difference between signs is the orientation of the hand, for example NOW (in some varieties) and BRITISH (the former having the handshape oriented palm up and the latter palm down).
The fourth parameter in sign language phonology is Movement. This parameter concerns exactly how a handshape moves in a sign. LIVE and FEEL have the same location, handshape and orientation, but the movement (repeated up and down Movement or short upwards Movement) marks them as distinct signs.
The final parameter concerns the non-manual features (NMFs) of the sign. This parameter includes facial expressions and lip patterns. There are some signs that share the same Location, Handshape, Orientation and Movement and are only differentiated by NMFs. By including English mouthing alongside the sign, we can clarify whether a sign means GARAGE or GERMANY. As well as mouthing, there are facial expressions in sign languages that distinguish between signs. For example, the signs DEPRESSED and RELIEVED are differentiated only by the facial expression displaying these two emotions. There is also an NMF that marks negation (head shakes, mouth turns down and eyebrows raise and furrow). The sign MILK with the negation NMF becomes NO-MILK. At sentence level, NMFs can also turn a plain form into a question (through raising of the eyebrows and head tilt) so ALL-RIGHT can become the question form ALL-RIGHT?
These five parameters are the same across all sign languages and, like spoken language phonology, each sign language has restrictions on the way in which these parameters may combine. Certain combinations of these parameters are phonotactically illegal (for example, some Handshapes are not made in certain Orientations). Orfanidou et al (2009) found that when they presented BSL signers with phonotactically illegal nonsense signs, signers often used phonotactic knowledge to correct them. This suggests that native signers, like native speakers, have an underlying understanding of the phonotactics of their language.
Although phonology may at first seem about as far away as possible from the study of sign languages, I hope this post has shown that spoken language terminology and concepts can be successfully applied to another language modality. If you enjoy reading about sign linguistics, have a look at BSL QED’s short linguistics notes on BSL for more.
Sutton-Spence, R., & Woll, B. (1999). The linguistics of British Sign Language: An introduction. Cambridge: Cambridge University Press.
Orfanidou, E., Adam, R., McQueen, J. M., & Morgan, G. (2009). Making sense of nonsense in British Sign Language (BSL): The contribution of different phonological parameters to sign recognition. Memory & Cognition, 37(3), 302–15.
Edited18/03/15 to revise and clarify section on NMFs