Showing posts with label authentic materials. Show all posts
Showing posts with label authentic materials. Show all posts

27 December 2012

Computer games as language study

Before the Christmas holidays I tried to give advice to my students about what to do to keep practicing their English in the holidays, and the thing I really recommended them to do was to play computer games, and I was sure I'd blogged about computer games before, but apparently not...

Anyway, computer games as language study?  Nonsense, right?  "Blam blam blam, pow pow zap kaboom" is the same in any language, after all.

True.  Not all games are of any use to a language learner as most have virtually zero language content.  Action games really aren't much use... unless you're playing with native speakers on the internet, but most people aren't going to do that when they can play with people who speak their own language (and also on a local server minimising the lag).

The games that I recommend to my students are point-and-click adventures, the likes of the old Lucasarts Monkey Island series (if you're familiar with those).  If you're not familiar with the genre, you basically go around talking to people and trying to find objects to solve puzzles.  There isn't usually any way to die, you just keep going until you find the solution and win the game.  These are particularly useful for several reasons:
  1. Dialogue is central to the game.
    In most action games, very little of the dialogue is needed to complete the game; in a point-and-click adventure, the dialogue is full of clues on how to solve the puzzles.  The learner is therefore forced to pay attention and to try to understand.
  2. Dialogue is partially repetitive and has a restricted vocabulary.
    On the simplest level, there are default phrases that are repeated whenever the player tries to perform a task that isn't part of the game (eg "I don't want to cut that", "It's too heavy").  More subtley, though, the vocabulary is very "tight" as the same words appear very frequently to give you clues as to how to solve the puzzles.
    When you're watching a film or reading a book, there is repetition of language, but not to the same extent.  And yet, because these games are designed for native speakers, the repetition is not so blatant and restrictive that it becomes boring, unlike many dedicated learners' resources.
  3. Subtitled dialogue.Most of the games in this genre have voices and text, although the earlier ones are text only.  Having the option to read the text as you're listening seems to help train people to recognise the spoken form of words they know how to read.  You can't really do that with film and TV, because the subtitles never match exactly what the actors say (subtitles have to be easy and quick to read) but in these games, the spoken and written dialogue is almost always exactly the same -- as a rough estimate, I'd say at least 95% of all dialogue matches.  This makes this the only class of "authentic materials" that offers the ability to read and listen at the same time.
  4. Slow pacing.
    The dialogue in these games is pretty easy to follow, as it's not subject to the usual sources of interference.  In films, people talk over each other and loud sound effects mix with the dialogue.  In a point-and-click adventure, everything is usually clear and distinct (and if it's not, you can usually turn down the volume on music and sound effects independently of the dialogue).  As a bonus, the pacing of the game gives you plenty of time to look up unknown vocabulary as it occurs -- the game naturally waits for you, so there's no need to pause and unpause.
There's a few games of this type available for free at scummvm.org, and there are others for sale at GOG.com and on Steam, in various languages.

I've used this approach myself, having bought the first two Runaway games when I arrived in Spain, and having downloaded a couple of games in French before moving there.

It won't teach you a language by itself, but as a form of practice when you're at the immediate level, it's very, very effective, and I would recommend it to anyone, even if you don't normally play computer games....

29 May 2012

Authentic: long v short pt 5

Today was a wee bit frustrating.  I spent a solid chunk of time trying to get the ntlk data module installed, and with it the file english.pickle that would have allowed me to do part-of-speech (POS) tagging.  This would have made it almost trivially easy to eliminate the proper nouns and get a genuine look at the real "words" that are of interest to the learner.

Ah well, it looks like it wasn't meant to be.

So I started working towards custom code to eliminate the proper nouns manually, something which would be handy in the future anyway.  The first step was to identify some candidates for further inspection, and seeing as I'm working with English, that's pretty easy: if it's not all in lower case, there's something funny about it.  I wrote the code to identify all the tokens (words) that contained capitals.  Yes, at this point I could have checked whether it was the start of the sentence or not, but that wouldn't have really helped, because proper nouns occur at the start of sentences too, so i'd still need to check.

When I generated my set of candidates, though, it was a little long.  For The 39 Steps, I was looking at 919 tokens to check manually, and that's a fairly short book.  As I'm doing this for fun, it seemed like checking that many would be a little bit boring, particularly in longer books.  (I later checked the candidate set for the 3 books in total, and it turned out to be over 3000 words, which is more than my time's worth.)

My first quick test then was to have a look at the difference in figures.  Eliminating every single item with any capitals in it drops the type:token ratio in The 39 Steps from 14.48% to 13.14% -- that's almost a a 10% drop (it's 1.35 percentage points, but it's 9.27 percent).  Before properly addressing the proper nouns, I wanted to see how big a difference this crude adjustment makes to the figures.  It seemed just a little too high to realistically be led by proper nouns alone.  But can that be?  I mean, how many words are likely to occur only at the start of sentences?

So on I went, hoping that the data I could generate at this stage would start to shed some light on this figure.

The first graph I produced showed me the running type:token ratios and introduction rates for both the full token set, and the token set with non-lowercase words eliminated:
The two pairs of lines follow each other pretty closely, getting closer together as they progress.  But in order to start getting a clear idea of what was going last time, I had to go to another level of abstraction and measure some useful differences.  So here is the difference between the running ratios for all words and lowercase only, and the corresponding difference in introduction rates:
Now you'd be forgiven for thinking that the difference is diminishing here -- I was fooled into thinking the same thing, but then I realised I was dealing with numbers here rather than proper stats, and I redid the analysis but with a difference in percentage:
The overall running type:token ratio does indeed decrease, but it halves (20% down to 10%) then stabilises.  The introduction rate, on the other hand, is all over the place -- there's no identifiable trend at all.  Even subsampling my data didn't give any clear and understandable trends (and since I'm using a desktop office package for my analysis it's a bit of faff to do the resampling automatically -- it's just further proof that I need to get myself familiar with the statistical analysis tools for Python (eg numpy), but my head's full with the NLTK stuff for now, so I'll leave the improved statistical stuff for
another time).  Here's the same graphs, but with 2000 word samples instead of 500 word samples:

So not promising, really.  Still no stable, identifiable trends.

Books as a series
But I had all the infrastructure in place now, so I figured I might as well rerun the analysis on the 3 books as a single body and see what came out.  Let's just go straight to the relative difference between the lines for all words and eliminating all words not entirely in lower case:
Oooh... now where did I leave those figures on where the individual books started...?  44625 and 152034, and there's a notable period of high difference (20-30%) from about 45000 words, and that massive spike you seem on the graph -- which is actually a 63.64% difference -- occurs from 152000-152500.

Bingo: we've got decent support for Thrissel's suggestion that a lot of proper nouns are introduced early on in... at least some novels.

Not the sort of information I was originally looking for, but actually quite interesting.  It's kind of turning the project in a slightly different direction than I had planned.  I'll just have to go with the flow.

What I did wrong today
One of the minor irritations of the day was when I started writing up my results, and after having done the coding, data generation and analysis, I realised a fairly simple refinement I could have made.  It was a real *palmface* moment: I could have simply taken my first list of candidate proper nouns and eliminated any candidates that also appeared completely in lower case.  Having done that, I would have been left with a much shorter list of candidates, and it may well have been worth my time manually checking the results.

>sigh<

But of course, that's as much the point of the exercise as anything: to work through the process and the problems and to start thinking about what can be done better.

It also occurs to me now that I also managed to eliminate every single occurrence of the word I from the books!  Quite a fundamental error, even if it only made a minute difference to the final ratios.

Perhaps I'm being a little too "hacky" in all this.  I'll have to pick up my game a bit soon....

28 May 2012

Authentics: long v short - pt 4

Well, I had a nice weekend and visited some friends in Edinburgh for a wedding.  The weather's too good to spend too much time inside, so I'll just write up a few more tests then go and enjoy the sunshine.

Multiple books in a series
Today's figures come from two of the books I've already mentioned -- John Buchan's The 39 Steps and Greenmantle, and the next book in the series: Mr Standfast.

Again, the graphs at different sample sizes show different parts of the dataset more clearly than others:
The graph as 1000 words is too unstable to clearly identify the end of the the first book, and the start of the second book is only identifiable because the introduction rate is greater than the running ratio for the first time in any of my tests.

The 2500 and 5000 word samples give us a clear spike for the end of the first book, but the end of the second book is obscured slightly by noise, and becomes clearer again in the 10000 and 25000 word sample sizes, although the end of the first book is completely lost by the time we reach the 25000 word sample graph.

Having done all that, I went back and verified the peaks matched the word counts -- The 39 Steps is 44625 words long, and Greenmantle is 107409 words long, so ends with the 152034th word.  The peaks on the graphs all occur shorlty after 45000 and 150000.

It was the first graph, from the 1000 word sample set, that piqued my curiosity.  Having spotted the two lines intersecting at the start of the third book, I decided to check the difference between the running type:token ratio and the introduction rate, and I graphed that.  At 1000 word samples, there was still too much noise:
However, given that I already knew what I was looking for, I could tell that it showed useful trends, and even just moving up to 2500 word samples made the trends pretty clear:
Going forward, I need to compare the difference between books in a series, books by the same author (but not in a series) and unrelated books, and I believe that the difference between the running type:token ratio and the introduction rate may be the best metric to use in comparing the three classes.

Problems with today's results
I can't rule out that the spikes in new language at the start of each book aren't heavily influenced by the volume of proper nouns, as Thrissel suggested, so I'm probably going to have to make an attempt at finding a quick way of identifying them.  The best way of doing this would be to write a script that identifies all words with initial caps in the original stream, then asks me if these are proper nouns or not.

By treating the three books as one continuous text in the analysis, it looks like I've inadvertently smoothed out the spike somewhat at the start of each book.  In future I should make sure individual samples are taken from one book at a time so that the distinction is preserved.

25 May 2012

Authentics: long vs short part 3

Excuse the slight change of title -- I figured the original long title was probably getting truncated in people's feeds, so I wanted to abbreviate it.  If you've been following my blog recently, you should have already seen my previous two posts on my little project; I am trying to investigate whether my normal advice that long fiction (novels or TV serials) is better than short fiction (short stories and feature films) for the learner.

Sample sizes
Anyway, as I said last time, I wanted to start comparing a fixed length of text, rather than variable-length chapters as my benchmark.  I was looking for a sampling length that would give a clear picture of the overall progression without having too much interference from little local fluctuations.  My first set of results suggests that this is a fool's errand.  The following set of images shows the graphs for the novel Greenmantle by John Buchan, with samples taken ever 1000, 2500, 5000 and 1000 words.
While using larger samples gives a much smoother line, it also unfortunately obliterates some of the most important detail in the graph, in that we start to lose the steep drop at the start -- that's information that's really crucial to my investigation, so I'll have to make put up with various humps and wiggles in the line for now.  However, that's not to say that the other graphs aren't interesting in and of themselves -- the little hump at around 50000-60000 words in the 5000 word sample version suggests that something important may be happening at this point in the story, causing a batch of new vocabulary to be introduced, or perhaps the introduction of a new character with a different style of speech.  Anyway, as interesting as that may be, it would be a diversion from the matter at hand.

Alternatively, I could move away from using linear sampling/projections and start charting using logarithmic or exponential data, and while now would be a good time to start refreshing my memory on that sort of statistical analysis, it also risks diverting me from the task at hand, and I'm following the Coursera.org machine learning course currently, so I should be able to get the computer to do the work itself in a few weeks anyway.  Besides, I've still not got myself a high-frequency word list, and the pattern might be completely different once I've eliminated common words of English from the equation.

So for now I'll stick to working with multiple sample sizes.  I'll admit to being a bit simplistic in my approach to this so far, as I ran my little Python program once for every sample size, rather than just running it once with the smallest sample size then resampling the data.

The program I'm using at the moment is pretty straightforward:

def collect_stats (token_list, step_size):
   i=0
   return_array=[];
   while (i      i += step_size
      running_types_total = number_of_types (token_list[:i])
      if (i         return_array.append([i,running_types_total])
      else :
         return_array.append([len(token_list),running_types_total])
   return return_array
This takes an NLTK token list (it would work with any simple list of strings too, though) and the size of samples to be taken, then builds up a list of lists [[a1,b1],[a2,b2],...] where each a is the number of the last word included in the sample, and each b is the number of unique tokens from the beginning of the text to the ath word.

The number_of_types function just returns len(set(w.lower() for w in token_list)).

This means that at every stage I have a running total of tokens, and it's only when I want to produce a graph that I calculate the number of new tokens in the given slice (= b(n) - b(n-1)), and there's therefore no reason why I can't skip several slices to decrease my sampling rate (eg b(n) - b(n-3)).

Next up
I've taken a running sample of three books from the same series -- The 39 Steps, Greenmantle and Mr Standfast, and run them through as one text, so I'll look at the output of that next, but I don't think it'll be much use until I've got something to compare with -- either/both of: a selection of novels by one author that aren't a series; and a selection of novels by different authors.

24 May 2012

Authentic materials for learners: long form or short form? (pt 2)

Well, as I was saying in part I, I've always claimed novels are easier for the learner than short stories, and I was wanting to back up my claims with some figures.  So for my initial investigation I fired up my copy of the free TextSTAT package and away I went.

I was talking about the type:token ratio last time, and that seemed as good a place as any to start.  I managed to skip the logical first step, which would have been to compare a novel and a short story, but I'll have to come back to that later.

What I started with was an Italian novel, but I got figures that were too high to be useful.  One of the problem with languages such as Italian is that they write some of their clitics in the same written word as the main word (EG "to know (someone)" -> conoscere; "to know me" -> conoscermi), increasing the type:token ratio significantly.  You've also got the problem that it has verb conjugations and it drops subject pronouns in most situations.  Overall Italian (and Spanish and Catalan, among others) would be a bad choice for a demonstration language.  Today, I'm using English as it's a very isolating language -- the only common inflections are past-tense-ed, second-person-present-s and plural-s.  This makes it easy to get a reasonably accurate measure of the lexical variety without any clever parsing.  I will most likely use French at some point too, because while it is not as straightforward as English in that sense (it's got a lot of verb conjugation going on), it doesn't have the same clitics problem as Italian, and the French don't drop their pronouns.
Today's findings: 1 - running ratios
I decided to look at how the type:token ratio changes as a text proceeds.  I wanted to measure this chapter by chapter, counting the types and tokens in chapter 1, then loading the second chapter into the concordance and checking the type:token ratio for chapters 1 & 2 combined, then 1, 2 & 3 etc.  I realised, however that it would be more efficient to load all chapters into memory at the same time and work down from the other end: all chapters, then close the last chapter and take the figures again, then close the second last chapter and take the figures again.
In the end, I got a nice little graph (using LibreOffice) that showed a marked tendency to decreasing type:token ratio as the books progressed:
The x-axis shows the chapter number, the y-axis shows the type:token ratio (remember, this is the type:token ratio for the entire book up to and including the numbered chapter).  Notice how the type:token ratio halves by around the 6th or 7th chapter.

So by one measure, the longer the novel is, the easier it would appear to be.
Today's findings: 2 - introduction rates
I figured I could go a bit deeper into this without generating any new data.  What I wanted to look at now was how much new material was introduced in each chapter -- ie. a ratio of new types to tokens. It's easy enough to do -- I could obtain the number of new types in any given chapter by deleting the running total at the previous chapter from the running total at the current chapter.
The graph I got was even more interesting than the last:
While the running ratio halves after 6 or 7 chapters, the introduction rate halves after only 2-4!  It certainly looks like each chapter will on average be easier than the last.
One curious feature is the large uptick at the end of the children's novel Laddie (green).  This illustrates one quirk that the learner should always bear in mind: kids books are often actually more complicated linguistically than adults' books, as the author on some level seeks to educate or improve the person reading.  The author of this book seems to have kept the language consistently simple through most of the book, but realising he was coming to the end, crammed in as much complexity as possible.

Another curious feature is that the figures claim no new vocabulary is introduced in the fourth chapter of The 39 Steps (yellow).  While this is theoretically possible, its more likely that it's ...ahem... experimenter error, which a quick look at the actual figures verifies: chapters 3 and 4 are listed in my output as being exactly the same length, which is more than a little unlikely.  It looks like I loaded the same chapter twice...
Further analysis
Notice that in both graphs, the figures are the same at chapter one.  This is to be expected, as every type encountered in the first chapter is encountered for the first time in the book (by definition).

So what happens if we stick the running ratio of type:token against the introduction rate of new types?

This:
So while the overall type:token ratio continues to fall notably from the 10th to the 20th chapter, suggesting decreasing difficulty, the introduction rate gets fairly erratic by around the 10th chapter (despite still tending downwards), so perhaps there is a limit after which it is not safe to assume that each chapter is a difficult as the last.

Perhaps the measure of efficiency is related to the difference between the running ratio and the introduction rate, and once that gap starts to narrow, there is no advantage?

Problems with today's findings
This was a first exploratory experiment, so I didn't conduct it with a whole lot of rigour.  Here are the main factors affecting todays results:
  1. I didn't eliminate common words -- it is impossible to see from the figures I have how many of the types introduced at any stages are ones we would expect learners to know already and how many will be genuinely new to them.
  2. When examining Pride and Prejudice and The 39 Steps, I hadn't told the concordancer to ignore case, so anything appearing at the start of a sentence and in the middle would be counted as two types -- eg that and That.  (It was the first time I'd used TextSTAT and I hadn't realised it defaulted to case-sensitive -- I won't make that mistake again.)
  3. The length of chapters varies significantly from book to book and even from chapter to chapter within books, so the lines are not to scale with each other, and each individual line is not in a continuous scale with itself.  The graphs, though presented in a line, are arguably not true line graphs, as they occur from samples arbitrarily dispersed.
Accounting for these problems in the future
  1. There are plenty of frequency lists on the net, so I'll be able to eliminate common words without any real difficulty.
  2. The case sensitivity issue, now that I'm aware of it, will not be a problem.
  3. When I ran the initial data, I was using TextSTAT as my installation of Python and NLTK was playing up (I had too many different versions of Python installed, and some of the shared libraries were conflicting).  I've now got Python to load NLTK without problems, so I can do almost any query I want.  Future queries will be sampled regularly after a specific number of words.
Experiments to carry out
At some point I'm going to want to go back and compare short stories with novels, but for now I'm going to head a little further down the path I'm on.

My first task is to work out a decent sampling interval: ever 1000 words? 5000? 10,000? 50,000?  I'll run a few trials and see what my gut reaction is -- that should be the next post.  (It might even prove that the chapter is the logical division anyway -- after all, it divides subjects, which would indicate different semantic domains...)
I also want to look at what happens when we look at sequels after each other.  Those of you familiar with John Buchan will notice that I've included such a pair as individual novels here -- The 39 Steps and Greenmantle.  I might include initial findings from this next time, as they'll determine my next step.
After this I'll either move on to looking at more pairs of original book + sequel (to look for a generalisable pattern), looking at longer serieses of books (to see if they get continually easier) or comparing book-and-sequel to two different books from the same author (to see if any perceived benefits from reading a book and its sequel are just coincidence and really only because of the author).

Caveat emptor
Remember, though, that this little study is never going to be scientifically rigorous, as I don't really currently have the time to deal with the volume of data required to make it truly representative.  However, it's nice to think how big a job this would have been before computers made this sort of research accessible to the hobbyist.  Many thanks to the guys who wrote the various tools I'm using -- your work is genuinely appreciated.

23 May 2012

Authentic materials for learners: long form or short form? (part I)

Anyone who gets sufficiently far on in learning a language is going to want to start reading, watching or listening to materials intended for natives.  This is what a lot of language teachers refer to as "authentic materials".  (Now, an alarming amount of so-called authentic materials in the classroom are actually very heavily doctored, but that's not the sort of authentic materials I'm going to talk about today.)

I have often claimed that it is better for the learner to use longer materials than short materials (eg when discussing films vs TV serieses in a previous post).  This wasn't an idea I came up with myself, but advice I'd been given when I was at high school, although I can't remember who first suggested it to me.

Anyway, I was told at the time that the first one or two hundred pages of a novel contain most of the language (in terms of grammar, vocabulary and turn-of-phrase) that will occur in the entire book.  It therefore follows that the first two hundred pages of any work are the most difficult, and therefore the longer the book is, the easier the ending will be, because you won't be confused by the language.  This also means that the book is acting as active revision, and that by the end of the book, you will have learnt most of the major vocabulary in it.

A 50 page short story would intuitively sound easier to read than a novel, but this isn't really the case, because you're dealing with something that is going to be littered with new words on every single page.

And what about a piece of flash fiction?  Realistically, we're not going to expect much repetition at all.  Compare with the short extracts of authentic works printed in many classroom language textbooks -- none of the "content words" that are specific to the story are likely to be repeated at all, so they will be looked up by the reader, then promptly forgotten about.

Although I was told this about reading books (as opposed to short stories), I believe this holds for any form of literature, fiction or non-fiction, regardless of medium.

A half-hour documentary will be self-reinforcing in a way that a 4 minute news report on the same topic won't be (on TV or radio).  An 8 hour long TV series similarly will reinforce its language than an 80 minute feature film, or particulary a 10 minute short film.

Anyway, I've been repeating this advice for years, and I've always said that my experience backs it up. Well, in the little gap between finishing my Gaelic course and starting my next job, I was wanting to do a little work with corpus analysis software and it occurred to me that this would be a great little exercise to get me back into the swing of things, so I downloaded several resources: TextSTAT, a concordancer package written in Python at the Free University of Berlin; AntConc, a Linux/Mac/Windows concordancer by Lawrence Antony at Waseda University in Japan; and the Natural Language Toolkit for python, which will allow me to write more flexible, custom queries on my data.

Type:token ratio
One of the most basic statistical measures of diversity in a text is the so-called "type:token ratio".  The number of "tokens" in a text is the number of individual words, the number of "types" is the number of different word forms.

For example, the phrase "the cat chased the dog" has 5 tokens, but only 4 types, because the is only counted once when determining the number of types.

Or again, "I told the woman to tell the man to tell you" has 11 tokens (11 words in the sentence), but as the, to and tell occur twice each, there's only 8 types in the sentence.

The type:token ratio is exactly what you'd expect if you're at all familiar with statistics: the number of types divided by the number of tokens.  In the first example, we have 4:5 = 4/5 = 0.8 (or 80%) and in the second we have 8:11 = 0.727272... (roughly 73%).

Notice how the type:token ratio on the longer sentence is lower than that on the shorter sentence -- in this case it's a matter of my choice of words, but as a general rule, type:token ratio decreases with the length of text examined, which only goes to justify the advice of favouring long-form over short-form materials for the learner.

However, that's still to be proven in practice.

So over the next few weeks, I'll be experimenting with a bunch of public domain texts from Project Gutenberg.  I'll be trying to investigate the basic premise of whether long-form fiction is intrinsically easier than short-form, then investigating whether this extends to reading several books by the same author as opposed to books by different authors, and how much of a difference it makes whether these books are part of a series or individual stories.

The size of this study is going to be very small, as the main goal for me is simply to gain a better understanding of the technology and to reason through the process of designing logically sound research in a corpus, so the conclusions won't be scientific proof of anything, but it will hopefully be interesting (to me at least).

If you're aware of any research that covers the areas I'm looking at, please feel free to drop a reference in the comments, and if you have anything to add or suggest, I'm all ears.

31 March 2012

Towards a model for sustainable language learning resources

The volume of language learning materials released over the years is unimaginable.  And yet, the resources available to the individual learner or teacher are incredibly limited.  Too much of the material that is available is bound to a particular course or learning task.

Nowadays people are making available for free, via the internet, material that is of use to the language learner, but unfortunately a lot of it is still bound to a particular task.

What we need is to see a critical mass of materials that can be used together by the creative teacher or learner.  There are a few current projects working towards this.  For example, there's the Leonardo Pools project, and notably among the Leonardo members, the Scottish Gaelic and English Guthan nan Eilean/Scottish Island Voices project.

What we've got here is videos that are available under an open license for teachers to use any way they chose.  (Sadly they've not given legally useful terms for the license, but it's a start...)  There's also a linked project around building online tools to use these resources, but to me that's a secondary issue.

What I said was that we need a "critical mass" of material, and what we're seeing from even the most prolific of the Pools teams isn't enough on its own to reach that level.  That's why we should all be thinking about what we could be doing to work towards that goal.  I'm not a fan of Benny Lewis's blog or his advice, but he's fairly prolific in making videos.  Now most of those videos aren't a great deal of use to the learner because they feature his own learner speech, but recently he posted a video of a presentation given in Quechua about traditional crafts, which is exactly the sort of thing that I would love to see more of.

So we could all go out and video natives speaking their own language.  Problem solved, right?

Not quite.  Who's going to use those materials?  How do we make sure people have the right to use them?  And what rights do we give them?

Should a commercial operation such as Lingq.com be allowed to use the material for free, in order to make a profit?

It's a tough one.  The limiting factor in most language courses is the volume of material, and the limiting factor with regards volume of material is the cost of production and licensing.  But still, should the public and publicly funded projects be paying for the production of the material and someone else be making the money?

Is language learning enough of a "social good" that the ability of others to profit without giving back is a necessary evil?  I'd love to hear your thoughts.

18 February 2011

Watching films as language study...

Well, I've been a little bit too technical and theoretical of late, so let's go for something more practical for a change.

A lot of people love the idea that you can learn a language just by watching films (such as Keith Lucas, discussed last week).  You can't.  Well, maybe there's one or two linguistic supermen out there who can, but for most of us, it won't work.

Can we get anything out of films?  Of course.
Can we get a lot out of films?  Hard to say.

First of all, if you're an absolute beginner, you're not going to understand anything watching the film without subtitles, and all proponents of target-language-only learning say that it's in understanding that we learn.

But unfortunately, once you start reading subtitles, you stop listening.  The brain, so I'm told, has only got one "language channel", and if you load it through the eyes, the brain tunes out the words hitting your ear so as not to mix up the two streams.  I'm sure you've tried talking to someone while reading or writing and found that you've written down a word from your conversation or suddenly said a word you've just read.

So once you start tuning out the sounds, your not going to learn much.

I first bought a DVD player in the January sales in 2005, with the express purpose of learning from foreign films.  My plan was quite typical: watch them with English subtitles, then later watch them again with the subtitles off.  Well, I never really did that -- I just kept buying and watching them with the subtitles on.  Not brilliant for my language skills, but now I've got one of the best DVD collections of anyone I know.  (Well, I know some people who have better collections, but at least mine's all originals!!)

However, after about a year, I started to notice little things at the start and end of sentences.  Little things like "you know...", "I see..." etc.  You know, little things that just seeped through before I started or after I finished reading the subtitles.  But I'm still only using one language channel.  People who can hold a conversation while reading a book aren't really doing two things at once, they're simply switching backwards and forwards between two tasks very rapidly, and this is what I started doing.  As a kid, I could never hold a conversation while reading, so it's not an innate talent on my part.  (My big sister always used to be able to do it.  I always assumed she was faking it or lying.)

Over the intervening years, I've been able to pick up more and more, but it seems to me that in a way I'm "primed" by the subtitles -- I'm anticipating how that would translate and what I hear is then matched against my expectations.

But really, the way to improve when you're good is to go without English subtitles (or whatever your native language is).  The first step to achieving that is to get material with target language subtitles.  The subtitles never match what is said on screen, so it's limited, but it does help you get tricky words.

Just now, I've been watching a French series Un Village Français. I tried watching it without the subtitles, but a few words slipped by me.  The first time I watched with subtitles on, I saw the word "scierie" and I realised it had to be "sawmill" ("scie" is "saw", and I knew the guy owns a sawmill from watching it before).  I'd watched two whole serieses without subtitles and never realised what this word was.  I hadn't even noticed that the word existed.  Two minutes with subtitles on, and I doubt I'll ever forget it.

But so far, so vague.
How did I start being able to listen while reading?  It's hard for me to say, as I wasn't really thinking about it at the time, but I believe it was when I started echoing my favourite actors to try and get the rhythm of the languages.  You can't do that without listening (obviously) and at first this got in the way of reading the subtitles and I ended up using the pause button a lot.  But having done that, it seems like my brain started realising that it had to listen and eventually I got there.

I only really noticed I was doing it when I went to see a French film and one of the characters was bemoaning the fact that kids today don't watch French cinema.  The subtitles talked about "rubbish from far away", the voice said "American crap".

But even after years, my "listening while reading" is still very limited.  It leaves me with a question I can't answer.  Do I get more out of watching with subtitles and hearing less of the speech or do I get more out of watching without subtitles and hearing more, even if I understand less?  It's impossible for me to measure this, and in the end the choice is made for me by circumstance, because if I have subtitles, I watch with subtitles.  If I don't, I watch without.

TV vs film for learners

But on a different tack, it's worth noting that watching serieses is far better for your language skills than watching films.  A film is relatively short, so there's little recycling of dialogue.  Each new film has potentially new accents and ways of speaking, but a 90 minute film finishes just as you're starting to get accustomed to the actors.

TV serieses, on the other had, offer several hours of dialogue written by the same scriptwriters, delivered by the same actors in the same accents, and covering the same topics.  The vocabulary and turn of phrase is repeated in throughout the length of the series, naturally reviewing and revising your learning. I've been following a particular series in Spanish for about two and a half years now, and I personally feel it has been immensely helpful to me.  Of course I've learnt a lot from other sources during the same time, but this has really aided my listening comprehension.

I already mentioned Un Village Français - I bought a two-series boxed set for around the cost of two full-priced feature films, and that's 10 hours of drama with 3 hours of historical documentary as bonus features for the price of 3 hours of film.  As I progressed through the series, I really did feel like I "tuned in" to the accents - there were things in the first few episodes that I should have understood (in terms of grammar and vocabulary, they were withing my boundaries) but that I didn't (because my ear wasn't picking up the detail of what the actors were saying).

So how do you make films and TV part of your learning strategy?

In the beginning, I don't think you really can.  At that stage, don't consider it "learning time", consider it "TV time".  Get used to the whole idea of subtitled foreign cinema with subtitles in your native language.  If you start to hear a word or two, great.  If you don't then it's no loss as this isn't "learning time".

I only really started getting serious with Spanish TV in the run up to my exams.  I'd studied a lot, I'd learnt a lot, but it still felt really disjoint.  I considered TV viewing as a type of revision -- I was hearing stuff I already knew, but used in many different ways.  I got used to the speed of natural speech in various accents, but I don't think I could have done that if I didn't already have a solid foundation in the grammar, because it reduced the amount of unknown material in the language.  In the end I picked up a couple of structural points too, and some good vocabulary, but mostly I mostly found that it took the language I knew in an academic context and made it more real and alive.

(And in the spirit of taking nobody's word for it and what I said in the follow-up, I'd like to point out that I can say definitively that I learned the Spanish construction "volver a hacer" from the Spanish series Águila Roja.  The fact that I can give a specific example suggests to me that I didn't learn much in this way.  Unfortunately, if you don't think about it, you can be misled into believing that remembering an example is proof of the effectiveness of a method.)

One thing I think would work well is to use a DVD player or computer video player that you can slow down.  I'd like to start watching foreign serieses with the first few episodes slowed by about 10 or 15% while I get used to the characters and their ways of speaking, then speed it up to normal speed for the rest of the series.  Unfortunately most of my serieses to date have been on-line without speed control, but I'll give it a go later in the year when I order in some French TV, and possibly Italian too.

24 December 2010

Dialogues from Day One.

I discussed dialogues briefly in an earlier post on expository and naturalistic language.  Fasulye suggested in the comment section that dialogues didn't necessarily lead to the use on unnaturalistic language.  OK, so I didn't say that it did -- the point I raised was that dialogues aren't a "magic bullet" that makes all language seem naturalistic.

However, that said, I'm not a big fan on dialogues anyway, so today I'm going to talk about how starting a course with dialogues from the very first lesson actually slows down progress for the learner.

My contention:
The need for a coherent dialogue forces the author to use language that the student isn't yet ready to understand.
The dialogue format forces the learner to move between such a variety of different language, that it forces the student to attempt to learn too many things at once.

I'll use as my example one of the ever-popular Teach Yourself books.

Lesson 1 TY Welsh opens with the following dialogue (my translation)
Matthew: Good morning.
Elen: Good morning. Who are you?
Matthew: I'm Matthew.
Elen: How's things? I'm Elen, the Welsh course tutor.
Matthew: I'm a learner, a very nervous learner!
Elen: Welcome to Lampeter, Matthew. Don't be nervous, everything will be fine.

What do we start off with?  It's those old favourites -- hello, what's your name etc.

But what does this teach us?

Let's have a look at the Welsh for "who are you" and "I'm Matthew":  "Pwy dych chi?" and "Matthew ydw i".

These two phrases are completely alien to the English speaker.  There is only one clue that the English speaker can use to try to make sense of this -- the name "Matthew".  A learner might assume that "pwy" and "ydw" are linked, but they're not -- "dych" goes with "ydw", even though the two are not visibly related.

This is the verb "to be", and this problem isn't unique to Welsh -- consider the English "are", "am" and "is".  So even when we look at dialogues from an entirely expository point of view, we have a problem that means we have too many unknowns for the new learner.

Consider the following (not a real example) as though it was in lesson one:
John: Are you tired?
Sally: Yes, I am tired.

You as a learner are asked to contrast the question with the answer, but we have a massive amount of variation in a very simple sentence.  First of all, we have the matter of the irregular verb forms, as above.  Secondly, the pronouns are radically different (as in most languages).  Finally, we have a change of word order.  Learners could confuse their verbs and pronouns, and miss the word order entirely.

OK, that's not a real lesson 1 example, but I've already given a worse example from the Welsh course - Pwy dych chi?.  In the Welsh, the word order doesn't change for the answer Matthew ydw i, but that's arguably as difficult for an English speaker as English word order is for speakers of a language that doesn't change order.  We also have no repeated recognisable word form to highlight any the word order in Welsh.  There is an awful lot of rules in play here, each interacting to make the full meaning of the sentence.  Without seeing these in isolation, the role of individual elements is obscured.

And it's even more complicated in French.  Many courses will introduce Comment t'appelles tu? and the response Je m'appelle Jean-Pierre (or whatever name).  This introduces the complication of the reflexive pronoun, which is a version of the object pronoun.  Well, actually, the reflexive pronoun is identical to the normal object pronoun for "me" and "you", which actually makes this more confusing.  While the change of word order for the question is theoretically the same as English, the lack of auxiliary do (eg Do you know?) in French questions makes it completely different to the untrained eye.  The fact that this places the object before the subject is particularly alien to the English speaker.  This is massively difficult, and so the learner is only expected to memorise or learn to recognise the phrase.  The assumption here is that by exposure to later examples, the learner will induce the underlying patterns, but this is something that dialogues are actually very bad at.

Dialogues by their nature attempt to model naturalistic conversations, and this leads them to include a very wide variety of language.  Unfortunately, variety means very little repetition, so there is very little material to induce the rules from.  It gets worse when the writer is trying particularly hard to be naturalistic, because many of the expository cues are lost.  Remember this from earlier?  I'm a learner, a very nervous learner!  Notice that this uses elision (the ommission of repeated words) for increase naturalisticness, but missing the opportunity to reinforce the structure "I am".

French courses rarely follow up the je m'appelle with any other reflexive constructions -- the only thing it is contrasted with is usually il/elle s'appelle (he/she/it is called).  The student is left knowing the phrase for a long time without being given the input to learn why it means what it means.  In fact, this risks interfering with normal (non-reflexive) object pronouns, because the learner is overexposed to the reflexive form, and unexposed to the base form for a long time.

The root cause of the problem

The language in a naturalistic dialogue is linked by context, and elision is a major feature of natural language.
In short, we actively avoid repeating language in a conversation.

This leaves us teaching language that is only bound by context, so is semantically reinforcing, but not syntactically reinforcing.

If we progress in a language by learning a new word, it opens up a few extra possibilities, but learning new grammatical structures can double our knowledge of the language.

So imagine you know "I like...", "I have..." and "cars", "trees" and "dogs" -- you can say 6 combinations.  If you next learn to say "cats", that's an additional two sentences -- "I like cats" and "I have cats" -- so 8 in total.

But if instead you learn the negation "don't", that doubles the number of sentences to 12.

Massive growth in beginner language is only possible if you focus on teaching language points that can be combined within a sentence to make bigger and more complicated sentences.  The dialogue format militates against this, and after one dialogue-based lesson, a learner is not likely to be able to produce even as much as is in the dialogues themselves.  Compare with the Michel Thomas courses where (even excluding the -ible/-able words) the learner has a range of expression that while limited still covers dozens of different possible sentences.  By building on this, the student experiences almost exponential growth.  That's cool.

10 December 2010

Whither comprehensible input?

In an earlier post discussing expository and naturalistic language, I linked to this video from a talk given by Stephen Krashen, discussing comprehensible input (CI):


So what exactly is comprehensible input?  A very clever man once said that "If you can't explain it simply, you don't understand it well enough."  That man was Albert Einstein, and never has a truer word been spoken.

I have never, ever had a simple explanation of what comprehensible input is.  I have had trivial explanations, but while these are simple, they explain nothing.  So I decided to go to my friendly neighbourhood search engine and look for that elusive definition, and ended up back in YouTubeLand.

First up, here's a teacher's explanation of CI, from a resource pack for teachers of modern languages:


The first thing of note is when she says "Sometimes easier to say what it's not" -- this is an immediate warning that we're in the field of not understanding in Einstein's terms.  She then follows this up with a list of quite trivially obvious things that aren't good teaching practice (just like Krashen's first demonstration with German).  Anyone watching the video should already be familiar with the idea that talking at full speed to a beginner is a no-no in the language classroom, so why is this of such note when discussing CI in particular?

As she continues, she gives an example of what she has developed as CI and says that "most sentences are about 4, maybe 5 words long."  This is a very impoverished model of language, as she is therefore reduced to generating a series of independent declarative statements.  The real meat of any language comes in the textual metafunction -- how words change each other's meaning.  All those little connectors that compare, contrast and relate individual phrases by demonstrating cause and effect, simultaneity, sequence of events and all that stuff that allows simple facts to add meaning to each other (my recent post against rote learning demonstrated how I used various facts and consequences to learn the order of 4 of the US presidents).

Textual metafunctions are notoriously difficult to learn by induction from context and do appear to need genuine conscious teaching -- the TEFL methodology normally gets round this by defining them as inherently "advanced" and teaching them through target language several years down the line.  However, doing this actually makes it harder to engage with authentic native materials, and the process of "grading" texts becomes ever more important.  CI would appear to keep students away from native language, not get them closer.

Ok, so just before the one minute mark, she recommends that material is 80% comprehensible.  This sounds very scientific, but at no point has she described how to measure percentage comprehensibility.  I'm not sure where this 80% comes from, but I've heard it bandied about quite often, and I suspect it's just another generalisation from the Pareto Principle.  (80% has become something of a superstition for our times -- the Pareto Principle has gone from being an observation of general behaviour in systems, to being considered an immutable law for success...)  But in this case specifically, what does she mean?  Well the general claim is that the reader will be able to infer no more than 20% of the meaning from the context provided by the other 80%.  This is all well and good, but this is not a simple game of numbers -- certainly language features will provide useful context for working out others, but other features won't.  If CI is to work, you need to look at which features interact in this way to reveal useful information about unknown language.  Again, how to do this is never really explained in a discussion of CI.  (It is discussed at some length by academic linguists, who develop some very sophisticated maps of how certain language features reveal others, but this remains in the journals, and never makes it into language teacher training materials.)

Now, at just after 3:30, the presenter goes quite far off the rails, telling us that texts that are "dry and uninteresting" are not CI. The whole video was announced as being on the topic of comprehensible input, but now she starts talking about something entirely different -- student engagement.  This is a recurring problem in teaching methodologies -- a fairly simple idea covering a small part of the methodology is used to describe the whole methodology.  Student engagement is very important, but engagement doesn't increase the "comprehensibility" of a text, it just increases the chances that the student will put in the effort required in order to understand it.   Both of these two variables must be considered at all times, so it is not helpful to try to present them as a single item.

So, on we go, and here's another great example of Einstein's definition of not understanding.

The definition revealed on the slides between 0:50 and 1:00 is anything but clear.  In fact, it isn't even a "definition" at all in the most common usage of the term -- the second and third bullet points, presented as part of the definition, are sidenotes on particular issues.


As she expands, the definition becomes rather circular.  The first "characteristic of comprehensible input" that she defines is "understandable" (2:14), which is the Germanic synonym of the Latinate "comprehensible".  The fact that she doesn't stop there means that she is, like the woman in the previous video, expanding the term "comprehensible input" in directions it wasn't intended to go in.  But like the previous in the previous video, we quickly get a "what it's not", and again it's a trivially obviously bad way of teaching that she throws in.  It does not help define CI.

As she expands on "understandable", she hits a particular danger zone, asking that you teach "using as many cognates as possible."  In the TEFL world, that means using "arrive" when you really mean "get to" or "depart" when you mean "leave".  Target-language-only teaching justifies itself by saying that we need to learn how words are used -- grading language by the introduction of cognates actively teaches language that is not natural, and the natural tendency for Romance language speakers to overuse terms such as "arrive" is something that has to be actively combatted in class, not reinforced.

Yes, using cognates is good when they can be used naturally.  The example she gives is a good one -- English "important" vs Spanish "importante" -- but she presents no guidelines or advice on how and when to use cognates.

Like in the previous video, an unrelated variable is thrown into the mix -- the "affective filter".  Again, this is not part of comprehensible input and only serves to distract from the discussion on what CI is/isn't.  She then starts talking about repetition.  Again, I'm confused as to how this fits into CI.  Surely the point isn't about "repetition" in the rote sense, but the sense of having it come up in 7 different texts before production.  But if she means that, she should say so.

Then she moves on to "visuals".  This is quite insidious, because now that we've establish the notion that texts must be "comprehensible" as a sort of divine law, we can address the fact that it is impossible to make texts comprehensible.  So step by step we redefine what is and isn't the text.  We introduce the notion that including "visuals" makes the text understandable, when in reality we are using visuals because the text is not comprehensible.  Why is it that visuals make a text "comprehensible", but a native-language equivalent word doesn't?  The video gives us the normal line (3:33): connecting a "new vocabulary word" [sic] to a picture "they will remember it quicker rather than having to connect the word in their target language to the word in their native language and then [exasperated tone] finally to the image produced in their mind."  The idea is thus presented that an image is more closely tied to the mental model of a concept than a word, but any look at the nature of abstraction and iconicity will show that this idea is built on very shaky ground (I'll explore this in greater depth in a later blog post).

The visuals is then supported by something else that all CI advocates propose: body language.  Unfortunately body language is linguistic, meaning there is native body language, there is target body language, and there is something that is neither.  Classroom body language is normally a constructed language, and it is not the target language.  Suddenly the target-language-only classroom has grown a third language.  And the student must learn that as well as the intended target language.  At least if the native language was used there'd be no need to learn a third code.

I was particulary intrigued by the rather odd claim that unspecified "brain research" has somehow proven that we can only learn 7 new words per day.  It's patently absurd -- you cannot "learn" any words in 1 day -- learning only takes place over the longer term, and I have never seen any language course that doesn't use less than 7 words in lesson one -- it would be a boring lesson indeed!

The demonstration given a 5:45 was something I did not expect; to me, this is nothing more than an audiolingual drill, which demonstrates something quite important:  language teachers don't normally change their techniques to match what research proves is effective (or claims to, at least), but instead find ways to justify what they already do in terms of the new idea.  Recasting the idea of audiolingual substitution as comprehensible input is dead easy -- if you want "N+1 comprehensibility (a woolly favourite of the CI crowd) then what could be more natural than taking a known structure and simply adding a new word?

The next video is an extreme example of the same phenomenon.  It is an excerpt from a relatively old book-and-tape English course.

This is a really heavily behaviorist course, but someone has recently decided it's CI and relabled it as such for the purposes of the video. I don't believe it's what Krashen intended the term to mean, but it certainly fits the definition he and others provide.  There's a lot of repetition, there are visuals supporting the text, the text uses context to support the comprehension of new structures and vocabulary...

...so we must conclude that "comprehensible input" is an overbroad term that doesn't really define anything specific.

Finally, here's a video presented as an example of "comprehensible input".


The language used in general is highly expository and not particularly naturalistic. The visuals she uses do not support the learning of grammar, they only assist in the learning of very concrete terms, particularly proper nouns.  Does the Mexican flag assist us in understanding that she's talking about her family? No.  It only helps us understand "Mexico", which is probably where most of the class are from.

Now, note at 0:58 she gets the students to say something -- a single word: south. She then proceeds to go through a string of questions with very simple one-word answers.  The goal here isn't to get the students talking, it's merely to verify that they've understood.

And this is where CI theory falls flat: it is based on Krashen's idea that "language acquisition" is a matter of simply absorbing and absorbing a language until you know the full language, but the streets of any major city are full of immigrants who learn to understand the language of their new homes, but never learn to speak it with any particular degree of accuracy.

CI fails because it necessarily focuses on concrete vocabulary and on learning to understand.  But these are far, far easier to learn than functional vocabulary (modal verbs, linkers etc) and learning to produce.  More than that, these are the things that students can learn outside of the classroom.  Surely it's the teacher's job to teach the hard stuff, the stuff that the student can't learn independently, and then let them do the rest on their own?

19 November 2010

Expository vs Naturalistic Language Examples

A couple of weeks ago, I was discussing authentic materials.  The main problem I identified was the lack of mutual reinforcement between individual texts (I hate that word, but I just can't find a suitable alternative...) meaning that very little language presented is retained.

So where did our modern love of "authentics" come from?

Authentic materials is actually one of the oldest tools in the language learner's toolbox.  Classical education has long focused on the reading of genuine Latin and Greek texts.  If you have a look at the Open University's course catalogue, you'll see that their classical language courses are called Reading Classical Greek and Reading Classical Latin, which is a pretty clear statement of the course goals.  The Greek course looks at a lot of literature in translation, but the Latin course is a perfect example of learning by authentic materials, as it looks at excerpts from Roman dramas and Cicero's speeches.

The use of authentic materials would even appear to go at the very least as far back as the heyday of the Roman Empire, where Greek was the fashionable language du jour.  Greek slaves were sold into rich Roman households where they would teach the children of the house to read and understand the works of writers such as Homer.

But despite two millenia as one of the most widely used tools in language learning, there are those who present the idea of using "real" language as a new and revolutionary idea.  In fact, many proponents of "real language" actively attack old ways of learning as ineffective and outdated.

But if we don't go straight for authentic material, what is there?

The very extreme opposite of authentic material is the stereotypical idea of trite sentences designed purely to demonstrate grammar points -- what I call expository language.

There are several classic examples of the absurdities that a purely expository approach leaves us with.

To the French person, the archetype is "My tailor is rich", which I'm told was the opening sentence of the original Assimil course.
In English, our traditional archetype is "La plume de ma tante" ("my aunt's pen", literally "the pen of my aunt") in such contrivances as "la plume de ma tante est sur le table".

Over a hundred years ago, people were already spending a lot of time attacking this approach.  The Danish language teacher Otto Jespersen wrote a book entitled How to Teach a Foreign Language (translated to English by Sophia Yhlen-Olsen Bertelsen) in which he put forth an argument for the so-called "direct" or "natural" method - ie that of teaching the language monolingually, by only speaking the target language.
"Disconnected words are but stones for bread;" he said, "one cannot say anything sensible with mere lists of words," and this is certainly true. "Indeed not even disconnected sentences ought to be used," he continued, "at all events, not in such a manner and to such an extent as in most books according to the old method," and while I wouldn't argue with this, we can see a little hint of what Decoo classes under the heading of "denigration of others" in his lecture On The Mortality of Language Learning Methods.

I'll reproduce some of Jespersen's examples, all taken from genuine courses of the time, for your benefit.
"My aunt is my mother's friend. My dear friend, you are speaking too rapidly. That is a good book. We are too old. This gentleman is quite sad. The boy has drowned many dogs."
Clearly there is no consistency or logic behind these, and it is hard to build up any sort of a bigger picture.

He then picks an example from a French book:
" Nous sommes a Paris, vous etes a Londres. Louise et Amelie, ou etes-vous? Nous avons trouvé la lettre sur la table. Avez-vous pris le livre ? Avons-nous eté a Berlin ? Amélie, vous etes triste. Louis, avez-vous vu Philippe? Sommes-nous a Londres ?"

And this is Jespersen's criticism of it:
"The speakers seem to have a strange sense of locality. First, they say that they themselves are in Paris, but the one (the ones?) that they are speaking with are in London (conversation by telephone?) ; then they cannot remember if they themselves have been in Berlin ; and at last they ask if they themselves are in London."

There is nothing in his criticism that really applies to any method, "old" or otherwise.  We are in fact looking at a criticism of choice of material.

I'd like to give a few examples that I think underline this point.

An Comunn Gaidhealach's Elementary Course of Gaelic was first published almost 100 years ago.  I picked up a reprint of the 1921 edition in a charity shop a couple of years back.  The first edition was written at the just after the high point of the "natural methods", and the revised edition was put together about 30 years after Jespersen's book, so it's quite likely that natural/direct thinking had an effect on both the original author and the author of the revised edition.  So let's have a look at some of the exercises in the book.

The first lesson has the following as a reading exercise (this is my translation of the original Gaelic)
The dog is at the door. The cat is on the floor. The swan is on the lake. The seal is on the rock. The man has a head. The cow and the bull are in the meadow.
There is a fort on the hill and there is a man in the fort. What is this? This is a hole. What is in the hole? There is a mouse in the hole. Where is the foal? The foal is in the stable. The boy is at the door with the cow....[etc]

This makes the mistake that Jespersen highlights of being disjointed and "jumping around" between subjects, but is certainly not as bad as his examples.  Jespersen's focus on the disjointedness misses the problems of the individual sentences. The author of the Gaelic book is trying to paint a picture, but he is writing expository text here -- his main goal is still to show the grammar, not to be natural.  Because of this, he ignores the problem of introducing new subjects with a definite article.  "The dog" and "the cat" are fine, because we are all acustomed to talking this way about family pets.  But "the swan" and "the seal" are more troublesome, as I'm likely to ask "which swan?"  The definite article assumes that we have a shared idea of a particular swan or seal.  We're more likely to say things like "there is a swan on the loch", as this doesn't assume any prior knowledge of the swan (I can now use the definite article, because I introduced the swan with "there is...").

The second paragraph is where this really starts to get troublesome, because we hit that old schoolboy motivation-killer: answer in sentences. "What is this? This is a hole." "Where is the foal? The foal is in the stable."  Point out to any teacher that natives don't answer in sentences and you'll get a simple and very logical answer: the reason for answering in sentences is to learn the grammar.  This is the very definition of expository language -- examples that exist purely to demonstrate a language point.

And here's where the "natural" and "direct" methods' justification starts to unravel.  When you're in a monolingual classroom, the simplest way to prompt a student to say something is by asking a question and demanding a fully formed response.  This means that your "natural" method is pretty much guaranteed to produce expository language and not naturalistic or authentic language.

"Answer in sentences" has pervaded language learning, and we see it not only in monolingual methods, but often the bilingual classroom will present new language with a native language explanation followed by monolingual practice.  Even methods using pure translation will often fall into this trap.  The original courses by Michel Thomas did not, but many of the courses written by others under the brand after his death do.  The Japanese course is a perfect example of expository language gone wrong.  The learner is asked to translate "do you want this?" and then "no, I want that."  Now there may not seem to be anything terribly wrong with this at first glance, but think about this: when I am talking to you, what is "this" to me is "that" to you.  This is even more problematic in Japanese, as it has a 3-way distinction equivalent to the Shakespearean "this" (near me), "that" (near you) and "yonder" (near neither of us).  The author is so fixated on the grammatical and lexical contrast between the two sentences that the physical logic of the dialogue is lost.  Again, the expository displaces the naturalistic, and the problem of meaningless and nonsensical language reappears.  Similar problems with here/there/yonder occur in almost all of the Pimsleur courses.  If you listen carefully, you'll often find yourself asking where the hotel is, only to be told it's "there", meaning where you are.

OK, so I have mostly given examples from bilingual courses or courses with explicit instruction.

One of the most vocal opponents of explicit instruction among the internet set is Stephen Kaufmann, Lingosteve on YouTube.  He is adamant that the only way to learn is by understanding bits of language.  He's put together a fairly sophisticated website dedicated to this idea, LingQ.  Kaufmann really hits that "denigration of others" that Decoo points out.  His whole argument is based on the same idea as Jespersen: he associates unnatural language with conscious methods.

But if we have a look at LingQ, will we find evidence of naturalistic or expository material?  Hmm....

Here's the first few lines of the first lesson in Portuguese (my translation):
"Welcome to LingQ.  My name is Mairo. What is your name? I live in Brazil. Where do you live? Do you want to learn Portuguese?..."

The conscious contrast between Mairo's personal information and his request for information from the learner is clearly expository.

And now an early Spanish lesson (again, my translation):
" Listen and repeat: What is your name? My name is Ana. What is his name? His name is Juan. What is her name? Her name is Maria. What age are you? I am 25 years old. What age is Juan? He is 22 years old. How old is Maria? She is 19 years old."
Here again we have clear expository goals: 1) question form vs statement form; 2) contrasting 1st, 2nd and 3rd person conjugations; 3) contrasting masculine and feminine pronouns in the 3rd person.

So even though we aren't going through any native-language instruction, we still get the problems that Jespersen was railing against.  The problem was not the medium of instruction, it was the material.

One form that is very widely used in both monolingual and translating courses is the dialogue.  Some of LingQ's texts are two-man podcasts.  Teach Yourself and Colloquial start each section with a dialog.  Assimil is based almost entirely on dialogues.  Dialogues often include the "answer in sentences" problem as described above, but not always.

The dialogue is said to give a natural context to the language, but sometimes this is assumed and the author ends up ignoring the naturalness of speech and produces a dialogue that is absurd almost to the point of meaninglessness, and becomes once more purely expository language.  This post was inspired by once such book: Beginner's Basque by Wim Jensen.  I can't say I was that hopeful when I picked it up -- it's by Hippocrene Books, who seem to specialise in cheap reprints -- but the first dialogue was worse than anything I have ever seen.  It comes with an English translation on the facing page, so I'll just use that (my comments are in italics.

Bernard: Good morning! I am Bernard. I am a boy. (Would anyone say this?  Certainly, the other person should be able to see that Bernard is a boy, so the effect is of someone with a learning disability.  Except that Bernard is not a boy.  The voice you here is of a man who would appear to be in his late twenties or early thirties.)
Johanna: Hello! I am Johanna. I am a girl. (Classic expository language -- using almost exactly the same structures with a word or two changed.  Again, the effect of learning difficulties comes through, and again, the voice actor is clearly an adult.)
Bernard: My name is Bernard. (Expository -- it restates known information needlessly, simply to demonstrate a different structure) I am Johanna's brother. (Woah there.  Who exactly is Bernard supposed to be talking to? I thought he was talking to Johanna, but there's no way he'd say this to her.)

Johanna: My name is Johanna. I am Bernard's sister. (Again we have an expository near-exact repetition, and again it really doesn't feel like Johanna's talking to Bernard.  Maybe they're introducing themselves to us?  Like a "piece to camera" in a video course?  It's not a particularly natural context though - it's what they call "breaking the fourth wall".)
Bernard: Johanna is a nice name. Your name is nice. (Nope, Bernard is clearly talking to Johanna.  But here again we have repeated information for contrast of structures, in this case attributive vs predicative adjectives.  Naturalisticness has been sacrificed again in favour of exposition.)
Johanna: Yes, it is nice, but Bernard is a nice name too. (And here we have a partial "answer in sentences" and more redundant echoing to demonstrate a particular form.)

Bernard: I am very glad. (??)
Johanna: See you!

This odd dynamic continues throughout the book.  The final dialogue in the book sees Johanna and Bernard discussing a family trip to the mountains.  From the dialogue, they clearly both know the plan, and take it in turns to say parts of it.  Who exactly are they presenting information to?  They are either saying things to each other they already know, or they're talking to you,

So really, dialogues are no kind of magic bullet.  Simply shifting your expository language into a dialogue does not automatically make it natural or meaningful.  Often it forces the author to be more consistent and coherent, but on the other hand, it can actually amplify the absurdity of some sentences by creating a clash between the expected behaviour in the context and the actual words of the participants.

But then we come to one of the most inexplicably popular figures in foreign language learning: Stephen Krashen.  Krashen was one of the big figures in the latest reincarnation of the direct/natural methods (and as Decoo says, in language, every method comes back again and again) and he was big on avoiding rules.  One of his justifications was getting people into "real" language "as soon as possible".  But as I said previously, supporters of authentic material allow it to be doctored and still call it authentic.  Krashen takes this self-deceit a fair bit further by that weaselly phrase "as soon as possible".  "As soon as possible" accepts that it's not possible right from the word go.  Have a quick look at a video of him in action, in a lecture he gave on his theories:

If you think about it, what did he start with?

He took a naturalistic piece of German and demonstrated that it wasn't an effective teaching strategy.  Then he presented a piece of very contrived expository language and called it "comprehensible input".  But it was not comprehensible.  Certain words and phrases were made very obvious, but you did not understand "what he said", but rather fragments of it.

So we go back to Jespersen's original argument -- that bilingual courses result in unnatural examples of the target language.  But monolingual courses are worse -- Krashen demonstrates quite aptly the opposite of his argument: that it is impossible to teach monolingually with natural language.  The one thing in favour of monolingual learning is that it does restrict the artificiality of the language -- the language must be unnatural to be understood, but it cannot be nonsensical or it will not be understood at all.

In that case, monolingual teaching is a bit of a crutch -- it gives us better results without having to fully address the problem.  But without these restrictions, and with a bit of brainpower, a bilingual course can do so much better.  It is extremely hard to elicit sentences like "do you know where it is?" and "I'm sorry, I didn't see you" in a monolingual classroom because of the non-specific function words, but these are extremely natural precisely because of those words; meanwhile they are actually very easy to prompt for by translation.  And once we're into function words, we move onto modality -- needs, desires etc.  These are very difficult to pick up from input, but in the Michel Thomas courses (the originals, not the potboilers produced posthumously), "wanting" appears 15 minutes into the course.  In Italian you'll be saying "I don't want to know", in German "What do you want to eat?" and in French "I would like to speak French". In the Spanish course it's actually held back until a full half hour into the course. *gasp*

Compare Krashen's demonstration with Thomas -- Krashen necessarily gives us easy words, because he relies on physical demonstration.  Thomas gives us words and structures that have vast conceptual meaning, but a very abstract, non-physical concept.  Krashen and his supporters would argue that because we are learning through translation, we are learning to translate.  Yet Krashen has never given any good demonstration of a reliable way to learn this very important functional language.  When it comes to grading authentics, it's the functional language that we generally need to remove to make it what he calls "comprehensible input", because it's inherently non-obvious.  If you want to get into native materials "as soon as possible", it's the non-obvious stuff that you need to teach/learn "as soon as possible".

So Jespersen is mostly wrong.  Yes, the worst examples of meaningless expository language could only occur in a bilingual course, but the cure is not to go monolingual, because only a bilingual translating course can employ genuinely natural language.