Posts from January 2017 (25)

January 30, 2017

Stat of the Week Competition: January 28 – February 3 2017

Each week, we would like to invite readers of Stats Chat to submit nominations for our Stat of the Week competition and be in with the chance to win an iTunes voucher.

Here’s how it works:

  • Anyone may add a comment on this post to nominate their Stat of the Week candidate before midday Friday February 3 2017.
  • Statistics can be bad, exemplary or fascinating.
  • The statistic must be in the NZ media during the period of January 28 – February 3 2017 inclusive.
  • Quote the statistic, when and where it was published and tell us why it should be our Stat of the Week.

Next Monday at midday we’ll announce the winner of this week’s Stat of the Week competition, and start a new one.

(more…)

Stat of the Week Competition Discussion: January 28 – February 3 2017

If you’d like to comment on or debate any of this week’s Stat of the Week nominations, please do so below!

January 28, 2017

Charms to soothe the savage beast

Q: Did you see dogs prefer reggae and soft rock?

A: Not rap?

Q: Rap? You mean because of the human voices? Or because of Snoop Dogg?

A: Um. Yes. Voices. Definitely the voices thing. Wouldn’t dream of the horrible pun.

Q: Anyway, how did they find out what sort of music the dogs liked? Did they give them buttons to push, like those experiments with rats?

A: No

Q: Did they see which speaker the dogs liked to sit near?

A: No.

Q: Can you work with me here?

A: They measured how relaxed the dogs were, by heart rate and whether they were lying down, and whether they were barking.

Q: The music they ‘liked most’ was really the music that made them lie down quietly and relax?

A: Yes.

Q: Have these people ever been teenagers?

A: To be fair, the research paper didn’t claim they were looking at preferences. That seems to be an invention of the press release.

Q: That would be the research paper that none of the stories linked, and most of them didn’t even hint at the existence of?

A: Yes, that one.

Q: So what were they really looking at?

A: The Scottish SPCA wants dogs to be quiet and relaxed (and presumptively  happy) in the kennels, while they’re waiting to find a new home.

Q: And soft rock and reggae were more relaxing than rap or thrash metal?

A: They didn’t look at all musical genres, just a few.  The dogs got a week of no music and a week with a different style each day (in random order, with music from Spotify).

Q: Soft rock and reggae were better than the other ones?

A: Well, Motown seemed to increase heart rate rather than decrease it, but the others were all pretty much the same.

Q: The others?

A:  Soft Rock, Reggae, Pop, Classical, Silence

Q: Wait, what? “Silence”?

A: Yes, a day of no music was about as good as a day of relaxing music.  It looks like variety might be the key. The researchers say

Interestingly, the physiological and behavioural changes observed in this study were maintained over the 5d of auditory stimulation, suggesting that providing a variety of different genres may help minimise habituation to auditory enrichment

Q: So what they really found is that playing dogs a variety of music relaxes them

A: Yes, but that’s not such a good headline.

January 23, 2017

Stat of the Week Competition: January 21 – 27 2017

Each week, we would like to invite readers of Stats Chat to submit nominations for our Stat of the Week competition and be in with the chance to win an iTunes voucher.

Here’s how it works:

  • Anyone may add a comment on this post to nominate their Stat of the Week candidate before midday Friday January 27 2017.
  • Statistics can be bad, exemplary or fascinating.
  • The statistic must be in the NZ media during the period of January 21 – 27 2017 inclusive.
  • Quote the statistic, when and where it was published and tell us why it should be our Stat of the Week.

Next Monday at midday we’ll announce the winner of this week’s Stat of the Week competition, and start a new one.

(more…)

Stat of the Week Competition Discussion: January 21 – 27 2017

If you’d like to comment on or debate any of this week’s Stat of the Week nominations, please do so below!

January 20, 2017

Recycling

Herald storySchool costs: $40,000 for ‘free’ state education

Last year’s Dom Post storyFamilies struggle to afford the rising cost of back-to-school requirements

Recycling last year’s StatsChat post:

So, it’s a non-random, unweighted survey, probably with a low response rate, among people signed up for an education-savings programme. You’d expect it to overestimate, but it’s not clear how much. Also

Figures have been rounded and represent the upper ranges that parents can reasonably expect to pay

It’s a real issue, but these particular numbers don’t deserve the publicity they get.

Measuring accuracy

Stuff  has a story “Scanner that can detect brain bleeds to be introduced in New Zealand.” The scanner uses infrared light to see relatively small accumulations of blood in the brain, with the aim of detecting bleeding quickly.  The traditional approach of looking at symptoms can often miss a bleed until after it’s done a lot of damage.

Accuracy is important for a device like this.  You don’t want to send lots of people off for CT scans, which are expensive and expose the patient to radiation, but you also don’t want to falsely reassure someone who really has a brain injury and who might then ignore later symptoms.

The story at Stuff claims 94% accuracy, but doesn’t say exactly what they mean by ‘accuracy’. Another story, at Gizmodo, says “A green light on the scanner gives the patient the all clear, and a red light shows a 90 per cent chance of haemorrhage.” The Gizmodo figures fit better with what’s on the manufacturer’s website, where they claim “Sensitivity = 88% / Specificity = 90.7%”.  That is, of people with (the right sort of) bleed, 88% will be detected, and of people without those bleeds, 90.7% will be cleared.

The Gizmodo story still confuses the forward and backwards probabilities. Out of 100 people with brain bleeds, 88 will get a red light on the machine. That’s not the same as their claim: that out of 100 people who get a red light on the machine, 90 have a bleed.

Suppose about 10% of the people it’s used on really do have brain bleeds. Out of an average 100 uses there would be 10 actual bleeds, 9 of whom would get a red light. There would be 90 without actual bleeds, about 9 of whom would get a red light.  So the red light would only indicate about a 50% chance of a haemorrhageThat’s still pretty good, especially as it can be done rapidly and safely, but it’s not 90%.

The other aspect of the story that’s not clear until you read the whole thing is what the news actually is. Based on the headline, you might think the point of the story is that someone’s started using this device in NZ, maybe in rugby or in ambulances, or is trialling it, or has at least ordered it.  But no.

No-one in New Zealand has yet got their hands on an infrascanner, but the hope is for it to be rolled out among major sporting bodies, public and private ambulance services, trauma centres and remote healthcare facilities.

 

 

January 18, 2017

Recognising te reo

Those of you on Twitter will have seen the little ‘translate this tweet’ suggestions that it puts up. If you’re from or in New Zealand you probably will have seen that reo Māori is often recognised by the algorithm as Latvian, presumably because Latvian also has long vowels indicated by macrons.   I’ve always been surprised by this, because Latvian looks so different.

It turns out I’m right.  Even looking just at individual letters, it’s very easy to distinguish the two.  I downloaded 74000 paragraphs of Latvian Wikipedia, a total of 6.5 million letters, and looked at how long the Latvians can go without using letters that don’t appear in te reo: specifically, s,z,j,v,d,c, g not as ng, the six accented consonants, and any consonant at the end of a word. On average, I only needed to wait five letters to know the language is Latvian rather than Māori, and 99% of the time it took less than 21 letters.

Another language that Twitter often guesses is Finnish. That makes more sense: many of the letters not used in Māori are also rare or absent in Finnish, and ‘g’ appears mostly as ‘ng’.   However, Finnish does have ‘s’, has ‘ä’ and ‘ö’, and ‘y’, and has words ending in consonants, so it should also be feasible to distinguish.

 

Update: Indonesian is another popular guess, but it has ‘d’,’j’,’y’,”b”, and it has lots of works ending with consonants.  The average time to rule out te reo is slightly longer, at nearly 6 characters, and the 99th percentile is 22 letters.  So if the algorithm can’t tell, it should probably guess it’s not Indonesian.

Update: For very short tweets, and those in mixed languages, nothing’s going to work, but this is about tweets where the answer is obvious to a human.

January 17, 2017

Briefly

  • There’s a planned course at the University of Washington “Calling Bullshit in the Age of Big Data”. Here’s the website with syllabus and readings, and the Twitter account.
  • Via a tweet from ‘Calling Bullshit’, there’s a computer science preprint looking at distinguishing ‘criminals’ from ‘normal people’ using photographs.  I usually wouldn’t comment here on research papers that haven’t made it to the news, but this sentence was irresistible
    “Unlike a human examiner/judge, a computer vision algorithm or classifier has absolutely no subjective baggages, having no emotions, no biases whatsoever due to past experience, race, religion, political doctrine, gender, age, etc.
    An aim of both the course and this blog is to increase the number of people who find this sort of claim ridiculous.
  • For map nerds: a detailed cartographic comparison of Google Maps and Apple Maps.
  • Data journalism: the Guardian looks at the spatial concentration of gun violence in the US.
  • There’s a quote circulating widely now on social media “Journalism is  printing what someone else does not want printed. Everything else is public relations.” It’s being attributed to Orwell. He didn’t say it — which I think matters in this context.
    According to Quote Investigator, versions of it described as an ‘old saying’ were around in US journalism in the early 20th century. Later, in 1930, Walter Winchell attributed a version to William Randolph Hearst. More recently, it has been attributed to Lord Northcliffe, a UK pioneer of tabloid journalism. It wasn’t attributed to Orwell until the 1990s, decades after he died.

And finally: this is actually true

January 16, 2017

Stat of the Week Competition: January 14 – 20 2017

Each week, we would like to invite readers of Stats Chat to submit nominations for our Stat of the Week competition and be in with the chance to win an iTunes voucher.

Here’s how it works:

  • Anyone may add a comment on this post to nominate their Stat of the Week candidate before midday Friday January 20 2017.
  • Statistics can be bad, exemplary or fascinating.
  • The statistic must be in the NZ media during the period of January 14 – 20 2017 inclusive.
  • Quote the statistic, when and where it was published and tell us why it should be our Stat of the Week.

Next Monday at midday we’ll announce the winner of this week’s Stat of the Week competition, and start a new one.

(more…)