Business Acumen Didn’t Help The Missing Link

First caveat, this research was all done before the dawn of the reproducibility movement, so take all of this with a grain of salt. Some day hopefully all discussion of science requires considering this. Anyway, onto the show.

There seems to be this common finding in psychology literature: people make accurate snap judgments about certain social qualities of others.

The term thin slice comes from a frequently cited article by Nalini Ambady and Robert Rosenthal, in which subjects evaluated thirty-second silent video clips of instructors teaching a class.’ Subsequent analysis found that these brief evaluations predicted the instructors’ end-of-semester student ratings. Their work built on earlier research that found a similar predictive power in job interviews,’ where the first impressions were critical for the eventual hiring decision.

You know, trust your gut and all that. Here’s some more of the same theme:

Judgments made after a 100-ms exposure correlated highly with judgments made in the absence of time constraints, suggesting that this exposure time was sufficient for participants to form an impression.

Great. So here’s my question: why the hell is it so hard to figure out whether a job candidate will be any good? Consider the findings of employment-related thin-slice research: it can predict whether the person is hired or not. It cannot predict if the person will be any good on the job. Here’s a quote from this paper:

DeGroot and Motowidlo (1999) found that actual performance ratings of managers in a news-publishing company were associated with naive raters’ judgments. Ratings of 10-second clips of interviews of 22 managers on liking, trust, competence, dominance, persuasiveness, influence and willingness to help the manager, when combined into a composite, were significantly positively correlated with job performance ratings by their supervisors.

In contrast, a study on the vocal characteristics of direct salespeople found a relation between certain microtraits and actual sales performance, but no relation between thin-slice trait judgments and actual sales performance (Peterson, Cannito & Brown, 1995). Twenty-one direct salespeople gave audiotaped, scripted sales-pitch introductions, which were five sentences and 64 words long. Speaking rate, average pause duration, loudness, variability and fundamental frequency (the vibration rate of the vocal folds in the throat) were measured and a sample of housewives who were representative of the target audience, rated the salespeople on questions related to the salespeople’s personality characteristics and rated their own receptivity toward the salespeople. The housewives were able to detect difference in speaking rates, but this was not related to differences in their perceptions of the salespeople. In addition, the housewives judgments of attitude and receptivity were not correlated with sales performance.

What can we learn from all this? Why don’t we have a basic intuition for business ability? It seems that business skills are simply not something that mattered in the evolutionary crucible. Idle speculation here but maybe things like social status and dominance are good things to snap judge because they help avoid costly conflict.

Business, though? Cro-Mangon man says: “Meh”.

Psychology Fights BS

I like the idea of psychology because it lets me anchor intuition about others in something more concrete. Perhaps unsurprisingly for a field so deeply susceptible to BS, psychology research is currently at the forefront of the reproducibility movement. Consider Ed Yong’s reporting work:

“Some people are concerned that this will damage psychology as a whole and the public will perceive an epidemic of fraud,” says Simonsohn. “I think that’s unfounded.” He notes that retractions are common in many fields, and cites the case of anaesthesiologist Yoshitaka Fujii, who was recently found to have fabricated data in at least 172 papers.

“We in psychology are actually trying to fix things,” he says. “It would be ironic if that led to the perception that we are less credible than other sciences are. My hope is that five years from now, other sciences will look to psychology as an example of proper reporting of scientific research.”

It’s important to realize that much of scientific research is BS. At least psychology, unlike economics, has real experiments to reproduce!

Lightning

I find it weird that we don’t know what causes lightning.

There are two hypotheses noted by Wikipedia:

Cloud particle collision hypothesis

According to this cloud particle charging hypothesis, charges are separated when ice crystals rebound off graupel. Charge separation appears to require strong updrafts which carry water droplets upward,supercooling them to between -10 and -40 °C (14 and -40 °F). These water droplets collide with ice crystals to form a soft ice-water mixture called graupel. Collisions between ice crystals and graupel pellets usually results in positive charge being transferred to the ice crystals, and negative charge to the graupel.[14]

Updrafts drive the less heavy ice crystals upwards, causing the cloud top to accumulate increasing positive charge.Gravity causes the heavier negatively charged graupel to fall toward the middle and lower portions of the cloud, building up an increasing negative charge. Charge separation and accumulation continue until the electrical potential becomes sufficient to initiate a lightning discharge, which occurs when the distribution of positive and negative charges forms a sufficiently strong electric field.[14]

Polarization mechanism hypothesis

The mechanism by which charge separation happens is still the subject of research. Another hypothesis is the polarization mechanism, which has two components:[34]

  1. Falling droplets of ice and rain become electrically polarized as they fall through the earth’s magnetic field;
  2. Colliding/rebounding cloud particles become oppositely charged.

There are several hypotheses for the origin of charge separation.[35][36][37]

Here’s a cool video of lightning captured at 7,207 images per second. Light travels at something like 300,000,000 meters per second so it doesn’t break down the ‘big flash’, but it shows a lot.

Here’s another quote:

How does lightning form? Evidently we’re still trying to figure it out! It all starts in the clouds where both ice crystals and hail stones form:

Scientists believe that as these hail stones fall back through the rising ice crystals, millions of tiny collisions occur. These collisions build up an electric charge which is stored in the cloud like a battery. ”A cloud is very much like a battery, but a battery with a much higher voltage than your typical flashlight battery… not 1.5 volts but 100 million volts.”

But what scientists don’t know is exactly how this electric charge generates lightning. “What remains a major meteorological mystery is how it is that ice particle collisions result in the generation of lightning. We’re very much in the middle ages on that problem.”

From the Discovery Channel’s “Raging Planet” series.

Big Data To Cure Cancer? Matter of Time

I almost can’t believe this is happening. Incredibly exciting. Get used to these kinds of projects.

In 2007, Ian Clements was given a year to live. He was diagnosed with terminal metastatic bladder cancer. Ian began charting, quantifying, and recording as much of his life as possible in an effort to learn which lifestyle behaviors have the greatest impact on his cancer.

Ian has fought his disease successfully for five years, and now he asks the Kaggle community to look at his data to see what significant correlations and connections we can find. We at Kaggle are humbled by his efforts and want to help Ian share his data with the wider world by hosting it on our website.

This is an exercise in collaborative data exploration rather than a standard Kaggle competition. The ideal result would be a model suggesting which lifestyle behaviors may have the greatest effect on Ian’s health, but any insights into his dataset are welcome. While we understand it may not be possible to extrapolate insights from this dataset to the overall population, it will nevertheless be very helpful for Ian in generating hypotheses and suggesting different behaviors. We hope that you will find it interesting to take a look and see what you can find.

Dear youths of the world: GET INTO THIS FIELD.

Medicine, *BIG* Data and $$$

Another tour de force from SBM. How does one summarize? I almost blogged this NYT article about treating Leukmia last week but felt I had nothing to add (go read it!). I should have known SBM would deliver the goods, though.

Here’s the big data part:

Taking the results of the sequencing of the entire genome and RNAseq data and analyzing them allows scientists to probe the genome and transcriptome of cancers in a way that was never before possible. It produces an enormous amount of data, too, terabytes from a single experiment. At cancer meetings I’ve been to, investigators frequently refer to a “firehose” of data, petabytes in magnitude.

I’ll offer comment on this part:

There’s no doubt that “individualized” medicine will become increasingly a part of modern medical care, with the individualization based on sequencing the genomes and transcriptomes of patients. In just a few years, the price of a complete genome sequence has fallen from hundreds of thousands of dollars to around $15,000. True, that doesn’t count all the analysis and that’s $15,000 per genome, which means at least $30,000 to sequence a normal and cancerous genome. There are, however, lots of things we do in medicine that cost $15,000. The price doesn’t have to come down much more before whole genome sequencing starts to look doable for individual patients. After all, gene tests like the OncoType DX cost on the order of $3,000 to $4,000, and we now order this test fairly routinely for patients with estrogen receptor-positive, node-negative breast cancer because in the end it saves a lot of patients from unnecessary chemotherapy.

The bottom line is that at some point every single person is going to get their genome sequenced. That’s about 4m newborns per year after the backlog of 330m+ people. But here’s the thing with cancer, it’s a genetic disease, which means that the cancer itself has a different genome than yours. Finding those differences is the entire point of genetic therapy.

So $3,500 per genome x 35,000 leukemia patients per year = $122m of new health care costs per year. No big deal, right? Well how about the 1.5m people who get diagnosed with all cancers per year?

Genome sequencing is going to be a gigantic business very, very soon. The health care cost curve is bending, all right.

Dear Geeks of Sport and Science: Feast On THIS

What would happen if you tried to hit a baseball pitched at 90% the speed of light?
– Ellen McManis
Let’s set aside the question of how we got the baseball moving that fast. We’ll suppose it’s a normal pitch, except in the instant the pitcher releases the ball, it magically accelerates to 0.9c. From that point onward, everything proceeds according to normal physics.:

The ideas of aerodynamics don’t apply here. Normally, air would flow around anything moving through it. But the air molecules in front of this ball don’t have time to be jostled out of the way. The ball smacks into them hard that the atoms in the air molecules actually fuse with the atoms in the ball’s surface. Each collision releases a burst of gamma rays and scattered particles.

That’s xkcd‘s new blog. And there’s this incredible illustration:

And much much more.

The most exciting blog idea I’ve come across in years? Yep.

Debunking Dr. Oz (Or: How Charlatans Do Science)

…when the sign in front of my local pharmacy started advertising “Green coffee beans – as seen on Dr. Oz” [as a weight loss wonder drug -DW], I tracked down the clip in question.

That’s Scott Gavura of Science-Based Medicine (SBM), who then goes on to drilling into a (the?) piece of evidence that could possibly support this supplement manufacturer’s claims. I have to admit I love reading the occasional holy smackdown of crackpot science. And today I got my fill:

…Green coffee extract (the brand “GCA”) was used in the study. The authors note that GCA has a standardized content of 45.9% chlorogenic acid, which is purported to be the active ingredient. Now contrary to what was said on the Dr. Oz show, chlorogenic acid is also in roasted coffee in significant amounts, so you don’t need to take green coffee extract to get a good dose.

…The study is entitled Randomized, double-blind, placebo-controlled, linear dose, crossover study to evaluate the efficacy and safety of a green coffee bean extract in overweight subjects. The lead author, Joe A Vinson, is a chemist at the University of Scranton, Pennsylvania. None of the three authors appear to be clinicians or medical professionals, and none appear to have published obesity-related research before, according to PubMed. The study was funded by a supplement manufacturer, Applied Food Sciences.

To start — this is a very tiny trial — just 16 patients (8 males, 8 females) with an average age of 33 years. The research location was a hospital in Bangalore, India. How these patients were recruited was not disclosed. Normally a trial would list detailed inclusion and exclusion criteria, and then describe how many patients were considered and the reasons for exclusion. This paper just reports the final number, and there is no information provided on why 16 was felt to be the desireable number. The average weight was 76.6kg (168 lbs) and the average body mass index (BMI) was 28.22. While the BMI on an individual basis may not be informative, when looking at a population, a score between 25 and 30 is usually accepted to mean overweight, but not obese. The details on how these measurements were taken were not well described — which is surprising, given this is a this is a pretty important part of the study.

And on and on it goes. Science lovers and crackpot-haters, do feast your eyes on the whole article.

Big Deal In Physics Today

Today CERN announced that they have probably found the Higgs Boson. I read this last night so I was expecting it:

They’re very likely going to announce that the 5-σ threshold, the “gold standard” for discovery, has been reached, and that we’ve discovered the Higgs boson.

Assuming things go as we expect, the speculation will turn to the question of what does it mean, and I’ve got some early analysis for you.

For a Higgs right around 125 GeV, which is where all preliminary analysis points, we expect that the Higgs will be produced at a certain rate, and that it will decay into various other particles at particular rates. What are those rates? There’s an excellent analysis in this paper, that shows what the standard model rate is for various possibilities, and what the preliminary data is for each detector, thus far.

If this is, in fact, where the Higgs appears to be, and the rates observed are consistent with the standard model predictions, and there are no other “new particle” announcements that come out on the 4th, then this is an amazing victory for the standard model.

And a nightmare scenario for everything else, including supersymmetry, extra dimensions, and string theory.

Here’s a bit more:

The great nightmare of people interested in particle physics is that the standard model works toowell. That the Higgs exists somewhere between 120 and 140 GeV, that there’s no supersymmetry or extra dimensions or composite Higgs or technicolor or anything surprising.

And this would mean that we’d never understand why the Higgs couples the way it does, or whythe particles in the Universe have the masses that they do; the best we’d be able to say is, “They just do.”

So I’m hopeful that they don’t find only the Higgs, because if they do, we could be living through the last hurrah of particle physics, and there’s too much that we still need to know! Here’s hoping…

I haven’t seen anything in any of the press releases commenting on whether this nightmare scenario is true or not.

The Time Smear

Soon after the advent of ticking clocks, scientists observed that the time told by them (and now, much more accurate clocks), and the time told by the Earth’s position were rarely exactly the same. It turns out that being on a revolving imperfect sphere floating in space, being reshaped by earthquakes and volcanic eruptions, and being dragged around by gravitational forces makes your rotation somewhat irregular. Who knew?

That’s the official Google blog.

Now that’s all well and good but eventually we started building systems that need to all talk to each other and agree on the time. If our arbitrary system of time keeping doesn’t/can’t exactly match the (changing) benchmark of the Earth’s position in space, what are we to do? In other words:

Very large-scale distributed systems, like ours, demand that time be well-synchronized and expect that time always moves forwards. Computers traditionally accommodate leap seconds by setting their clock backwards by one second at the very end of the day. But this “repeated” second can be a problem. For example, what happens to write operations that happen during that second? Does email that comes in during that second get stored correctly?

Well, Google does something called a Time Smear:

The solution we came up with came to be known as the “leap smear.” We modified our internal NTP servers to gradually add a couple of milliseconds to every update, varying over a time window before the moment when the leap second actually happens. This meant that when it became time to add an extra second at midnight, our clocks had already taken this into account, by skewing the time over the course of the day. All of our servers were then able to continue as normal with the new year, blissfully unaware that a leap second had just occurred.

Cool! More discussion on the topic from HN here.