By Gordon Hull

One of the things that marketers like about big data is that they can personalize ads.  That operation is getting increasingly sophisticated.  We’ve known for a while that basic personality traits (like introversion/extraversion) can be predicted from Facebook likes.  I missed this paper when it came out, but some of the same authors as the initial Facebook “likes” paper have now done the inevitable follow-up.  By targeting Facebook ads on the basis of openness and extraversion (where the correlation with likes is fairly robust), they were able to make users 1.54 times more likely to make a purchase (thus about 40% more likely) than with non-“psychologically-tailored” advertising.  The study size, as with most research on FB, is enormous – the study reached some 3.5 million users (a point the authors use to try to defuse some objections).  The authors duly note the murky ethical issues that emerge: on the one hand, you could proactively try to assist people who show signs of depression; on the other hand, weaknesses like susceptibility to addictive gambling or dubious political targeting could be exploited.  That observation is pretty obvious now, in the wake of Cambridge Analytics.  Two other takeaways stood out more to me.  First, the study predicted personality on the basis of only one like.  As the authors note, that means the study likely underestimates the potential effect of psychological targeting.  Second, the authors emphasize that a point that a lot of us have been trying to make for a while: that these results undermine a lot of the current regulatory strategy for privacy. I’ll let them speak:

“The psychological targeting procedure described in this manuscript challenges the extent to which existing and proposed legislation can protect individual privacy in the digital age. While previous research shows that having direct access to an individual’s digital footprint makes it possible to accurately predict intimate traits, the current study demonstrates that such inferences can be made even without having direct access to individuals’ data. Although we used indirect group-level targeting in a way that was anonymous at the individual level and thus preserved—rather than invaded—participants’ privacy, the same approach could also be used to reveal individuals’ intimate traits without their awareness. For example, a company could advertise a link to a product or a questionnaire on Facebook, targeting people who follow a Facebook Like that is highly predictive of introversion. Simply following such a link reveals the trait to the advertiser, without the individuals being aware that they have exposed this information. To date, legislative approaches in the US and Europe have focused on increasing the transparency of how information is gathered and ensuring that consumers have a mechanism to “opt out” of tracking. Crucially, none of the measures currently in place or in discussion address the techniques described in this paper: Our empirical experiments were performed without collecting any individual-level information whatsoever on our subjects yet revealed personal information that many would consider deeply private. Consequently, current approaches are ill equipped to address the potential abuse of online information in the context of psychological targeting” (internal citations omitted).

Marketing has of course become increasingly targeted over time.  Early magazine ads imagined their readership into very rough categories (“homeowning housewife”) and tailored ads to an imagined construction of what such a reader was like; this later moved into empirical work on the tastes of such individuals at increasing levels of granularity.  But data represents something different insofar as it vastly increases the amount that can be known about targets, and it enables the sort of predictive modeling that the work on FB likes demonstrates.  That result was anticipated in a 2015 paper that proposed that systems could dynamically adapt nudges to the type of ‘persuasion principle’ that was most effective on the individual being nudged. Thus, some people would see a message instructing them what to do; others will receive one claiming that a social consensus supports the desired behavior, etc.   As Adam Arvidsson was able to presciently propose fifteen years ago:

“As consumers, mobile and supposedly self-reliant, we are subject to a virtually ever-present ‘panoptic sort’; a surveillant gaze that expands far beyond the walls of the prisons, factories and hospitals and schools where, according to Foucault the panoptic model first developed. The physical, social and cultural mobility of social life, the moving about between environments and activities that has become a key characteristic of post-modern life, has also become a source of value to be realised on the market for commodified information.” (457; internal citations omitted)

The current research on FB likes as a tool for targeting shows this is not that hard at all. You can find out everything you need to know about somebody based on what they say they like on Facebook.  One of the likes associated with openness?  “Philosophy.”  Enjoy your ads!

Posted in , , ,

6 responses to “The Big Data Train Chugs On, Predictive Personality Edition”

  1. dmf Avatar

    do we know (and if so how) these nudges supposedly translate/correlate to off the screen behaviors?
    https://www.wired.com/story/the-noisy-fallacies-of-psychographic-targeting/

    Like

  2. Gordon Hull Avatar

    So that’s the million dollar question. My hunch is that yes, this can influence off-screen behaviors, but that the psychographic stuff that Cambridge Analytica is peddling is probably overhyped. If we grant that you can get big-5 personality types from “likes,” then you’d need to be able to go from those to political leaning. My sense is that there is literature to that effect – openness correlates with being liberal, etc. I don’t know how good it is, and it sounds like there may be some issues with that literature. It seems to me that by the time you get to the idea that you’re going to change somebody’s vote, that’s a lot of steps to get through, and I can’t imagine that the result is that robust, especially since different people respond to different kinds of political ads. That’s the case against, anyway.
    I think a more likely scenario is that the real goal was to drive turnout in selected groups. Facebook users can be nudged to vote at slightly higher rates:
    https://www.nature.com/news/facebook-experiment-boosts-us-voter-turnout-1.11401
    (that’s a link to the writeup – the actual study links at the end of it, but it’s paywalled)
    You wouldn’t be able to make a huge difference in turnout, but in a close election if you sent nudges to people whose political identity you were pretty sure of (the earlier of the two Kosinski studies says that’s predictable with fairly high accuracy using likes), then you might generate enough more votes for your candidate to tip a close election.
    All of which is to say that I think there probably is a real-world impact here in that people’s behavior can be subtly changed, either for voting or buying stuff. The amount of change is fairly low – but you don’t need to change a lot of people’s behavior for, say, an ad-targeting effort to be worth it. And all of this is at such massive scale that even a very low percentage adds up to a fair number of people.

    Like

  3. dmf Avatar

    thanks can certainly see how at such scales these things can add up just worried that the hype is far from the engineering (social as well as machine), never been convinced by the research around personality tests and the like (no reliable predictive powers demonstrated) and the fears/hype around sales/media is all too familiar,
    more worried about the ways in which the numbers games (searches etc) are being manipulated, the algorithmic biases Frank Pasquale and co. are tracking, and the sorts of monopolistic/data-sovereignty issues Evgeny raises:
    https://2018.festivaleconomia.eu/-/geopolitica-e-geoeconomia-dell-intelligenza-artificia-1

    Like

  4. Dave Dayanan Avatar

    Thanks for clarifying my Mind Gordon. Great Article.

    Like

  5. Gordon Hull Avatar

    Agreed that Frank Pasquale’s work is great! (full disclosure: he’s a coauthor on a recent paper about employee wellness programs)
    The stuff about algorithmic bias is really scary, and the work on it is starting to be really good. If you don’t know them yet, try these papers:
    Solon Barocas and Andrew Selbst, ‘Big Data’s Disparate Impact,” https://ssrn.com/abstract=2477899
    Selbst, “Disparate Impact in Big Data Policing,” https://ssrn.com/abstract=2819182
    Margaret Hu, “Algorithmic Jim Crow,” https://ssrn.com/abstract=3071791
    There’s also Cathy O’Neill’s book, Weapons of Math Destruction and the Virginia Eubanks book, *Automating Inequality.”
    I’ll have to follow the subtitles (alas, I have no Italian) on Morozov – I generally like his work too.

    Like

  6. dmf Avatar

    thanks I’ll have to track down yer paper with FP (is it OA somewhere?), Evgeny’s talk is available there in his english (scroll over the tiny headphone symbol) be interested in what you make of it, his partner in thoughtcrime (@francesca_bria) is going to be at an interesting looking conference soon: http://barcelona.makerfaire.com/conversations/
    I’ve corresponded with Cathy some about these issues she’s very sharp and has been in the belly of the beast which is helpful, and some interesting crossovers forming with folks like the Open Markets Institute.

    Like

Leave a comment