Main menu

Here Is How You Can Sell Your Old Gadgets Online

BestsellerMagazine.com - CATEGORY Headline Report: TITLE

What this can lead to is communities and, eventually, a nation infuriated by things others don’t know about. The information that makes us angriest becomes the information least likely to be questioned. We wind up stewing over things that, by design, few others can correct, engage with or learn from. A Jeffersonian public square where lots of viewpoints go to mingle, debate and compromise, this is not.

The reason you don’t know about all of these things is the same reason that we as citizens should be worried: the whole system operates in the darkness of proprietary data and algorithmic processes at internet companies. Unless they tell us how they use the data they collect about us and design their targeting algorithms, we can only guess.

It’s important to distinguish what is new about this process. There are similarities here to the world of broadcast political ads — which itself is a cesspool. If a campaign produces a profoundly noxious ad, it might get some extra coverage in the news about how the ad was especially outrageous, meaning more people see it but at no additional cost. But producing and buying airtime for a TV ad is a lot more expensive and reaches a lot fewer people than if a political organization can make a toxic ad go viral on Facebook, Twitter or YouTube and reach millions of people. Plus, when that organization does this on TV, it is transparent to everyone what it’s doing because the ad is on TV, and the organization is required by law to put its name on the ad and survive regulatory scrutiny. If a political organization does the same thing on a social media platform, it is, again, only visible to the people you targeted and those they share it with. And labels showing who bought the ad are often not all that they could be. (The companies say they are fixing this — and if they don’t, then regulation probably will compel it.)

TV stations also don’t have nearly as much detailed information about what makes their viewers react. But social media sites do. This gives political groups and campaigns incredible power — and is the secret sauce of the Internet advertising business, since it also proffers commercial advertisers the same ability. Here’s how it works.

First, the campaign collects as many email addresses from as many places as possible from potential supporters. Sometimes they take the voter file itself (the public list of all registered voters) and use data-mining techniques to match names and home addresses with email addresses. Next, the campaign uploads that massive list of email addresses into a social media service. Facebook, for example, can match the email addresses to individual users to create a “Custom Audience.” This is where Cambridge Analytica had a huge advantage, since they had the private Facebook data themselves and did not need to rely on guesswork to match email addresses and Facebook pages to actual voters. This audience can then be sliced and diced into different demographic groups, right down to people’s political and cultural preferences and biases. Here again, Cambridge Analytica may well have used their own private data stash to figure out ways to target specific voters with specific messages by studying their past behavior on Facebook. These filtered groups can then be tested to see which people respond well to which messages. From there, Facebook has another tool called “Lookalike Audiences” — as do other sites – that will find people that are similar to those designated in any given slice of the Custom Audience. Then the campaign buys ads that deliver the messages that Facebook data confirms people want to hear — turning up the outrage and sensation factor to get attention (ideally for free).

All of this does not add up to sites like Facebook and Twitter intentionally undermining Hillary Clinton. It is simply the nature of ad tech and social media: use personal data to divide up classes of the American population like barn animals, then feed us highly personalized messages designed to push our particular buttons so well that we share them and they go viral, thus keeping people on the site longer. Social media rewards provocation — again, without repercussion, since we usually only share content with our friends in a way that is largely invisible to the broader public. Morality and integrity count little in online advertising.

The real question here isn’t which campaign got the advantage. The real question is whether this micro-targeted free-for-all should be allowed in the political sphere at all in the way it is currently designed —with very little transparency about who is pulling these strings and how they are doing it.

When Russian trolls used social media to manipulate the voting public — apparently even more cost-effectively than the Trump campaign — it triggered a national scandal along with demands that the social media companies be held accountable for letting it happen. But when our own political parties do this to us, we often turn a blind eye. It’s politics as usual. Perhaps it is time to reconsider how and when we should set appropriate restrictions on the use of social media for political communications, especially as another set of national elections is just around the corner.

Beyond basic commitments to assure the privacy and security of their personal data, voters have a right to know who is trying to send them political messages and how they are doing it. They should know who bought the ads, how much they spent and what particular demographic audiences were targeted. They should be able to look at all the ads run by the people trying to reach them. This should notbe a database available somewhere on the Internet that normal users would never visit. It should be pushed forward as a part of the ad, so that it is easy to click and see the data right there. Beyond that, social media sites should take action against any political communicator that tries to break the rules. Facebook has responded to the Cambridge Analytica story to say that the technique used to extract data from 50 million users is no longer allowed. But no one is sure exactly what types of sensitive personal data are already out there or who possesses them — making it likely that efforts at voter exploitation will continue. For this reason, the companies should continue to develop better algorithmic detection systems to discover attempts to mislead prospective voters and act against them before or soon after they are disseminated.

But the American public must be wary, since the drive toward total transparency is unlikely to come from the politicians currently in power — or the tech companies themselves, even after they adopt some laudable measures. If there is anything we should learn from the Cambridge Analytica revelations, it is that unless things change, we can expect the spread of disinformation and the systemic manipulation of voters to happen all over again, not only in U.S. national elections but throughout the world. Because if there’s one thing everyone can agree on, it’s that these tools are effective.