Facebook’s Future Plans for Data Collection Beyond All Imagination

Created on Saturday, 31 May 2014 09:44

Facebook's dark plans for the future are given away in its patent applications.

“No one knows who will live in this cage in the future, or whether at the end of this tremendous development, entirely new prophets will arise, or there will be a great rebirth of old ideas and ideals, or, if neither, mechanized petrification, embellished with a sort of convulsive self-importance. For of the fast stage of this cultural development, it might well be truly said: ‘Specialists without spirit, sensualists without heart; this nullity imagines that it has attained a level of civilization never before achieved.’”

—Max Weber, 1905

 On November 12, 2013, Facebook, Inc. filed its 178th patent application for a consumer profiling technique the company calls “inferring household income for users of a social networking system.”

“The amount of information gathered from users,” explain Facebook programmers Justin Voskuhl and Ramesh Vyaghrapuri in their patent application, “is staggering — information describing recent moves to a new city, graduations, births, engagements, marriages, and the like.” Facebook and other so-called tech companies have been warehousing all of this information since their respective inceptions. In Facebook’s case, its data vault includes information posted as early as 2004, when the site first went live. Now in a single month the amount of information forever recorded by Facebook —dinner plans, vacation destinations, emotional states, sexual activity, political views, etc.— far surpasses what was recorded during the company’s first several years of operation. And while no one outside of the company knows for certain, it is believed that Facebook has amassed one of the widest and deepest databases in history. Facebook has over 1,189,000,000 “monthly active users” around the world as of October 2013, providing considerable width of data.

Facebook has also stored away trillions and trillions of missives and images and logged other data about the lives of this billion plus statistical sample of humanity. Adjusting for bogus or duplicate accounts it all adds up to about 1/7th of humanity from which some kind of data has been recorded.

According to Facebook’s programmers like Voskuhl and Vyaghrapuri, of all the clever uses they have already applied this pile of data toward, Facebook has so far “lacked tools to synthesize this information about users for targeting advertisements based on their perceived income.” Now they have such a tool thanks to the retention and analysis of variables the company’s positivist specialists believe are correlated with income levels.

They’ll have many more tools within the next year to run similar predictions. Indeed, Facebook, Google, Yahoo, Twitter, and the hundreds of smaller tech lesser-known tech firms that now control the main portals of social, economic, and political life on the web (which is now to say everywhere as all economic and much social activity is made cyber) are only getting started. The Big Data analytics revolutions have barely begun, and these firms are just beginning to tinker with rational-instrumental methods of predicting and manipulating human behavior.

There are few, if any, government regulations restricting their imaginations at this point. Indeed, the U.S. President himself is a true believer in Big Data; the brain of Obama’s election team was a now-famous “cave” filled with young Ivy League men (and a few women) sucking up electioneering information and crunching demographic and consumer data to target individual voters with appeals timed to maximize the probability of a vote for the new Big Blue, not IBM, but the Democratic Party’s candidate of “Hope” and “Change.” The halls of power are enraptured by the potential of rational-instrumental methods paired with unprecedented access to data that describes the social lives of hundreds of millions.

Facebook’s intellectual property portfolio reads like cliff notes summarizing the aspirations of all corporations in capitalist modernity; to optimize efficiency in order to maximize profits and reduce or externalize risk. Unlike most other corporations, and unlike previous phases in the development of rational bureaucracies, Facebook and its tech peers have accumulated never before seen quantities of information about individuals and groups. Recent breakthroughs in networked computing make an analysis of these gigantic data sets fast and cheap. Facebook’s patent holdings are just a taste of what’s arriving here and now.

The way you type, the rate, common mistakes, intervals between certain characters, are all unique, like your fingerprint, and there are already cyber robots that can identify you as you peck away at keys. Facebook has even patented methods of individual identification with obviously cybernetic overtones, where the machine becomes an appendage of the person. U.S. Patents 8,306,256, 8,472,662, and 8,503,718, all filed within the last year, allow Facebook’s web robots to identify a user based on the unique pixelation and other characteristics of their smartphone’s camera. Identification of the subject is the first step toward building a useful data set to file among the billion or so other user logs. Then comes analysis, then prediction, then efforts to influence a parting of money.

Many Facebook patents pertain to advertising techniques that are designed and targeted, and continuously redesigned with ever-finer calibrations by robot programs, to be absorbed by the gazes of individuals as they scroll and swipe across their Facebook feeds, or on third-party web sites.

Speaking of feeds, U.S. Patent 8,352,859, Facebook’s system for “Dynamically providing a feed of stories about a user of a social networking system” is used by the company to organize the constantly updated posts and activities inputted by a user’s “friends.” Of course, embedded in this system are means of inserting advertisements. According to Facebook’s programmers, a user’s feeds are frequently injected with “a depiction of a product, a depiction of a logo, a display of a trademark, an inducement to buy a product, an inducement to buy a service, an inducement to invest, an offer for sale, a product description, trade promotion, a survey, a political message, an opinion, a public service announcement, news, a religious message, educational information, a coupon, entertainment, a file of data, an article, a book, a picture, travel information, and the like.” That’s a long list for sure, but what gets injected is more often than not whatever will boost revenues for Facebook.

The advantage here, according to Facebook, is that “rather than having to initiate calls or emails to learn news of another user, a user of a social networking website may passively receive alerts to new postings by other users.” The web robot knows best. Sit back and relax and let sociality wash over you, passively. This is merely one of Facebook’s many “systems for tailoring connections between various users” so that these connections ripple with ads uncannily resonant with desires and needs revealed in the quietly observed flow of e-mails, texts, images, and clicks captured forever in dark inaccessible servers of Facebook, Google and the like. These communications services are free in order to control the freedom of data that might otherwise crash about randomly, generating few opportunities for sales.

Where this fails Facebook ratchets up the probability of influencing the user to behave as a predictable consumer. “Targeted advertisements often fail to earn a user’s trust in the advertised product,” explain Facebook’s programmers in U.S. Patent 8,527,344, filed in September of this year. “For example, the user may be skeptical of the claims made by the advertisement. Thus, targeted advertisements may not be very effective in selling an advertised product.” Facebook’s computer programmers who now profess mastery over sociological forces add that even celebrity endorsements are viewed with skepticism by the savvy citizen of the modulated Internet. They’re probably right.

Facebook’s solution is to mobilize its users as trusted advertisers in their own right. “Unlike advertisements, most users seek and read content generated by their friends within the social networking system; thus,” concludes Facebook’s mathematicians of human inducement, “advertisements generated by a friend of the user are more likely to catch the attention of the user, increasing the effectiveness of the advertisement.” That Facebook’s current So-And-So-likes-BrandX ads are often so clumsy and ineffective does not negate the qualitative shift in this model of advertising and the possibilities of un-freedom it evokes.

Forget iPhones and applications, the tech industry’s core consumer product is now advertising. Their essential practice is mass surveillance conducted in real-time through continuous and multiple sensors that pass, for most people, entirely unnoticed. The autonomy and unpredictability of the individual —in Facebook’s language the individual is the “user”— is their fundamental business problem. Reducing autonomy via surveillance and predictive algorithms that can placate existing desires, and even stimulate and mold new desires is the tech industry’s reason for being. Selling their capacious surveillance and consumer stimulus capabilities to the highest bidder is the ultimate end.

Sounds too dystopian? Perhaps, and this is by no means the world we live in, not yet. It is, however, a tendency rooted in the tech economy. The advent of mobile, hand-held, wirelessly networked computers, called “smartphones,” is still so new that the technology and its services feel like a parallel universe, a new layer of existence added upon our existing social relationships, business activities, and political affiliations. In many ways it feels liberating and often playful. Our devices can map geographic routes, identify places and things, provide information about almost anything in real-time, respond to our voices, and replace our wallets. Who hasn’t consulted “Dr. Google” to answer a pressing question? Everyone and everything is seemingly within reach and there is a kind of freedom to this utility.

Most of Facebook’s “users” have only been registered on the web site since 2010, and so the quintessential social network feels new and fun, and although perhaps fraught with some privacy concerns, it does not altogether fell like a threat to the autonomy of the individual. To say it is a cliche sci-fi nightmare narrative of tech-bureaucracy, and we all tell one another that the reality is more complex.

Privacy continues, however, to be too narrowly conceptualized as a liberal right against incursions of government, and while the tech companies have certainly been involved in a good deal of old-fashioned mass surveillance for the sake of our federal Big Brother, there’s another means of dissolving privacy that is more fundamental to the goals of the tech companies and more threatening to social creativity and political freedom.

Georgetown University law professor Julie Cohen notes that pervasive surveillance is inimical to the spaces of privacy that are required for liberal democracy, but she adds importantly, that the surveillance and advertising strategies of the tech industry go further.

“A society that permits the unchecked ascendancy of surveillance infrastructures, which dampen and modulate behavioral variability, cannot hope to maintain a vibrant tradition of cultural and technical innovation,” writes Cohen in a forthcoming Harvard Law Review article.

“Modulation” is Cohen’s term for the tech industry’s practice of using algorithms and other logical machine operations to mine an individual’s data so as to continuously personalize information streams. Facebook’s patents are largely techniques of modulation, as are Google’s and the rest of the industry leaders. Facebook conducts meticulous surveillance on users, collects their data, tracks their movements on the web, and feeds the individual specific content that is determined to best resonate with their desires, behaviors, and predicted future movements. The point is to perfect the form and function of the rational-instrumental bureaucracy as defined by Max Weber: to constantly ratchet up efficiency, calculability, predictability, and control. If they succeed in their own terms, the tech companies stand to create a feedback loop made perfectly to fit each and every one of us, an increasingly closed system of personal development in which the great algorithms in the cloud endlessly tailor the psychological and social inputs of humans who lose the gift of randomness and irrationality.

“It is modulation, not privacy, that poses the greater threat to innovative practice,” explains Cohen. “Regimes of pervasively distributed surveillance and modulation seek to mold individual preferences and behavior in ways that reduce the serendipity and the freedom to tinker on which innovation thrives.” Cohen has pointed out the obvious irony here, not that it’s easy to miss; the tech industry is uncritically labeled America’s hothouse of innovation, but it may in fact be killing innovation by disenchanting the world and locking inspiration in a cage.

If there were limits to the reach of the tech industry’s surveillance and stimuli strategies it would indeed be less worrisome. Only parts of our lives would be subject to this modulation, and it could therefore benefit us. But the industry aspires to totalitarian visions in which universal data sets are constantly mobilized to transform an individual’s interface with society, family, the economy, and other institutions. The tech industry’s luminaries are clear in their desire to observe and log everything and use every “data point” to establish optimum efficiency in life as the pursuit of consumer happiness. Consumer happiness is, in turn, a step toward the rational pursuit of maximum corporate profit. We are told that the “Internet of things” is arriving, that soon every object will have embedded within it a computer that is networked to the sublime cloud, and that the physical environment will be made “smart” through the same strategy of modulation so that we might be made free not just in cyberspace, but also in the meatspace.

Whereas the Internet of the late 1990s matured as an archipelago of innumerable disjointed and disconnected web sites and databases, today’s Internet is gripped by a handful of giant companies that observe much of the traffic and communications, and which deliver much of the information from an Android phone or laptop computer, to distant servers, and back. The future Internet being built by the tech giants —putting aside the Internet of things for the moment— is already well into its beta testing phase. It’s a seamlessly integrated quilt of web sites and apps that all absorb “user” data, everything from clicks and keywords to biometric voice identification and geolocation.

United States Patent 8,572,174, another of Facebook’s recent inventions, allows the company to personalize a web page outside of Facebook’s own system with content from Facebook’s databases. Facebook is selling what the company calls its “rich set of social information” to third party web sites in order to “provide personalized content for their users based on social information about those users that is maintained by, or otherwise accessible to, the social networking system.” Facebook’s users generated this rich social information, worth many billions of dollars as recent quarterly earnings of the company attest.

In this way the entire Internet becomes Facebook. The totalitarian ambition here is obvious, and it can be read in the securities filings, patent applications, and other non-sanitized business documents crafted by the tech industry for the financial analysts who supply the capital for further so-called innovation. Everywhere you go on the web, with your phone or tablet, you’re a “user,” and your social network data will be mined every second by every application, site, and service to “enhance your experience,” as Facebook and others say. The tech industry’s leaders aim to expand this into the physical world, creating modulated advertising and environmental experiences as cameras and sensors track our movements.

Facebook and the rest of the tech industry fear autonomy and unpredictability. The ultimate expression of these irrational variables that cannot be mined with algorithmic methods is the absence from the networks of surveillance in which data is collected.

One of Facebook’s preventative measures is United States Patent 8,560,962, “promoting the participation of low-activity users in the social networking system.” This novel invention devised by programmers in Facebook’s Palo Alto and San Francisco offices involves a “process of inducing interactions,” that are meant to maximize the amount of “user-generated content” on Facebook by getting lapsed users to return, and stimulating all users to produce more and more data. User-generated content is, after all, worth billions. Think twice before you hit “like” next time, or tap that conspicuously placed “share” button; a machine likely put that content and interaction before your eyes after a logical operation determined it to have the highest probability of tempting you to add to the data stream, thereby increasing corporate revenues.

Facebook’s patents on techniques of modulating “user” behavior are few compared to the real giants of the tech industry’s surveillance and influence agenda. Amazon, Microsoft, and of course Google hold some of the most fundamental patents using personal data to attempt to shape an individual’s behavior into predictable consumptive patterns. Smaller specialized firms like Choicestream and Gist Communications have filed dozens more applications for modulation techniques. The rate of this so-called innovation is rapidly telescoping.

Perhaps we do know who will live in the iron cage. It might very well be a cage made of our own user-generated content, paradoxically ushering in a new era of possibilities in shopping convenience and the delivery of satisfactory experiences even while it eradicates many degrees of chance, and pain, and struggle (the motive forces of human progress) in a robot-powered quest to have us construct identities and relationships that yield to prediction and computer-generated suggestion. Defense of individual privacy and autonomy today is rightly motivated by the reach of an Orwellian security state (the NSA, FBI, CIA). This surveillance changes our behavior by chilling us, by telling us we are always being watched by an authority. Authority thereby represses in us whatever might happen to be defined as “crime,” or any anti-social behavior at the moment. But what about the surveillance that does not seek to repress us, the watching computer eyes and ears that instead hope to stimulate a particular set of monetized behaviors in us with the intimate knowledge gained from our every online utterance, even our facial expressions and finger movements?

Darwin Bond-Graham, a contributing editor to CounterPunch, is a sociologist and author who lives and works in Oakland, CA. His essay on economic inequality in the “new” California economy appears in the July issue of CounterPunch magazine. He is a contributor to Hopeless: Barack Obama and the Politics of Illusion

Source: Alternet.org


Additional comments from me:

The corporations that own the social media outlets capitalize (yes, CAPITALIZE) on peoples, oftentimes insatiable, need be seen, to be heard, and to be popular (..Facebook, iPhone or better yet iEverything, YouTube). From early childhood on most human beings have been programmed to believe that you only matter if you win, if you’re "out there" and if lots of other people pay attention to you.

While I acknowledge that social media can serve a useful purpose as a means of staying in touch or connecting with others of like mind, friends, and loved ones, I strongly advise that you do not live your life on those outlets.

Cyber connections cannot replace actual human interaction and your 1700 Facebook ‘friends’, 5000 Twitter followers and 500+ LinkedIn connections, will not all be there with a bowl of soup when you’ve caught a cold. 

And I know, it begs the question would you even want them to since you couldn’t possibly consume such quantities, but I think I've made my point.

 

Mind Control Theories And Techniques Used By The Mass Media

Created on Saturday, 05 June 2010 16:18

Programming Through Mass Media

Mass media are media forms designed to reach the largest audience possible. They include television, movies, radio, newspapers, magazines, books, records, video games and the internet.

Many studies have been conducted in the past century to measure the effects of mass media on the population, in order to discover the best techniques to influence it. From those studies emerged the science of Communications, which is used in marketing, public relations and politics. Mass communication is a necessary tool to ensure the functionality of a large democracy; it is also a necessary tool for a dictatorship.

There used to be a variety of viewpoints, ideas and opinions in popular culture. The consolidation of media corporations has, however, produced a standardization of the cultural industry. Ever wondered why all recent music sounds the same and all recent movies look the same? The following is part of the answer:

Media Ownership

The number of corporations owning the majority of U.S. media outlets went from 50 to 6 in less than 20 years. Here are the top corporations evolving around the world and the assets they own:

 

The Standardization of Human Thought

The merger of media companies in the last decades generated a small oligarchy of media conglomerates. The TV shows you follow, the music you listen to, the movies you watch and the newspapers you read are basically all produced by SIX corporations. 

The owners of those conglomerates have close ties with the so-called "world’s elite" and, in many ways, they are "the elite". By owning all of the possible outlets having the potential to reach the masses, these conglomerates have the power to create in the minds of the people a single and cohesive world view, engendering a “standardization of human thought”.

Even movements or styles that are considered marginal are, in fact, extensions of mainstream thinking. Mass media produce their own rebels who definitely look the part but are still part of the establishment and do not question any of it. Artists, creations and ideas that do not fit the mainstream way of thinking are mercilessly rejected and forgotten by the conglomerates, which in turn makes them virtually disappear from society itself. However, ideas that are deemed to be valid and desirable to be accepted by society are skillfully marketed to the masses, in order to make them become the self-evident norm.

In 1928, Edward Bernays already saw the immense potential of motion pictures to standardize thought:

“The American motion picture is the greatest unconscious carrier of propaganda in the world today. It is a great distributor for ideas and opinions. The motion picture can standardize the ideas and habits of a nation. Because pictures are made to meet market demands, they reflect, emphasize, and even exaggerate broad popular tendencies, rather than stimulate new ideas and opinions. The motion picture avails itself only of ideas and facts which are in vogue. As the newspaper seeks to purvey news, it seeks to purvey entertainment.”
– Edward Bernays, Propaganda

These facts were flagged as dangers to human freedom in the 1930s by thinkers of the school of Frankfurt such as Theodor Adorno and Herbert Marcuse. They identified three main problems with the cultural industry. The industry can:

  1. reduce human beings to the state of mass by hindering the development of emancipated individuals, who are capable of making rational decisions;
  2. replace the legitimate drive for autonomy and self-awareness by the safe laziness of conformism and passivity; and
  3. validate the idea that men actually seek to escape the absurd and cruel world in which they live, by losing themselves in a hypnotic state of self-satisfaction.

The notion of escapism is even more relevant today with the advent of online video games, 3D movies and home theaters. The masses, constantly seeking state-of-the-art entertainment, will resort to high-budget products that can only be produced by the biggest media corporations of the world. These products contain carefully calculated messages and symbols which are nothing more and nothing less than entertaining propaganda. The public has been trained to LOVE its propaganda to the extent that it spends its hard-earned money to be exposed to it. Propaganda (used in a political, cultural, and commercial sense) is no longer the coercive or authoritative communication form found in dictatorships: it has become the synonym of entertainment and pleasure.

A single piece of media often does not have a lasting effect on the human psyche. Mass media, however, by its omnipresent nature, creates a living environment we evolve in on a daily basis. It defines the norm and excludes the undesirable. The same way carriage horses wear blinders so they can only see what is right in front of them, the masses can only see where they are supposed to go.

Manipulation Techniques

The drive to sell products and ideas to the masses has lead to an unprecedented amount of research on human behavior and on the human psyche. Cognitive sciences, psychology, sociology, semiotics, linguistics, and other related fields were and still are extensively researched through well-funded studies.

The results of those studies are applied to advertisements, movies, music videos and other media in order to make them as influential as possible. The art of marketing is highly calculated and scientific because it must reach both the individual and the collective consciousness. In high-budget cultural products, a video is never "just a video".

Images, symbols and meanings are strategically placed in order to generate the desired effect.

Desensitization

In the past, when changes were imposed on populations, they would take to the streets, protest and even riot. The main reason for this clash was due to the fact that the change was clearly announced by the rulers and understood by the population. It was sudden and its effects could clearly be analyzed and evaluated. Today, when the elite needs a part of the agenda to be accepted by the public, it is done through desensitization. The agenda, which might go against the public best interests, is slowly, gradually, and repetitively introduced to the world through movies (by involving it within the plot), music videos (which make it cool and sexy) and the news (which presents it as "a solution" to today’s problems). After several years of exposing the masses to a particular agenda, the elite openly presents the concept to the world and, due to mental programming, it is greeted with general indifference and is passively accepted.

Predictive programming is often found in the science fiction genre. It presents a specific image of the future – the one that is desired by the elite – and ultimately becomes in the minds of men an inevitability. A decade ago, the public was being desensitized to war against the Arab world.

Today, the population is gradually being exposed to the existence of mind control, transhumanism, and an Illuminati elite. Emerging from the shadows, those concepts are now everywhere in popular culture.

Edited from: The Vigilant Citizen

Related

Starsuckers

Psywar

Weapons Of Mass Deception