How your dad’s music influences your taste

[div class=attrib]From Sonos:[end-div]

There’s no end to the reasons why you listen to the music you do today, but we’re willing to bet that more than a few of you were subjected to your father’s music at some point in the past (or present). So that leads to the question: what do dear old dad’s listening habits say about the artists in your repertoire? In honor of Father’s Day, we tried our hand at finding out.

[div class=attrib]More from the Source here.[end-div]

The Technology of Personalization and the Bubble Syndrome

A decade ago in another place and era during my days as director of technology research for a Fortune X company I tinkered with a cool array of then new personalization tools. The aim was simple, use some of these emerging technologies to deliver a more customized and personalized user experience for our customers and suppliers. What could be wrong with that? Surely, custom tools and more personalized data could do nothing but improve knowledge and enhance business relationships for all concerned. Our customers would benefit from seeing only the information they asked for, our suppliers would benefit from better analysis and filtered feedback, and we, the corporation in the middle, would benefit from making everyone in our supply chain more efficient and happy. Advertisers would be even happier since with more focused data they would be able to deliver messages that were increasingly more precise and relevant based on personal context.

Fast forward to the present. Customization, or filtering, technologies have indeed helped optimize the supply chain; personalization tools and services have made customer experiences more focused and efficient. In today’s online world it’s so much easier to find, navigate and transact when the supplier at the other end of our browser knows who we are, where we live, what we earn, what we like and dislike, and so on. After all, if a supplier knows my needs, requirements, options, status and even personality, I’m much more likely to only receive information, services or products that fall within the bounds that define “me” in the supplier’s database.

And, therein lies the crux of the issue that has helped me to realize that personalization offers a false promise despite the seemingly obvious benefits to all concerned. The benefits are outweighed by two key issues: erosion of privacy and the bubble syndrome.

Privacy as Commodity

I’ll not dwell too long on the issue of privacy since in this article I’m much more concerned with the personalization bubble. However, as we have increasingly seen in recent times privacy in all its forms is becoming a scarce, and tradable commodity. Much of our data is now in the hands of a plethora of suppliers, intermediaries and their partners, ready for continued monetization. Our locations are constantly pinged and polled; our internet browsers note our web surfing habits and preferences; our purchases generate genius suggestions and recommendations to further whet our consumerist desires. Now in digital form this data is open to legitimate sharing and highly vulnerable to discovery by hackers, phishers and spammers and any with technical or financial resources.

Bubble Syndrome

Personalization technologies filter content at various levels, minutely and broadly, both overtly and covertly. For instance, I may explicitly signal my preferences for certain types of clothing deals at my favorite online retailer by answering a quick retail survey or checking a handful of specific preference buttons on a website.

However, my previous online purchases, browsing behaviors, time spent of various online pages, visits to other online retailers and a range of other flags deliver a range of implicit or “covert” information to the same retailer (and others). This helps the retailer filter, customize and personalize what I get to see even before I have made a conscious decision to limit my searches and exposure to information. Clearly, this is not too concerning when my retailer knows I’m male and usually purchase size 32 inch jeans; after all why would I need to see deals or product information for women’s shoes.

But, this type of covert filtering becomes more worrisome when the data being filtered and personalized is information, news, opinion and comment in all its glorious diversity. Sophisticated media organizations, information portals, aggregators and news services can deliver personalized and filtered information based on your overt and covert personal preferences as well. So, if you subscribe only to a certain type of information based on topic, interest, political persuasion or other dimension your personalized news services will continue to deliver mostly or only this type of information. And, as I have already described, your online behaviors will deliver additional filtering parameters to these news and information providers so that they may further personalize and narrow your consumption of information.

Increasingly, we will not be aware of what we don’t know. Whether explicitly or not, our use of personalization technologies will have the ability to build a filter, a bubble, around us, which will permit only information that we wish to see or that which our online suppliers wish us to see. We’ll not even get exposed to peripheral and tangential information — that information which lies outside the bubble. This filtering of the rich oceans of diverse information to a mono-dimensional stream will have profound implications for our social and cultural fabric.

I assume that our increasingly crowded planet will require ever more creativity, insight, tolerance and empathy as we tackle humanity’s many social and political challenges in the future. And, these very seeds of creativity, insight, tolerance and empathy are those that are most at risk from the personalization filter. How are we to be more tolerant of others’ opinions if we are never exposed to them in the first place? How are we to gain insight when disparate knowledge is no longer available for serendipitous discovery? How are we to become more creative if we are less exposed to ideas outside of our normal sphere, our bubble?

For some ideas on how to punch a few holes in your online filter bubble read Eli Pariser’s practical guide, here.

Filter Bubble image courtesy of TechCrunch.

Cosmic Smoothness

Simulations based on the standard cosmological model, as shown here, indicate that on very large distance scales, galaxies should be uniformly distributed. But observations show a clumpier distribution than expected. (The length bar represents about $2.3$ billion light years.)[div class=attrib]From American Physical Society, Michael J. Hudson:[end-div]

The universe is expected to be very nearly homogeneous in density on large scales. In Physical Review Letters, Shaun Thomas and colleagues from University College London analyze measurements of the density of galaxies on the largest spatial scales so far—billions of light years—and find that the universe is less smooth than expected. If it holds up, this result will have important implications for our understanding of dark matter, dark energy, and perhaps gravity itself.

In the current standard cosmological model, the average mass-energy density of the observable universe consists of 5% normal matter (most of which is hydrogen and helium), 23% dark matter, and 72% dark energy. The dark energy is assumed to be uniform, but the normal and dark matter are not. The balance between matter and dark energy determines both how the universe expands and how regions of unusually high or low matter density evolve with time.

The same cosmological model predicts the statistics of the nonuniform structure and their dependence on spatial scale. On scales that are small by cosmological standards, fluctuations in the matter density are comparable to its mean, in agreement with what is seen: matter is clumped into galaxies, clusters of galaxies, and filaments of the “cosmic web.” On larger scales, however, the contrast of the structures compared to the mean density decreases. On the largest cosmological scales, these density fluctuations are small in amplitude compared to the average density of the universe and so are well described by linear perturbation theory (see simulation results in Fig. 1). Moreover, these perturbations can be calibrated at early times directly from the cosmic microwave background (CMB), a snapshot of the universe from when it was only 380,000 years old. Despite the fact that only 5% of the Universe is well understood, this model is an excellent fit to data spanning a wide range of spatial scales as the fluctuations evolved from the time of the CMB to the present age of the universe, some 13.8 billion years. On the largest scales, dark energy drives accelerated expansion of the universe. Because this aspect of the standard model is least understood, it is important to test it on these scales.

Thomas et al. use publicly-released catalogs from the Sloan Digital Sky Survey to select more than 700,000 galaxies whose observed colors indicate a significant redshift and are therefore presumed to be at large cosmological distances. They use the redshift of the galaxies, combined with their observed positions on the sky, to create a rough three-dimensional map of the galaxies in space and to assess the homogeneity on scales of a couple of billion light years. One complication is that Thomas et al. measure the density of galaxies, not the density of all matter, but we expect that fluctuations of these two densities about their means to be proportional; the constant of proportionality can be calibrated by observations on smaller scales. Indeed, on small scales the galaxy data are in good agreement with the standard model. On the largest scales, the fluctuations in galaxy density are expected to be of order a percent of the mean density, but Thomas et al. find fluctuations double this prediction. This result then suggests that the universe is less homogeneous than expected.

This result is not entirely new: previous studies based on subsets of the data studied by Thomas et al. showed the same effect, albeit with a lower statistical significance. In addition, there are other ways of probing the large-scale mass distribution. For example, inhomogeneities in the mass distribution lead to inhomogeneities in the local rate of expansion. Some studies have suggested that, on very large scales, this expansion too is less homogeneous than the model predictions.

Future large-scale surveys will produce an avalanche of data. These surveys will allow the methods employed by Thomas et al. and others to be extended to still larger scales. Of course, the challenge for these future surveys will be to correct for the systematic effects to even greater accuracy.

[div class=attrib]More from theSource here.[end-div]

Lemonade without the Lemons: New Search Engine Looks for Uplifting News

[div class=attrib]From Scientific American:[end-div]

Good news, if you haven’t noticed, has always been a rare commodity. We all have our ways of coping, but the media’s pessimistic proclivity presented a serious problem for Jurriaan Kamp, editor of the San Francisco-based Ode magazine—a must-read for “intelligent optimists”—who was in dire need of an editorial pick-me-up, last year in particular. His bright idea: an algorithm that can sense the tone of daily news and separate the uplifting stories from the Debbie Downers.

Talk about a ripe moment: A Pew survey last month found the number of Americans hearing “mostly bad” news about the economy and other issues is at its highest since the downturn in 2008. That is unlikely to change anytime soon: global obesity rates are climbing, the Middle East is unstable, and campaign 2012 vitriol is only just beginning to spew in the U.S. The problem is not trivial. A handful of studies, including one published in the Clinical Psychology Review in 2010, have linked positive thinking to better health. Another from the Journal of Economic Psychology the year prior found upbeat people can even make more money.

Kamp, realizing he could be a purveyor of optimism in an untapped market, partnered with Federated Media Publishing, a San Francisco–based company that leads the field in search semantics. The aim was to create an automated system for Ode to sort and aggregate news from the world’s 60 largest news sources based on solutions, not problems. The system, released last week in public beta testing online and to be formally introduced in the next few months, runs thousands of directives to find a story’s context. “It’s kind of like playing 20 questions, building an ontology to find either optimism or pessimism,” says Tim Musgrove, the chief scientist who designed the broader system, which has been dubbed a “slant engine”. Think of the word “hydrogen” paired with “energy” rather than “bomb.”

Web semantics developers in recent years have trained computers to classify news topics based on intuitive keywords and recognizable names. But the slant engine dives deeper into algorithmic programming. It starts by classifying a story’s topic as either a world problem (disease and poverty, for example) or a social good (health care and education). Then it looks for revealing phrases. “Efforts against” in a story, referring to a world problem, would signal something good. “Setbacks to” a social good, likely bad. Thousands of questions later every story is eventually assigned a score between 0 and 1—above 0.95 fast-tracks the story to Ode’s Web interface, called OdeWire. Below that, a score higher than 0.6 is reviewed by a human. The system is trained to only collect themes that are “meaningfully optimistic,” meaning it throws away flash-in-the-pan stories about things like sports or celebrities.

[div class=attrib]More from theSource here.[end-div]

Self-Published Author Sells a Million E-Books on Amazon

[div class=attrib]From ReadWriteWeb:[end-div]

Since the Kindle’s launch, Amazon has heralded each new arrival into what it calls the “Kindle Million Club,” the group of authors who have sold over 1 million Kindle e-books. There have been seven authors in this club up ’til now – some of the big names in publishing: Stieg Larsson, James Patterson, and Nora Roberts for example.

But the admission today of the eighth member of this club is really quite extraordinary. Not because John Locke is a 60 year old former insurance salesman from Kentucky with no writing or publishing background. But because John Locke has accomplished the feat of selling one million e-books as a completely self-published author.

Rather than being published by major publishing house – and all the perks that have long been associated with that (marketing, book tours, prime shelf space in retail stores) – Locke has sold 1,010,370 Kindle books (as of yesterday) having used Kindle Direct Publishing to get his e-books into the Amazon store. No major publisher. No major marketing.

Locke writes primarily crime and adventure stories, including Vegas Moon, Wish List, and the New York Times E-Book Bestseller, Saving Rachel. Most of the e-books sell for $.99, and he says he makes 35 cents on every sale. That sort of per book profit is something that authors would never get from a traditional book deal.

[div class=attrib]More from theSource here.[end-div]

Book Review: Solar. Ian McEwan

Solar is a timely, hilarious novel from the author of Atonement that examines the self-absorption and (self-)deceptions of Nobel Prize-winning physicist Michael Beard. With his best work many decades behind him Beard trades on his professional reputation to earn continuing financial favor, and maintain influence and respect amongst his peers. And, with his personal life in an ever-decreasing spiral, with his fifth marriage coming to an end, Beard manages to entangle himself in an impossible accident which has the power to re-shape his own world, and the planet in the process.

Ian McEwan’s depiction of Michael Beard is engaging and thoroughly entertaining. Beard hops from relationship to relationship in his singular quest for “love”, but very much on his own terms. And, this very centric view of himself extends to his own science, where his personal contributions don’t seem to be all that they appear. Satire and climate science makes a stylish and witty combination in the hands of McEwan.

Book Review: The Social Animal. David Brooks

David Brooks brings us a detailed journey through the building blocks of the self in his new book, The Social Animal: A Story of Love, Character and Achievement. With his insight and gift for narrative Brooks weaves an engaging and compelling story of Erica and Harold. Brooks uses the characters of Erica and Harold as platforms on which he visualizes the results of numerous psychological, social and cultural studies. Placed in contemporary time the two characters show us a holistic picture in practical terms of the unconscious effects of physical and social context on behavioral and character traits. The narrative takes us through typical life events and stages: infancy, childhood, school, parenting, work-life, attachment, aging. At each stage, Brooks illustrates his views of the human condition by selecting a flurry of facts and anecdotal studies.

The psychologist in me would say that this is a rather shallow attempt at synthesizing profoundly complex issues. Brooks certainly makes use of many studies from the brain and social sciences, but never dwells long enough to give us a detailed sense of major underlying implications or competing scientific positions. So too, the character development of Erica and Harold lacks the depth and breadth one would expect — Brooks fails to explore much of what typically seems to motivate human behavior: greed, ambition, lust, violence, empathy.  Despite these flaws in the execution of the idea, Brooks’ attempt is praiseworthy; perhaps in the hands of a more skilled social scientist, or Rousseau who used this technique much more effectively, this type of approach would gain a better grade.

Book Review: The Drunkard’s Walk: How Randomness Rules Our Lives. Leonard Mlodinow

Leonard Mlodinow weaves a compelling path through the world of statistical probability showing us how the laws of chance affect our lives on personal and grande scales. Mlodinow skillfully illustrates randomness and its profound implications by presenting complex mathematical constructs in language for the rest of us (non-mathematicians), without dumbing-down this important subject.

The book defines many of the important mathematical concepts behind randomness and exposes the key fallacies that often blind us as we wander through life on our “drunkard’s walk”. The law of large numbers, the prosecutor’s fallacy, conditional probability, the availability bias and bell curves were never so approachable.

Whether it’s a deluded gambler, baseball star on a “winning streak” or a fortunate CEO wallowing in the good times, Mlodinow debunks the common conceptions that skill, planning and foresight result in any significant results beyond pure chance. With the skill of a storyteller Mlodinow shows us how polls, grades, ratings and even measures of corporate success are far less objective and reliable than we ought to believe. Lords of Wall Street take notice, the secrets of your successes are not all that they seem.

Prophecy Fail

[div class=attrib]From Slate:[end-div]

What happens to a doomsday cult when the world doesn’t end?

Preacher and evangelical broadcaster Harold Camping has announced that Jesus Christ will return to Earth this Saturday, May 21, and many of his followers are traveling the country in preparation for the weekend Rapture. They’re undeterred, it seems, by Mr. Camping’s dodgy track record with end-of-the-world predictions. (Years ago, he argued at length that the reckoning would come in 1994.) We’ve yet to learn what motivates people like him to predict (and predict again) the end of the world, but there’s a long and unexpected psychological literature on how the faithful make sense of missed appointments with the apocalypse.

The most famous study into doomsday mix-ups was published in a 1956 book by renowned psychologist Leon Festinger and his colleagues called When Prophecy Fails. A fringe religious group called the Seekers had made the papers by predicting that a flood was coming to destroy the West Coast. The group was led by an eccentric but earnest lady called Dorothy Martin, given the pseudonym Marian Keech in the book, who believed that superior beings from the planet Clarion were communicating to her through automatic writing. They told her they had been monitoring Earth and would arrive to rescue the Seekers in a flying saucer before the cataclysm struck.
Festinger was fascinated by how we deal with information that fails to match up to our beliefs, and suspected that we are strongly motivated to resolve the conflict—a state of mind he called “cognitive dissonance.” He wanted a clear-cut case with which to test his fledgling ideas, so decided to follow Martin’s group as the much vaunted date came and went. Would they give up their closely held beliefs, or would they work to justify them even in the face of the most brutal contradiction?

The Seekers abandoned their jobs, possessions, and spouses to wait for the flying saucer, but neither the aliens nor the apocalypse arrived. After several uncomfortable hours on the appointed day, Martin received a “message” saying that the group “had spread so much light that God had saved the world from destruction.” The group responded by proselytizing with a renewed vigour. According to Festinger, they resolved the intense conflict between reality and prophecy by seeking safety in numbers. “If more people can be persuaded that the system of belief is correct, then clearly, it must, after all, be correct.”

[div class=attrib]More from theSource here:[end-div]

Are We Intelligent Matter or Incarnate Spirit?

[div class=attrib]From Evolutionary Philosophy:[end-div]

One of the most confounding philosophical questions involves our understanding of who we really are. Are we intelligent matter – stuff that got smart – or are we incarnate spirit – smarts that grew stuff around it? This question is inherent in the very nature of our experience of being human. We have bodies and we have the experience of consciousness – mind and matter, body and soul. Which one is more us, which came first, and which is really running the show?

The great religious traditions of the west have tended towards the outlook that we are spiritual beings who became flesh. First there was God, pure spirit and from God came us. Our more recent scientific understanding of reality has lead many to believe that we are matter that evolved into life and intelligence. Now, of course, there are always those who land somewhere in between these extremes – probably most people reading this blog for instance – still this is the divide that has generally separated science from religion and idealists from materialists.

If, in fact, we are essentially spirit that has taken form it would mean that in some significant way human beings are separate from the universe. We have some source of intelligence and will that is free from the rest of nature, that acts in nature while maintaining a foothold in some transcendent outside reference point. In this view, the core of our being stands apart from and above the laws of nature and we are therefore uniquely autonomous and responsible as the source of our own action in the universe.

[div class=attrib]More from theSource here.[end-div]

Test-tube truths

[div class=attrib]From Eurozine:[end-div]

In his new book, American atheist Sam Harris argues that science can replace theology as the ultimate moral authority. Kenan Malik is sceptical of any such yearning for moral certainty, be it scientific or divine.

“If God does not exist, everything is permitted.” Dostoevsky never actually wrote that line, though so often is it attributed to him that he may as well have. It has become the almost reflexive response of believers when faced with an argument for a godless world. Without religious faith, runs the argument, we cannot anchor our moral truths or truly know right from wrong. Without belief in God we will be lost in a miasma of moral nihilism. In recent years, the riposte of many to this challenge has been to argue that moral codes are not revealed by God but instantiated in nature, and in particular in the brain. Ethics is not a theological matter but a scientific one. Science is not simply a means of making sense of facts about the world, but also about values, because values are in essence facts in another form.

Few people have expressed this argument more forcefully than the neuroscientist Sam Harris. Over the past few years, through books such as The End of Faith and Letter to a Christian Nation, Harris has gained a considerable reputation as a no-holds-barred critic of religion, in particular of Islam, and as an acerbic champion of science. In his new book, The Moral Landscape: How Science Can Determine Human Values, he sets out to demolish the traditional philosophical distinction between is and ought, between the way the world is and the way that it should be, a distinction we most associate with David Hume.

What Hume failed to understand, Harris argues, is that science can bridge the gap between ought and is, by turning moral claims into empirical facts. Values, he argues, are facts about the “states of the world” and “states of the human brain”. We need to think of morality, therefore, as “an undeveloped branch of science”: “Questions about values are really questions about the wellbeing of conscious creatures. Values, therefore, translate into facts that can be scientifically understood: regarding positive and negative social emotions, the effects of specific laws on human relationships, the neurophysiology of happiness and suffering, etc.” Science, and neuroscience in particular, does not simply explain why we might respond in particular ways to equality or to torture but also whether equality is a good, and torture morally acceptable. Where there are disagreements over moral questions, Harris believes, science will decide which view is right “because the discrepant answers people give to them translate into differences in our brains, in the brains of others and in the world at large.”

Harris is nothing if not self-confident. There is a voluminous philosophical literature that stretches back almost to the origins of the discipline on the relationship between facts and values. Harris chooses to ignore most of it. He does not wish to engage “more directly with the academic literature on moral philosophy”, he explains in a footnote, because he did not develop his arguments “by reading the work of moral philosophers” and because he is “convinced that every appearance of terms like ‘metaethics’, ‘deontology’, ‘noncognitivism’, ‘antirealism’, ’emotivism’, etc directly increases the amount of boredom in the universe.”

[div class=attrib]More from theSource here.[end-div]

How Self-Control Works

[div class=attrib]From Scientific American:[end-div]

The scientific community is increasingly coming to realize how central self-control is to many important life outcomes. We have always known about the impact of socioeconomic status and IQ, but these are factors that are highly resistant to interventions. In contrast, self-control may be something that we can tap into to make sweeping improvements life outcomes.

If you think about the environment we live in, you will notice how it is essentially designed to challenge every grain of our self-control. Businesses have the means and motivation to get us to do things NOW, not later. Krispy Kreme wants us to buy a dozen doughnuts while they are hot; Best Buy wants us to buy a television before we leave the store today; even our physicians want us to hurry up and schedule our annual checkup.

There is not much place for waiting in today’s marketplace. In fact you can think about the whole capitalist system as being designed to get us to take actions and spend money now – and those businesses that are more successful in that do better and prosper (at least in the short term).  And this of course continuously tests our ability to resist temptation and exercise self-control.

It is in this very environment that it’s particularly important to understand what’s going on behind the mysterious force of self-control.

[div class=attrib]More from theSource here.[end-div]

America: Paradoxical icon of the new

[div class=attrib]From Eurozine:[end-div]

Blaming the American Way of Life for the ills of post-industrial European society is a poor excuse for Europeans’ own partiality to consumer pleasures, writes Petr Fischer. On a positive note, American individualism could teach Europe a thing or two about social solidarity.

“Business–Answer–Solution” reads the advertising banner of the subsidiary of a foreign company in the centre of Prague. At first sight, the banner is not particularly interesting, in this case meaning that it is not particularly surprising. Surprising things are those that capture our attention, that shock us in their particular way. This corporate motto repeats the famous, infinitely repeated mantra of aggressive global capitalism, its focus purely pragmatic: give us a problem and we will come up with a solution that profits both you and us. “Win-win capitalism”, one could say in today’s international newspeak.

What is interesting – in other words disconcerting – is the fact that the banner covers the window of a small shop situated directly behind the National Museum, a building that – as in every other European city – symbolizes a certain perception of historicity cultivated on the old continent at least since the nineteenth century. The National Museum preserves the history of the Czech nation, and the people who work in it analyse and reflect on Czech national existence, its peculiarity, uniqueness, difference or connectedness. This activity is not governed by the pragmatic slogan of performance, of completed things, of faits accomplis; rather, it is ruled by a different three words, directed at thinking and its incessant, uncertain movement: Discussion–Question–Searching.

Both slogans represent two sides of the same coin of western civilization, two sides that, so far, have been more or less separate. The first represents the straightforward American way, leveraging everything along the way, everything at hand that can help business; the latter represents the difficult, reflective way of the old continent, left by its American child so that it could later be changed according to America’s picture. The fact that the multinational company’s motto is located just “behind” the building that, synecdochically, expresses the basic historic orientation of all European nations, is symbolic. “Behind”, meta in Greek, describes, in the European tradition, something that transcends everything we can arrive at though normal reasoning. In Aristotle’s canon, so the philosophical legend has it, such was the name of the texts found in the library behind the thinker’s treatise on physics. However metaphysics has since come to signify a system of thought that transcends the world of tangible facts and things, that represents some invisible internal order of the world. Business–Answer–Solution, the catchword of American pragmatism, is, as its location behind the National Museum suggests, perhaps the only really functioning metaphysics of today’s world.

New is always better

Since its discovery, America has been referred to as the New World. But what exactly is new about it for the Europeans? In De la démocratie en Amérique, Alexis de Tocqueville – one of the first to systematically analyse American institutions, republican political systems, and above all what today is called the “American way of life” – concluded that the newness of America consist mostly of a kind of neophilia, a love of all that is new.

“The Americans live in a country of wonders, everything around them is in incessant motion, and every motion seems to be progress,” says de Tocqueville. “The image of the new is closely connected with the image of the better. They see no limits set by Nature on man’s efforts; in American eyes, that which does not exist is what no one has yet tried.” In this extension of the purest Enlightenment optimism, the new is associated with a higher, moral quality. The gaze of the man turns toward the future, the past ceases to be important because, in the rush towards the new, the better, it loses its value, becomes inferior. The essential is what will be, or rather, what part of the future can be realized “now”.

[div class=attrib]More from theSource here.[end-div]

How Free Is Your Will?

[div class=attrib]From Scientific American:[end-div]

Think about the last time you got bored with the TV channel you were watching and decided to change it with the remote control. Or a time you grabbed a magazine off a newsstand, or raised a hand to hail a taxi. As we go about our daily lives, we constantly make choices to act in certain ways. We all believe we exercise free will in such actions – we decide what to do and when to do it. Free will, however, becomes more complicated when you try to think how it can arise from brain activity.

Do we control our neurons or do they control us? If everything we do starts in the brain, what kind of neural activity would reflect free choice? And how would you feel about your free will if we were to tell you that neuroscientists can look at your brain activity, and tell that you are about to make a decision to move – and that they could do this a whole second and a half before you yourself became aware of your own choice?

Scientists from UCLA and Harvard — Itzhak Fried, Roy Mukamel and Gabriel Kreiman — have taken an audacious step in the search for free will, reported in a new article in the journal Neuron. They used a powerful tool – intracranial recording – to find neurons in the human brain whose activity predicts decisions to make a movement, challenging conventional notions of free will.

Fried is one of a handful of neurosurgeons in the world who perform the delicate procedure of inserting electrodes into a living human brain, and using them to record activity from individual neurons. He does this to pin down the source of debilitating seizures in the brains of epileptic patients. Once he locates the part of the patients’ brains that sparks off the seizures, he can remove it, pulling the plug on their neuronal electrical storms.

[div class=attrib]More from theSource here.[end-div]

A New Tool for Creative Thinking: Mind-Body Dissonance

[div class=attrib]From Scientific American:[end-div]

A New Tool for Creative Thinking: Mind-Body Dissonance

Did you ever get the giggles during a religious service or some other serious occasion?  Did you ever have to smile politely when you felt like screaming?  In these situations, the emotions that we are required to express differ from the ones we are feeling inside.  That can be stressful, unpleasant, and exhausting.  Normally our minds and our bodies are in harmony.  When facial expressions or posture depart from how we feel, we experience what two psychologists at Northwestern University, Li Huang and Adam Galinsky, call mind–body dissonance.  And in a fascinating new paper, they show that such awkward clashes between mind and body can actually be useful: they help us think more expansively.

Ask yourself, would you say that a camel is a vehicle?  Would you describe a handbag as an item of clothing?  Your default answer might be negative, but there’s a way in which the camels can be regarded as forms of transport, and handbags can certainly be said to dress up an outfit.  When we think expansively, we think about categories more inclusively, we stop privileging the average cases, and extend our horizons to the atypical or exotic.  Expansive thought can be regarded a kind of creativity, and an opportunity for new insights.

Huang and Galinsky have shown that mind–body dissonance can make us think expansively.  In a clever series of studies, they developed a way to get people’s facial expressions to depart from their emotional experiences.  Participants were asked to either hold a pen between their teeth, forcing an unwitting smile, or to affix two golf tees in a particular position on their foreheads, unwittingly forcing an expression of sadness.  While in these facial configurations subjects were asked to recall happy and sad events or listen to happy and sad music.

[div class=attrib]More from theSource here.[end-div]

Jevons Paradox: Energy Efficiency Increases Consumption?

Energy efficiency sounds simple, but it’s rather difficult to measure. Sure when you purchase a shiny, new more energy efficient washing machine compared with your previous model you’re making a personal dent in energy consumption. But, what if in aggregate overall consumption increases because more people want that energy efficient model? In a nutshell, that’s Jevons Paradox, named after a 19th-century British economist, William Jevons. He observed that while the steam engine consumed energy more efficiently from coal, it also stimulated so much economic growth that coal consumption actually increased. Thus, Jevons argued that improvements in fuel efficiency tend to increase, rather than decrease, fuel use.

John Tierney over at the New York Times brings Jevons into the 21st century and discovers that the issues remain the same.

[div class=attrib]From the New York Times:[end-div]

For the sake of a cleaner planet, should Americans wear dirtier clothes?

This is not a simple question, but then, nothing about dirty laundry is simple anymore. We’ve come far since the carefree days of 1996, when Consumer Reports tested some midpriced top-loaders and reported that “any washing machine will get clothes clean.”

In this year’s report, no top-loading machine got top marks for cleaning. The best performers were front-loaders costing on average more than $1,000. Even after adjusting for inflation, that’s still $350 more than the top-loaders of 1996.

What happened to yesterday’s top-loaders? To comply with federal energy-efficiency requirements, manufacturers made changes like reducing the quantity of hot water. The result was a bunch of what Consumer Reports called “washday wash-outs,” which left some clothes “nearly as stained after washing as they were when we put them in.”

Now, you might think that dirtier clothes are a small price to pay to save the planet. Energy-efficiency standards have been embraced by politicians of both parties as one of the easiest ways to combat global warming. Making appliances, cars, buildings and factories more efficient is called the “low-hanging fruit” of strategies to cut greenhouse emissions.

But a growing number of economists say that the environmental benefits of energy efficiency have been oversold. Paradoxically, there could even be more emissions as a result of some improvements in energy efficiency, these economists say.

The problem is known as the energy rebound effect. While there’s no doubt that fuel-efficient cars burn less gasoline per mile, the lower cost at the pump tends to encourage extra driving. There’s also an indirect rebound effect as drivers use the money they save on gasoline to buy other things that produce greenhouse emissions, like new electronic gadgets or vacation trips on fuel-burning planes.

[div class=attrib]Read more here.[end-div]

[div class=attrib]Image courtesy of Wikipedia, Popular Science Monthly / Creative Commons.[end-div]

A radical pessimist’s guide to the next 10 years

[div class=attrib]The Globe and Mail:[end-div]

The iconic writer reveals the shape of things to come, with 45 tips for survival and a matching glossary of the new words you’ll need to talk about your messed-up future.

1) It’s going to get worse

No silver linings and no lemonade. The elevator only goes down. The bright note is that the elevator will, at some point, stop.

2) The future isn’t going to feel futuristic

It’s simply going to feel weird and out-of-control-ish, the way it does now, because too many things are changing too quickly. The reason the future feels odd is because of its unpredictability. If the future didn’t feel weirdly unexpected, then something would be wrong.

3) The future is going to happen no matter what we do. The future will feel even faster than it does now

The next sets of triumphing technologies are going to happen, no matter who invents them or where or how. Not that technology alone dictates the future, but in the end it always leaves its mark. The only unknown factor is the pace at which new technologies will appear. This technological determinism, with its sense of constantly awaiting a new era-changing technology every day, is one of the hallmarks of the next decade.

4)Move to Vancouver, San Diego, Shannon or Liverpool

There’ll be just as much freaky extreme weather in these west-coast cities, but at least the west coasts won’t be broiling hot and cryogenically cold.

5) You’ll spend a lot of your time feeling like a dog leashed to a pole outside the grocery store – separation anxiety will become your permanent state

6) The middle class is over. It’s not coming back

Remember travel agents? Remember how they just kind of vanished one day?

That’s where all the other jobs that once made us middle-class are going – to that same, magical, class-killing, job-sucking wormhole into which travel-agency jobs vanished, never to return. However, this won’t stop people from self-identifying as middle-class, and as the years pass we’ll be entering a replay of the antebellum South, when people defined themselves by the social status of their ancestors three generations back. Enjoy the new monoclass!

7) Retail will start to resemble Mexican drugstores

In Mexico, if one wishes to buy a toothbrush, one goes to a drugstore where one of every item for sale is on display inside a glass display case that circles the store. One selects the toothbrush and one of an obvious surplus of staff runs to the back to fetch the toothbrush. It’s not very efficient, but it does offer otherwise unemployed people something to do during the day.

8) Try to live near a subway entrance

In a world of crazy-expensive oil, it’s the only real estate that will hold its value, if not increase.

9) The suburbs are doomed, especially thoseE.T. , California-style suburbs

This is a no-brainer, but the former homes will make amazing hangouts for gangs, weirdoes and people performing illegal activities. The pretend gates at the entranceways to gated communities will become real, and the charred stubs of previous white-collar homes will serve only to make the still-standing structures creepier and more exotic.

10) In the same way you can never go backward to a slower computer, you can never go backward to a lessened state of connectedness

11) Old people won’t be quite so clueless

No more “the Google,” because they’ll be just that little bit younger.

12) Expect less

Not zero, just less.

13) Enjoy lettuce while you still can

And anything else that arrives in your life from a truck, for that matter. For vegetables, get used to whatever it is they served in railway hotels in the 1890s. Jams. Preserves. Pickled everything.

14) Something smarter than us is going to emerge

Thank you, algorithms and cloud computing.

15) Make sure you’ve got someone to change your diaper

Sponsor a Class of 2112 med student. Adopt up a storm around the age of 50.

16) “You” will be turning into a cloud of data that circles the planet like a thin gauze

While it’s already hard enough to tell how others perceive us physically, your global, phantom, information-self will prove equally vexing to you: your shopping trends, blog residues, CCTV appearances – it all works in tandem to create a virtual being that you may neither like nor recognize.

17) You may well burn out on the effort of being an individual

You’ve become a notch in the Internet’s belt. Don’t try to delude yourself that you’re a romantic lone individual. To the new order, you’re just a node. There is no escape

18) Untombed landfills will glut the market with 20th-century artifacts

19) The Arctic will become like Antarctica – an everyone/no one space

Who owns Antarctica? Everyone and no one. It’s pie-sliced into unenforceable wedges. And before getting huffy, ask yourself, if you’re a Canadian: Could you draw an even remotely convincing map of all those islands in Nunavut and the Northwest Territories? Quick, draw Ellesmere Island.

20)

North America can easily fragment quickly as did the Eastern Bloc in 1989

Quebec will decide to quietly and quite pleasantly leave Canada. California contemplates splitting into two states, fiscal and non-fiscal. Cuba becomes a Club Med with weapons. The Hate States will form a coalition.

21) We will still be annoyed by people who pun, but we will be able to show them mercy because punning will be revealed to be some sort of connectopathic glitch: The punner, like someone with Tourette’s, has no medical ability not to pun

22) Your sense of time will continue to shred. Years will feel like hours

23) Everyone will be feeling the same way as you

There’s some comfort to be found there.

24) It is going to become much easier to explain why you are the way you are

Much of what we now consider “personality” will be explained away as structural and chemical functions of the brain.

25) Dreams will get better

26)

Being alone will become easier

27)Hooking up will become ever more mechanical and binary

28) It will become harder to view your life as “a story”

The way we define our sense of self will continue to morph via new ways of socializing. The notion of your life needing to be a story will seem slightly corny and dated. Your life becomes however many friends you have online.

29) You will have more say in how long or short you wish your life to feel

Time perception is very much about how you sequence your activities, how many activities you layer overtop of others, and the types of gaps, if any, you leave in between activities.

30) Some existing medical conditions will be seen as sequencing malfunctions

The ability to create and remember sequences is an almost entirely human ability (some crows have been shown to sequence). Dogs, while highly intelligent, still cannot form sequences; it’s the reason why well-trained dogs at shows are still led from station to station by handlers instead of completing the course themselves.

Dysfunctional mental states stem from malfunctions in the brain’s sequencing capacity. One commonly known short-term sequencing dysfunction is dyslexia. People unable to sequence over a slightly longer term might be “not good with directions.” The ultimate sequencing dysfunction is the inability to look at one’s life as a meaningful sequence or story.

31) The built world will continue looking more and more like Microsoft packaging

“We were flying over Phoenix, and it looked like the crumpled-up packaging from a 2006 MS Digital Image Suite.”

32) Musical appreciation will shed all age barriers

33) People who shun new technologies will be viewed as passive-aggressive control freaks trying to rope people into their world, much like vegetarian teenage girls in the early 1980s

1980: “We can’t go to that restaurant. Karen’s vegetarian and it doesn’t have anything for her.”

2010: “What restaurant are we going to? I don’t know. Karen was supposed to tell me, but she doesn’t have a cell, so I can’t ask her. I’m sick of her crazy control-freak behaviour. Let’s go someplace else and not tell her where.”

34) You’re going to miss the 1990s more than you ever thought

35) Stupid people will be in charge, only to be replaced by ever-stupider people. You will live in a world without kings, only princes in whom our faith is shattered

36) Metaphor drift will become pandemic

Words adopted by technology will increasingly drift into new realms to the point where they observe different grammatical laws, e.g., “one mouse”/“three mouses;” “memory hog”/“delete the spam.”

37) People will stop caring how they appear to others

The number of tribal categories one can belong to will become infinite. To use a high-school analogy, 40 years ago you had jocks and nerds. Nowadays, there are Goths, emos, punks, metal-heads, geeks and so forth.

38)Knowing everything will become dull

It all started out so graciously: At a dinner for six, a question arises about, say, that Japanese movie you saw in 1997 (Tampopo), or whether or not Joey Bishop is still alive (no). And before long, you know the answer to everything.

39) IKEA will become an ever-more-spiritual sanctuary

40) We will become more matter-of-fact, in general, about our bodies

41) The future of politics is the careful and effective implanting into the minds of voters images that can never be removed

42) You’ll spend a lot of time shopping online from your jail cell

Over-criminalization of the populace, paired with the triumph of shopping as a dominant cultural activity, will create a world where the two poles of society are shopping and jail.

43) Getting to work will provide vibrant and fun new challenges

Gravel roads, potholes, outhouses, overcrowded buses, short-term hired bodyguards, highwaymen, kidnapping, overnight camping in fields, snaggle-toothed crazy ladies casting spells on you, frightened villagers, organ thieves, exhibitionists and lots of healthy fresh air.

44) Your dream life will increasingly look like Google Street View

45) We will accept the obvious truth that we brought this upon ourselves

Douglas Coupland is a writer and artist based in Vancouver, where he will deliver the first of five CBC Massey Lectures – a ‘novel in five hours’ about the future – on Tuesday.

[div class=attrib]More from theSource here.[end-div]

Contain this!

[div class=attrib]From Eurozine:[end-div]

WikiLeaks’ series of exposés is causing a very different news and informational landscape to emerge. Whilst acknowledging the structural leakiness of networked organisations, Felix Stalder finds deeper reasons for the crisis of information security and the new distribution of investigative journalism.

WikiLeaks is one of the defining stories of the Internet, which means by now, one of the defining stories of the present, period. At least four large-scale trends which permeate our societies as a whole are fused here into an explosive mixture whose fall-out is far from clear. First is a change in the materiality of communication. Communication becomes more extensive, more recorded, and the records become more mobile. Second is a crisis of institutions, particularly in western democracies, where moralistic rhetoric and the ugliness of daily practice are diverging ever more at the very moment when institutional personnel are being encouraged to think more for themselves. Third is the rise of new actors, “super-empowered” individuals, capable of intervening into historical developments at a systemic level. Finally, fourth is a structural transformation of the public sphere (through media consolidation at one pole, and the explosion of non-institutional publishers at the other), to an extent that rivals the one described by Habermas with the rise of mass media at the turn of the twentieth century.

Leaky containers

Imagine dumping nearly 400 000 paper documents into a dead drop located discreetly on the hard shoulder of a road. Impossible. Now imagine the same thing with digital records on a USB stick, or as an upload from any networked computer. No problem at all. Yet, the material differences between paper and digital records go much further than mere bulk. Digital records are the impulses travelling through the nervous systems of dynamic, distributed organisations of all sizes. They are intended, from the beginning, to circulate with ease. Otherwise such organisations would fall apart and dynamism would grind to a halt. The more flexible and distributed organisations become, the more records they need to produce and the faster these need to circulate. Due to their distributed aspect and the pressure for cross-organisational cooperation, it is increasingly difficult to keep records within particular organisations whose boundaries are blurring anyway. Surveillance researchers such as David Lyon have long been writing about the leakiness of “containers”, meaning the tendency for sensitive digital records to cross the boundaries of the institutions which produce them. This leakiness is often driven by commercial considerations (private data being sold), but it happens also out of incompetence (systems being secured insufficiently), or because insiders deliberately violate organisational policies for their own purposes. Either they are whistle-blowers motivated by conscience, as in the case of WikiLeaks, or individuals selling information for private gain, as in the case of the numerous employees of Swiss banks who recently copied the details of private accounts and sold them to tax authorities across Europe. Within certain organisation such as banks and the military, virtually everything is classified and large number of people have access to this data, not least mid-level staff who handle the streams of raw data such as individuals’ records produced as part of daily procedure.

[div class=attrib]More from theSource here.[end-div]

Map of the World’s Countries Rearranged by Population

[div class=attrib]From Frank Jacobs / BigThink:[end-div]

What if the world were rearranged so that the inhabitants of the country with the largest population would move to the country with the largest area? And the second-largest population would migrate to the second-largest country, and so on?

The result would be this disconcerting, disorienting map. In the world described by it, the differences in population density between countries would be less extreme than they are today. The world’s most densely populated country currently is Monaco, with 43,830 inhabitants/mi² (16,923 per km²) (1). On the other end of the scale is Mongolia, which is less densely populated by a factor of almost exactly 10,000, with a mere 4.4 inhabitants/mi² (1.7 per km²).

The averages per country would more closely resemble the global average of 34 per mi² (13 per km²). But those evened-out statistics would describe a very strange world indeed. The global population realignment would involve massive migrations, lead to a heap of painful demotions and triumphant promotions, and produce a few very weird new neighbourhoods.

Take the world’s largest country: Russia. It would be taken over by its Asian neighbour and rival China, the country with the world’s largest population. Overcrowded China would not just occupy underpopulated Siberia – a long-time Russian fear – but also fan out all the way across the Urals to Russia’s westernmost borders. China would thus become a major European power. Russia itself would be relegated to Kazakhstan, which still is the largest landlocked country in the world, but with few hopes of a role on the world stage commensurate with Russia’s clout, which in no small part derives from its sheer size.

Canada, the world’s second-largest country, would be transformed into an Arctic, or at least quite chilly version of India, the country with the world’s second-largest population. The country would no longer be a thinly populated northern afterthought of the US. The billion Indians north of the Great Lakes would make Canada a very distinct, very powerful global player.

Strangely enough, the US itself would not have to swap its population with another country. With 310 million inhabitants, it is the third most populous nation in the world. And with an area of just over 3.7 million mi² (slightly more than 9.6 million km²), it is also the world’s third largest country (2). Brazil, at number five in both lists, is in the same situation. Other non-movers are Yemen and Ireland. Every other country moves house. A few interesting swaps:

  • Countries with relatively high population densities move to more spacious environments. This increases their visibility. Look at those 94 million Filipinos, for example, no longer confined to that small archipelago just south of China. They now occupy the sprawling Democratic Republic of the Congo, the 12th largest country in the world, and slap bang in the middle of Africa too.
  • The reverse is also true. Mongolia, that large, sparsely populated chunk of a country between Russia and China, is relegated to tiny Belgium, whose even tinier neighbour Luxembourg is populated by 320,000 Icelanders, no longer enjoying the instant recognition provided by their distinctly shaped North Atlantic island home.
  • Australia’s 22.5 million inhabitants would move to Spain, the world’s 51st largest country. This would probably be the furthest migration, as both countries are almost exactly antipodean to each other. But Australians would not have to adapt too much to the mainly hot and dry Spanish climate.
  • But spare a thought for those unfortunate Vietnamese. Used to a lush, tropical climate, the 85 million inhabitants of Vietnam would be shipped off to icy Greenland. Even though that Arctic dependency of Denmark has warmed up a bit due to recent climate changes, it would still be mainly snowy, empty and freezing. One imagines a giant group huddle, just to keep warm.
  • Jamaica would still be island-shaped – but landlocked, as the Jamaicans would move to Lesotho, an independent enclave completely surrounded by South Africa – or rather, in this strange new world, South Korea. Those South Koreans probably couldn’t believe their bad luck. Of all the potential new friends in the world, who gets to be their northern neighbour but their wacky cousin, North Korea? It seems the heavily militarised DMZ will move from the Korean peninsula to the South African-Botswanan border.
  • The UK migrates from its strategically advantageous island position off Europe’s western edge to a place smack in the middle of the Sahara desert, to one of those countries the name of which one always has to look up (3). No longer splendidly isolated, it will have to share the neighbourhood with such upstarts as Mexico, Myanmar, Thailand and – good heavens – Iran. Back home, its sceptered isles are taken over by the Tunisians. Even Enoch Powell didn’t see that one coming.
  • Some countries only move a few doors down, so to speak. El Salvador gets Guatemala, Honduras takes over Nicaragua, Nepal occupies Birma/Myanmar and Turkey sets up house in Iran. Others wake up in a whole new environment. Dusty, landlocked Central African Republic is moving to the luscious island of Sri Lanka, with its pristine, ocean-lapped shores. The mountain-dwelling Swiss will have to adapt to life in the flood-infested river delta of Bangladesh.
  • Geography, they say, is destiny (4). Some countries are plagued or blessed by their present location. How would they fare elsewhere? Take Iraq, brought down by wars both of the civil and the other kind, and burdened with enough oil to finance lavish dictatorships and arouse the avidity of superpowers. What if the 31.5 million Iraqis moved to the somewhat larger, equally sunny country of Zambia – getting a lot of nice, non-threatening neighbours in the process?

Rearranged maps that switch the labels of the countries depicted, as if in some parlour game, to represent some type of statistical data, are an interesting subcategory of curious cartography. The most popular example discussed on this blog is the map of the US, with the states’ names replaced by that of countries with an equivalent GDP (see #131). Somewhat related, if by topic rather than technique, is the cartogram discussed in blog post #96, showing the world’s countries shrunk or inflated to reflect the size of their population.

Many thanks to all who sent in this map: Matt Chisholm, Criggie, Roel Damiaans, Sebastian Dinjens, Irwin Hébert, Allard H., Olivier Muzerelle, Rodrigo Oliva, Rich Sturges, and John Thorne. The map is referenced on half a dozen websites where it can be seen in full resolution (this one among them), but it is unclear where it first originated, and who produced it (the map is signed, in the bottom right hand corner, by JPALMZ).

—–

(1) Most (dependent) territories and countries in the top 20 of Wikipedia’s population density ranking have tiny areas, with populations that are, in relation to those of other countries, quite negligeable. The first country on the list with both a substantial surface and population is Bangladesh, in 9th place with a total population of over 162 million and a density of 1,126 inhabitants/mi² (56 per km²).

(2) Actually, the US contends third place with China. Both countries have almost the same size, and varying definitions of how large they are. Depending on whether or not you include Taiwan and (other) disputed areas in China, and overseas territories in the US, either country  can be third of fourth on the list.

(3) Niger, not to be confused with nearby Nigeria. Nor with neighbouring Burkina Faso, which used to be Upper Volta (even though there never was a Lower Volta except, perhaps, Niger. Or Nigeria).

(4) The same is said of demography. And of a bunch of other stuff.

[div class=attrib]More from theSource here.[end-div]

The Top Ten Daily Consequences of Having Evolved

[div class=attrib]From Smithsonian.com:[end-div]

Natural selection acts by winnowing the individuals of each generation, sometimes clumsily, as old parts and genes are co-opted for new roles. As a result, all species inhabit bodies imperfect for the lives they live. Our own bodies are worse off than most simply because of the many differences between the wilderness in which we evolved and the modern world in which we live. We feel the consequences every day. Here are ten.

1. Our cells are weird chimeras
Perhaps a billion years ago, a single-celled organism arose that would ultimately give rise to all of the plants and animals on Earth, including us. This ancestor was the result of a merging: one cell swallowed, imperfectly, another cell. The predator provided the outsides, the nucleus and most of the rest of the chimera. The prey became the mitochondrion, the cellular organ that produces energy. Most of the time, this ancient symbiosis proceeds amicably. But every so often, our mitochondria and their surrounding cells fight. The result is diseases, such as mitochondrial myopathies (a range of muscle diseases) or Leigh’s disease (which affects the central nervous system).

2. Hiccups
The first air-breathing fish and amphibians extracted oxygen using gills when in the water and primitive lungs when on land—and to do so, they had to be able to close the glottis, or entryway to the lungs, when underwater. Importantly, the entryway (or glottis) to the lungs could be closed. When underwater, the animals pushed water past their gills while simultaneously pushing the glottis down. We descendants of these animals were left with vestiges of their history, including the hiccup. In hiccupping, we use ancient muscles to quickly close the glottis while sucking in (albeit air, not water). Hiccups no longer serve a function, but they persist without causing us harm—aside from frustration and occasional embarrassment. One of the reasons it is so difficult to stop hiccupping is that the entire process is controlled by a part of our brain that evolved long before consciousness, and so try as you might, you cannot think hiccups away.

3. Backaches
The backs of vertebrates evolved as a kind of horizontal pole under which guts were slung. It was arched in the way a bridge might be arched, to support weight. Then, for reasons anthropologists debate long into the night, our hominid ancestors stood upright, which was the bodily equivalent of tipping a bridge on end. Standing on hind legs offered advantages—seeing long distances, for one, or freeing the hands to do other things—but it also turned our backs from an arched bridge to an S shape. The letter S, for all its beauty, is not meant to support weight and so our backs fail, consistently and painfully.

[div class=attrib]More from theSource here.[end-div]

MondayPoem: The Lie

[div class=attrib]By Robert Pinsky for Slate:[end-div]

Denunciation abounds, in its many forms: snark (was that word invented or fostered in a poem, Lewis Carroll’s “The Hunting of the Snark“?), ranking-out, calling-out, bringing-down, blowing-up, flaming, scorching, trashing, negative campaigning, skepticism, exposure, nailing, shafting, finishing, diminishing, down-blogging. Aggressive moral denunciation—performed with varying degrees of justice and skill in life, in print, on the Web, in politics, on television and radio, in book-reviewing, in sports, in courtrooms and committee meetings—generates dismay and glee in its audience. Sometimes, for many of us, dismay and glee simultaneously, in an uneasy combination.

A basic form of denunciation is indicated by the slightly archaic but useful expression giving the lie.

No one has ever given the lie more memorably, explicitly, and universally than Sir Walter Raleigh (1552-1618) in “The Lie.” The poem, among other things, demonstrates the power of repetition and refrain. The power, too, of plain rather than fancy or arcane words—for example, blabbing.

I remember being enchanted—a bit excessively, I now think—when I first read “The Lie” by a single wonderful image early on: “Say to the court it glows/ And shines like rotten wood.” The mental picture of an opalescent, greenish glow on a moldy softwood plank—that phosphorescent decay—knocked me out (to use an expression from those student days). It was a period when images were highly prized, and my teachers encouraged me to prize images, the deeper the better. Well, though I may have been unreflectingly guided by fashion, at least I had the brains to appreciate this great image of Raleigh’s.

But now that superb rotten wood feels like an incidental or ancillary beauty to me, one moment in a larger force. What propels this poem is not its images but its masterful breaking down of an idea into social and moral components: the brilliant, considered division into hammer-blows of example and refrain while the pace and content vary around that central pulse. “Driving home the point” could not have a more apt demonstration.

Raleigh’s manic, extended thoroughness; his resourceful rhyming; his relentless, wide gaze that takes in love and zeal, wit and wisdom, and, ultimately, also includes his own soul’s “blabbing”—this is form as audible conviction: conviction of a degree and kind attainable only by a poem.

“The Lie”

Go, soul, the body’s guest,
….Upon a thankless arrant;
Fear not to touch the best;
….The truth shall be thy warrant:
….….Go, since I needs must die,
….….And give the world the lie.

Say to the court it glows
….And shines like rotten wood,
Say to the church it shows
….What’s good, and doth no good:
….….If church and court reply,
….….Then give them both the lie.

Tell potentates, they live
….Acting, by others’ action;
Not lov’d unless they give;
….Not strong, but by affection.
….….If potentates reply,
….….Give potentates the lie.

Tell men of high condition,
….That manage the estate,
Their purpose is ambition;
….Their practice only hate.
….….And if they once reply,
….….Then give them all the lie.

Tell them that brave it most,
….They beg for more by spending,
Who in their greatest cost
….Like nothing but commending.
….….And if they make reply,
….….Then give them all the lie.

Tell zeal it wants devotion;
….Tell love it is but lust;
Tell time it meets but motion;
….Tell flesh it is but dust:
….….And wish them not reply,
….….For thou must give the lie.

Tell age it daily wasteth;
….Tell honour how it alters;
Tell beauty how she blasteth;
….Tell favour how it falters:
….….And as they shall reply,
….….Give every one the lie.

Tell wit how much it wrangles
….In tickle points of niceness;
Tell wisdom she entangles
….Herself in over-wiseness:
….….And when they do reply,
….….Straight give them both the lie.

Tell physic of her boldness;
….Tell skill it is prevention;
Tell charity of coldness;
….Tell law it is contention:
….….And as they do reply,
….….So give them still the lie.

Tell fortune of her blindness;
….Tell nature of decay;
Tell friendship of unkindness;
….Tell justice of delay:
….….And if they will reply,
….….Then give them all the lie.

Tell arts they have no soundness,
….But vary by esteeming;
Tell schools they want profoundness,
….And stand too much on seeming.
….….If arts and schools reply,
….….Give arts and schools the lie.

Tell faith it’s fled the city;
….Tell how the country erreth;
Tell manhood, shakes off pity;
….Tell virtue, least preferreth.
….….And if they do reply,
….….Spare not to give the lie.

So when thou hast, as I
….Commanded thee, done blabbing;
Because to give the lie
….Deserves no less than stabbing:
….….Stab at thee, he that will,
….….No stab thy soul can kill!

—Sir Walter Raleigh

[div class=attrib]More from theSource here.[end-div]

Search Engine History

It’s hard to believe that internet based search engines have been in the mainstream consciousness for around twenty years now. It seems not too long ago that we were all playing Pong and searching index cards at the local library. Infographics Labs puts the last twenty years of search in summary for us below.

[div class-attrib]From Infographic Labs:[end-div]

Search Engine History

Infographic: Search Engine History by Infographiclabs

Andre Geim: in praise of graphene

[div class=attrib]From Nature:[end-div]

Nobel laureate explains why the carbon sheets deserved to win this year’s prize.

This year’s Nobel Prize in Physics went to the discoverers of the one-atom-thick sheets of carbon known as graphene. Andre Geim of the University of Manchester, UK, who shared the award with his colleague Konstantin Novoselov, tells Nature why graphene deserves the prize, and why he hasn’t patented it.

In one sentence, what is graphene?

Graphene is a single plane of graphite that has to be pulled out of bulk graphite to show its amazing properties.

What are these properties?

It’s the thinnest possible material you can imagine. It also has the largest surface-to-weight ratio: with one gram of graphene you can cover several football pitches (in Manchester, you know, we measure surface area in football pitches). It’s also the strongest material ever measured; it’s the stiffest material we know; it’s the most stretchable crystal. That’s not the full list of superlatives, but it’s pretty impressive.

A lot of people expected you to win, but not so soon after the discovery in 2004. Were you expecting it?

I didn’t think it would happen this year. I was thinking about next year or maybe 2014. I slept quite soundly without much expectation. Yeah, it’s good, it’s good.

Graphene has won, but not that much has actually been done with it yet. Do you think it was too soon?

No. The prize, if you read the citation, was given for the properties of graphene; it wasn’t given for expectations that have not yet been realized. Ernest Rutherford’s 1908 Nobel Prize in Chemistry wasn’t given for the nuclear power station — he wouldn’t have survived that long — it was given for showing how interesting atomic physics could be. I believe the Nobel prize committee did a good job.

Do you think that carbon nanotubes were unfairly overlooked?

It’s difficult to judge; I’m a little afraid of being biased. If the prize had been given for bringing graphene to the attention of the community, then it would have been unfair to take it away from carbon nanotubes. But it was given for graphene’s properties, and I think carbon nanotubes did not deliver that range of properties. Everyone knows that — in terms of physics, not applications — carbon nanotubes were not as successful as graphene.

Why do you think graphene has become so popular in the physics community?

I would say there are three important things about graphene. It’s two-dimensional, which is the best possible number for studying fundamental physics. The second thing is the quality of graphene, which stems from its extremely strong carbon–carbon bonds. And finally, the system is also metallic.

What do you think graphene will be used for first?

Two or three months ago, I was in South Korea, and I was shown a graphene roadmap, compiled by Samsung. On this roadmap were approximately 50 dots, corresponding to particular applications. One of the closest applications with a reasonable market value was a flexible touch screen. Samsung expects something within two to three years.

[div class=attrib]More from theSource here.[end-div]

Small Change. Why the Revolution will Not be Tweeted

[div class=attrib]From The New Yorker:[end-div]

At four-thirty in the afternoon on Monday, February 1, 1960, four college students sat down at the lunch counter at the Woolworth’s in downtown Greensboro, North Carolina. They were freshmen at North Carolina A. & T., a black college a mile or so away.

“I’d like a cup of coffee, please,” one of the four, Ezell Blair, said to the waitress.

“We don’t serve Negroes here,” she replied.

The Woolworth’s lunch counter was a long L-shaped bar that could seat sixty-six people, with a standup snack bar at one end. The seats were for whites. The snack bar was for blacks. Another employee, a black woman who worked at the steam table, approached the students and tried to warn them away. “You’re acting stupid, ignorant!” she said. They didn’t move. Around five-thirty, the front doors to the store were locked. The four still didn’t move. Finally, they left by a side door. Outside, a small crowd had gathered, including a photographer from the Greensboro Record. “I’ll be back tomorrow with A. & T. College,” one of the students said.

By next morning, the protest had grown to twenty-seven men and four women, most from the same dormitory as the original four. The men were dressed in suits and ties. The students had brought their schoolwork, and studied as they sat at the counter. On Wednesday, students from Greensboro’s “Negro” secondary school, Dudley High, joined in, and the number of protesters swelled to eighty. By Thursday, the protesters numbered three hundred, including three white women, from the Greensboro campus of the University of North Carolina. By Saturday, the sit-in had reached six hundred. People spilled out onto the street. White teen-agers waved Confederate flags. Someone threw a firecracker. At noon, the A. & T. football team arrived. “Here comes the wrecking crew,” one of the white students shouted.

By the following Monday, sit-ins had spread to Winston-Salem, twenty-five miles away, and Durham, fifty miles away. The day after that, students at Fayetteville State Teachers College and at Johnson C. Smith College, in Charlotte, joined in, followed on Wednesday by students at St. Augustine’s College and Shaw University, in Raleigh. On Thursday and Friday, the protest crossed state lines, surfacing in Hampton and Portsmouth, Virginia, in Rock Hill, South Carolina, and in Chattanooga, Tennessee. By the end of the month, there were sit-ins throughout the South, as far west as Texas. “I asked every student I met what the first day of the sitdowns had been like on his campus,” the political theorist Michael Walzer wrote in Dissent. “The answer was always the same: ‘It was like a fever. Everyone wanted to go.’ ” Some seventy thousand students eventually took part. Thousands were arrested and untold thousands more radicalized. These events in the early sixties became a civil-rights war that engulfed the South for the rest of the decade—and it happened without e-mail, texting, Facebook, or Twitter.

The world, we are told, is in the midst of a revolution. The new tools of social media have reinvented social activism. With Facebook and Twitter and the like, the traditional relationship between political authority and popular will has been upended, making it easier for the powerless to collaborate, coördinate, and give voice to their concerns. When ten thousand protesters took to the streets in Moldova in the spring of 2009 to protest against their country’s Communist government, the action was dubbed the Twitter Revolution, because of the means by which the demonstrators had been brought together. A few months after that, when student protests rocked Tehran, the State Department took the unusual step of asking Twitter to suspend scheduled maintenance of its Web site, because the Administration didn’t want such a critical organizing tool out of service at the height of the demonstrations. “Without Twitter the people of Iran would not have felt empowered and confident to stand up for freedom and democracy,” Mark Pfeifle, a former national-security adviser, later wrote, calling for Twitter to be nominated for the Nobel Peace Prize. Where activists were once defined by their causes, they are now defined by their tools. Facebook warriors go online to push for change. “You are the best hope for us all,” James K. Glassman, a former senior State Department official, told a crowd of cyber activists at a recent conference sponsored by Facebook, A. T. & T., Howcast, MTV, and Google. Sites like Facebook, Glassman said, “give the U.S. a significant competitive advantage over terrorists. Some time ago, I said that Al Qaeda was ‘eating our lunch on the Internet.’ That is no longer the case. Al Qaeda is stuck in Web 1.0. The Internet is now about interactivity and conversation.”

These are strong, and puzzling, claims. Why does it matter who is eating whose lunch on the Internet? Are people who log on to their Facebook page really the best hope for us all? As for Moldova’s so-called Twitter Revolution, Evgeny Morozov, a scholar at Stanford who has been the most persistent of digital evangelism’s critics, points out that Twitter had scant internal significance in Moldova, a country where very few Twitter accounts exist. Nor does it seem to have been a revolution, not least because the protests—as Anne Applebaum suggested in the Washington Post—may well have been a bit of stagecraft cooked up by the government. (In a country paranoid about Romanian revanchism, the protesters flew a Romanian flag over the Parliament building.) In the Iranian case, meanwhile, the people tweeting about the demonstrations were almost all in the West. “It is time to get Twitter’s role in the events in Iran right,” Golnaz Esfandiari wrote, this past summer, in Foreign Policy. “Simply put: There was no Twitter Revolution inside Iran.” The cadre of prominent bloggers, like Andrew Sullivan, who championed the role of social media in Iran, Esfandiari continued, misunderstood the situation. “Western journalists who couldn’t reach—or didn’t bother reaching?—people on the ground in Iran simply scrolled through the English-language tweets post with tag #iranelection,” she wrote. “Through it all, no one seemed to wonder why people trying to coordinate protests in Iran would be writing in any language other than Farsi.”

Some of this grandiosity is to be expected. Innovators tend to be solipsists. They often want to cram every stray fact and experience into their new model. As the historian Robert Darnton has written, “The marvels of communication technology in the present have produced a false consciousness about the past—even a sense that communication has no history, or had nothing of importance to consider before the days of television and the Internet.” But there is something else at work here, in the outsized enthusiasm for social media. Fifty years after one of the most extraordinary episodes of social upheaval in American history, we seem to have forgotten what activism is.

[div class=attrib]More from theSource here.[end-div]

Art. Does it have to be BOLD to be good?

The lengthy corridors of art history over the last five hundred years are decorated with numerous bold and monumental works. Just to name a handful of memorable favorites you’ll see a pattern emerge: Guernica (Pablo Picasso), The Persistence of Memory (Salvador Dali), The Dance (Henri Matisse), The Garden of Earthly Delights (Heironymous Bosch). Yes, these works are bold. They’re bold in the sense that they represented a fundamental shift from the artistic sensibilities and ideas of their times. These works stirred the salons and caused commotion among the “cognosenti” and the chattering classes. They implored (or decried) the establishment to take notice of new forms, new messages, new perspectives.

And, now here we are in the 21st century, floating in a bottomless bowl of a bold media soup; 24-hour opinion and hyperbole; oversized interactive billboards, explosive 3D movies, voyeuristic reality TV, garish commercials, sexually charged headlines and suggestive mainstream magazines. The provocative images, the loudness, the vividness, the anger – it’s all bold and it’s vying for your increasingly fragmented and desensitized attention. But, this contemporary boldness seems more aligned with surface brightness and bigness than it is with depth of meaning. The boldness of works by earlier artists such as Picasso, Dali, Bosch came from depth of meaning rather than use of neon paints or other bold visual noise.

So, what of contemporary art over the last couple of decades? Well, a pseudo-scientific tour of half-a-dozen art galleries featuring the in-the-moment works of art may well tell you the same story – it’s mostly bold as well. What’s been selling at the top art auction houses? Bold. What’s been making headlines in the art world? Bold.

The trend is and has been set for a while: it has to be brighter, louder, bigger. Indeed, a recent feature article in the New York Times on the 25th Paris Biennale seems to confirm this trend in Western art. (Background: The Biennale is home to around a hundred of the world’s most exclusive art galleries, those that purport to set the art world’s trends, make or break emerging artists and most importantly (for them) set “market” prices.) The article’s author, Souren Melikian, states:

Perception is changing. Interest in subtle nuances is receding as our attention span shortens. Awareness of this trend probably accounts for the recent art trade emphasis on clarity and monumentality and the striking progression of 20th-century modernity.

Well, I certainly take no issue with the observation that “commercial” art has become much more monumental and less subtle, especially over the last 40 years. By it’s very nature for most art to be successful in today’s market overflowing with noise, distraction and mediocrity it must draw someone’s fragmented and limited attention, and sadly, it does this by being bold, bright or big! However, I strongly disagree that “clarity” is a direct result of this new trend in boldness. I could recite a list as long as my arm of paintings and other art works that show remarkable clarity even though they are merely subtle.

Perhaps paradoxically, brokers and buyers of bold seem exclusively to associate boldness with a statement of modernity, compositional complexity, and layered meaning. The galleries at the Biennale seem to be confusing subtlety with dullness, simplicity and shallowness. Yet, the world is full of an equal number of works that exhibit just as much richness, depth and emotion as their bolder counterparts despite their surface subtlety. There is room for reflection and nuanced mood; there is room for complexity and depth in meaning from simple composition; there is room for pastels in this over-saturated, bold neon world.

As Bob Duggan eloquently states, at BigThink:

The meek, such as 2009 Turner Prize winner Richard Wright (reviewed recently by me here) may yet inherit the earth, but only in a characteristically quiet way. Hirst’s jewel-encrusted skulls will always grab headlines, but Wright’s simpler, pensive work can engage hearts and minds in a more fulfilling way. And why is it important that the right thing happens and the Wrights win out over the Hirsts? Because art remains one of the few havens for thought in our noise- and light-polluted world.

So, I’m encouraged to see that I am not yet a lost and lone voice in this noisy wilderness of bold brashness. Oh, and in case you’re wondering what a meaningfully complex yet subtle painting looks like, gaze at Half Light by Dana Blanchard above.

Commonplaces of technology critique

[div class=attrib]From Eurozine:[end-div]

What is it good for? A passing fad! It makes you stupid! Today’s technology critique is tomorrow’s embarrassing error of judgement, as Katrin Passig shows. Her suggestion: one should try to avoid repeating the most commonplace critiques, particularly in public.

In a 1969 study on colour designations in different cultures, anthropologist Brent Berlin and linguist Paul Kay described how the sequence of levels of observed progression was always the same. Cultures with only two colour concepts distinguish between “light” and “dark” shades. If the culture recognizes three colours, the third will be red. If the language differentiates further, first come green and/or yellow, then blue. All languages with six colour designations distinguish between black, white, red, green, blue and yellow. The next level is brown, then, in varying sequences, orange, pink, purple and/or grey, with light blue appearing last of all.

The reaction to technical innovations, both in the media and in our private lives, follows similarly preconceived paths. The first, entirely knee-jerk dismissal is the “What the hell is it good for?” (Argument No.1) with which IBM engineer Robert Lloyd greeted the microprocessor in 1968. Even practices and techniques that only constitute a variation on the familiar – the electric typewriter as successor to the mechanical version, for instance – are met with distaste in the cultural criticism sector. Inventions like the telephone or the Internet, which open up a whole new world, have it even tougher. If cultural critics had existed at the dawn of life itself, they would have written grumpily in their magazines: “Life – what is it good for? Things were just fine before.”

Because the new throws into confusion processes that people have got used to, it is often perceived not only as useless but as a downright nuisance. The student Friedrich August Köhler wrote in 1790 after a journey on foot from Tübingen to Ulm: “[Signposts] had been put up everywhere following an edict of the local prince, but their existence proved short-lived, since they tended to be destroyed by a boisterous rabble in most places. This was most often the case in areas where the country folk live scattered about on farms, and when going on business to the next city or village more often than not come home inebriated and, knowing the way as they do, consider signposts unnecessary.”

The Parisians seem to have greeted the introduction of street lighting in 1667 under Louis XIV with a similar lack of enthusiasm. Dietmar Kammerer conjectured in the Süddeutsche Zeitung that the regular destruction of these street lamps represented a protest on the part of the citizens against the loss of their private sphere, since it seemed clear to them that here was “a measure introduced by the king to bring the streets under his control”. A simpler explanation would be that citizens tend in the main to react aggressively to unsupervised innovations in their midst. Recently, Deutsche Bahn explained that the initial vandalism of their “bikes for hire” had died down, now that locals had “grown accustomed to the sight of the bicycles”.

When it turns out that the novelty is not as useless as initially assumed, there follows the brief interregnum of Argument No.2: “Who wants it anyway?” “That’s an amazing invention,” gushed US President Rutherford B. Hayes of the telephone, “but who would ever want to use one of them?” And the film studio boss Harry M. Warner is quoted as asking in 1927, “Who the hell wants to hear actors talk?”.

[div class=attrib]More from theSource here.[end-div]

MondayPoem: The Chimney Sweeper

[div class=attrib]By Robert Pinsky for Slate:[end-div]

Here is a pair of poems more familiar than many I’ve presented here in the monthly “Classic Poem” feature—familiar, maybe, yet with an unsettling quality that seems inexhaustible. As in much of William Blake’s writing, what I may think I know, he manages to make me wonder if I really do know.

“Blake’s poetry has the unpleasantness of great poetry,” says T.S. Eliot (who has a way of parodying himself even while making wise observations). The truth in Eliot’s remark, for me, has to do not simply with Blake’s indictment of conventional churches, governments, artists but with his general, metaphysical defiance toward customary ways of understanding the universe.

The “unpleasantness of great poetry,” as exemplified by Blake, is rooted in a seductively beautiful process of unbalancing and disrupting. Great poetry gives us elaborately attractive constructions of architecture or music or landscape—while preventing us from settling comfortably into this new and engaging structure, cadence, or terrain. In his Songs of Innocence and Experience, Shewing the Two Contrary States of the Human Soul, Blake achieves a binary, deceptively simple version of that splendid “unpleasantness.”

In particular, the two poems both titled “The Chimney Sweeper” offer eloquent examples of Blake’s unsettling art. (One “Chimney Sweeper” poem comes from the Songs of Innocence; the other, from the Songs of Experience.) I can think to myself that the poem in Songs of Innocence is more powerful than the one in Songs of Experience, because the Innocence characters—both the “I” who speaks and “little Tom Dacre”—provide, in their heartbreaking extremes of acceptance, the more devastating indictment of social and economic arrangements that sell and buy children, sending them to do crippling, fatal labor.

By that light, the Experience poem entitled “The Chimney Sweeper,” explicit and accusatory, can seem a lesser work of art. The Innocence poem is implicit and ironic. Its delusional or deceptive Angel with a bright key exposes religion as exploiting the credulous children, rather than protecting them or rescuing them. The profoundly, utterly “innocent” speaker provides a subversive drama.

But that judgment is unsettled by second thoughts: Does the irony of the Innocence poem affect me all the more—does it penetrate without seeming heavy?—precisely because I am aware of the Experience poem? Do the explicit lines “They clothed me in the clothes of death,/ And taught me to sing the notes of woe” re-enforce the Innocence poem’s meanings—while pointedly differing from, maybe even criticizing, that counterpart-poem’s ironic method? And doesn’t that, too, bring another, significant note of dramatic outrage?

Or, to put it the question more in terms of subject matter, both poems dramatize the way religion, government, and custom collaborate in social arrangements that impose cruel treatment on some people while enhancing the lives of others (for example, by cleaning their chimneys). Does the naked, declarative quality of the Experience poem sharpen my understanding of the Innocence poem? Does the pairing hold back or forbid my understanding’s tendency to become self-congratulatory or pleasantly resolved? It is in the nature of William Blake’s genius to make such questions not just literary but moral.

“The Chimney Sweeper,” from Songs of Innocence

When my mother died I was very young,
And my father sold me while yet my tongue
Could scarcely cry ” ‘weep! ‘weep! ‘weep! ‘weep!’ ”
So your chimneys I sweep & in soot I sleep.

There’s little Tom Dacre, who cried when his head
That curled like a lamb’s back, was shaved: so I said,
“Hush, Tom! never mind it, for when your head’s bare
You know that the soot cannot spoil your white hair.”

And so he was quiet, & that very night,
As Tom was a-sleeping he had such a sight!
That thousands of sweepers, Dick, Joe, Ned & Jack,
Were all of them locked up in coffins of black.

And by came an Angel who had a bright key,
And he opened the coffins & set them all free;
Then down a green plain, leaping, laughing, they run,
And wash in a river and shine in the Sun.

Then naked & white, all their bags left behind,
They rise upon clouds and sport in the wind.
And the Angel told Tom, if he’d be a good boy,
He’d have God for his father & never want joy.

And so Tom awoke; and we rose in the dark,
And got with our bags & our brushes to work.
Though the morning was cold, Tom was happy & warm;
So if all do their duty, they need not fear harm.

—William Blake

[div class=attrib]More from theSource here.[end-div]

Google’s Earth

[div class=attrib]From The New York Times:[end-div]

“I ACTUALLY think most people don’t want Google to answer their questions,” said the search giant’s chief executive, Eric Schmidt, in a recent and controversial interview. “They want Google to tell them what they should be doing next.” Do we really desire Google to tell us what we should be doing next? I believe that we do, though with some rather complicated qualifiers.

Science fiction never imagined Google, but it certainly imagined computers that would advise us what to do. HAL 9000, in “2001: A Space Odyssey,” will forever come to mind, his advice, we assume, eminently reliable — before his malfunction. But HAL was a discrete entity, a genie in a bottle, something we imagined owning or being assigned. Google is a distributed entity, a two-way membrane, a game-changing tool on the order of the equally handy flint hand ax, with which we chop our way through the very densest thickets of information. Google is all of those things, and a very large and powerful corporation to boot.

We have yet to take Google’s measure. We’ve seen nothing like it before, and we already perceive much of our world through it. We would all very much like to be sagely and reliably advised by our own private genie; we would like the genie to make the world more transparent, more easily navigable. Google does that for us: it makes everything in the world accessible to everyone, and everyone accessible to the world. But we see everyone looking in, and blame Google.

Google is not ours. Which feels confusing, because we are its unpaid content-providers, in one way or another. We generate product for Google, our every search a minuscule contribution. Google is made of us, a sort of coral reef of human minds and their products. And still we balk at Mr. Schmidt’s claim that we want Google to tell us what to do next. Is he saying that when we search for dinner recommendations, Google might recommend a movie instead? If our genie recommended the movie, I imagine we’d go, intrigued. If Google did that, I imagine, we’d bridle, then begin our next search.

We never imagined that artificial intelligence would be like this. We imagined discrete entities. Genies. We also seldom imagined (in spite of ample evidence) that emergent technologies would leave legislation in the dust, yet they do. In a world characterized by technologically driven change, we necessarily legislate after the fact, perpetually scrambling to catch up, while the core architectures of the future, increasingly, are erected by entities like Google.

William Gibson is the author of the forthcoming novel “Zero History.”

[div class=attrib]More from theSource here.[end-div]