Teaching a Machine to Love  XOR

xorsketch

The XOR function outputs true if one of the two inputs are true

The exclusive or function, also known as XOR (but never going by both names simultaneously), has a special relationship to artificial intelligence in general, and neural networks in particular. This is thanks to a prominent book from 1969 by Marvin Minsky and Seymour Papert entitled “Perceptrons: an Introduction to Computational Geometry.” Depending on who you ask, this text was single-handedly responsible for the AI winter due to its critiques of the state of the art neural network of the time. In an alternative view, few people ever actually read the book but everyone heard about it, and the tendency was to generalize a special-case limitation of local and single-layer perceptrons to the point where interest and funding for neural networks evaporated. In any case, thanks to back-propagation, neural networks are now in widespread use and we can easily train a three-layer neural network to replicate the XOR function.

In words, the XOR function is true for two inputs if one of them, but not both, is true. When you plot the XOR as a graph, it becomes obvious why the early perceptron would have trouble getting it right more than half the time.

sketch2dxor

There’s not a way to draw a straight 2D line on the graph and separate the true and false outputs for XOR, red and green in the sketch above. Go ahead and try. The same is going to be true trying to use a plane to separate a 3D version and so on to higher dimensions.

sketch3dxor

That’s a problem because a single layer perceptron can only classify points linearly. But if we allow ourselves a curved boundary, we can separate the true and false outputs easily, which is exactly what we get by adding a hidden layer to a neural network.

xorwhiddenlayer

The truth-table for XOR is as follows:

Input Output
00 0
01 1
10 1
11 0

If we want to train a neural network to replicate the table above, we use backpropagation to flow the output errors backward through the network based on the neuron activations of each node. Based on the gradient of these activations and the error in the layer immediately above, the network error can be optimized by something like gradient descent. As a result, our network can now be taught to represent a non-linear function. For a network with two inputs, three hidden units, and one output the training might go something like this:

trainingxor

Update (2017/03/02) Here’s the gist for making the gif above:

Advertisements

Things to Think About From 2016

thescinder2016wordle

A word cloud of theScinder’s output for 2016, made with wordle.net

CRISPR/Cas9

This subject includes throwbacks to 2015, when I did most of my writing about CRISPR/Cas9. That’s not to say 2016 didn’t contain any major genetic engineering news. In particular scientists are continue to move ahead with the genetic modification of human embryos.

If you feel like I did before I engaged in some deeper background reading, you can catch up with my notes on the basics. I used the protein structures for existing gene-editing techniques to highlight the differences between the old-school gene editing techniques and editing with cas9. I also compared the effort it takes to modify a genome with cas9 to how difficult it was using zinc-finger nucleases, the previous state-of-the-art (spoiler: it amounts to days of difference).

TLDR: The advantage of genetic engineering with Cas9 over previous methods is the difference between writing out a sequence of letters and solving complex molecular binding problems.

aLIGO and the detection of gravitational waves

Among the most impressive scientific breakthroughs of the previous hundred years or so, a bunch of clever people with very sensitive machines announced they’ve detected the squidge-squodging of space. A lot of the LIGO data is available from the LIGO Open Science Center, and this is a great way to learn signal processing techniques in Python. I synchronized the sound of gravitational wave chirp GW150914 to a simulated visualization (from SXS) of a corresponding black hole inspiral and the result is the following video. You can read my notes about the process here. I also modified the chirp to play the first few notes of the “Super Mario Brothers” theme.

Machine Learning

I’ve just started an intensive study of the subject, but machine learning continues to dip its toes into everything to do with modern human life. We have a lot of experience with meat-based learning programs, which should give us some insight into how to avoid common pitfalls. The related renewed interest in artificial intelligence should make the next few years interesting. If we do end up with a “hard” general artificial intelligence sometime soon, it might make competition a bit tough, if you could call it competition at all.

Devote a few seconds of thought to the twin issues of privacy and data ownership.

Mars

2016 also marked a renewed interest in manned space exploration, largely because of the announcement from space enthusiast Elon Musk that he’s really stoked to send a few people to Mars. NASA is still interested in Mars as well, and might be a good partner to temper Musk’s enthusiasm. In the Q&A at about 1:21 in the video below, Musk seems to suggest a willingness to die as the primary prerequisite for his first batch of settlers. There’s some known unavoidable and unknown unknowable dangers in the venture, but de-prioritizing survivability as a mission constraint runs a better chance of delaying manned exploration as long as it remains as expensive as Musk optimistically expects.

Here’s some stuff that’s a little a lot less serious about living on Mars.

It doesn’t grab the headlines with such vigor, but Jeff Bezo’s Blue Origins had an impressive year: retiring their first rocket after five flights and exceeding the mission design in a final test of a launch escape system.
Blue Origin is also working on an orbital launch system called New Glenn, in honor of the first astronaut from the USA to orbit the earth.

In that case, where are we headed?

The previous year provided some exciting moments to really trip the synapses, but we had some worrying turns as well. The biggest challenges of the next few decades will all have technical components, and understanding them doesn’t come for free. Humanity is learning more about biology at more fundamental levels, and medicine won’t look the same in ten years. A lot of people seem unconcerned that we probably won’t make the 2 degrees Celsius threshold for limiting climate change, although not worrying about something doesn’t mean it won’t kill anyone. Scientists and engineers have been clever enough to develop machine learners to assist our curiosity, and it’s exciting to think that resurgent interest in AI might give us someone to talk to soon. Hopefully they’ll be better conversationalists than the currently available chatbots, and a second opinion on the nature of the universe could be useful. It’s not going to be easy to keep up with improving automation, and humans will have to think about what working means to them.

Take some time to really dig into these subjects. You probably already have some thoughts and opinions on some of them, so try to read a contrary take. If you can’t think of evidence that might change your mind, you don’t deserve your conclusions.

Remember that science, technological development, and innovation have a much larger long-term effect on humans and our place in the universe than the petty machinations of human fractionation. So keep learning, figure out something new, and remember that if you possess general intelligence you can approach any subject. On the other hand, autogenous annihilation is one of the most plausible solutions to the Fermi Paradox. This is no time to get Kehoed

Introduction to ML

mlbrainmap

You can’t swing a cat through a newsfeed these days without hitting a story about the extent of learning algorithms in everyday life. As many readers are no doubt aware, there’s a type of computation that’s brought new perspective to a wide swath of common and previously intractable problems, and unlike a more deterministic programmatic approach, this new approach to is based on learning. That’s right, instead of a linear sequence of instructions these algorithms figure out what to do for themselves.

There’s probably not a single aspect of our modern world that isn’t affected by or directly based on the new field of learning-based methods. But this isn’t just any type of learning: Meat Learning, or ML for short, affects almost every essential experience in the modern world. Choosing what book to read next? Meat learning. What’s for lunch? ML. Driving cars? ML conquered that task years ago (despite a high-frequency of catastrophic failure). Whether you’ve been aware of it or not, and for better or for worse, we’ve truly entered the age of the learning algorithm.

Many of the basic ideas and underlying structures of modern ML have been around for millennia, if not epochs, but it’s only been in the last few million years that ML has really puffed up its breast-feathers. Almost all ML algorithms in common usage today run on a variant of the same hardware: a fatty grey mass known to the professionals that work with them as a “meat computer.”

Rather than an improvement in the basic design or clever new architectures, the enabling breakthrough for modern ML has been largely based on the sheer mass of computational meat (archaically known as grey matter) incorporated in bleeding edge modern platforms. This new type of learner, of which a wide swath of variants all share the same designation of “human,” hasn’t collectively been around for long (geologically speaking), but has already had a huge impact. It is nearly impossible to go anywhere on the planet without running into some sign of human meat-learners’ work, play, or self-destructive irony.

Previous generations of meat-learners had about half the mass of computational meat allocated for each kg of body mass compared to the current state-of-the-art human models, at least after engaging regularization factors to make the human variants feel better about themselves. This vast increase in computational mass in a meat computer is almost entirely comprised of warm, squishy subunits intended to mimic the activity of artificial neural networks. Like an iceberg, the majority of these networks are hidden from view, which makes them very mysterious (some would say nonsensical) to work with. More on that later on.

hiddenlayers

The official definition of a meat learning platform, according to ML pioneer Sammy Arthur is “…something that can learn, at least a little bit, but it has to be made of meat. The soft squishy stuff.” Some practitioners would say that the modern definition has grown since Arthur’s definitive statements. Although ML was at first considered a subset of the field of organic intelligence (OI for short), it is now widely acknowledged that ML has outgrown its humble beginnings as a specialization of OI, becoming a fully-fledged field in its own right. Also, it is quite clear that many of the modern systems running the cutting-edge human variants don’t possess all the attributes for qualification as bona fide general organic intelligence.

Unlike traditional programming, wherein the outcome depends on a succession of conditional statements, meat learners adapt to a given problem by being “trained” on large sets of data, aka, learning. To train a meat learning platform on solving a problem requires a large amount of input data, and in some cases, acquiring the right data can be quite the challenge. After learning on a sufficient number of training sets, a ML program becomes capable of independently completing a wide variety of tasks, including both stuff and things.

This unique learning approach means that meat learners often find solutions to problems that are unusual and sometimes downright surprising (or stupid). Due to the nature of how ML systems are trained, completing a task may include steps that appear to make very little sense until, finally, the program somehow converges to an interesting solution. Sometimes. Despite numerous advantages, these capabilities come at a cost.

Training an ML algorithm requires a vast input of training data, often taking twenty years or so before an ML system is ready for even the simplest of tasks. Due to the speed issue, most developers train an ML platform for multiple tasks in parallel, which seems to work out more or less OK. One potential disadvantage is that in most cases the vast majority of the squishy neural network performing ML computations is made up of hidden layers, the submerged part of the iceberg alluded to above. As a consequence, nobody is quite sure how they do what they do. In practice this isn’t much of a problem when a given ML is working, but it makes problems almost impossible to debug. Unlike a proper artificial neural network, not even the architecture of the system is available for design and adjustment. The only way to adjust a given ML system is by modifying the training inputs, often with unpredictable results. To make matters worse, in most cases learners heavily weight early training inputs, assigning little to no weight to those encountered as a more mature ML system. In all cases, once a particular ML circuit becomes entrenched in a wider ML architecture, it becomes subject to a “learned inertia” that makes it difficult for a learner to adapt previously learned strategies and can lead to typing in all caps.

Over-reliance on hidden layers and learned inertia aren’t the only problems associated with ML. Although meat-learners tend to be quite flexible, there are certain types of computation that ML just isn’t suited for. If your task involves Bayesian inference, for instance, you can just about forget about training an ML human platform to suit. Additionally, ML performance can be unacceptably erratic. A meat learner that previously performed reasonably well on a given task might perform terribly the next time around just because they are tired, drunk, or bored, and they usually are at least one of those three things.

ML has long been plagued by hardware problems as well. ML platforms tend to run at speeds of just a few tens of Hertz, and the upkeep of such a system can be expensive. Both dynamic and static memory suffer from volatility and infidelity, and these problems only get worse as the system matures. ML is inherently crufty: many aspects of the hardware have been retained from the very early days of the first chordate models, in turn borrowing much of their fundamentals from even earlier systems. The basis for switching action potentials, the discrete building block of meat learning activity, in microalgae is so similar to that in state-of-the-art meat learners that they can be functionally swapped. This highlights the archaic mechanics at the foundation of meat learning, but sheds some insight by making meat neurons (even in hidden layers) amenable to switching on or off with light exposure.

Despite myriad problems, ML seems to be here to stay and is liable to play a large role for at least another ten years or so. ML is slow, but gets around the problem by utilizing massive parallelization. ML takes years to train, but once it has learned a fair amount from training inputs it can perform a wide variety of tasks. ML is erratic and highly prone to various emotional and drive-reduction states, but most ML systems get around this issue by pretending everything is fine. Hidden layers make it difficult to determine the inner workings of a modern ML platform, but many agree that’s for the best anyway.

Whether ML developers overcome the existing issues or not may ultimately be irrelevant: for the nonce, meat learning in its many forms is the best and only irrevocable tool we’ve got.

Update 2016/12/23: Comment on action potentials added to paragraph 11

­A Skeptic Over Coffee – Young Blood

dsc_0005

A tragic tale of a star-crossed pair,
science vs. a journalist’s flare

When reporting on scientific topics, particularly when describing individual papers, how important is it for the popular coverage to have anything to do with the source material? Let’s take a look at a recent science paper from Justin Rebo and others in Nature Communications and the accompanying coverage by Claire Maldarelli at Popular Science

Interest in parabiosis has increased recently due to coverage of scientific papers describing promising results in mice and the high-profile of some parabiosis enthusiasts. Parabiosis, from the Latin for “living beside”, typically has involved stitching two mice together. After a few days the fused tissue provides blood exchange through a network of newly formed capillaries.

The most recent investigation into the healing effects of youthful blood exchange from Rebo et al. expands the equipment list used for blood exchange beyond the old technique of duct-taping two animals together surgically joining two animals. Instead of relying on the animals to grow new capillary beds for blood exchange to occur, the authors of the new paper used a small pump to exchange a few drops of blood at a time until both mice had approximately the same proportion of their own blood and that of a donor and vice-versa.

According to the coverage from Popular Science:

While infusing blood from a younger mouse into an older mouse had no effect on the elderly mouse in the latest study, infusing blood from an older mouse into a younger one caused a host of problems in organs and other tissues.

Just a few paragraphs further Maldarelli quotes Conboy (last author on the paper) as saying “‘This study tells us that young blood, by itself, cannot work as medicine’.” In contrast, in the paper the authors state that “Importantly, our work on rodent blood exchange establishes that blood age has virtually immediate effects on regeneration of all three germ layer derivatives.” and later that “. . . extracorporeal blood manipulation provides a modality of rapid translation to human clinical intervention.”[1] There seems to be a bit of disagreement between the version of Conboy on the author list of the scientific article and the version of Conboy quoted in the PopSci coverage of the same article.

We also learned from Maldarelli that the tests reported in the paper were performed a month after completing the blood exchange procedure, but the longest duration from blood exchange to the experiment’s end (sacrifice for post-mortem tissue analysis) was 6 days after blood exchange.

I came across the PopSci coverage when it appeared on a meta-news site that highlights popular web articles, so it’s safe to assume I wasn’t the first to read it. Shouldn’t the coverage of scientific articles reported in the lay press have more in common with the source material than just buzzwords? The science wasn’t strictly cut and dried: not every marker or metric responded in the same way to the old/young blood exchange, and while I agree that we shouldn’t be encouraging anyone to build a blood-exchange rejuvenation pod in their garage, the findings of the article fell a long way from the conclusions reported in the lay-article: that young blood had no effect on the physiology of old mice. This is to say nothing about the quality of the paper itself and the confidence we should assign to the experimental results in the first place: with 12 mice total* and a p-value cutoff of 0.05 (1 out of every 20 experiments will appear significant at random), I’d take the original results with a grain of salt as well.

This is the face of science we show the public, and it’s unreliable. It is no easy task for journalists to accurately report and interpret scientific research. Deadlines are tight, and writers face competition and pressure from cheap amateur blogs and regurgitation feeds. “What can I do to help?” you ask. As a consumer of information you can demand scientific literacy in the science news you consume. Ask for writers to convey confidence and probability in a consisent way that can be understood and compared to other results by non-specialists. As a bare minimum, science and the press that covers it should at least have more in common than the latest brand of esoteric jargon.

If we only pay attention to the most outlandish scientific results, then most scientific results will be outlandish.

*The methods describe a purchase of 6 old and 6 young mice. However, elsewhere in the paper the groups are said to contain 8 mice each. Thus it is not clear how many mice in total were used in these experiments, and how they managed to create 12 blood exchange pairings for both control and experimental groups without re-using the same mice.

[1] Rebo, J. et al. A single heterochronic blood exchange reveals rapid inhibition of multiple tissues by old blood. Nat. Commun. 7, 13363 doi: 10.1038/ncomms13363 (2016).

A month on Mars

The year is 2035, and the new space race is well underway.

Jeffrey Aussat straightened his back under the Martian sun. He stretched as he leaned onto the handle of his space-shovel, raising his hand to wipe the sweat from his tired brow. Of course this made him feel stupid, as it had every time since they landed. His clumsy hand, gloved up and looking for all the world just like the hand of Gozer the Destructor, stops short as it meets the clear glass of his visor. Jeff curses himself at the unavoidable fact that, despite nearly a (Mars) month since they arrived on the spaceship Clever Reference, he still couldn’t get used to the simplest things. Like the need to have this damn fish-bowl on every time he goes outside.

Jeff curses himself again as his shovel snaps in half. Losing focus during retrospection and self-pity, he somehow must have applied an off-axis load onto the carbon fiber handle. A few moments respite for his weary, microgravity-weakened bones, had turned into disaster. On Mars, the gravity may be slight but the days sure are long, but they don’t tell you that in the brochure.

Jeff now found himself up a recurring slope lineae without a planetary-protection cleared drill bit. Jeff and his partner had started out their ‘stead with 32 shovels, and in just a few weeks every single on had fallen prey to some combination of user error and catastrophic failure. Every building in their inflatable homestead creation kit was designed to be placed underground, damping temperature swings and blocking some of the deadly radiation pouring down on Mars surface. Specifically, the buildings needed to have a huge amount of ground piled on top of them to keep the humans alive, and without a working shovel they couldn’t move regolith quickly enough to make their new home habitable. Due to some shady logistics, they wouldn’t receive their “mule”- a heavy lifting robot- until the next colonization flotilla arrived, roughly two years on.

Jeff holds the transmit button on his radio as he slumps down in the shade of his space-wheelbarrow, half-piled high with regolith and also made from carbon fiber. “Becky, I think we have a problem,” he said.

After a short intermission of static, Becky replied with a sigh, “You’ve got a leak in your suit again, don’t you?” Getting used to the strange Martian gravity after playing zero-G ping pong for three months, Jeff had often ended up tumbling down to hands and knees during the first weeks of their stay, a stress the suits were well-designed to withstand. Repeated joint flexion of the suit fabric with embedded Martian dust, however, rapidly opened up a community of near-microscopic pinholes that were almost impossible to find and patch.

“No, not this time. It’s the shovel.”

“The last shovel?”

Jeff paused. “… Yeah.” This was bad. They would have to resort to much less efficient regolith maneuvering techniques, working only at night and sleeping under the raw materials in the shed to limit radiation exposure. After the recurring problem with clumsiness-induced suit leaks, Becky’s patience was sure to be running out on him. The trip over had already placed enough stress on their relationship. “Is the 3D printer working yet? Maybe we can print a new one, or print a repair splint for one of the frayed shovel shafts.”

Silence followed for nearly a minute. She was either checking the printer status or seriously considering filing flight plans to leave. “I’m afraid the printer’s still down. The print nozzle was damaged during the last maintenance test.”

“Oh.” Jeff replied. He didn’t finish converting the thought running through his head to speech: so we’re screwed then.

“No problem. I’ll order a fresh crate from Amazon.”

“What?” This was either a joke, a hoax, or lifesaving news.

“Check your email. They’ve opened up a new distribution center on Phobos. Bezos built it up and staffed it without telling anybody.”

“You’ve got to be kidding me.”

“No joke. I need a few extra items to qualify for free shipping, do you need anything?”

“I’m sure we can think of something. I’ll return to the compound with the regolith I’ve collected and we can run an inventory.” Jeff tossed the broken shovel on top of the regolith in the enormous wheelbarrow. The designers had figured that, if everything on Mars would weigh so much less than on Earth, all the tools should be designed to be that much larger. The result was a suite of construction and farming tools that were cartoonishly two and a half times too large when fully assembled. As Jeff wheeled the barrow around to face the glint from the compound’s solar panels, he felt his mood pick up. They were going to be OK after all.

“There’s something else going on that’s a bit weird.” Becky said.

Jeff skipped a step, catching himself on the wheelbarrow handles to prevent impregnating the knees on his suit with more abrasive dust. “What is it?” he asked.

“You remember that huge rover from 2020?”

Jeff made a vague confirmatory noise “Uh . . . the Scrutiny, was it?”

“Yeah, that’s the one. It’s attacking the water scavenging plant.”

“What? Why? I thought that thing was supposed to be retired by now, parked somewhere near Jezero delta?”

“Well it’s here, and it’s pushing the water plant over. The LEDs are putting out some sort of morse code, I’m still trying to figure it out.” Becky explained.

“How long until it damages the water plant?” Jeff inquired.

“At this rate, probably a couple of weeks. They didn’t move very fast back then.”

Jeff felt the spring return to his step. Two weeks was enough time to contact the mission controllers to get some help debugging the rovers strange behavior. As he realized the problem was tractable, the physical sensation of a weight lifted from his shoulders. Also, the motility assist systems on his suit had finally finished calibrating.

“Too bad they didn’t set up the distro center in time for Mars One.” Jeff joked

“Too soon, Jeff, that’s not funny.” Becky said coldly.

The Mars One mission had ended in a tragicomic maelstrom of cannibalism and incidental lyophilization. The cameras, intended to live-broadcast the travails of the crew around the clock, were among the last systems still running on the capsule. Although the sponsors had long disavowed any relationship to the mission, anyone with a standard transceiver and a darkly morbid curiosity could ping the ship and tune in to the dismal situation. A series of planned challenges/mission planning fiascos ultimately meant they never got onto the correct Mars rendezvous trajectory. In their current orbit, apoapsis would never quite reach Mars orbit, nor would periapsis ever bring them close enough for an earthly recapture. Ironically, what remained of the crew and craft would probably outlast them all. The perfectly preserved astronauts would remained unchanged for millennia in their wayward but stable orbit, like confused Pharaohs circling the portal to the netherworld.

A skeptic over coffee: sick of lab meetings

rhinovirus

This post brought to you by a dedicated community of human Rhinovirus ( pdb model 1AYM).

Imagine the following dialogue between researchers:

Wayne the Brain: “Third one this week ::Cough:: I am literally sick of lab meetings.”
Wankdorf: “Oh I feel ya. There are way too many lab meetings. It’s a real waste of time, but that’s the cost of pulling from so many different realms of expertise in interdisciplinary projects.”
Wayne the Brain: “No no no, I am literally sick of lab meetings. All the exposure is really taking a toll on my health. ”
Wankdorf: “Why didn’t you say so?! Stay away, you purveyor of vile pestilence! ::cough::”

I hope, dear reader, that you spotted the root cause of their misunderstanding. Wayne (the Brain) was hypothesizing a suspected transmission rate while simultaneously advertising his own condition as definitely infected and possibly contagious. Wankdorf (unsurprisingly) misinterprets the statement by applying a more colloquial definition of the term “literally.” It’s not clear whether infection of the second researcher could have been avoided and the spread of the disease slowed had they practised more effective communication, but that scenario is plausible given what we know.

Of course this is an extreme example, and the consequences may not always be so dire. The most frustrating part of the above exchange and subsequent misunderstanding is that neither participant was strictly wrong in the definition they assumed for “literally.” This word now literally can be used to say “in the truest sense of the words” and the exact opposite, and my brain literally imploded when I learned about the new definition.

If you don’t believe me, check out the definition in both the Cambridge and Merriam-Webster online dictionaries. I’ve screenshotted the definitions to preserve this embarrassment for posterity:

merriamwebsterliterally

cambridgeliterally

Language is dynamic, some (Wankdorf etc.,) would even say that it is dynamical. Hence it doesn’t make you appear smarter to bore your friends by talking about Romans every time they say “decimate.” Language is constantly changing in response to the selective pressures of popular usage, subject to many factors as people and cultures interact.

Similar to many other examples of evolution, humans affect the way a language changes by taking note of and modifying the selective pressures they individually exert. The consequences may be particularly important in science, where English is the common tongue but not in general the first language of most practitioners. I expect that modern English will evolve to encompass multiple forms based on usage. Native speakers sat on the British Isles, laying in North America, and so on will continue to retain and invent complexity and idiosyncrasy, while international English will come to resemble a utilitarian version of Up-Goer Five English, paring off superfluous complexities while retaining the most effective elements to become as simple as possible, but no simpler. It’s possible that international English will even retain sarcasm.

Pop quiz: what’s your favourite English speaker idiosyncrasies used in this article?

A skeptic over coffee: who owns you your data?

AskDNA

“Everyone Belongs to Everyone Else”

-mnemomic marketing from Aldous Huxley’s Brave New World

A collaboration between mail-order genomics company 23andMe and pharmaceutical giant Pfizer reported 15 novel genes linked to depression in a genome-wide association study published in Nature. The substantial 23andMe user base and relative prevalence of the mental illness provided the numbers necessary to find correlations between a collection of single nucleotide polymorphisms (SNPs) and the condition.

This is a gentle reminder that even when the service isn’t free, you very well may be the product. It’s not just Google and Facebook whose business plans hinge on user data. From 23andMe’s massive database of user genetic information to Tesla’s fleet learning Autopilot (and many more subtle examples that don’t make headlines), you’re bound to be the input to a machine learning algorithm somewhere.

On the one hand, it’s nice to feel secure in a little privacy now and again. On the other, blissful technological utopia? If only the tradeoffs were so clear. Note that some (including bearded mo. bio. maestro George Church) say that privacy is a thing of the past, and that openness is the key (the 23andMe study participants consented that their data be used for research). We’ve known for a while that it’s possible to infer the sources of anonymous genome data from publicly available metadata.

The data of the every person are fueling the biggest changes of our time in transportation, technology, healthcare and commerce, and there’s a buck (or a trillion) to be made there. It remains to be seen if the benefits will mainly be consolidated by those who already control large pieces of the pie or to fall largely to the multitudes making up the crust (with plenty of opportunities for crumb-snatchers). On the bright side, if your data make up a large enough portion of machine learning inputs for the programs that eventually coalesce into an omnipotent AI, maybe there’ll be a bit of you in the next generation superorganism.