Innovation as a Dialectic

These days, innovation is not an opportunity but a mandate. By innovate, I mean apply technology to do old things differently (faster, cheaper, more efficiently) or do new things that were not previously possible. There are many arguments one could put forth to critique our unshaken faith in progress and growth. Let’s table these critiques and take it as given that innovation is a good thing. Let’s also restrict our scope to enterprise innovation rather than broad consumer or societal innovation.

I’ve probably seen over 100 different organizations’ approaches to “digital” innovation over the past year in my role leading business development for Fast Forward Labs. While there are general similarities, often influenced by the current popularity of lean startup methodology or design thinking, no two organizations approach innovation identically. Differences arise from the simple fact that organizations are made of people; people differ; individual motives, thoughts, and actions combine together in complex ways to create emergent behavior at the group level; amazingly interesting and complex things result (it’s a miracle that a group of even 50 people can work together as a unit to generate value that greatly exceeds that aggregate value from each individual contributor); past generations of people in organizations pass down behavior and habits to future generations through a mysterious process called culture; developments in technology (amidst other things) occur outside of the system* that is the organization, and then the organization does what it can to tweak the system to accept (or reject) these external developments; some people feel threatened and scared, others are excited and stimulated; the process never ends, and technology changes faster than people’s ability to adopt it…

Observing all these environments, and observing them with the keenness and laser-focused attention that only arises when one wants to influence their behavior, when one must empathize deeply enough to become the person one is observing, to adopt their perspective nearly completely - their ambitions, their fears, their surprises, their frustrations - so as to be able to then convince them that spending money on our company’s services is a sound thing to do (but it’s not only mercenary: people are ends in themselves, not a means to an end. Even if I fail to sell, I am always motivated and energized by the opportunity to get to know yet another individual’s mind and heart), I’ve come to accept a few axioms:

  1. Innovation is hard. Way harder than all the meaningless marketing makes it seem.
  2. There is no one way to innovate. The approach depends on organizational culture, structure, history, product, and context.
  3. Inventions are not solutions. Just because something is technically possible doesn’t mean people will appreciate its utility clearly enough to want to change their habits.
  4. Most people adopt new technologies through imitation, not imagination, often following the normal distribution of Geoffrey Moore’s technology adoption lifecycle.
  5. We always overestimate the change that will occur in the next two years and underestimate the change that will occur in the next ten. (Bill Gates)

Now, research in artificial intelligence is currently moving at a fast clip. At least once a month some amazing theoretical breakthrough (like systems beating Smash Brothers or Texas Hold’em Poker champions) is made by a university computer science department or the well-funded corporate research labs that have virtually replaced academia (this is worrisome for inequality of income and opportunity, and worthy of further discussion in a future post). Executives at organizations that aren’t Google, Facebook, Amazon, Apple, and Microsoft see the news and ask their tech leadership whether any of this is worth paying attention to, if it’s just a passing fad for geeks or if it’s poised to change the economy as we know it. If they do take AI seriously, the next step is to figure out how to go about applying it in their businesses. That’s where things get interesting.

There are two opposite ways to innovate with data and algorithms. The first is a top down approach that starts with new technological capabilities and looks for ways to apply them. The second is a bottom up approach that starts with an existing business problem or pain point and looks for possible technical solutions to the problem.

Top down approaches to innovation are incredibly exciting. They require unbridled creativity, imagination, exploration, coupled with the patience and diligence to go from idea to realized experiment, complete with whatever proof is required to convince others that a theoretical milestone has actually been achieved. Take computer vision as an example. Just a few years ago, the ability of computers to automatically recognize the objects in images - to be shown a picture without any metadata and say, That’s a cat! That’s a dog! That’s a girl riding a bicycle in the park on a summer day! (way more complicated technically) - was mediocre at best. Researchers achieved ok performance using classic classification algorithms, but nothing to write home about. But then, with a phase shift worthy of a Kuhnian scientific revolution, the entire community standardized on a different algorithmic technique that had been shunned by most of the machine learning community for some time. These artificial neural networks, whose underlying architecture is loosely inspired by the synapses that connect neurons in our brain’s visual cortex, did an incredible job transforming image pixels into series of numbers called vectors that could then be twisted, turned, and processed until they reliably matched with linguistic category labels. The results captivated popular press attention, especially since they were backed by large marketing machines at companies like Google and Facebook (would AI be as popular today if research labs were still tucked away in academic ivory towers?). And since then we at Fast Forward Labs have helped companies build systems that curate video content for creatives at publishing houses and identify critical moments in surgical processes to improve performance of remote operators of robotic tools.

That said, many efforts to go from capability to application fail. First, most people struggle to draw analogies between one domain (recognizing cats on the internet) and another (classifying whether a phone attempting to connect to wifi is stationary or moving in a vehicle). The analogies exist at a level of abstraction most people don’t think at - you have to view the data as a vector and think about the properties of the vector to evaluate whether the algorithm would be a good tool for the job - and that seem so different from our standard perceptual habits. Second, it’s technically difficult to scale a mathematical model to actually work for a real-world problem, and many data scientists and researchers lack the software engineering skills required to build real products. Third, drawing on Geoff Moore’s adoption lifecycle, many people lack patience to work with early technologies that don’t immediate do what they’re supposed to do. Applying a new technology often requires finding a few compassionate early adopters who are willing to give feedback to improve a crappy tool. Picking the wrong early users can kill a project. Fourth, organizations that are risk averse like to wait until their peers have tested and tried a new thing before they go about disrupting day-to-day operations. As they say, no one gets fired for hiring McKinsey or IBM. And finally, people are busy. It’s hard to devote the time and attention needed to understand a new capability well enough to envision its potential applicability. Most people can barely keep up with their current workload, and deprioritizing possibility to keep up with yesterday’s tasks.

Which is why it can often be more effective to approach innovation by solving a real problem instead of inventing a problem for a solution. This bottom up approach is more closely aligned with design thinking. The focus is on the business: what people do, how they do it, and where technology may be able to help them do it better. The approach works best when led by a hybrid business-technical person whose job it is to figure out which problems to solve - based on predicted impact and relatively low technical difficulty so progress can be made quickly - and muster the right technology to solve them - by building something internally or buying a third-party product. People tend to have more emotional skin in the game with innovation driven by problem solving because they feel greater ownership: they know the work intimately, and may be motivated by the recognition of doing something better and faster. The risk, particularly when using technology to automate a current task, is that people will fear changing their habits (although we are amazingly adaptable to new tools) or, worse, that they will be replaced by a machine.

The core difficulty with innovating by solving problems is that the best solution to the most valuable business problem is almost always technically boring. Technical research teams want to explore and make cool stuff; they don’t want to apply their time and energy to building a model using math they learned in undergrad to a problem that makes their soul cringe. While it seems exciting to find applications for deep learning in finance, for example, linear regression models are likely the best technical solution for 85% of problems. They are easy to build and interpretable, making it easier for users to understand why a tool outputs the answers it does, easier for developers to identify and fix bugs, and easier for engineers to deploy the model on existing commodity hardware without having to invest in new servers with GPUs. The other risk of starting with business problems is that in can lead to mediocre solutions. Lacking awareness of what’s possible, teams may settle for what they know rather than what’s best. If you’ve never seen an automobile, you’ll count how many horses you need to make it across the country faster.

As the title suggests, therefore, the best approach is to treat innovation as a dialectic capable of balancing creative capabilities with pragmatic solutions. Innovation is a balancing act uniting contradictory forces: the promise of tomorrow against the pull of today, the whimsy of technologists against the stubbornness of end users, the curiosity of potential against the calculation of risk. These contradictions are the growing pains of innovation, and the organizations that win are those that embrace them as part of growing up.

*Even an internal R&D department may be considered outside the system of the organization, particularly if R&D acts as a separate and independent unit that is not integrated into day-to-day operations. What an R&D team develops still has to be integrated into an existing equilibrium.

 

Progress and Relative Definitions

Over the past year,  I’ve given numerous talks attempting to explain what artificial intelligence (AI) is to non-technical audiences. I’d love to start these talks with a solid, intuitive definition for AI, but have come to believe a good definition doesn’t exist. Back in September, I started one talk by providing a few definitions of intelligence (plain old, not artificial - a distinction which itself requires clarification) from people working in AI:

“Intelligence is the computational part of the ability to achieve goals in the world.” John McCarthy, a 20th-century computer scientist who helped found the field of AI

“Intelligence is the use of information to make decisions which save energy in the pursuit of a given task.” Neil Lawrence, a younger professor at the University of Sheffield

“Intelligence is the quality that enables an entity to function appropriately and with foresight in its environment.” Nils Nilsson, an emeritus professor from Stanford’s engineering department

I couldn’t help but accompany these definitions with Robert Musil’s maxim definition of stupidity (if not my favorite author, Musil is certainly up there in the top 10):

“Act as well as you can and as badly as you must, but in doing so remain aware of the margin of error of your actions!” Robert Musil, a 20th century Austrian novelist

There are other definitions for intelligence out there, but I intentionally selected these four because they all present intelligence as related to action, as related to using information wisely to do something in the world. Another potential definition of intelligence would be to make truthful statements about the world, the stuff of the predicate logic we use to say that an X is an X and a Y is a Y. Perhaps sorting manifold, complex inputs into different categories, the tasks of perception and the mathematical classifications that mimic perception, is a stepping stone to what eventually becomes using information to act.

At any rate, there are two things to note.

First, what I like about Musil’s definition, besides the wonderfully deep moral commentary of sometimes needing to act as badly as you must, is that he includes as part of his definition of stupidity (see intelligence) a call to remain aware of margins of error. There is no better training in uncertainty than working in artificial intelligence. Statistics-based AI systems (which means most contemporary systems) provide approximate best guesses, playing Marco Polo, as my friend Blaise Aguera y Arcas says, until they get close enough for government work; some systems output “maximum likely” answers, and others (like the probabilistic programming tools my colleagues at Fast Forward Labs just researched) output full probability distributions, with affiliated confidence rates for each point in the distribution, which we then have to interpret to gauge how much we should rely on the AI to inform our actions. I’ll save other thoughts about the strange unintuitive nature of thinking probabilistically another time (hopefully in a future post about Michael Lewis’s latest book, The Undoing Project.)

Second, these definitions of intelligence don’t help people understand AI. They may be a step above the buzzword junk that litters the internet (all the stuff about pattern recognition magic that will change your business that leads people outside the field to believe that all machine learning is unsupervised, whereas unsupervised learning is an active and early area of research), but they don’t leave audiences feeling like they’ve learned anything useful and meaningful. I’ve found it’s more effective to walk people through some simple linear or logistic regression models to give them an intuition of what the math actually looks like. They may not leave with minds blown away at the possibilities, but they do leave with the confident clarity of having learned something that makes sense.

As it feels like a fruitless task to actually define AI, I (and my colleague Hilary Mason, who used this first) instead like to start my talks with a teaser definition to get people thinking:

“AI is whatever we can do that computers can’t…yet.” Nancy Fulda, a science fiction writer, on Writing Excuses

This definition doesn’t do much to help audiences actually understand AI either. But it does help people understand why it might not make sense to define a given technology category - especially one advancing so quickly - in the first place. For indeed, an attempt to provide specific examples of the things AI systems can and cannot do would eventually - potentially even quickly - be outdated. AI, as such, lies within the horizons of near future possibility. Go too far ahead and you get science fiction. Go even further an you get madness or stupidity. Go too behind and you get plain old technology. Self-driving cars are currently an example of AI because we’re just about there. AlphaGo is an example of AI because it came quicker than we thought. Building a system that uses a statistical language model that’s not cutting edge may be AI for the user of the system but feel like plain old data science to the builder of the system, as for the user it’s on the verge of the possible, and for the builder it’s behind the curve of the possible. As Gideon Lewis-Kraus astutely observed in his very well written exposé on Google’s new translation technology, Google Maps would seem like magic to someone in the 1970s even though it feels commonplace to us today.

So what’s the point? Here’s a stab. It can be challenging to work and live in a period of instability, when things seem to be changing faster than definitions - and corollary social practices like policies and regulations - can keep up with. I personally like how it feels to be work in a vortex of messiness and uncertainty (despite my anxious disposition). I like it because it opens up the possibility for relativist, non-definitions to be more meaningful than predicate truths, the possibility to realize that the very technology I work on can best be defined within the relative horizons of expectation. And I think I like that because (and this is a somewhat tired maxim but hey, it still feels meaningful) it’s the stuff of being human. As Sartre said and as Heidegger said before him, we are beings for whom existence precedes essence. There is no model of me or you sitting up there in the Platonic realm of forms that gets realized as we live in the world. Our history is undefined, leading to all sorts of anxieties, worries, fears, pain, suffering, all of it, and all this suffering also leads the the attendant feelings of joy, excitement, wonder (scratch that, as I think wonder is aligned with perception), and then we look back on what we’ve lived and the essence we’ve become and it feels so rich because it’s us. Each of us Molly Bloom, able to say yes yes and think back on Algeciras, not necessarily because Leo is the catch of the century, but because it’s with him that we’ve spent the last 20 years of our lives.

The image is Paul Klee’s Angelus Novus, which Walter Benjamin described as “an angel looking as though he is about to move away from something he is fixedly contemplating,” an angel looking back at the chain of history piling ruin upon ruin as the storm of progress hurls him into the future.