Be Where Your Business Is

This post previously appeared on the readwrite blog.

 

A CEO running a B-to-B startup in needs to live in the city where their business is – or else they’ll never scale.


I was having breakfast with Erin, an ex-student, just off a red-eye flight from New York. She’s built a 65-person startup selling enterprise software to the financial services industry. Erin had previously worked in New York for one of those companies and had a stellar reputation in the industry. As one would expect, with banks and hedge funds as customers, the majority were based in the New York metropolitan area.

Where Are Your Biggest Business Deals?
Looking a bit bleary-eyed, Erin explained, “Customers love our product, and I think we’ve found product/market fit. I personally sold the first big deals and hired the VP of sales who’s building the sales team in our New York office. They’re growing the number of accounts and the deal size, but it feels like we’re incrementally growing a small business, not heading for exponential growth. I know the opportunity is much bigger, but I can’t put my finger on what’s wrong.”

Erin continued, “My investors are starting to get impatient. They’re comparing us to another startup in our space that’s growing much faster. My VP of Sales and I are running as fast as we can, but I’ve been around long enough to know I might be the ex-CEO if we can’t scale.”

While Erin’s main sales office is in New York, next to her major prospects and customers, Erin’s company was headquartered in Silicon Valley, down the street from where we were having breakfast. During the Covid pandemic, most of her engineering team worked remotely. Her inside sales team (Sales Development and Business Development reps) used email, phone, social media and Zoom for prospecting and generating leads. At the same time, her account executives were able to use Zoom for sales calls and close and grow business virtually.

There’s a Pattern Here
Over breakfast, I listened to Erin describe what at first seemed like a series of disconnected events.

First, a new competitor started up. Initially, she wasn’t concerned as the competitor’s product had only a subset of the features that Erin’s company did. However, the competitor’s headquarters was based in New York, and their VP of Sales and CEO were now meeting face-to-face with customers, most of whom had returned to their offices. While Erin’s New York-based account execs were selling to the middle tier management of organizations, the CEO of her competitor had developed relationships with the exec staff of potential customers. She lamented, “We’ve lost a couple of deals because we were selling at the wrong level.”

Second, Erin’s VP of sales had just bought a condo in Miami to be next to her aging parents, so she was commuting to NY four days a week and managing the sales force from Miami when she wasn’t in New York. Erin sighed, “She’s as exhausted as I am flying up and down the East Coast.”

Third, Erin’s account execs were running into the typical organizational speedbumps and roadblocks that closing big deals often encounter. However, solving them via email, Zoom and once-a-month fly-in meetings wasn’t the same as the NY account execs being able to say, “Hey, our VP of Sales and CEO are just down the street. Can we all grab a quick coffee and talk this over?” Issues that could have been solved casually and quickly ballooned into ones that took more work and sometimes a plane trip for her VP of Sales or Erin to solve.

By the time we had finished breakfast it was clear to me that Erin was the one putting obstacles in front of her path to scale. Here’s what I observed and suggested.

Keep Your Eye on The Prize
While Erin had sold the first deals herself, she needed to consider whether each deal happened because as CEO, she could call on the company’s engineers to pivot the product. Were the account execs in New York trying to execute a sales model that wasn’t yet repeatable and scalable without the founder’s intervention? Had a repeatable and scalable sales process truly been validated? Or did each sale require a heroic effort?

Next, setting up their New York office without Erin or her VP of Sales physically living in New York might have worked during Covid but was now holding her company back. At this phase of her company the goal of the office shouldn’t be to add new accounts incrementally – but should be how to scale – repeatably. Hiring account execs in an office in New York let Erin believe that she had a tested, validated, and repeatable sales playbook that could rapidly scale the business. The reality was that without her and the VP of Sales living and breathing the business in New York, they were trying to scale a startup remotely.

Her early customers told Erin that her company had built a series of truly disruptive financial service products. But now, the company was in a different phase – it needed to build and grow the business exponentially. And in this phase, her focus as a CEO needed to change – from searching for product/market fit to driving exponential growth.

Exponential Growth Requires Relentless Execution
Because most of her company’s customers were concentrated in a single city, Erin and her VP of Sales needed to be there – not visiting in a hotel room. I suggested that:

  • Erin had to quickly decide if she wanted to be the one to scale the business. If not, her investors were going to find someone who could.
  • If so, she needed to realize that she had missed an important transition in her company. In a high-dollar B-to-B business, building and scaling sales can’t be done remotely. And she was losing ground every day. Her New York office needed a footprint larger than she was. It needed business development and marketing people rapidly creating demand.
  • Her VP of Sales might be wonderful, but with the all the travel the company is only getting her half-time. Erin needs a full-time head of sales in New York. Time to have a difficult conversation.
  • Because she was behind, Erin needed to rent an apartment in New York for a year, and spend the next six months there and at least two weeks a month after that. Her goal was to:
    • 1) Validate that there was a repeatable sales process. It not, build one
    • 2) Build a New York office that could create a sales and marketing footprint without her presence. Only then could she cut back her time in the City.
  • Finally, she needed to consider that if her customers were primarily in New York and the engineers were working remotely, why weren’t the company headquarters in New York?

I Hate New York
As we dug into these issues, I was pretty surprised to hear her say, “I spent a big part of my career in New York. I thought coming out to Stanford and the West Coast meant I could leave the bureaucracy of large companies and that culture behind. Covid let me do that for a few years. I guess now I’m just avoiding jumping back into an environment I thought I had left.”

We lingered over coffee as I suggested it was time for her to take stock of what’s next. She had something rare – a services company that provided real value with products that early customers loved. Her staff didn’t think they were joining a small business, neither did her investors. If she wasn’t prepared to build something to its potential, what was her next move?

Lessons Learned

  • For a startup, the next step after finding product/market fit is finding a repeatable and scalable sales process
  • This requires a transition to the relentless execution of creating demand and exponentially growing sales
  • If your customers are concentrated in a city or region, you need to be where your customers are
  • The CEO needs to lead this growth focus
  • And then hand it off to a team equally capable and committed

Technology, Innovation, and Great Power Competition  – 2022 Wrap Up

We just wrapped up the second year of our Technology, Innovation, and Great Power Competition class – now part of our Stanford Gordian Knot Center for National Security Innovation.

Joe FelterRaj Shah and I designed the class to 1) give our students an appreciation of the challenges and opportunities for the United States in its enduring strategic competition with the People’s Republic of China, Russia and other rivals, and 2) offer insights on how commercial technology (AI, machine learning, autonomy, cyber, quantum, semiconductors, access to space, biotech, hypersonics, and others) are radically changing how we will compete across all the elements of national power e.g. diplomatic, informational, military, economic, financial, intelligence and law enforcement (our influence and footprint on the world stage).


Why This Class?
The return of strategic competition between great powers became a centerpiece of the 2017 National Security Strategy and 2018 National Defense Strategy. The 2021 Interim National Security Guidance and the administration’s recently released 2022 National Security Strategy make clear that China has rapidly become more assertive and is the only competitor potentially capable of combining its economic, diplomatic, military, and technological power to mount a sustained challenge to a stable and open international system. And as we’ve seen in the Ukraine, Russia remains determined to wage a brutal war to play a disruptive role on the world stage.

Prevailing in this competition will require more than merely acquiring the fruits of this technological revolution; it will require a paradigm shift in the thinking of how this technology can be rapidly integrated into new capabilities and platforms to drive new operational and organizational concepts and strategies that change and optimize the way we compete.

Class Organization
The readings, lectures, and guest speakers explored how emerging commercial technologies pose challenges and create opportunities for the United States in strategic competition with great power rivals with an emphasis on the People’s Republic of China. We focused on the challenges created when U.S. government agencies, our federal research labs, and government contractors no longer have exclusive access to these advanced technologies.

This course included all that you would expect from a Stanford graduate-level class in the Masters in International Policy – comprehensive readings, guest lectures from current and former senior officials/experts, and written papers. What makes the class unique however, is that this is an experiential policy class. Students formed small teams and embarked on a quarter-long project that got them out of the classroom to 1) identify a priority national security challenge, and then to 2) validate the problem and propose a detailed solution tested against actual stakeholders in the technology and national security ecosystem.

The class was split into three parts. Part 1, weeks 1 through 4 covered international relations theories, strategies and policies around Great Power Competition specifically focused on the People’s Republic of China (PRC) and the Communist Peoples Party (CCP). Part 2, weeks 5 through 8, dove into the commercial technologies: semiconductors, space, cyber, AI and Machine Learning, High Performance Computing, and Biotech. In between parts 1 and 2 of the class, the students had a midterm individual project. It required them to write a 2,000-word policy memo describing how a U.S. competitor is using a specific technology to counter U.S. interests and a proposal for how the U.S. should respond. (These policy memos were reviewed by Tarun Chhabra, the Senior Director for Technology and National Security at the National Security Council.)

Each week the students had to read 5-10 articles (see class readings here.) And each week we had guest speakers on great power competition, and technology and its impact on national power and lectures/class discussion.

Guest Speakers
In addition to the teaching team, the course drew on the experience and expertise of guest lecturers from industry and from across U.S. Government agencies to provide context and perspective on commercial technologies and national security.

Our class opened with three guest speakers; former U.S. Secretary of Defense James Mattis and the CIA’s CTO and COO Nand Mulchandani and Andy Makridis. The last class closed with a talk by Google ex-Chairman Eric Schmidt.

In the weeks in-between we had teaching team lectures followed by speakers that led discussions on the critical commercial technologies. For semiconductors, the White House Coordinator for the CHIPS Act – Ronnie Chatterji, and the CTO of Applied Materials – Om Nalamasu. For commercial tech integration and space, former Defense Innovation Unit (DIU) Director Mike Brown and B. General Bucky Butow – Director of the Space Portfolio. For Artificial Intelligence, Lt. Gen. (Ret) Jack Shanahan, former director of the Joint Artificial Intelligence Center. And for synthetic biology Stanford Professor Drew Endy – President, BioBricks Foundation.

Team-based Experiential Project
The third part of the class was unique – a quarter-long, team-based project. Students formed teams and developed hypotheses of how commercial technologies can be used in new and creative ways to help the U.S. wield its instruments of national power. And consistent with all our Gordian Knot Center classes, they got out of the classroom and interviewed 20+ beneficiaries, policy makers, and other key stakeholders testing their hypotheses and proposed solutions. At the end of the quarter, each of the teams gave a final “Lessons Learned” presentation and followed up with a 3,000 to 5,000-word team-written paper.

By the end of the class all the teams realized that the problem they had selected had morphed into something bigger, deeper, and much more interesting.

Team 1: Climate Change

Original Problem Statement: What combinations of technologies and international financial relationships should the US prioritize to mitigate climate change?

Final Problem Statement: How should the US manage China’s dominance in solar panels?

If you can’t see the presentation click here.

We knew that these students could write a great research paper. As we pointed out to them, while you can be the smartest person in the building, it’s unlikely that 1) all the facts are in the building, 2) you’re smarter than the collective intelligence sitting outside the building.

Jonah Cader: “Technology, Innovation and Great Power Competition (TIGPC) is that rare combination of the theoretical, tactical, and practical. Over 10 weeks, Blank, Felter, and Shah outline the complexities of modern geopolitical tensions and bring students up the learning curves of critical areas of technological competition, from semiconductors to artificial intelligence. Each week of the seminar is a crash course in a new domain, brought to life by rich discussion and an incredible slate of practitioners who live and breathe the content of TIGPC daily. Beyond the classroom, the course plunges students into getting “out of the building” to iterate quickly while translating learnings to the real world. Along the way the course acts as a strong call to public service.”

Team 2: Networks

Original Problem Statement: How might we implement a ubiquitous secure global access to the internet in order to help circumvent censorship in authoritarian regimes?

 Final Problem Statement: How can we create an open, free Internet and maintain effective lines of communication in Taiwan in preparation for a potential invasion?

If you can’t see the presentation click here

By week 2 of the class students formed teams around a specific technology challenge facing a US government agency and worked throughout the course to develop their own proposals to help the U.S. compete more effectively through new operational concepts, organizations, and/or strategies.


Jason KimThis course doesn’t just discuss U.S. national security issues. It teaches students how to apply an influential and proven methodology to rapidly develop solutions to our most challenging problems.”


Team 3: Acquisition

Original Problem Statement: How can the U.S. Department of Defense match or beat the speed of great power competitors in acquiring and integrating critical technologies?

Final Problem Statement: How can the U.S. Department of Defense deploy alternative funding mechanisms in parallel to traditional procurement vehicles to enable and incentivize the delivery of critical next-generation technology in under 5 years?

If you can’t see the presentation click here

We wanted to give our students hands-on experience on how to deeply understand a problem at the intersection of our country’s diplomacy, information, its military capabilities, economic strength, finance, intelligence, and law enforcement and dual-use technology. First by having them develop hypotheses about the problem; next by getting out of the classroom and talking to relevant stakeholders across government, industry, and academia to validate their assumptions; and finally by taking what they learned to propose and prototype solutions to these problems.


Matt Kaplan: “The TIGPC class was a highlight of my academic experience at Stanford. Over the ten week quarter, I learned a tremendous amount about the importance of technology in global politics from the three professors and from the experts in government, business, and academia who came to speak. The class epitomizes some of the best parts of my time here: the opportunity to learn from incredible, caring faculty and to work with inspiring classmates. Joe, Steve, Raj instilled in my classmates and me a fresh sense of excitement to work in public service.”

 Team 4: Wargames

Original Problem Statement: The U.S. needs a way, given a representative simulation, to rapidly explore a strategy for possible novel uses of existing platforms and weapons.

Final Problem Statement: Strategic wargames stand to benefit from a stronger integration of AI+ML but are struggling to find adoption and usage. How can this be addressed?

If you can’t see the presentation click here

We want our students to build the reflexes and skills to deeply understand a problem by gathering first-hand information and validating that the problem they are solving is the real problem, not a symptom of something else. Then, students began rapidly building minimal viable solutions (policy, software, hardware …) as a way to test and validate their understanding of both the problem and what it would take to solve it.


Etienne Reche-Ley: “Technology, Innovation and Great Power Competition gave me an opportunity to dive into a real world national security threat to the United States and understand the implications of it within the great power competition. Unlike any other class I have taken at Stanford, this class allowed me to take action on our problem about networks, censorship and the lack of free flow of information in authoritarian regimes, and gave me the chance to meet and learn from a multitude of experts on the topic. I finished this class with a deep understanding of our problem, a proposed actionable solution and a newfound interest in the intersection of technology and innovation as it applies to national defense. I am very grateful to have been part of this course, and it has inspired me to go a step further and pursue a career related to national security.”

Team 6: Disinformation

Original Problem Statement: Disinformation is a national security threat.

Final Problem Statement: The U.S.’s ability to close the disinformation response kill chain is hampered by a lack of coordination between U.S. government  agencies,  no clear ownership of the disinformation problem, and a lack of  clear guidelines on public-private partnerships.

If you can’t see the presentation click here

One other goal of the class was to continue to validate and refine our pedagogy of combining a traditional lecture class with an experiential project. We did this by tasking the students to 1) use what they learned from the lectures and 2) then test their assumptions outside the classroom, the external input they received would be a force multiplier. It would make the lecture material real, tangible and actionable. And we and they would end up with something quite valuable.


Shreyas Lakhtakia: “TIGPC is an interdisciplinary class like no other. It is a fabulous introduction to some of the most significant tech and geopolitical challenges and questions of the 21st century. The class, like the topics it covers, is incredible and ambitious – it’s a great way to level up your understanding of not just international policy, political theory and technology policy but also deep tech and the role of startups in projecting national power. If you’re curious about the future of the world and the role of the US in it, you won’t find a more unique course, a more dedicated teaching team or better speakers to hear from than this!”

Team 7: Quantum Technology

Original Problem Statement: China’s planned government investment in quantum dwarfs that of the U.S. by a factor of 10.

Final Problem Statement: The US quantum ecosystem does not generate enough awareness of opportunities to pursue careers in quantum that could catalyze industry growth.

If you can’t see the presentation click here

We knew we were asking a lot from our students. We were integrating a lecture class with a heavy reading list with the best practices of hypothesis testing from Lean Launchpad/Hacking for Defense/I-Corps. But I’ve yet to bet wrong in pushing students past what they think is reasonable. Most rise way above the occasion.


 Team 9: Lithium-Ion Batteries

Original Problem Statement: Supply and production of lithium-ion batteries is centered in China. How can the U.S. become competitive?

Final Problem Statement: China controls the processing of critical materials used for lithium-ion batteries. To regain control the DOE needs to incentivize short and long-term strategies to increase processing of critical materials and decrease dependence on lithium-ion batteries.

If you can’t see the presentation click here


All of our students put in extraordinary amount of work. Our students came from a diverse set of background and interests – from undergraduate sophomores to 5th year PhD’s – in a mix including international policy, economics, computer science, business, law and engineering. Some will go on to senior roles in State, Defense, policy or other agencies. Others will join or found the companies building new disruptive technologies. They’ll be the ones to determine what the world-order will look like for the rest of the century and beyond. Will it be a rules-based order where states cooperate to pursue a shared vision for a free and open region and where the sovereignty of all countries large and small is protected under international law? Or will it be an autocratic and dystopian future coerced and imposed by a neo-totalitarian regime?

This class changed the trajectory of many of our students. A number expressed newfound interest in exploring career options in the field of national security. Several will be taking advantage of opportunities provided by the Gordian Knot Center for National Security Innovation to further pursue their contribution to national security.

This course and our work at Stanford’s Gordian Knot Center would not be possible without the unrelenting support and guidance from Ambassador Mike McFaul and Professor Riitta Katila, GKC founding faculty and Principal Investigators, and the tenacity of David Hoyt, Gordian Knot Center Assistant Director.

Lessons Learned

  • We combined lecture and experiential learning so our students can act on problems not just admire them
    • The external input the students received was a force multiplier
    • It made the lecture material real, tangible and actionable
    • Pushing students past what they think is reasonable results in extraordinary output. Most rise way above the occasion
  • The class creates opportunities for our best and brightest to engage and address challenges at the nexus of technology, innovation and national security
    • The final presentations and papers from the class are proof that will happen

Why The Pentagon Can’t Count: It’s Time to Reinvent the Audit

This article previously appeared in War on the Rocks.

In the past, headlines about the Pentagon failing its financial audit again would never have caught my attention. But having been in the middle of this conversation when I served on one of the Defense Department’s advisory boards, I understand why the Pentagon can’t count. The experience taught me a valuable lesson about innovation and imagination in large organizations, and the difference visionary leadership – or the lack of it – can make.

With audit costs approaching a billion dollars a year the Pentagon had an opportunity to lead in modernizing auditing. Instead it opted for more of the same.

Auditing the Department of Defense
By law, the Department of Defense has to provide Congress and the public with an assessment of where it spends its money and to provide transparency of its operations. A financial audit counts what the Department of Defense has, where it has it, and if they know where its money is being spent.

Auditing the Department of Defense is a massive undertaking. For one thing, it is the country’s largest employer, with 2.9 million people (1.3 million on active duty, 800,000 in the reserve components, and 770,000 civilians.) The audit has to count the location and condition of every piece of military equipment, property, inventory, and supplies. And there are a lot of them. The department has 643,900 assets, from buildings, to pipelines, roads, and fences located on over 4,860 sites, as well as 19,700 aircraft and over 290 battle force ships. To complicate the audit, the department has 326 different and separate financial management systems, 4,700 data warehouses and over 10,000 different and disconnected data management systems.

(BTW, just like in the private sector, financial audits and audits of contracts are separate. While the DoD Office of Inspector General is responsible for these financial audits of trillions of dollars of assets and liabilities, the Defense Contract Audit Agency is responsible for auditing the hundreds of billions of dollars of acquisition contracts. They have the same issues.)

This is the fifth year the Department has undergone a financial statement audit – and failed it. The audit was not a trivial effort, it required 1,600 auditors – 1,450 from public accounting firms and 150 from the Office of Inspector General. In 2019, the audit cost $428 million in auditing costs ($186 million to the auditors along with $242 million to audit support) and another $472 million to fix the issues the audit discovered.

Let’s Invent the Future of Audit
The Defense of Department’s 40-plus advisory boards are staffed by outsiders who can provide independent perspectives and advice. I sat on one of these boards, and our charter was to leverage private sector lessons to improve audit quality.

With defense spending on auditing approaching a billion dollars a year, it was clear it would take a decade or more to catch up to the audit standards of private companies. But no single company or even entire industry was spending this much money on auditing. And remarkably, the Defense Department seemed intent on doing the same thing year after year, just with more people and with a few more tools and processes to get incrementally better. It dawned on me that if we tried to look over the horizon, the department could audit faster, cheaper, and more effectively by inventing the future tools and techniques rather than repeating the past.

Nothing in our charter asked the advisory board to invent the future. But I found myself asking, “What if we could?” What if we could provide the defense department with new technology, new approaches to auditing, analytics practices, audit research, and standards, all while creating audit and data management research and a new generation of finance applications and vendors?

The Pentagon Once Led Business Innovation
I reminded my fellow advisory board members that in 1959, at the dawn of the computer age, the Defense Department was the largest user of computers for business applications.

However, there was no common business programing language. So rather than wait for one, the Defense Department led the effort to create one – the COBOL programming language. And 20 years later, it did the same for the ADA programming language.

With that history in mind, I proposed we lead again. And that we start an initiative for the 5th generation of audit practices (the Audit 5.0 Initiative) with machine learning, predictive analytics, Intelligent sampling and predictions. This initiative would also include automating ETL, predictive analytics, fraud detection, and a new generation of audit standards.

I pointed out that this program wouldn’t need more funds since the Department of Defense could allocate 10% of the $428M we were spending on auditors and fund SBIR (Small Business Innovation Research) programs in auditing/data management/finance to generate 5-10 new startups in this space each year. Simultaneously we could fund academic research, to incentivize research on Machine Learning as applied to Audit 5.0 challenges in finance, auditing and data management.

In addition, we could create new audit standards by working with existing government audit standards bodies such as (The Generally Accepted Government Auditing Standards (GAGAS), Yellow Book, the GAO’s Standards for Internal Control in the Federal Government, Green Book and the Federal Accounting Standards Advisory Board (FASAB). We could collaborate with civilian audit standard bodies (ASB (Auditing Standards Board) and PCAOB (Public Company Accounting Oversight Board). Working together, the defense department could create the next generation of machine-driven and semiautomated standards. Furthermore, it could help the Independent Public Accounting firms (KPMG, EY, PwC, Deloitte, et al) create a new practice and make them partners in the Audit 5.0 initiative.

By investing 10 percent of the existing auditing budget over the next few years, these activities would create a defense audit center of excellence that would fund academic centers for advanced audit research, standup “future of audit” programs that would create new 5-10 startups each year, be the focal point for government an industry finance and audit standards, and create public-private partnerships rather than mandates.

Spinning up these activities up would dramatically reduce the department’s audit costs, standardize its financial management environment, and provide confidence in their budget, auditability, and transparency. And as a bonus, it would create a new generation of finance, audit and data management startups, funded by private capital.

The Road Not Taken
I was in awe of my fellow advisory board members. They had spent decades in senior roles in finance and accounting in both the public and private sectors. Yet, when I pitched this idea, they politely listened to what I had to say and then moved on to their agenda – providing the DoD with Incremental improvements.

At the time I was disappointed, but not surprised. An advisory board is only as good as what it’s being chartered and staffed to do. If they are being asked to provide a 10 percent incremental advice, they’ll do so. But if they’re asked for revolutionary i.e. 10x advice, they can change the world. But that requires a different charter, leadership, people, innovation, and imagination.

In the end, the Department of Defense, the largest purchaser of accounting services in the world, whiffed a chance to be the leader in creating the next generation of audit tools and services, not only for financial audits, but for the hundreds of billions of dollars of acquisition contracts the Defense Contract Audit Agency audits. By now the department could have audit tools driven by machine learning algorithms, ferreting out fraud by vendors or contractors and anticipating programs that are at risk.

Lessons Learned

  • If you only get what you ask for you haven’t hired people with imagination
    • America’s defense leaders ought to ask and act for transformational, contrarian and disruptive advice
    • And ensure they have the will and organizations to act on it
  • Move requests for advice for incremental improvements to the consulting firms that currently serve the Defense Department
  • Defense leaders need to consider whether spending a billion dollars a year for an audit is causing the department to become appreciably more efficient or better managed
    • Or whether there might be a better way

The 6th Lean Innovation Educators Summit – Education and Innovation in the Age of Chaos and Disruption

Join Jerry Engel, Pete Newell, and Steve Weinstein for the sixth edition of the Lean Innovation Educators Summit December 14, 1-4 pm Eastern Time, 10 am-1 pm Pacific Time. Register here.

This virtual gathering will bring together entrepreneurship educators from around the world who are putting Lean Innovation to work in their classrooms, accelerators, venture studios, and student-driven ventures.

The summit topic is “Education and Innovation in the Age of Chaos and Disruption.

Our students will be facing the challenges of a world that’s rapidly changing, chaotic and uncertain. A world undergoing climate change, supply chain disruptions, political instability and continual technology innovation and disruption. It’s incumbent on us as educators to provide the next generation of innovators with the tools and mindset to meet these challenges.

Among the questions we’ll address in this short summit:

  • How do we as entrepreneurship and innovation educators best prepare the next generation?
  • What role should our institutions help us do this?
  • What are the other systems and partnerships that we need to take advantage of?

We will have concurrent breakout sessions so participants have the opportunity to choose their own path to explore. We’ll then going to pivot to hear from colleagues across three broad categories of innovation:

  • Curriculum – We’ll discuss how best to equip educators with the tools they need to cultivate and guide student teams around solving mission-driven problems.
  • Ecosystems – We’ll explore partnerships that engage and inform positive student engagement and outcomes and how to support diversity of thought, background.
  • Trends – The rate of technological disruption shows no sign of slowing down. Climate change was a hypothesis for our generation but will be the facts on the ground for our students. The struggle between great powers and a fluid global landscape will accelerate. All of these will shape what future curriculums our students need and educators must deliver.

Alexander Osterwalder, the creator of the business model canvas and Strategyzer co-founder will join the discussion about the intersection of education, innovation and entrepreneurship

During the breakout sessions, you will have the opportunity to contribute to the conversation via Chat, Q&A, and an online community bulletin board. We will close out the Summit with Alex Osterwalder’s fireside chat moderated by Dr. Jerry Engel.

How to register

When you register, you will receive a link to an online collaboration space where you can submit questions, challenges and feedback. This feedback will inform the content of the presentations, post-event white papers, and the curriculum delivered to our educator community.

This session is free but limited to Innovation educators. Register here and learn more on our website: We look forward to gathering as a community of educators to shape the future of Lean Innovation Education.

The Three Pillars of World-class Corporate Innovation

My good friend Alexander Osterwalder, the inventor of the business model canvas (one of foundations of the Lean Methodology) has written a playbook (along with his associate partner Tendayi Viki,) From Innovation Theater to Growth Engine to explain how to build and implement repeatable innovation processes inside a company. 

Here’s their introduction to the key concepts inside the playbook.


Over 75% of executives report that innovation is a top three priority at their companies. However, only 20% of executives indicate that their companies are ready to innovate at scale. This is the challenge for contemporary organizations: How to develop a world-class ecosystem that can drive repeatable innovation at scale.

The playbook describes the three pillars of corporate innovation: Innovation Portfolios, Innovation Programs and a Culture of Innovation. Under each pillar, the playbook describes three questions that leaders and teams can ask to evaluate whether their company has the right innovation ecosystem in place.

Innovation Portfolio: what are your company’s portfolio of innovation projects?

  • Are your company’s innovation efforts exploring or exploiting business modes?
  • Does your company have a balanced portfolio of projects that cover efficiency, sustaining and transformative innovation?
  • What is the health of your innovation funnel or pipeline?

Explore: Search for new value propositions and business models by designing and testing new business ideas rather than execution. 

Exploit: Manage existing business models by scaling emerging businesses, renovating declining ones and protecting the successful ones.


Innovation Programs: how are your company’s innovation programs are structured and managed.

  • Do your leaders get excited about the wrong innovation programs?
  • What results are your innovation programs producing?
  • Are your company’s innovation programs interconnected in a strategic way?

To close the innovation capability gap, companies can evaluate their innovation programs by asking whether they’reinnovation theater or producing tangible results for the company.

  • Value Creation: Creating new products, services, value propositions and business models. These programs invest in and manage innovation projects that create value by producing new growth or cost savings.
  • Culture Change: Transforming the company to establish an innovation culture. This may include new processes, metrics, incentive systems, or changing organizational structures. These transformations help the company innovate in a consistent and repeatable way.

Innovation Culture: What are the blockers and enablers of innovation in your company –

  • How much time does your leadership spend on innovation?
  • Where does innovation live in your organization and how much power does it have?
  • What is your kill rate for innovation projects?

To overcome the innovation capability gap, companies need to create a culture that enables the right behaviors to produce world-class innovative outcomes. A reliable indicator of the quality of your innovation culture is how innovation teams would describe it. Is it a culture that is dominated by blockers of innovation or enablers of innovation?

  • Leadership Support: How can corporate leaders have the biggest impact on innovation in terms of time spent, strategic guidance, and resource allocation.
  • Organizational Design: How to give innovation legitimacy and power, the right incentives, and clear policies for collaboration with the core business.
  • Innovation Practice: How to develop people’s innovation skills and experience and acquire the right innovation talent. How to ensure that we are using the right tools, processes, and metrics to test and adapt ideas in order to reduce risk.

 Lessons Learned

  • The three pillars of an innovation ecosystem:
    • Innovation Portfolios
    • Innovation Programs
    • a Culture of Innovation
  • Download the Osterwalder Playbook here

A Simple Map for Innovation at Scale

An edited version of this article previously appeared in the Boston Consulting Group’s strategy think tank website.

I spent last week at a global Fortune 50 company offsite watching them grapple with disruption. This 100+-year-old company has seven major product divisions, each with hundreds of products. Currently a market leader, they’re watching a new and relentless competitor with more money, more people and more advanced technology appear seemingly out of nowhere, attempting to grab customers and gain market share.

This company was so serious about dealing with this threat (they described it as “existential to their survival”) that they had mobilized the entire corporation to come up with new solutions. This wasn’t a small undertaking, because the threats were coming from multiple areas in multiple dimensions; How do they embrace new technologies? How do they convert existing manufacturing plants (and their workforce) for a completely new set of technologies? How do they bring on new supply chains? How do they become present on new social media and communications channels? How do they connect with a new generation of customers who had no brand loyalty? How to they use the new distribution channels competitors have adopted? How do they make these transitions without alienating and losing their existing customers, distribution channels and partners? And how do they motivate their most important asset – their people – to operate with speed, urgency, and passion?

The company believed they had a handful of years to solve these problems before their decline would become irreversible. This meeting was a biannual gathering of all the leadership involved in the corporate-wide initiatives to out-innovate their new disruptors. They called it the “Tsunami Initiative” to emphasize they were fighting the tidal wave of creative destruction engulfing their industry.

To succeed they realized this isn’t simply coming up with one new product. It meant pivoting an entire company – and its culture. The scale of solutions needed dwarf anything a single startup would be working on.

The company had hired a leading management consulting firm that helped them select 15 critical areas of change the Tsunami Initiative was tasked to work on. My hosts, John and Avika, at the offsite were the co-leads overseeing the 15 topic areas. The consulting firm suggested that they organize these 15 topic areas as a matrix organization, and the ballroom was filled with several hundred people from across their company –  action groups and subgroups with people from across the company: engineering, manufacturing, market analysis and collection, distribution channels, and sales. Some of the teams even included some of their close partners. Over a thousand more were working on the projects in offices scattered across the globe.

John and Avika had invited me to look at their innovation process and offer some suggestions.

Are these the real problems?
This was one of the best organized innovation initiatives I have seen. All 15 topic had team leads presenting poster sessions, there were presenters from the field sales and partners emphasizing the urgency and specificity of the problems, and there were breakout sessions where the topic area teams brainstormed with each other. After the end of the day people gathered around the firepit for informal conversations. It was a testament to John and Avika’s leadership that even off duty people were passionately debating how to solve these problems. It was an amazing display of organizational esprit de corps.

While the subject of each of the 15 topic areas had been suggested by the consulting firm, it was in conjunction with the company’s corporate strategy group, and the people who generated these topic area requirements were part of the offsite. Not only were the requirements people in attendance but so was a transition team to facilitate the delivery of the products from these topic teams into production and sales.

However, I noticed that several of the requirements from corporate strategy seemed to be priorities given to them from others (e.g. here are the problems the CFO or CEO or board thinks we ought to work on) or likely here are the topics the consulting firm thought they should focus on) and/or were from subject matter experts (e.g. I’m the expert in this field. No need to talk to anyone else; here’s what we need). It appeared the corporate strategy group was delivering problems as fixed requirements, e.g. deliver these specific features and functions the solution ought to provide.

Here was a major effort involving lots of people but missing the chance to get the root cause of the problems.

I told John and Avika that I understood some requirements were known and immutable. However, when all of the requirements are handed to the action teams this way the assumption is that the problems have been validated, and the teams do not need to do any further exploration of the problem space themselves.

Those tight bounds on requirements constrain the ability of the topic area action teams to:

  • Deeply understand the problems – who are the customers, internal stakeholders (sales, other departments) and beneficiaries (shareholders, etc.)? How to adjudicate between them, priority of the solution, timing of the solutions, minimum feature set, dependencies, etc.
  • Figure out whether the problem is a symptom of something more important
  • Understand whether the problem is immediately solvable, requires multiple minimum viable products to test several solutions, or needs more R&D

I noticed that with all of the requirements fixed upfront, instead of having a freedom to innovate, the topic area action teams had become extensions of existing product development groups. They were getting trapped into existing mindsets and were likely producing far less than they were capable of. This is a common mistake corporate innovation teams tend to make.

I reminded them that when team members get out of their buildings and comfort zones, and directly talk to, observe, and interact with the customers, stakeholders and beneficiaries, it allows them to be agile, and the solutions they deliver will be needed, timely, relevant and take less time and resources to develop. It’s the difference between admiring a problem and solving one.

As I mentioned this, I realized having all fixed requirements is a symptom of something else more interesting – how the topic leads and team members were organized. From where I sat, it seemed there was a lack of a common framework and process. 

Give the Topic Areas a Common Framework
I asked John and Avika if they had considered offering the topic action team leaders and their team members a simple conceptual framework (one picture) and common language. I suggested this would allow the teams to know when and how to “ideate” and incorporate innovative ideas that accelerate better outcomes. The framework would use the initial corporate strategy requirements as a starting point rather than a fixed destination. See the diagram.

I drew them a simple chart and explained that most problems start in the bottom right box.

These are “unvalidated” problems. Teams would use a customer discovery process to validate them. (At times some problems might require more R&D before they can be solved.) Once the problems are validated, teams move to the box on the bottom left and explore multiple solutions. Both boxes on the bottom are where ideation and innovation-type of problem/solution brainstorming are critical. At times this can be accelerated by bringing in the horizon 3, out-of-the-box thinkers that every company has, and let them lend their critical eye to the problem/solution.

If a solution is found and solves the problem, the team heads up to the box on the top left.

But I explained that very often the solution is unknown. In that case think about having the teams do a “technical terrain walk.” This is the process of describing the problem to multiple sources (vendors, internal developers, other internal programs) debriefing on the sum of what was found. A terrain walk often discovers that the problem is actually a symptom of another problem or that the sources see it as a different version of the problem. Or that an existing solution already exists or can be modified to fit.

But often, no existing solution exists. In this case, teams could head to the box on the top right and build Minimal Viable Products – the smallest feature set to test with customers and partners. This MVP testing often results in new learnings from the customers, beneficiaries, and stakeholders –  for example, they may tell the topic developer that the first 20% of the deliverable is “good enough” or the problem has changed, or the timing has changed, or it needs to be compatible with something else, etc. Finally, when a solution is wanted by customers/beneficiaries/stakeholders and is technically feasible, then the teams move to the box on the top left.

The result of this would be teams rapidly iterating to deliver solutions wanted and needed by customers within the limited time the company had left.

Creative destruction
Those companies that make it do so with an integrated effort of inspired and visionary leadership, motivated people, innovative products, and relentless execution and passion.

Watching and listening to hundreds of people fighting the tsunami in a legendary company was humbling.

I hope they make it.

Lessons Learned

  • Creative destruction and disruption will happen to every company. How will you respond?
  • Topic action teams need to deeply understand the problems as the customer understands them, not just what the corporate strategy requirements dictate
    • This can’t be done without talking directly to the customers, internal stakeholders, and partners
  • Consider if the corporate strategy team should be more facilitators than gatekeepers
  • A light-weight way to keep topic teams in sync with corporate strategy is to offer a common innovation language and problem and solution framework

Mapping the Unknown – The Ten Steps to Map Any Industry

A journey of a thousand miles begins with a single step

 Lǎozi 老子

I just had lunch with Shenwei, one of my ex-students who had just taken a job in a mid-sized consulting firm.  After a bit of catching up I offered he was looking a bit lost. “I just got handed a project to help our firm enter a new industry – semiconductors. They want me to map out the space so we can figure out where we can add value.

When I asked what they already knew about it, they tossed me a tall stack of industry and stock analyst reports, company names, web sites, blogs. I started reading through a bunch of it and I’m drowning in data but don’t know where to start. I feel like I don’t know a thing.”

I told Shenwei I was happy for him because he had just been handed an awesome learning opportunity – how to rapidly understand and then map any new market. He gave me a “easy for you to say” look, but before he could object I handed him a pen and a napkin and asked him to write down the names of companies and concepts he read about that have anything to do with the semiconductor business – in 30 seconds. He quickly came up with a list with 9 names/terms. (See Mapping – First Pass)

“Great, now we have a start. Now give me a few words that describe what they do, or mean, or what you don’t know about them.”

Don’t let the enormity of unknowns frighten you. Start with what you do know.

After a few minutes he came up with a napkin sketch that looked like the picture in Mapping – Second Pass. 
Now we had some progress.

I pointed out he now had a starter list that not only contained companies but the beginning of a map of the relationships between those companies. And while he had a few facts, others were hypotheses and concepts. And he had a ton of unanswered questions.

We spent the next 20 minutes deconstructing that sketch and mapping out the Second Pass list as a diagram (see Mapping – Third Pass.)

As you keep reading more materials, you’ll have more questions than facts. Your goal is to first turn the questions into testable hypotheses (guesses). Then see if you can find data that turns the hypotheses into facts. For a while the questions will start accumulating faster than the facts. That’s OK.

Note that even with just the sparse set of information Shenwei had, in the bottom right-hand corner of his third mapping pass, a relationship diagram of the semiconductor industry was beginning to emerge.

Drawing a diagram of the relationships of companies in an industry can help you deeply understand how the industry works and who the key players are. Start building one immediately. As you find you can’t fill in all the relationships, the gaps outlining what you need to learn will become immediately visible.

As the information fog was beginning to lift, I could see Shenwei’s confidence returning. I pointed out that he had a real advantage that his assignment was in a known industry with lots of available information. He quickly realized that he could keep adding information to the columns in the third mapping pass as he read through the reports and web sites.

Google and Google Scholar are your best friends. As you discover new information increase your search terms.

My suggestion was to use the diagram in the third mapping pass as the beginning of a wall chart – either physically (or virtually if he could keep it in all in his head). And every time he learned more about the industry to update the relationship diagram of the industry and its segments. (When he pointed out that there were existing diagrams of the semiconductor industry he could copy, I suggested that he ignore them. The goal was for him to understand the industry well enough that he could draw his own map ab initio – from the beginning. And if he did so, he might create a much better one.)

When lunch was over Shenwei asked if it was OK if he checked in with me as he learned new things and I agreed. What he didn’t know was that this was only the first step in a ten-step industry mapping process.

Epilog

Over the next few weeks Shenwei shared what he had learned and sent me his increasingly refined and updated industry relationship map. (The 4th  mapping pass showed up 48 hours later.)In exchange I shared with him the news that he was on step one of a ten step industry mapping program. Other the next few weeks he quickly built on the industry map to answer questions 2 through 10 below.

Two weeks later he handed his leadership an industry report that covered the ten steps below and contained a sophisticated industry diagram he created from scratch. A far cry from his original napkin sketch!

Six months later his work on this project convinced his company that there was a large opportunity in the semiconductor space, and they started a new practice with him in it. His work won him the “best new employee” award.

The Ten Steps to Map any Industry

Start by continuously refining your understanding of the industry by diagramming it. List all the new words you encounter and create a glossary in your own words. Start collecting the best sources of information you’ve read.

Basic Industry Understanding

  1. Diagram the industry and its segments
    1. Start with anything
    2. Build your learning by successive iteration
    3. Who are the key suppliers to each segment?
    4. How does this industry feed into the larger economy?
  2. Create a glossary of industry unique terms
    1. Can you explain them to others? Are there analogies to other markets?
  3. Who are the industry experts in each segment? For the entire industry?
    1. Economic experts? E.g. industry analysts, universities, think tanks
    2. Technology experts? E.g. universities, think tanks
    3. Geographic experts?
  4. Key Conferences, blogs, web sites, etc.
    1. What are the best opensource data feeds?
    2. What are the best paid resources?
Overlay numbers, dollars, market share, Compound Annual Growth Rate (CAGR) on all parts of the industry diagram. That will inform velocity and direction of the market.

Detailed Industry Understanding

  1. Who are the market leaders? New entrants? In revenue, market share and growth rate
    1. In the U.S.
    2. Western countries
    3. China
  2. Understand the technology flows
    1. Who builds on top of who
    2. Who is critical versus who can be substituted
  3. Understand the economic flows
    1. Who buys from who in this industry
    2. Who buys the output from this industry.
    3. How cyclical is demand?
    4. What are the demand drivers?
    5. How do companies inside each segment get funded? Any differences in capital requirements? Ease of starting, etc.
  4. If applicable, understand the personnel flow for each segment
    1. Do people move just between their segments or up and down through the entire industry?
    2. Where do they get trained?
The beginner’s forecasting method is to simply extrapolate current growth rates forward. But in today’s technology markets, discontinuities are coming fast and furious. Are there other technologies from adjacent markets will impact this one? (e.g. AI, Quantum, High performance computing,…?). Are there other global or national economic initiatives that could change the shape of the market?

Forecasting

  1. What’s changed in the last 10 years? 5 years?
    1. Diagram the past incarnations of the industry
  2. What’s going to change in the next 5 years?
    1. Any big insight on disruption?
    2. New entrants?
    3. New technology?
    4. New foreign suppliers?
    5. Diagram your model of the industry in 5 years

National Industrial Policy – Private Capital and The America’s Frontier Fund Steps Up

This article previously appeared in The National Interest.

Last month the U.S. passed the CHIPS and Science Act, one of the first pieces of national industrial policy – government planning and intervention in a specific industry — in the last 50 years, in this case for semiconductors. After the celebratory champagne has been drunk and the confetti floats to the ground it’s helpful to put the CHIPS Act in context and understand the work that government and private capital have left to do.

Today the United States is in great power competition with China. It’s a contest over which nation’s diplomatic, information, military and economic system will lead the world in the 21st century. And the result is whether we face a Chinese dystopian future or a democratic one, where individuals and nations get to make their own choices. At the heart of this contest is leadership in emerging and disruptive technologies – running the gamut from semiconductors and supercomputers to biotech and blockchain and everything in between.

National Industrial Policy – U.S. versus China
Unlike the U.S., China manages its industrial policy via top-down 5-year plans. Their overall goal is to turn China into a technologically advanced and militarily powerful state that can challenge U.S. commercial and military leadership. Unlike the U.S., China has embraced the idea that national security is inexorably intertwined with commercial technology (semiconductors, drones, AI, machine learning, autonomy, biotech, cyber, semiconductors, quantum, high-performance computing, commercial access to space, et al.)  They’ve made what they call military/civil fusion – building a dual-use ecosystem by tightly coupling their commercial technology companies with their defense ecosystem.

China has used its last three 5-year plans to invest in critical technologies (semiconductors, supercomputers, Al/ML, quantum, access to space, biotech.) as a national priority. They have built a sophisticated public/private financing ecosystem to support these plans. The Chinese technology funding ecosystem includes regional investment funds that exceed 700 billion dollars (what they call their Civil/Military Guidance Funds). These are investment vehicles in which central and local government agencies make investments that are combined with private venture capital and State-Owned Enterprises in areas of strategic importance. They are tightly coupling critical civilian companies to their defense ecosystem to help them develop military weapons and strategic surprises. (Tai Ming Cheung’s book is the best description of the system.)

The U.S. has nothing comparable.

In contrast, for the last several decades, planning in the U.S. economy was left to “the market.” Driven by economic theory from the Chicago School of Economics, its premise is that free markets best allocate resources in an economy and that minimal, or even no, government intervention is best for economic prosperity. We ran our economy on this theory as a bipartisan experiment in the U.S. for the last several decades. Optimizing profit above else led to wholesale offshoring of manufacturing and entire industries in order to lower costs. Investors shifted to making massive investments in industries with the quickest and greatest returns without long-term capital investments (e.g. social media, ecommerce, gaming) instead of in hardware, semiconductors, advanced manufacturing, transportation infrastructure, etc. The result was that by default, private equity and venture capital were the de facto decision makers of U.S. industrial policy.

With the demise of the Soviet Union and the U.S. as the sole superpower, this “profits first” strategy was “good enough” as there was no other nation that could match our technical superiority. That changed when we weren’t paying attention.

China’s Ambition and Strategic Surprises
In the first two decades of the 21st century, while the U.S. was focused on combating non-nation states (ISIS, Al-Qaeda…) U.S. policymakers failed to understand China’s size, scale, ambition, and national commitment to surpass the U.S. as the global leader in technology. Not just in “a” technology but in all of those that are critical to both our national and economic security in this century.

China’s top-down national industrial policy means we are being out-planned, outmanned, and outspent. By some estimates, China could be the leader in a number of critical technology areas sooner than we think. While Chinese investment in technology at times has been redundant and wasteful, the sum of these tech investments has resulted in a series of strategic surprises to the U.S.– hypersonics, ballistic missiles with maneuverable warheads as aircraft carrier killers, fractional orbital bombardment systems, rapid advances in space, semiconductors, supercomputers, and biotech …with more surprises likely – all with the goal to gain superiority over the U.S. both commercially and militarily.

Limits and Obstacles to China’s Dominance
However, America has advantages that China lacks: capital markets that can be incented not coerced, untapped innovation talent willing to help, labor markets that can be upskilled, university and corporate research that still excels, etc. At the same time, a few cracks are showing in China’s march to technology supremacy; their detention of some of their most successful entrepreneurs and investors, a crackdown on “superfluous” tech (gaming, online tutoring) and a slowdown of listings on the China’s version of NASDAQ, the Shanghai Stock Exchange’s STAR Market – may signal that the party is reining in its “anything goes” approach to pass the U.S.  Simultaneously the U.S. Commerce department has begun to prohibit export of critical equipment and components that China has needed to build their tech ecosystem.

Billionaires and Venture Capital Funding Defense Innovation
In the U.S. DoD’s traditional suppliers of defense tools, technologies, and weapons – the prime contractors and federal labs – are no longer the leaders in many of these emerging and disruptive  technologies.  And while the Department of Defense has world-class people and organizations it’s for a world that no longer exists. (Its inability to rapidly acquire and deploy commercial systems requires an organizational redesign on the scale of Goldwater/Nichols Act, not a reform.)

Technology innovation in many areas now falls to commercial companies. In lieu of a coherent U.S. national investment strategy across emerging and disruptive technologies (think of the CHIPS Act times ten), billionaires in the U.S. have started their own initiatives – Elon Musk – SpaceX and Starlink (reusable rockets and space-based broadband internet), Palmer Lucky –  Anduril (AI and Machine Learning for defense), Peter Theil – Palantir (data analytics). And in the last few years a series of defense-focused venture funds – Shield Capital, Lux Capital, and others – have emerged.

However, depending on billionaires interested in defense is not a sustainable strategy, and venture capital invests in businesses that can become profitable in 10 years or less. This means that technologies that might take decades to mature (fusion, activities in space, new industrial processes, …) get caught up and die in a “Valley of Death.” Attempts to bridge this Valley of Death often find technology companies relying on Government capital. These programs (DIU, In-Q-Tel, AFWERX, et al), are limited in scope, time and success at scale. These government investment programs have largely failed to scale these emerging and disruptive technologies for four reasons:

  • Government agencies have limited access to top investment talent to help them make sophisticated technical investment decisions
  • Government agencies lack the commercialization skills to help founders turn technical ideas into commercial ventures.
  • While the Dept of Defense has encouraged starting new ventures, it has failed to match it with the acquisition dollars to scale them. There’s no DoD coherent/committed strategy to create a new generation of prime contractors around these emerging and disruptive technologies.
  • No private or government funds operates as “patient capital” – investing in critical deep technologies that may take more than a decade to mature and scale

America’s Frontier Fund
Today one private capital fund is attempting to solve this problem. Gilman Louie, the founder of In-Q-Tel, has started America’s Frontier Fund (AFF.) This new fund will invest in key critical deep technologies to help the U.S. keep pace with the Chinese onslaught of capital focused on this area. AFF plans to raise one billion dollars in “patient private capital” from both public and private sources and to be entirely focused on identifying critical technologies and strategic investing. Setting up their fund as a non-profit allows them to focus on long-term investments for the country, not just what’s expedient to maximize profits. It will ensure these investments grow into large commercial and dual-use companies focused on the national interest.

They’ve built an extraordinary team of experienced venture capitalists (I’ve known Gilman Louie and Steve Weinstein for decades), a world-class chief scientist, a startup incubation team, and they come with a unique and deep understanding of the intersection of national security and emerging and disruptive technologies.

AFF is the most promising effort I have seen in tackling the long-term challenges of funding and scaling emerging and disruptive technologies head-on.

At stake is whether the rest of the 21st century will be determined by an authoritarian government wiling to impose a dystopian future on the world, or free nations able to determine their own future.

These are tough problems to solve, and no single fund is can take on the massive investments China is making, but it’s possible that the AFF’s market driven approach, when combined with the government’s halting steps reengaging in industrial policy, can tip the scale back in our favor.

Here’s hoping they succeed.

Finding and Growing the Islands of Innovation inside a large company – Action Plan for A New CTO

This post previously appeared in Fast Company.

How does a newly hired Chief Technology Officer (CTO) find and grow the islands of innovation inside a large company?

How not to waste your first six months as a new CTO thinking you’re making progress when the status quo is working to keep you at bay?


I just had coffee with Anthony, a friend who was just hired as the Chief Technology Officer (CTO) of a large company (30,000+ people.) He previously cofounded several enterprise software startups, and his previous job was building a new innovation organization from scratch inside another large company. But this is the first time he was the CTO of a company this size.

Good News and Bad
His good news was that his new company provides essential services and regardless of how much they stumbled they were going to be in business for a long time. But the bad news was that the company wasn’t keeping up with new technologies and new competitors who were moving faster. And the fact that they were an essential service made the internal cultural obstacles for change and innovation that much harder.

We both laughed when he shared that the senior execs told him that all the existing processes and policies were working just fine. It was clear that at least two of the four divisions didn’t really want him there. Some groups think he’s going to muck with their empires. Some of the groups are dysfunctional. Some are, as he said, “world-class people and organizations for a world that no longer exists.”

So the question we were pondering was, how do you quickly infiltrate a large, complex company of that size? How do you put wins on the board and get a coalition working? Perhaps by getting people to agree to common problems and strategies? And/or finding the existing organizational islands of innovation that were already delivering and help them scale?

The Journey Begins
In his first week the exec staff had pointed him to the existing corporate incubator. Anthony had long come to the same conclusion I had, that highly visible corporate incubators do a good job of shaping culture and getting great press, but most often their biggest products were demos that never get deployed to the field. Anthony concluded that the incubator in his new company was no exception. Successful organizations recognize that innovation isn’t a single activity (incubators, accelerators, hackathons); it is a strategically organized end-to-end process from idea to deployment.

In addition, he was already discovering that almost every division and function was building groups for innovation, incubation and technology scouting. Yet no one had a single road map for who was doing what across the enterprise. And more importantly it wasn’t clear which, if any, of those groups were actually continuously delivering products and services at high speed.  His first job was to build a map of all those activities.

Innovation Heroes are Not Repeatable or Scalable
Over coffee Anthony offered that in a company this size he knew he would find “innovation heroes” – the individuals others in the company point to who single-handedly fought the system and got a new product, project or service delivered (see article here.) But if that was all his company had, his work was going to be much tougher than he thought, as innovation heroics as the sole source of deployment of new capabilities are a sign of a dysfunctional organization.

Anthony believed one of his roles as CTO was to:

  • Map and evaluate all the innovation, incubation and technology scouting activities
  • Help the company understand they need innovation and execution to occur simultaneously. (This is the concept of an ambidextrous organization (seethis HBR article).)
  • Educate the company that innovation and execution have different processes, people, and culture. They need each other – and need to respect and depend on each other
  • Create an innovation pipeline – from problem to deployment – and get it adopted at scale

Anthony was hoping that somewhere three, four or five levels down the organization were the real centers of innovation, where existing departments/groups – not individuals – were already accelerating mission/delivering innovative products/services at high speed. His challenge was to

find these islands of innovation and who was running them and understand if/how they

  • Leveraged existing company competencies and assets
  • Understand if/how they co-opted/bypassed existing processes and procedures
  • Had a continuous customer discovery to create products that customers need and want
  • Figured out how to deliver with speed and urgency
  • And if they somehow had made this a repeatable process

If these groups existed, his job as CTO was to take their learning and:

  • Figure out what barriers the innovation groups were running into and help build innovation processes in parallel to those for execution
  • Use their work to create a common language and tools for innovation around rapid acceleration of existing mission and delivery
  • Make permanent delivering products and services at speed with a written innovation doctrine and policy
  • Instrument the process with metrics and diagnostics

Get out of the office
So with another cup of coffee the question we were trying to answer was, how does a newly hired CTO find the real islands of innovation in a company his size?

A first place to start was with the innovation heroes/rebels. They often know where all the innovation bodies were buried. But Anthony’s insight was he needed to get out of his 8th floor office and spend time where his company’s products and services were being developed and delivered.

It was likely that most innovative groups were not simply talking about innovation, but were the ones who rapidly delivering innovative solutions to customer’s needs.

One Last Thing
As we were finishing my coffee Anthony said, “I’m going to let a few of the execs know I’m not out for turf because I only intend to be here for a few years.” I almost spit out the rest of my coffee. I asked how many years the division C-level staff has been at the company. “Some of them for decades” he replied.  I pointed out that in a large organization saying you’re just “visiting” will set you up for failure, as the executives who have made the company their career will simply wait you out.

As he left, he looked at a bit more concerned than we started. “Looks like I have my work cut out for me.”

Lessons Learned

  • Large companies often have divisions and functions with innovation, incubation and technology scouting all operating independently with no common language or tools
  • Innovation heroics as the sole source of deployment of new capabilities are a sign of a dysfunctional organization
  • Innovation isn’t a single activity (incubators, accelerators, hackathons); it is a strategically organized end-to-end process from idea to deployment
  • Somewhere three, four or five levels down the organization are the real centers of innovation – accelerating mission/delivering innovative products/services at high speed
  • The CTO’s job is to:
    • create a common process, language and tools for innovation
    • make them permanent with a written innovation doctrine and policy
  • And don’t ever tell anyone you’re a “short timer”

Artificial Intelligence and Machine Learning– Explained

Artificial Intelligence is a once-in-a lifetime commercial and defense game changer

(download a PDF of this article here)

Hundreds of billions in public and private capital is being invested in Artificial Intelligence (AI) and Machine Learning companies. The number of patents filed in 2021 is more than 30 times higher than in 2015 as companies and countries across the world have realized that AI and Machine Learning will be a major disruptor and potentially change the balance of military power.

Until recently, the hype exceeded reality. Today, however, advances in AI in several important areas (here, here, here, here and here) equal and even surpass human capabilities.

If you haven’t paid attention, now’s the time.

Artificial Intelligence and the Department of Defense (DoD)
The Department of Defense has thought that Artificial Intelligence is such a foundational set of technologies that they started a dedicated organization- the JAIC – to enable and implement artificial intelligence across the Department. They provide the infrastructure, tools, and technical expertise for DoD users to successfully build and deploy their AI-accelerated projects.

Some specific defense related AI applications are listed later in this document.

We’re in the Middle of a Revolution
Imagine it’s 1950, and you’re a visitor who traveled back in time from today. Your job is to explain the impact computers will have on business, defense and society to people who are using manual calculators and slide rules. You succeed in convincing one company and a government to adopt computers and learn to code much faster than their competitors /adversaries. And they figure out how they could digitally enable their business – supply chain, customer interactions, etc. Think about the competitive edge they’d have by today in business or as a nation. They’d steamroll everyone.

That’s where we are today with Artificial Intelligence and Machine Learning. These technologies will transform businesses and government agencies. Today, 100s of billions of dollars in private capital have been invested in 1,000s of AI startups. The U.S. Department of Defense has created a dedicated organization to ensure its deployment.

But What Is It?
Compared to the classic computing we’ve had for the last 75 years, AI has led to new types of applications, e.g. facial recognition; new types of algorithms, e.g. machine learning; new types of computer architectures, e.g. neural nets; new hardware, e.g. GPUs; new types of software developers, e.g. data scientists; all under the overarching theme of artificial intelligence. The sum of these feels like buzzword bingo. But they herald a sea change in what computers are capable of doing, how they do it, and what hardware and software is needed to do it.

This brief will attempt to describe all of it.

New Words to Define Old Things
One of the reasons the world of AI/ML is confusing is that it’s created its own language and vocabulary. It uses new words to define programming steps, job descriptions, development tools, etc. But once you understand how the new world maps onto the classic computing world, it starts to make sense. So first a short list of some key definitions.

AI/ML – a shorthand for Artificial Intelligence/Machine Learning

Artificial Intelligence (AI) – a catchall term used to describe “Intelligent machines” which can solve problems, make/suggest decisions and perform tasks that have traditionally required humans to do. AI is not a single thing, but a constellation of different technologies.

Machine Learning (ML) – a subfield of artificial intelligence. Humans combine data with algorithms (see here for a list) to train a model using that data. This trained model can then make predications on new data (is this picture a cat, a dog or a person?) or decision-making processes (like understanding text and images) without being explicitly programmed to do so.

Machine learning algorithms – computer programs that adjust themselves to perform better as they are exposed to more data. The “learning” part of machine learning means these programs change how they process data over time. In other words, a machine-learning algorithm can adjust its own settings, given feedback on its previous performance in making predictions about a collection of data (images, text, etc.).

Deep Learning/Neural Nets – a subfield of machine learning. Neural networks make up the backbone of deep learning. (The “deep” in deep learning refers to the depth of layers in a neural network.) Neural nets are effective at a variety of tasks (e.g., image classification, speech recognition). A deep learning neural net algorithm is given massive volumes of data, and a task to perform – such as classification. The resulting model is capable of solving complex tasks such as recognizing objects within an image and translating speech in real time. In reality, the neural net is a logical concept that gets mapped onto a physical set of specialized processors. See here.)

Data Science – a new field of computer science. Broadly it encompasses data systems and processes aimed at maintaining data sets and deriving meaning out of them. In the context of AI, it’s the practice of people who are doing machine learning.

Data Scientists – responsible for extracting insights that help businesses make decisions. They explore and analyze data using machine learning platforms to create models about customers, processes, risks, or whatever they’re trying to predict.

What’s Different? Why is Machine Learning Possible Now?
To understand why AI/Machine Learning can do these things, let’s compare them to computers before AI came on the scene. (Warning – simplified examples below.)

Classic Computers

For the last 75 years computers (we’ll call these classic computers) have both shrunk to pocket size (iPhones) and grown to the size of warehouses (cloud data centers), yet they all continued to operate essentially the same way.

Classic Computers – Programming
Classic computers are designed to do anything a human explicitly tells them to do. People (programmers) write software code (programming) to develop applications, thinking a priori about all the rules, logic and knowledge that need to be built in to an application so that it can deliver a specific result. These rules are explicitly coded into a program using a software language (Python, JavaScript, C#, Rust, …).

Classic Computers –  Compiling
The code is then compiled using software to translate the programmer’s source code into a version that can be run on a target computer/browser/phone. For most of today’s programs, the computer used to develop and compile the code does not have to be that much faster than the one that will run it.

Classic Computers – Running/Executing Programs
Once a program is coded and compiled, it can be deployed and run (executed) on a desktop computer, phone, in a browser window, a data center cluster, in special hardware, etc. Programs/applications can be games, social media, office applications, missile guidance systems, bitcoin mining, or even operating systems e.g. Linux, Windows, IOS. These programs run on the same type of classic computer architectures they were programmed in.

Classic Computers – Software Updates, New Features
For programs written for classic computers, software developers receive bug reports, monitor for security breaches, and send out regular software updates that fix bugs, increase performance and at times add new features.

Classic Computers-  Hardware
The CPUs (Central Processing Units) that write and run these Classic Computer applications all have the same basic design (architecture). The CPUs are designed to handle a wide range of tasks quickly in a serial fashion. These CPUs range from Intel X86 chips, and the ARM cores on Apple M1 SoC, to the z15 in IBM mainframes.

Machine Learning

In contrast to programming on classic computing with fixed rules, machine learning is just like it sounds – we can train/teach a computer to “learn by example” by feeding it lots and lots of examples. (For images a rule of thumb is that a machine learning algorithm needs at least 5,000 labeled examples of each category in order to produce an AI model with decent performance.) Once it is trained, the computer runs on its own and can make predictions and/or complex decisions.

Just as traditional programming has three steps – first coding a program, next compiling it and then running it – machine learning also has three steps: training (teaching), pruning and inference (predicting by itself.)

Machine Learning – Training
Unlike programing classic computers with explicit rules, training is the process of “teaching” a computer to perform a task e.g. recognize faces, signals, understand text, etc. (Now you know why you’re asked to click on images of traffic lights, cross walks, stop signs, and buses or type the text of scanned image in ReCaptcha.) Humans provide massive volumes of “training data” (the more data, the better the model’s performance) and select the appropriate algorithm to find the best optimized outcome. (See the detailed “machine learning pipeline” section for the gory details.)

By running an algorithm selected by a data scientist on a set of training data, the Machine Learning system generates the rules embedded in a trained model. The system learns from examples (training data), rather than being explicitly programmed. (See the “Types of Machine Learning” section for more detail.) This self-correction is pretty cool. An input to a neural net results in a guess about what that input is. The neural net then takes its guess and compares it to a ground-truth about the data, effectively asking an expert “Did I get this right?” The difference between the network’s guess and the ground truth is its error. The network measures that error, and walks the error back over its model, adjusting weights to the extent that they contributed to the error.)

Just to make the point again: The algorithms combined with the training data – not external human computer programmers – create the rules that the AI uses. The resulting model is capable of solving complex tasks such as recognizing objects it’s never seen before, translating text or speech, or controlling a drone swarm.

(Instead of building a model from scratch you can now buy, for common machine learning tasks, pretrained models from others and here, much like chip designers buying IP Cores.)

Machine Learning Training – Hardware
Training a machine learning model is a very computationally intensive task. AI hardware must be able to perform thousands of multiplications and additions in a mathematical process called matrix multiplication. It requires specialized chips to run fast. (See the AI semiconductor section for details.)

Machine Learning – Simplification via pruning, quantization, distillation
Just like classic computer code needs to be compiled and optimized before it is deployed on its target hardware, the machine learning models are simplified and modified (pruned) to use less computing power, energy, and  memory before they’re deployed to run on their hardware.

Machine Learning – Inference Phase
Once the system has been trained it can be copied to other devices and run. And the computing hardware can now make inferences (predictions) on new data that the model has never seen before.

Inference can even occur locally on edge devices where physical devices meet the digital world (routers, sensors, IOT devices), close to the source of where the data is generated. This reduces network bandwidth issues and eliminates latency issues.

Machine Learning Inference – Hardware
Inference (running the model) requires substantially less compute power than training. But inference also benefits from specialized AI chips. (See the AI semiconductor section for details.)

Machine Learning – Performance Monitoring and Retraining
Just like classic computers where software developers do regular software updates to fix bugs and increase performance and add features, machine learning models also need to be updated regularly by adding new data to the old training pipelines and running them again. Why?

Over time machine learning models get stale. Their real-world performance generally degrades over time if they are not updated regularly with new training data that matches the changing state of the world. The models need to be monitored and retrained regularly for data and/or concept drift, harmful predictions, performance drops, etc. To stay up to date, the models need to re-learn the patterns by looking at the most recent data that better reflects reality.

One Last Thing – “Verifiability/Explainability”
Understanding how an AI works is essential to fostering trust and confidence in AI production models.

Neural Networks and Deep Learning differ from other types of Machine Learning algorithms in that they have low explainability. They can generate a prediction, but it is very difficult to understand or explain how it arrived at its prediction. This “explainability problem” is often described as a problem for all of AI, but it’s primarily a problem for Neural Networks and Deep Learning. Other types of Machine Learning algorithms – for example decision trees or linear regression– have very high explainability. The results of the five-year DARPA Explainable AI Program (XAI) are worth reading here.

So What Can Machine Learning Do?

It’s taken decades but as of today, on its simplest implementations, machine learning applications can do some tasks better and/or faster than humans. Machine Learning is most advanced and widely applied today in processing text (through Natural Language Processing) followed by understanding images and videos (through Computer Vision) and analytics and anomaly detection. For example:

Recognize and Understand Text/Natural Language Processing
AI is better than humans on basic reading comprehension benchmarks like SuperGLUE and SQuAD and their performance on complex linguistic tasks is almost there. Applications: GPT-3, M6, OPT-175B, Google Translate, Gmail Autocomplete, Chatbots, Text summarization.

Write Human-like Answers to Questions and Assist in Writing Computer Code
An AI can write original text that is indistinguishable from that created by humans. Examples GPT-3, Wu Dao 2.0 or generate computer code. Example GitHub Copilot, Wordtune

Recognize and Understand Images and video streams
An AI can see and understand what it sees. It can identify and detect an object or a feature in an image or video. It can even identify faces. It can scan news broadcasts or read and assess text that appears in videos. It has uses in threat detection –  airport security, banks, and sporting events. In medicine to interpret MRI’s or to design drugs. And in retail to scan and analyze in-store imagery to intuitively determine inventory movement. Examples of ImageNet benchmarks here and here

Turn 2D Images into 3D Rendered Scenes
AI using “NeRFs “neural radiance fields” can take 2d snapshots and render a finished 3D scene in realtime to create avatars or scenes for virtual worlds, to capture video conference participants and their environments in 3D, or to reconstruct scenes for 3D digital maps. The technology is an enabler of the metaverse, generating digital representations of real environments that creators can modify and build on. And self driving cars are using NeRF’s to render city-scale scenes spanning multiple blocks.

Detect Changes in Patterns/Recognize Anomalies
An AI can recognize patterns which don’t match the behaviors expected for a particular system, out of millions of different inputs or transactions. These applications can discover evidence of an attack on financial networks, fraud detection in insurance filings or credit card purchases; identify fake reviews; even tag sensor data in industrial facilities that mean there’s a safety issue. Examples here, here and here.

Power Recommendation Engines
An AI can provide recommendations based on user behaviors used in ecommerce to provide accurate suggestions of products to users for future purchases based on their shopping history. Examples: Netflix, TikTok, CrossingMinds and Recommendations AI

Recognize and Understand Your Voice
An AI can understand spoken language. Then it can comprehend what is being said and in what context. This can enable chatbots to have a conversation with people. It can record and transcribe meetings. (Some versions can even read lips to increase accuracy.) Applications: Siri/Alexa/Google Assistant. Example here

Create Artificial Images
AI can ​create artificial ​images​ (DeepFakes) that ​are​ indistinguishable ​from​ real ​ones using Generative Adversarial Networks.​ Useful in ​entertainment​, virtual worlds, gaming, fashion​ design, etc. Synthetic faces are now indistinguishable and more trustworthy than photos of real people. Paper here.

Create Artist Quality Illustrations from A Written Description
AI can generate images from text descriptions, creating anthropomorphized versions of animals and objects, combining unrelated concepts in plausible ways. An example application is Dall-E

Generative Design of Physical Products
Engineers can input design goals into AI-driven generative design software, along with parameters such as performance or spatial requirements, materials, manufacturing methods, and cost constraints. The software explores all the possible permutations of a solution, quickly generating design alternatives. Example here.

Sentiment Analysis
An AI leverages deep natural language processing, text analysis, and computational linguistics to gain insight into customer opinion, understanding of consumer sentiment, and measuring the impact of marketing strategies. Examples: Brand24, MonkeyLearn

What Does this Mean for Businesses?

Skip this section if you’re interested in national security applications

Hang on to your seat. We’re just at the beginning of the revolution. The next phase of AI, powered by ever increasing powerful AI hardware and cloud clusters, will combine some of these basic algorithms into applications that do things no human can. It will transform business and defense in ways that will create new applications and opportunities.

Human-Machine Teaming
Applications with embedded intelligence have already begun to appear thanks to massive language models. For example – Copilot as a pair-programmer in Microsoft Visual Studio VSCode. It’s not hard to imagine DALL-E 2 as an illustration assistant in a photo editing application, or GPT-3 as a writing assistant in Google Docs.

AI in Medicine
AI applications are already appearing in radiology, dermatology, and oncology. Examples: IDx-DR,OsteoDetect, Embrace2.  AI Medical image identification can automatically detect lesions, and tumors with diagnostics equal to or greater than humans. For Pharma, AI will power drug discovery design for finding new drug candidates. The FDA has a plan for approving AI software here and a list of AI-enabled medical devices here.

Autonomous Vehicles
Harder than it first seemed, but car companies like Tesla will eventually get better than human autonomy for highway driving and eventually city streets.

Decision support
Advanced virtual assistants can listen to and observe behaviors, build and maintain data models, and predict and recommend actions to assist people with and automate tasks that were previously only possible for humans to accomplish.

Supply chain management
AI applications are already appearing in predictive maintenance, risk management, procurement, order fulfillment, supply chain planning and promotion management.

Marketing
AI applications are already appearing in real-time personalization, content and media optimization and campaign orchestration to augment, streamline and automate marketing processes and tasks constrained by human costs and capability, and to uncover new customer insights and accelerate deployment at scale.

Making business smarter: Customer Support
AI applications are already appearing in virtual customer assistants with speech recognition, sentiment analysis, automated/augmented quality assurance and other technologies providing customers with 24/7 self- and assisted-service options across channels.

AI in National Security

Much like the dual-use/dual-nature of classical computers AI developed for commercial applications can also be used for national security.

AI/ML and Ubiquitous Technical Surveillance
AI/ML have made most cities untenable for traditional tradecraft. Machine learning can integrate travel data (customs, airline, train, car rental, hotel, license plate readers…,) integrate feeds from CCTV cameras for facial recognition and gait recognition, breadcrumbs from wireless devices and then combine it with DNA sampling. The result is automated persistent surveillance.

China’s employment of AI as a tool of repression and surveillance of the Uyghurs is a reminder of a dystopian future of how totalitarian regimes will use AI-enabled ubiquitous surveillance to repress and monitor its own populace.

AI/ML on the Battlefield
AI will enable new levels of performance and autonomy for weapon systems. Autonomously collaborating assets (e.g., drone swarms, ground vehicles) that can coordinate attacks, ISR missions, & more.

Fusing and making sense of sensor data (detecting threats in optical /SAR imagery, classifying aircraft based on radar returns, searching for anomalies in radio frequency signatures, etc.) Machine learning is better and faster than humans in finding targets hidden in a high-clutter background. Automated target detection and fires from satellite/UAV.

For example, an Unmanned Aerial Vehicle (UAV) or Unmanned Ground Vehicles with on board AI edge computers could use deep learning to detect and locate concealed chemical, biological and explosive threats by fusing imaging sensors and chemical/biological sensors.

Other examples include:

Use AI/ML countermeasures against adversarial, low probability of intercept/low probability of detection (LPI/LPD) radar techniques in radar and communication systems.

Given sequences of observations of unknown radar waveforms from arbitrary emitters without a priori knowledge, use machine learning to develop behavioral models to enable inference of radar intent and threat level, and to enable prediction of future behaviors.

For objects in space, use machine learning to predict and characterize a spacecrafts possible actions, its subsequent trajectory, and what threats it can pose from along that trajectory. Predict the outcomes of finite burn, continuous thrust, and impulsive maneuvers.

AI empowers other applications such as:

AI/ML in Collection
The front end of intelligence collection platforms has created a firehose of data that have overwhelmed human analysts. “Smart” sensors coupled with inference engines that can pre-process raw intelligence and prioritize what data to transmit and store –helpful in degraded or low-bandwidth environments.

Human-Machine Teaming in Signals Intelligence
Applications with embedded intelligence have already begun to appear in commercial applications thanks to massive language models. For example – Copilot as a pair-programmer in Microsoft Visual Studio VSCode. It’s not hard to imagine an AI that can detect and isolate anomalies and other patterns of interest in all sorts of signal data faster and more reliably than human operators.

AI-enabled natural language processing, computer vision, and audiovisual analysis can vastly reduce manual data processing. Advances in speech-to-text transcription and language analytics now enable reading comprehension, question answering, and automated summarization of large quantities of text. This not only prioritizes the work of human analysts, it’s a major force multiplier

AI can also be used to automate data conversion such as translations and decryptions, accelerating the ability to derive actionable insights.

Human-Machine Teaming in Tasking and Dissemination
AI-enabled systems will automate and optimize tasking and collection for platforms, sensors, and assets in near-real time in response to dynamic intelligence requirements or changes in the environment.

AI will be able to automatically generate machine-readable versions of intelligence products and disseminate them at machine speed so that computer systems across the IC and the military can ingest and use them in real time without manual intervention.

Human-Machine Teaming in Exploitation and Analytics
AI-enabled tools can augment filtering, flagging, and triage across multiple data sets. They can identify connections and correlations more efficiently and at a greater scale than human analysts, and can flag those findings and the most important content for human analysis.

AI can fuse data from multiple sources, types of intelligence, and classification levels to produce accurate predictive analysis in a way that is not currently possible. This can improve indications and warnings for military operations and active cyber defense.

AI/ML Information warfare
Nation states have used AI systems to enhance disinformation campaigns and cyberattacks. This included using “DeepFakes” (fake videos generated by a neural network that are nearly indistinguishable from reality). They are harvesting data on Americans to build profiles of our beliefs, behavior, and biological makeup for tailored attempts to manipulate or coerce individuals.

But because a large percentage of it is open-source AI is not limited to nation states, AI-powered cyber-attacks, deepfakes and AI software paired with commercially available drones can create “poor-man’s smart weapons” for use by rogue states, terrorists and criminals.

AI/ML Cyberwarfare
AI-enabled malware can learn and adapt to a system’s defensive measures, by probing a target system to look for system configuration and operational patterns and customize the attack payload to determine the most opportune time to execute the payload so to maximize the impact. Conversely, AI-enabled cyber-defensive tools can proactively locate and address network anomalies and system vulnerabilities.

Attacks Against AI – Adversarial AI
As AI proliferates, defeating adversaries will be predicated on defeating their AI and vice versa. As Neural Networks take over sensor processing and triage tasks, a human may only be alerted if the AI deems it suspicious. Therefore, we only need to defeat the AI to evade detection, not necessarily a human.

Adversarial attacks against AI fall into three types:

AI Attack Surfaces
Electronic Attack (EA), Electronic Protection (EP), Electronic Support (ES) all have analogues in the AI algorithmic domain. In the future, we may play the same game about the “Algorithmic Spectrum,” denying our adversaries their AI capabilities while defending ours. Other can steal or poison our models  or manipulate our training data.

What Makes AI Possible Now?

 Four changes make Machine Learning possible now:

  1. Massive Data Sets
  2. Improved Machine Learning algorithms
  3. Open-Source Code, Pretrained Models and Frameworks
  4. More computing power

Massive Data Sets
Machine Learning algorithms tend to require large quantities of training data in order to produce high-performance AI models. (Training OpenAI’s GPT-3 Natural Language Model with 175 billion parameters takes 1,024 Nvidia A100 GPUs more than one month.) Today, strategic and tactical sensors pour in a firehose of images, signals and other data. Billions of computers, digital devices and sensors connected to the Internet, producing and storing large volumes of data, which provide other sources of intelligence. For example facial recognition requires millions of labeled images of faces for training data.

Of course more data only helps if the data is relevant to your desired application. Training data needs to match the real-world operational data very, very closely to train a high-performing AI model.

Improved Machine Learning algorithms
The first Machine Learning algorithms are decades old, and some remain incredibly useful. However, researchers have discovered new algorithms that have greatly sped up the fields cutting-edge. These new algorithms have made Machine Learning models more flexible, more robust, and more capable of solving different types of problems.

Open-Source Code, Pretrained Models and Frameworks
Previously, developing Machine Learning systems required a lot of expertise and custom software development that made it out of reach for most organizations. Now open-source code libraries and developer tools allow organizations to use and build upon the work of external communities. No team or organization has to start from scratch, and many parts that used to require highly specialized expertise have been automated. Even non-experts and beginners can create useful AI tools. In some cases, open-source ML models can be entirely reused and purchased. Combined with standard competitions, open source, pretrained models and frameworks have moved the field forward faster than any federal lab or contractor. It’s been a feeding frenzy with the best and brightest researchers trying to one-up each other to prove which ideas are best.

The downside is that, unlike past DoD technology development – where the DoD leads it, can control it, and has the most advanced technology (like stealth and electronic warfare), in most cases the DoD will not have the most advanced algorithms or models. The analogy for AI is closer to microelectronics than it is EW. The path forward for the DoD should be supporting open research, but optimizing on data set collection, harvesting research results, and fast application. 

More computing power – special chips
Machine Learning systems require a lot of computing power. Today, it’s possible to run Machine Learning algorithms on massive datasets using commodity Graphics Processing Units (GPUs). While many of the AI performance improvements have been due to human cleverness on better models and algorithms, most of the performance gains have been the massive increase in compute performance.  (See the semiconductor section.)

More computing power – AI In the Cloud
The rapid growth in the size of machine learning models has been achieved by the move to large data center clusters. The size of machine learning models are limited by time to train them. For example, in training images, the size of the model scales with the number of pixels in an image. ImageNet Model sizes are 224×224 pixels. But HD (1920×1080) images require 40x more computation/memory. Large Natural Language Processing models – e.g. summarizing articles, English-to-Chinese translation like OpenAI’s GPT-3 require enormous models. GPT-3 uses 175 billion parameters and was trained on a cluster with 1,024 Nvidia A100 GPUs that cost ~$25 million! (Which is why large clusters exist in the cloud, or the largest companies/ government agencies.) Facebook’s Deep Learning and Recommendation Model (DLRM) was trained on 1TB data and has 24 billion parameters. Some cloud vendors train on >10TB data sets.

Instead of investing in massive amounts of computers needed for training companies can use the enormous on-demand, off-premises hardware in the cloud (e.g. Amazon AWS, Microsoft Azure) for both training machine learning models and deploying inferences.

We’re Just Getting Started
Progress in AI has been growing exponentially. The next 10 years will see a massive improvement on AI inference and training capabilities. This will require regular refreshes of the hardware– on the chip and cloud clusters – to take advantage. This is the AI version of Moore’s Law on steroids – applications that are completely infeasible today will be easy in 5 years.

What Can’t AI Do?

While AI can do a lot of things better than humans when focused on a narrow objective, there are many things it still can’t do. AI works well in specific domain where you have lots of data, time/resources to train, domain expertise to set the right goals/rewards during training, but that is not always the case.

For example AI models are only as good as the fidelity and quality of the training data. Having bad labels can wreak havoc on your training results. Protecting the integrity of the training data is critical.

In addition, AI is easily fooled by out-of-domain data (things it hasn’t seen before). This can happen by “overfitting” – when a model trains for too long on sample data or when the model is too complex, it can start to learn the “noise,” or irrelevant information, within the dataset. When the model memorizes the noise and fits too closely to the training set, the model becomes “overfitted,” and it is unable to generalize well to new data. If a model cannot generalize well to new data, then it will not be able to perform the classification or prediction tasks it was intended for. However, if you pause too early or exclude too many important features, you may encounter the opposite problem, and instead, you may “underfit” your model. Underfitting occurs when the model has not trained for enough time, or the input variables are not significant enough to determine a meaningful relationship between the input and output variables.

AI is also poor at estimating uncertainty /confidence (and explaining its decision-making). It can’t choose its own goals. (Executives need to define the decision that the AI will execute.  Without well-defined decisions to be made, data scientists will waste time, energy and money.) Except for simple cases an AI can’t (yet) figure out cause and effect or why something happened. It can’t think creatively or apply common sense.

AI is not very good at creating a strategy (unless it can pull from previous examples and mimic them, but then fails with the unexpected.) And it lacks generalized intelligence e.g. that can generalize knowledge and translate learning across domains.

All of these are research topics actively being worked on. Solving these will take a combination of high-performance computing, advanced AI/ML semiconductors, creative machine learning implementations and decision science. Some may be solved in the next decade, at least to a level where a human can’t tell the difference.

Where is AI in Business Going Next?

Skip this section if you’re interested in national security applications

Just as classic computers were applied to a broad set of business, science and military applications, AI is doing the same. AI is exploding not only in research and infrastructure (which go wide) but also in the application of AI to vertical problems (which go deep and depend more than ever on expertise). Some of the new applications on the horizon include Human AI/Teaming (AI helping in programming and decision making), smarter robotics and autonomous vehicles, AI-driven drug discovery and design, healthcare diagnostics, chip electronic design automation, and basic science research.

Advances in language understanding are being pursued to create systems that can summarize complex inputs and engage through human-like conversation, a critical component of next-generation teaming.

Where is AI and National Security Going Next?

In the near future AI may be able to predict the future actions an adversary could take and the actions a friendly force could take to counter these. The 20th century model loop of Observe–Orient–Decide and Act (OODA) is retrospective; an observation cannot be made until after the event has occurred. An AI-enabled decision-making cycle might be ‘sense–predict–agree–act’: AI senses the environment; predicts what the adversary might do and offers what a future friendly force response should be; the human part of the human–machine team agrees with this assessment; and AI acts by sending machine-to-machine instructions to the small, agile and many autonomous warfighting assets deployed en masse across the battlefield.

An example of this is DARPA’s ACE (Air Combat Evolution) program that is developing a warfighting concept for combined arms using a manned and unmanned systems. Humans will fight in close collaboration with autonomous weapon systems in complex environments with tactics informed by artificial intelligence.

A Once-in-a-Generation Event
Imagine it’s the 1980’s and you’re in charge of an intelligence agency. SIGINT and COMINT were analog and RF. You had worldwide collection systems with bespoke systems in space, air, underwater, etc. And you wake up to a world that shifts from copper to fiber. Most of your people, and equipment are going to be obsolete, and you need to learn how to capture those new bits. Almost every business processes needed to change, new organizations needed to be created, new skills were needed, and old ones were obsoleted. That’s what AI/ML is going to do to you and your agency.

The primary obstacle to innovation in national security is not technology, it is culture. The DoD and IC must overcome a host of institutional, bureaucratic, and policy challenges to adopting and integrating these new technologies. Many parts of our culture are resistant to change, reliant on traditional tradecraft and means of collection, and averse to risk-taking, (particularly acquiring and adopting new technologies and integrating outside information sources.)

History tells us that late adopters fall by the wayside as more agile and opportunistic governments master new technologies.

Carpe Diem.

Want more Detail?

Read on if you want to know about Machine Learning chips, see a sample Machine Learning Pipeline and learn about the four types of Machine Learning.

 

Artificial Intelligence/Machine Learning Semiconductors

Skip this section if all you need to know is that special chips are used for AI/ML.

AI/ML, semiconductors, and high-performance computing are intimately intertwined  – and progress in each is dependent on the others.  (See the “Semiconductor Ecosystem” report.)

Some machine learning models can have trillions of parameters and require a massive number of specialized AI chips to run. Edge computers are significantly less powerful than the massive compute power that’s located at data centers and the cloud. They need low power and specialized silicon.

Why Dedicated AI Chips and Chip Speed Matter
Dedicated chips for neutral nets (e.g. Nvidia GPUs, Xilinx FPUs, Google TPUs) are faster than conventional CPUs for three reasons: 1) they use parallelization, 2) they have larger memory bandwidth and 3) they have fast memory access.

There are three types of AI Chips:

  • Graphics Processing Units (GPUs) – Thousands of cores, parallel workloads, widespread use in machine learning
  • Field-Programmable Gate Arrays (FPGAs) – Good for algorithms; compression, video encoding, cryptocurrency,  genomics, search. Needs specialists to program
  • Application-Specific Integrated Circuits (ASICs) – custom chips e.g. Google TPU’s

Matrix multiplication plays a big part in neural network computations, especially if there are many layers and nodes. Graphics Processing Units (GPUs) contain 100s or 1,000s of cores that can do these multiplications simultaneously. And neural networks are inherently parallel which means that it’s easy to run a program across the cores and clusters of these processors. That makes AI chips 10s or even 1,000s of times faster and more efficient than classic CPUs for training and inference of AI algorithms. State-of-the-art AI chips are dramatically more cost-effective than state-of-the-art CPUs as a result of their greater efficiency for AI algorithms.

Cutting-edge AI systems require not only AI-specific chips, but state-of-the-art AI chips. Older AI chips incur huge energy consumption costs that quickly balloon to unaffordable levels. Using older AI chips today means overall costs and slowdowns at least an order of magnitude greater than for state-of- the-art AI chips.

Cost and speed make it virtually impossible to develop and deploy cutting-edge AI algorithms without state-of-the-art AI chips. Even with state-of-the-art AI chips, training a large AI algorithm can cost tens of millions of dollars and take weeks to complete. With general-purpose chips like CPUs or older AI chips, this training would take much longer and cost orders of magnitude more, making staying at the R&D frontier impossible. Similarly, performing inference using less advanced or less specialized chips could involve similar cost overruns and take orders of magnitude longer.

In addition to off-the-shelf AI chips from Nvidia, Xlinix and Intel, large companies like Facebook, Google, Amazon, have designed their own chips to accelerate AI. The opportunity is so large that there are hundreds of AI accelerator startups designing their own chips, funded by 10’s of billions of venture capital and private equity. None of these companies own a chip manufacturing plant (a fab) so they all use a foundry (an independent company that makes chips for others) like TSMC in Taiwan (or SMIC in China for for its defense related silicon.)

A Sample of AI GPU, FPGA and ASIC AI Chips and Where They’re Made

IP (Intellectual Property) Vendors Also Offer AI Accelerators
AI chip designers can buy AI IP Cores – prebuilt AI accelerators from Synopsys (EV7x,) Cadence (Tensilica AI,) Arm (Ethos,) Ceva (SensPro2, NeuPro), Imagination (Series4,) ThinkSilicon (Neox,) FlexLogic (eFPGA,) Edgecortix and others.

Other AI Hardware Architectures
Spiking Neural Networks (SNN) is a completely different approach from Deep Neural Nets. A form of Neuromorphic computing it tries to emulate how a brain works. SNN neurons use simple counters and adders—no matrix multiply hardware is needed and power consumption is much lower. SNNs are good at unsupervised learning – e.g. detecting patterns in unlabeled data streams. Combined with their low power they’re a good fit for sensors at the edge. Examples: BrainChip, GrAI Matter, Innatera, Intel.

Analog Machine Learning AI chips use analog circuits to do the matrix multiplication in memory. The result is extremely low power AI for always-on sensors. Examples: Mythic (AMP,) Aspinity (AML100,) Tetramem.

Optical (Photonics) AI Computation promise performance gains over standard digital silicon, and some are nearing production. They use intersecting coherent light beams rather than switching transistors to perform matrix multiplies. Computation happens in picoseconds and requires only power for the laser. (Though off-chip digital transitions still limit power savings.) Examples: Lightmatter, Lightelligence, Luminous, Lighton.

AI Hardware for the Edge
As more AI moves to the edge, the Edge AI accelerator market is segmenting into high-end chips for camera-based systems and low-power chips for simple sensors. For example:

AI Chips in Autonomous vehicles, Augmented Reality and multicamera surveillance systems These inference engines require high performance. Examples: Nvidia (Orin,) AMD (Versal,) Qualcomm (Cloud AI 100,) and acquired Arriver for automotive software.

AI Chips in Cameras for facial recognition, surveillance. These inference chips require a balance of processing power with low power. Putting an AI chip in each camera reduces latency and bandwidth. Examples: Hailo-8, Ambarella CV5S,  Quadric (Q16), (RealTek 3916N).

Ultralow-Power AI Chips Target IoT Sensors – IoT devices require very simple neural networks and can run for years on a single battery. Example applications: Presence detection, wakeword detection, gunshot detection… Examples: Syntiant (NDP,) Innatera, BrainChip

Running on the edge devices are deep learning models such as OmniMLFoghorn, specifically designed for edge accelerators.

AI/ML Hardware Benchmarks
While there are lots of claims about how much faster each of these chips are for AI/ML there are now a set of standard benchmarks –  MLCommons. These benchmarks were created by Google, Baidu, Stanford, Harvard and U.C. Berkeley.

One Last Thing – Non-Nvidia AI Chips and the “Nvidia Software Moat”
New AI accelerator chips have to cross the software moat that Nvidia has built around their GPU’s. As popular AI applications and frameworks are built on Nvidia CUDA software platform,  if new AI Accelerator vendors want to port these applications to their chips they have to build their own drivers, compiler, debugger, and other tools.

Details of a machine learning pipeline

This is a sample of the workflow (a pipeline) data scientists use to develop, deploy and maintain a machine learning model (see the detailed description here.)

The Types of Machine Learning

skip this section if you want to believe it’s magic.

Machine Learning algorithms fall into four classes:

  1. Supervised Learning
  2. Unsupervised Learning
  3. Semi-supervised Learning
  4. Reinforcement Learning

They differ based on:

  • What types of data their algorithms can work with
  • For supervised and unsupervised learning, whether or not the training data is labeled or unlabeled
  • How the system receives its data inputs

Supervised Learning

  • A “supervisor” (a human or a software system) accurately labels each of the training data inputs with its correct associated output
  • Note that pre-labeled data is only required for the training data that the algorithm uses to train the AI mode
  • In operation in the inference phase the AI will be generating its own labels, the accuracy of which will depend on the AI’s training
  • Supervised Learning can achieve extremely high performance, but they require very large, labeled datasets
  • Using labeled inputs and outputs, the model can measure its accuracy and learn over time
  • For images a rule of thumb is that the algorithm needs at least 5,000 labeled examples of each category in order to produce an AI model with decent performance
  • In supervised learning, the algorithm “learns” from the training dataset by iteratively making predictions on the data and adjusting for the correct answer.
  • While supervised learning models tend to be more accurate than unsupervised learning models, they require upfront human intervention to label the data appropriately.

Supervised Machine Learning – Categories and Examples:

  • Classification problems – use an algorithm to assign data into specific categories, such as separating apples from oranges. Or classify spam in a separate folder from your inbox. Linear classifiers, support vector machines, decision trees and random forest are all common types of classification algorithms.
  • Regression– understands the relationship between dependent and independent variables. Helpful for predicting numerical values based on different data points, such as sales revenue projections for a given business. Some popular regression algorithms are linear regression, logistic regression and polynomial regression.
  • Example algorithms include: Logistic Regression and Back Propagation Neural Networks

Unsupervised Learning

  • These algorithms can analyze and cluster unlabeled data sets. They discover hidden patterns in data without the need for human intervention (hence, they are “unsupervised”)
  • They can extract features from the data without a label for the results
  • For an image classifier, an unsupervised algorithm would not identify the image as a “cat” or a “dog.” Instead, it would sort the training dataset into various groups based on their similarity
  • Unsupervised Learning systems are often less predictable, but as unlabeled data is usually more available than labeled data, they are important
  • Unsupervised algorithms are useful when developers want to understand their own datasets and see what properties might be useful in either developing automation or change operational practices and policies
  • They still require some human intervention for validating the output 

Unsupervised Machine Learning – Categories and Examples

  • Clustering groups unlabeled data based on their similarities or differences. For example, K-means clustering algorithms assign similar data points into groups, where the K value represents the size of the grouping and granularity. This technique is helpful for market segmentation, image compression, etc.
  • Association finds relationships between variables in a given dataset. These methods are frequently used for market basket analysis and recommendation engines, along the lines of “Customers Who Bought This Item Also Bought” recommendations.
  • Dimensionality reduction is used when the number of features  (or dimensions) in a given dataset is too high. It reduces the number of data inputs to a manageable size while also preserving the data integrity. Often, this technique is used in the preprocessing data stage, such as when autoencoders remove noise from visual data to improve picture quality.
  • Example algorithms include: Apriori algorithm and K-Means

Difference between supervised and unsupervised learning

The main difference: Labeled data

  • Goals: In supervised learning, the goal is to predict outcomes for new data. You know up front the type of results to expect. With an unsupervised learning algorithm, the goal is to get insights from large volumes of new data. The machine learning itself determines what is different or interesting from the dataset.
  • Applications: Supervised learning models are ideal for spam detection, sentiment analysis, weather forecasting and pricing predictions, among other things. In contrast, unsupervised learning is a great fit for anomaly detection, recommendation engines, customer personas and medical imaging.
  • ComplexitySupervised learning is a simple method for machine learning, typically calculated through the use of programs like R or Python. In unsupervised learning, you need powerful tools for working with large amounts of unclassified data. Unsupervised learning models are computationally complex because they need a large training set to produce intended outcomes.
  • Drawbacks: Supervised learning models can be time-consuming to train, and the labels for input and output variables require expertise. Meanwhile, unsupervised learning methods can have wildly inaccurate results unless you have human intervention to validate the output variables.

Semi-Supervised Learning

  • “Semi- Supervised” algorithms combine techniques from Supervised and Unsupervised algorithms for applications with a small set of labeled data and a large set of unlabeled data.
  • In practice, using them leads to exactly what you would expect, a mix of some of both of the strengths and weaknesses of Supervised and Unsupervised approaches
  • Typical algorithms are extensions to other flexible methods that make assumptions about how to model the unlabeled data. An example is Generative Adversarial Networks trained on photographs can generate new photographs that look authentic to human observers (deep fakes)

Reinforcement Learning

  • Training data is collected by an autonomous, self-directed AI agent as it perceives its environment and performs goal-directed actions
  • The rewards are input data received by the AI agent when certain criteria are satisfied.
  • These criteria are typically unknown to the agent at the start of training
  • Rewards often contain only partial information. They don’t signal which inputs were good or not
  • The system is learning to take actions to maximize its receipt of cumulative rewards
  • Reinforcement AI can defeat humans– in chess, Go…
  • There are no labeled datasets for every possible move
  • There is no assessment of whether it was a “good or bad move
  • Instead, partial labels reveal the final outcome “win” or “lose”
  • The algorithms explore the space of possible actions to learn the optimal set of rules for determining the best action that maximize wins

Reinforcement Machine Learning – Categories and Examples

  • Algorithm examples include: DQN (Deep Q Network), DDPG (Deep Deterministic Policy Gradient), A3C (Asynchronous Advantage Actor-Critic Algorithm), NAF (Q-Learning with Normalized Advantage Functions), …
  • AlphaGo, a Reinforcement system played 4.9 million games of Go in 3 days against itself to learn how to play the game at a world-champion level
  • Reinforcement is challenging to use in the real world, as the real world is not as heavily bounded as video games and time cannot be sped up in the real world
  • There are consequences to failure in the real world

(download a PDF of this article here)

Sources:


%d bloggers like this: