Will the Real AI Please Stand Up? - InformationWeek

InformationWeek is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them.Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

IoT
IoT
Data Management // AI/Machine Learning
Commentary
5/20/2019
07:00 AM
Steve Herrod, Managing Director, General Catalyst
Steve Herrod, Managing Director, General Catalyst
Commentary
50%
50%

Will the Real AI Please Stand Up?

Make sure that your use of machine learning has more substance than hype. Here's a framework you can use to help you cut through the noise.

If you’re tuned into the innovation scene right now as an investor, journalist, or the average Jane, you may be wondering if we’re at the precipice of the most exciting time ever to be alive… like ever. I mean, if unbridled excitement of pitch decks and media coverage are to be believed, artificial intelligence (AI) and, more specifically, machine learning (ML), promises to be bigger than jet packs, crypto, and CBD oil all rolled up into one. Algorithms will soon(ish) be able to drive my car, beef up my cybersecurity, AND make me a more perfect chicken pot pie.

As investors who see a high volume of new, early-stage companies, we needed to create a framework for quickly understanding:

  • When are we seeing an actual application of ML?
  • Does the application of ML appreciably improve the product?
  • Will this product make it to market in a realistic timeframe for a venture investment?

So, I created a “Real AI Rubric” to cut through the noise and to pressure test the tech and ideas I'm seeing every day. With this, we’re able to get into a team’s thought process, learn how deep it is, and, in some cases, separate the tech from the marketing claims.

How does your solution improve as you have access to more/different data?

In most cases, ML-driven solutions improve with access to the right data and lots of it. There are several forms of training and models, but at the end of the day, ML solutions find hidden patterns in data and use those patterns to predict answers to subsequent queries. In fact, a very substantial amount of time of “AI work” goes into choosing and cleaning data sources that you’ll use to train your ML model. And even more of that data is used in the testing and tuning of your model. You get the idea.

Technical teams using ML should be able to clearly articulate how access to new and/or improved data sets can get them to better results faster versus if they’d taken a more traditional programming approach of capturing an algorithm in code and then refining that code to get to achieve the same goal -- better results faster.

How is your algorithm different than a linear regression?

Just a few years back, big data was the AI of the moment. Companies were spelunking through tons of structured and unstructured data to come to smarter conclusions using linear regression. A very common approach was (and still is) creating a simple linear regression of numerical data and then using this to predict outcomes. Linear regressions can be powerful and can absolutely be the right application of mathematics in a problem or product. However, it is not ML, nor is it any other form of AI and it shouldn’t be positioned as such. Knowing that a team knows the difference is as important as the difference itself.

What part of your solution could benefit from the use of GPUs, TPUs, or other accelerators?

One reason that Nvidia stock went on such a tear is that the same primitives used in graphics directly apply to a lot of heavily mathematical algorithms. They’re certainly used for statistics acceleration, but additionally, almost all of the ML frameworks have been modified to take advantage of this hardware. The point of this question isn’t just to see how well they understand the role of AI-focused hardware but to see how they’re thinking about the performance of their solution as more and more users and more and more data arrive.

Which open source packages are you using most heavily?

Not surprisingly, open source rules the roost when it comes to modern AI programming, and certain languages and frameworks like Tensorflow, Pytorch, Caffe2 have grown to be the norm. As with so many developer-focused debates, different camps arise with differing reasons for their choice -- and almost always with passion. There are plenty of comparisons available. What’s important here to investors evaluating your pitch isn’t that you’re using one framework over another but that you investigated the options and can thoughtfully explain what you’ve gone with and why.

What’s more, many of these frameworks are now hosted offerings in one of the public clouds and that leads me to my next question:

Are you using AI cloud services or cloud models and if not, why not?

There’s a substantial amount of AI work that goes on directly in the above frameworks, but increasingly the public clouds can hide these details from users and allow them to use a fully hosted API to get an answer. Why set up your own machine learning infrastructure to train and use image recognition models when Google’s Vision APIs are enough? Why work on your own Natural Language Processing (NLP) models if Amazon’s Comprehend can do the job? There are certainly times and reasons when building your own implementations make sense, and I like to hear teams discuss this precise tradeoff.

Which is more of a limiter to your solution? Computation time for training, or computation time for inference?

If the machine learning in an early stage company is more marketing than tech, the teams won’t make it past the first five questions. In cases where a team has convinced me of the basics, we’ll go deeper into their understanding of things like the limitations of each of the main phases of deep learning: Training as you build a model and inference when you use that model. Like several of the questions above, there’s no right answer here as it’s a function of the application. It’s a test of technical chops as much as it is of critical thinking.

If you’re employing -- or planning to employ -- machine learning in a new commercial solution that has the potential to revolutionize a product, process, or even an entire sector, the bottom line is, make sure that your use of ML is more substance than hype.

Steve Herrod, Ph.D., is a managing director at General Catalyst, a venture capital firm with approximately $5 billion in capital raised. At GC, Herrod focuses on investments in security, infrastructure, and SaaS technologies including Datto, Illumio, Menlo Security, and Contrast Security. Prior to joining General Catalyst, he was chief technology officer and SVP of R&D at VMware. Herrod earned a Ph.D. in Computer Science from Stanford University and a BA from the University of Texas at Austin.

 

The InformationWeek community brings together IT practitioners and industry experts with IT advice, education, and opinions. We strive to highlight technology executives and subject matter experts and use their knowledge and experiences to help our audience of IT ... View Full Bio
We welcome your comments on this topic on our social media channels, or [contact us directly] with questions about the site.
Comment  | 
Print  | 
More Insights
Slideshows
How to Land a Job in Cloud Computing
Cynthia Harvey, Freelance Journalist, InformationWeek,  6/19/2019
Commentary
How to Convince Wary Customers to Share Personal Information
John Edwards, Technology Journalist & Author,  6/17/2019
Commentary
The Art and Science of Robot Wrangling in the AI Era
Guest Commentary, Guest Commentary,  6/11/2019
White Papers
Register for InformationWeek Newsletters
Video
Current Issue
A New World of IT Management in 2019
This IT Trend Report highlights how several years of developments in technology and business strategies have led to a subsequent wave of changes in the role of an IT organization, how CIOs and other IT leaders approach management, in addition to the jobs of many IT professionals up and down the org chart.
Slideshows
Flash Poll