The Road to Cloud Nirvana: The Madrona Venture Group’s View on Serverless

S. Somasegar – Managing Director, Madrona Venture Group

The progression over the last 20 years from on-premise servers, to virtualization, to containerization, to microservices, to event-driven functions and now to serverless computing is allowing software development to become more and more abstracted from the underlying hardware and infrastructure. The combination of serverless computing, microservices, event-driven functions and containers truly form a distributed computing environment that enables developers to build and deploy at-scale distributed applications and services. This abstraction between applications and hardware allows companies and developers to focus on their applications and customers—not worrying about scaling, managing, and operating servers or runtimes.

In today’s cloud world, more and more companies are moving towards serverless products like AWS Lambda to run application backends, respond to voice and chatbot requests, and process streaming data because of the benefits of scaling, availability, cost, and most importantly, the ability to innovate faster because developers no longer need to manage servers. We believe that microservices and serverless functions will form the fabric of the intelligent applications of the future. The massive movement towards containers has validated the market demand for hardware abstraction and the ability to “write once, run anywhere,” and serverless computing is the next stage of this evolution.

Madrona’s Serverless Investing Thesis

Dan Li – Principal, Madrona Venture Group

Today, developers can use products like AWS Lambda, S3, and API Gateway in conjunction with services like Algorithmia, to assemble the right data sources, machine learning models, and business logic to quickly build prototypes and production-ready intelligent applications in a matter of hours. As more companies move towards this mode of application development, we expect to see a massive amount of innovation around AI and machine learning, application of AI to vertically-focused applications, and new applications for IOT devices driven by the ability for companies to build products faster than ever.

For all the above-mentioned reasons, Madrona has made several investments in companies building tools for microservices and serverless computing in the last year and we are continuing to look for opportunities in this space as the cloud infrastructure continues to evolve rapidly.Nevertheless, with the move towards containerization and serverless functions, it can be much harder to monitor application performance, debug applications, and ensure that applications have the correct security and policy settings. For example, SPIFFE (Secure Production Identity Framework for Everyone) provides some great context for the kinds of identity and trust-related work that needs to happen for people to be able to build, share, and consume micro-services in a safe and secure manner.

Below, you’ll hear from three of the startups in our portfolio and how they are building tools to enable developers and enterprises to adopt serverless approaches, or leveraging serverless technologies to innovate faster and better serve their customers.

Portfolio Company Use Cases

Algorithmia Logo

Algorithmia empowers every developer and company to deploy, manage, and share their AI/ML model portfolio with ease. Algorithmia began as the solution to co-founders Kenny Daniel and Diego Oppenheimer’s frustrations at how inaccessible AI/ML algorithms were. Kenny was tired of seeing his algorithms stuck in an unused portion of academia and Diego was tired of recreating algorithms he knew already existed for his work at Microsoft.

Kenny and Diego created Algorithmia as an open marketplace for algorithms in 2013 and today it services over 60,000 developers. From the beginning, Algorithmia has relied on serverless microservices, and this has allowed the company to quickly expand its offerings to include hosting AI/ML models and full enterprise AI Layer services.

AI/ML models are optimally deployed as serverless microservices, which allows them to quickly and effectively scale to handle any influx of data and usage. This is also the most cost-efficient method for consumers who only have to pay for the compute time they use. This empowers data scientists to consume and contribute algorithms at will. Every algorithm committed to the Algorithmia Marketplace is named, tagged, cataloged, and searchable by use case, keyword, or title. This has enabled Algorithmia to become an AWS Lambda Code Library Partner.

In addition to the Algorithm Marketplace, Algorithmia uses the serverless AI Layer to power two additional services: Hosting AI/ML Models and Enterprise Services where they work with government agencies, financial institutions, big pharma, and retail. The AI layer is cloud, stack, and language agnostic. It serves as a data connector, pulling data from any cloud or on-premises server. Developers can input their algorithms in any language (Python, Java, Scala, NodeJS, Rust, Ruby, and R), and a universal REST API will be automatically generated. This allows any consumer to call and chain algorithms in any combination of languages. Operating under a Kubernetes-orchestrated Docker system allows Algorithmia’s services to operate with the highest degree of efficiency.

As companies add AI/ML capabilities across their organizations, they have the opportunity to escape the complications that come with a monolithic application and begin to implement a serverless microservice architecture. Algorithmia provides the expertise and infrastructure to help them be successful.

Pulumi Logo

Pulumisaw an opportunity in 2017 to fundamentally reimagine how developers build and manage modern cloud systems, thanks in large part to the rise in serverless computing intersecting with advances in containers and managed cloud infrastructure in production. By using programming languages and tools that developers are already familiar with, rather than obscure DSLs and less capable, home-grown templating solutions, Pulumi’s customers are able to focus on application development and business logic rather than infrastructure.

As an example, one of Pulumi’s Enterprise customers was able to move from a dedicated team of DevOps Engineers to a combined Engineering organization—reducing their cloud infrastructure scripts to 1/100th the size in a language the entire team already knew and empowering their developers – and is now substantially more productive than ever before in building and continuously deploying new capabilities. The resulting system uses the best of what the modern cloud has to offer—dozens of AWS Lambdas for event-driven tasks, replacing a costly and complex queuing system, several containers that can run in either ECS or Kubernetes, and several managed AWS services like Amazon CloudFront, Amazon Elasticsearch Service, and Amazon ElastiCache—and now runs at a fraction of the cost before migrating to Pulumi. They have been able to spin up entirely new environments in minutes where it used to take weeks.

Before the recent culmination of serverless, containers, and hosted cloud infrastructure, such an approach simply would not have been possible. In fact, we at Pulumi believe that the real magic is in these approaches living in harmony with one another. Each has its own strengths: containers are great for complex stateful systems, often taking existing codebases and moving them to the cloud; serverless functions are perfect for ultra-low-cost event- and API-oriented systems; and hosted infrastructure lets you focus on your application-specific requirements, instead of reinventing the wheel by manually hosting something that your cloud provider can do better and cheaper. Arguably, each is “serverless” in its own way because infrastructure and servers fade into the background. This disruptive sea change has enabled Pulumi to build a single platform and management suite that fully realizes this entire spectrum of technologies.

The future is bright for serverless- and container-oriented cloud architectures, and Pulumi is excited to be right at the center of it helping customers to realize the incredible benefits.

IOpipe co-founders Erica Windisch and Adam Johnson went from virtualizing servers at companies like Docker, to going “all in” on serverless in 2016. Erica and Adam identified serverless as the next revolution in infrastructure, coming roughly 10 years after the launch of AWS EC2. With a shift in computing moving towards a serverless world, there are new challenges that emerge. From dozens of production Lambda user interviews, Erica and Adam identified that one of the major challenges in adopting serverless was a lack of visibility and instrumentation. In 2016, Erica and Adam co-founded IOpipe to focus on helping companies build, ship, and run serverless applications, faster.

IOpipe is an application operations platform built for serverless architectures running AWS Lambda. Through the collection of high fidelity telemetry within Lambda invocations, users can quickly correlate important data points to discover anomalies and identify issues. IOpipe is a cloud-based SaaS offering that offers tracing, profiling, metrics, logs, alerting, and debugging tools to power up operations and development teams.

IOpipe enables developers to debug code faster by providing real-time visibility into their functions as they develop them. Developers can dig deep into what’s really happening under the hood with tools such as profiling and tracing. Once they’re in production, IOpipe then provides a rich set of observability tools to help bubble up issues before they affect end-users. IOpipe saw customers who previously spent days debugging tough issues in production now able to find the root cause in minutes using IOpipe.

Since launching the IOpipe service in Q3 of 2017, the company has seen customers ranging from SaaS startups to large enterprises enabling their developers to build and ship Lambda functions into production at an incredibly rapid pace. What previously took one customer 18 months can now be done in just two months.

IOpipe works closely with AWS as an advanced tier partner, enabling AWS customers to embrace serverless architectures with power-tools such as IOpipe.

IOpipe – A Cloud Native Company

As the cloud increases its footprint in the enterprise, and ‘cloud native’ (aka serverless) development scales up, Madrona is continuing to uncover opportunities that make it easier to innovate and work in this new environment. Today we are proud to announce our seed investment in IOpipe. We co- led this round and teamed up with our colleagues at NEA and Underscore VC to fund the growth of this innovative company. IOpipe has moved its headquarters to Seattle to be near the center of gravity of the public cloud and the company will add team members in the area to support the company’s growth.

We first met Adam Johnson, the founder and CEO of IOpipe, at AWS re:Invent last fall– at TWO different events. After a quick follow-up pitch at the airport, we were hooked. IOpipe is providing important insight into how serverless applications run – and uses that info to diagnose what isn’t going right. AWS Lambda is a strategically important cloud service focused on this market and it has grown substantially into a major piece of how these applications area created. Developers embed IOpipe into their application and can see how functions are performing in realtime via the IOpipe dashboard.

For both developers and operations engineers, this is crucial information that will enable more organizations to develop cloud native serverless applications. IOpipe has already helped analyze and monitor over a billion events from users to date.

IOpipe is also an example of how accelerators are helping drive technology innovation and business. The company was a graduate of NYC Techstars – the founders moved to NYC for the program – which helped them flesh out their ideas, build a product and secure customers. We are supporter of Techstars through the Seattle chapter and love the chance to work with companies early on as they build and find their product market fit.

AWS re:Invent – the Big Announcements and Implications

The momentum continues to build and scale in leaps and bounds. That’s the overwhelming observation and feeling at the end of the 5th annual conference that Amazon hosted in Las Vegas last week for AWS (Amazon Web Services).

Here are some of the key take-aways that we think will have the highest industy impact.

Event-driven Functions and Serverless Computing

Serverless has definitely arrived. As expected, there were a number of new capabilities announced around Lambda, including C# language support, AWS Lambda@Edge to create a “CDN for Compute” and AWS Step Functions to coordinate the components of distributed applications using visual workflows through state machines. Beyond this, it was clear that Lambda, and the serverless approach overall, is being broadly woven into the fabric of AWS services.

In the world of event-driven functions, thinking about a standard way for people to publish events that make it easy to consume those events is going to be critical. Whichever platform gets there first will likely see a tremendous amount of traction.

Innovation in Machine and Deep Learning

AWS has had a machine learning service for a while now, and it was interesting to see a whole slew of new machine learning, deep learning and AI suite of services including Amazon Image Rekognition, Amazon Polly (Text to Speech deep learning service) and Amazon Lex (Natural Language Understanding engine inside Alexa that is now available as a service).

While the concrete use cases are still relatively spare, we – like Amazon – believe this functionality will be integrated into the functionality of virtually all applications in the future.

It is also clear that the proprietary data used to train models are what create differentiated and unique intelligent apps. The distinction between commodity and proprietary data is going to be critical as algorithms become more of a commodity.

Enterprise Credibility

In past years, whether it was intended or unintended, the perception was that taking a bet on AWS meant taking a bet on the public cloud. In other words, there was an unintended consequence of AWS as “all in on public cloud or nothing”. With the VMWare partnership, which was announced a couple months ago, but solidified on stage with VMWare’s CEO, Amazon clearly is supporting the hybrid infrastructure that many enterprises will be dealing with for years to come.

Equally noteworthy was the appearance of Aneel Bhusri, CEO of Workday, on stage to announce that Workday was moving to AWS as their primary cloud for production workloads. Clearly no longer just the realm of primarily dev and test, this is perhaps the strongest statement yet that the public cloud – and AWS in particular – is enterprise production capable.

Moving Up a Layer From a Set of Discrete Services to Solution-based Services

One big theme that showed through this year at AWS was the movement from a set of discrete services to complete solutions both for developers and for operators of applications and services. The beauty of this all is that AWS continues to move forward on this path in a way that is highly empowering for developers and operators.

This approach really shone through during Werner Vogels keynote on Day 2. He laid out AWS’ approach for the “modern data architecture” and then announced how the new service AWS Glue (fully managed data catalog and ETL service) covers all the missing pieces in terms of their end-to-end solution for a modern data architecture on AWS.

Eat the Ecosystem

One of the implications of AWS’ continued growth towards complete solutions is that they continue to eat into the domain of their partner ecosystem. This has been an implied theme in years past, but the pace is accelerating.

Some of the examples that drew the biggest notice:

• AWS X-Ray (analyze and debug distributed applications in production) which aims directly at current monitoring companies like New Relic, AppDynamics and Datadog
• AWS Lightsail (virtual private servers made easy) that, at $5/month, will put significant pressure on companies like Digital Ocean and Linode
• Rekognition (image recognition, part of AI suite described above) that provides a service very similar to Clarifai, who had actually been on a slide just a few prior to the service announcement!

No one should be surprised that AWS’ accelerating expansion will step on the toes of partners. An implication, as @benkepes tweeted, is that the best way to partner and extend AWS is to go very deep for a given use case because AWS will eventually provide the most common horizontal scenarios.

Partner Success = AWS Success

Although some of the new services conflicted with partner offerings, the other side of the coin was that AWS continues to embrace partners and is vested in partners’ success. They clearly understand that having their partners be successful ultimately contributes to more success for AWS. Having customers like Salesforce, WorkDay and Twilio take a complete bet on AWS , making the product of a partner like Chef be available as a fully-managed service on AWS, having a partner like Netflix excited to switch off their last datacenter as they are completely on AWS, and having a company like VMWare embrace AWS as their public cloud partner are some of the great examples of how Amazon is systematically working to ensure that their partners remain successful on AWS, all of which accrues more value and consumption of AWS.

Summary

The cloud opportunity is gigantic and there is room for multiple large players to have a meaningful position strength. However, as of today, Amazon is not just the clear leader but continues to stride forward in an amazing way.

First Published by Geekwire.