I make my living in the software business. I started out as a software developer, moved up to systems engineering and product management, and then on to solutions architecture.
It has made for a great career, but I’ll let you in on a little secret. My true passion is cars and racing. That’s right, I’m a race car guy at heart.
Spending time in both worlds, I started thinking about the changing requirements we often get when we’re creating applications or solutions. Having built and modified a few cars of my own, I also considered what a similar approach would be like for teams building race cars. Challenging is the word that comes to mind. Here’s why.
Let’s say our sponsor comes to us with a simple directive. “Build the fastest car you can.” So, you go and build a top-fuel dragster. Designed for short, straight-line races from a standing start, your light, high-powered car has the quickest acceleration in the world, reaching speeds of over 339 mph (546 kph) in less than three seconds.
In the world of containerized apps, that raw power and speed is the equivalent of major scaling capabilities.
It’s good, but the sponsor wants more. They want the car to be able to handle curves in the track, so the aerodynamics, suspension and steering need to be different. And it needs to be able to handle much longer races, so it needs to be fuel efficient. It also needs real-time monitoring of things like its tires, brakes, clutch, transmission, etc.
You’re essentially being asked to turn your dragster into a Formula One race car. In software, it’s the equivalent of handling completely different functions and workloads by adding all new functionality.
You go back to the drawing board in your garage to see if there’s a way to modify your dragster to meet the new requirement. But as any car person will tell you, it’s just not possible to morph a dragster into an F1 kit. So, you start from scratch and build a car that Jackie Stewart would be proud of.
The sponsor is happy for a few minutes, but then comes back with the need for your car to compete in the Pikes Peak International Hill Climb Race (a rain or shine event). So, you need to retool with an engine that can handle among other things the rapidly descending air density on a track with rapidly rising altitudes. Or is electric the way to go? Last year, a new course record was set by an electric car. If your experience so far has been in fuel cars, how will you handle the very different challenges of electric design?
You get the idea; I don’t want to overdo the car-building analogy. But if you and your team are building applications and solutions today, you’re likely having to deal with these types of fundamental, unanticipated requirement changes. Maybe its competitive pressures or advantage, an acquisition, maybe a complete pivot to fundamentally new technology.
In the old days, these sorts of radical changes in direction would utterly disrupt the development process. But today, using cloud-native architectures and dynamically assembled microservices, one can change a dragster app into an F1 solution, and then pivot your design so the resulting vehicle can get to the top of the mountain first.
But here’s the rub. You can’t do any of this type of transformative work without understanding what the new direction will require in terms of performance and architectural changes. What was once a finely-tuned and largely upfront effort, now needs to be tuned again. What are the peaks and valleys of high traffic, throughput, latency, or concurrent users going to do to your application? How do you best optimize for those scenarios? When the pivot comes, how do you yet again validate the proper optimization? Is the new architecture eveIn relevant to the new tasks at hand?
What if your toolkit included a solution that simplifies and automates the process of continuous Kubernetes optimization?
– Brad Ascar Sr. Solutions Architect, Carbon Relay
I am really excited to join the Carbon Relay team as chief sales officer and wanted to take a few moments to share why.
After eight very fulfilling years at Acquia, I was looking for a rapidly growing company that I could help accelerate. For anyone who has been fortunate enough to participate in building successful businesses there are a few common traits that you look for. Those are clearly in place at Carbon Relay. Let me explain more.
Firstly, the team. This team is really smart. As a go-to-market leader, it’s wonderful to find a company that is made up of exceptionally clever people. The combination of product management expertise, machine learning scientists and software engineers is impressive. Related to this is the culture that permeates through the organization: integrity, trust and empowerment. On top of that is the unwavering support of Insight Partners, which has made a strong commitment to the firm. Having their expertise to draw upon and collaborate with only reinforces the opportunity.
Secondly, the technology. Red Sky Ops was born out of the very real challenges that arise in operating and running Kubernetes at scale. What began in the lab with countless iterations, has emerged as an incredibly innovative technology that uses machine learning to continuously tune and optimize applications running on Kubernetes. It’s that algorithm, along with our team’s deep understanding of use of Kubernetes in production, which is the competitive advantage we bring to the market. And as we grow Carbon Relay and more organizations participate, they will have the opportunity to benefit from the increased learning of the product.
Thirdly, the timing. Since Google released Kubernetes as open source through the Cloud Native Computing Foundation in 2015, the platform has been rapidly adopted by organizations around the world. It delivers unmatched scale and reliability of workloads. Today the industry is seeing the steep rise in the numbers of organizations participating and adopting Kubernetes. For just one measure of the enthusiasm around Kubernetes, consider that the KubeCon+CloudNativeCon event in San Diego last year attracted 12,000 attendees, a 50% increase on the previous year.
At Carbon Relay we have a tremendous opportunity to help organizations around the world intelligently optimize their Kubernetes application performance. This will mean that application service level agreements can be met, that inherent risks in scaling operations can be mitigated and that developer productivity can deliver far higher yields. On top of this, costs can be optimized to ensure efficiency and effectiveness. In looking forward, we envisage a world where every customer gets value from including Red Sky Ops as a step in their CI/CD pipeline process: every application tuned and optimized intelligently and easily.
As we build the sales team, it’s clear that the opportunity is enormous. And my commitment is to build a team that is courageous, trustworthy and energized. I am confident that we can achieve tremendous success for our customers, partners, investors and team mates. I look forward to having you join me in the journey.
Earlier this week we were delighted to issue an announcement with Insight Partners detailing the firm’s $63M investment in Carbon Relay. The funding represents the beginning of a new and accelerated phase of Carbon Relay’s growth. It allows us to intensify development of our machine learning and data science platform, while building out our ability to get the product to more partners and customers.
We created Carbon Relay five years ago because our team is passionate about applying artificial intelligence and deep reinforcement learning to solve some of the toughest problems in IT. Building the company, we leveraged devops processes and depended heavily on cloud-native technologies, including the Kubernetes container orchestration engine. As we deepened our own use of Kubernetes, we quickly learned that it’s not just tough to manage—optimizing applications running on Kubernetes threatened to consume the time of some of our best technologists.
Turns out most organizations using it have the same challenge, as do Kubernetes service providers and even cloud providers. In direct response, Carbon Relay created our Red Sky Ops platform last year to take on the challenge of not simply managing, but optimizing the performance of applications running in Kubernetes environments. The result for customers and partners is more application uptime, less costly resource allocation and—maybe most important—less devops and netops time spent responding to alerts.
Demand for Red Sky Ops pushed Carbon Relay to the next level as a company. We’re thankful for the overwhelming response it has received from the cloud-native community and look forward to applying the investment from Insight Partners to pushing development of the platform. To support development, as well as sales, marketing and operations, we’re intensifying our focus on hiring in Boston and Washington, DC.
In response to an enormous amount of interest in seeing our product, Carbon Relay is also kicking off a series of live webinars with our CTO, Ofer Idan. He’ll demo Red Sky Ops and answer questions in real time, so if you’ve been looking for a technical overview of the platform, please register and join us. The first in the series will take place next Tuesday, February 18, at noon PT / 3 p.m. ET.
As we look ahead, we’re deepening our engagement with the vibrant open source cloud-native community around Kubernetes. Next month we’ll be at KubeCon+CloudNativeCon in Amsterdam to showcase the latest enhancements to Red Sky Ops and hear first hand from developers, engineers and devops pros how they’re managing and optimizing containerized applications. If you’ll be there as well, we hope you’ll stop by and tell us what’s on your mind.
This has been a landmark week for the Carbon Relay team and for the business we’re growing together. I’d like to thank each of our team members, partners and customers for your support. I must also recognize our early investors, who put their faith in us based on our vision and early technology. I’m grateful for your invaluable guidance and support.
Thanks to all who have joined us on this journey—I can’t wait to see where we go next.
Insight Partners investment will fuel the growth and market presence of Carbon Relay as demand grows for the Red Sky Ops machine learning platform for optimizing application performance on Kubernetes
BOSTON—February 11, 2020—Carbon Relay and Insight Partners today announced a $63 million transaction to accelerate the growth of its Red Sky Ops solution for optimizing application performance in Kubernetes environments.
Carbon Relay pioneered the use of machine learning and data science to automate and manage the complexity of finding optimal configurations for applications running in Kubernetes—eliminating the need for manual optimization by DevOps, networking, and IT professionals. With Red Sky Ops, enterprises, cloud providers, and Kubernetes distribution providers ensure better stability for their applications in production, free their most talented technologists from manually configuring applications, and end the costly practice of overprovisioning compute and storage resources to ensure application stability.
“We’re pleased to work with the Carbon Relay team to enable the rapid expansion of the company’s groundbreaking platform,” said Michael Triplett, Managing Director at Insight Partners. “As Kubernetes moves into mainstream use, development and operations teams across industries are attracted to its flexibility. At the same time, they’re dismayed by how difficult it is to use. Carbon Relay has brought together a team of world-class machine learning and engineering experts to bring to market an innovative solution to the challenge of Kubernetes complexity. The Insight team is excited to collaborate with Carbon Relay as the company scales to address this unique opportunity.”
Matt Provo, Co-founder and CEO of Carbon Relay, commented: “With Red Sky Ops, we created the first AIOps solution for optimizing applications in Kubernetes environments to enhance application performance, while also dramatically reducing infrastructure costs. We’ve been gratified by the interest in the platform from across the cloud-native and Kubernetes communities. Working with Insight Partners will enable us to rapidly scale the Red Sky Ops platform at a time when thousands of organizations are looking to enjoy the full benefits of moving to a microservices architecture. The Carbon Relay team looks forward to the next phase of growth of Red Sky Ops and could not be more thrilled to work alongside Mike and the Insight Partners team.”
Solving Kubernetes’ Complexity with Red Sky Ops
Red Sky Ops makes it easy for development and operations teams to manage millions of possible combinations of application variables and configuration settings. With the Carbon Relay platform, they can automatically identify and implement the best configurations for each application in any cloud environment. Customers benefit from lower infrastructure costs and reduced alerts distracting DevOps experts.
Red Sky Ops allows teams to create application deployment experiments and then run them under load to identify optimal application configurations. It automatically creates accurate, comprehensive suggestions for optimal configurations and the best possible application performance on a continuous basis.
About Carbon Relay
Carbon Relay brings together world-class data scientists and software engineers to enable businesses to drive breakthrough IT and operations efficiency. Two traits are woven throughout our team, our products and everything we do: awe of the power of AI to solve complex business problems and a passionate commitment to the environment. Our Red Sky platform uses machine learning to drive major application performance gains and cost reductions in complex environments. Carbon Relay was founded in 2015 and is based in Boston and Washington, DC. Learn more at https://www.carbonrelay.com.
About Insight Partners
Insight Partners is a leading global venture capital and private equity firm investing in high-growth technology and software companies that are driving transformative change in their industries. Founded in 1995, Insight currently has over $20 billion of assets under management and has cumulatively invested in more than 300 companies worldwide. Our mission is to find, fund, and work successfully with visionary executives, providing them with practical, hands-on growth expertise to foster long-term success. Across our people and our portfolio, we encourage a culture around a core belief: growth equals opportunity. For more information on Insight and all its investments, visit www.insightpartners.com or follow us on Twitter @insightpartners.
Managing applications running in Kubernetes can be far more complex and time-consuming than most DevOps, networking and IT professionals expect. The platform’s flexibility is both its strength and weakness. Kubernetes allows experts to tune it to support their organization’s needs, yet it can be so tough that many teams find it frustrating to the point of being unmanageable. Teams are often forced to opt to dramatically overprovision compute and storage resources to ensure application performance, running up unsustainable costs.
To address this complexity, Carbon Relay created Red Sky Ops, an AIOps platform for deploying, scaling and managing containerized applications in Kubernetes environments. It uses machine learning to automatically determine the optimal configuration for apps running in Kubernetes, eliminating the need for ineffective manual optimization.
Using ML-powered experimentation, Red Sky Ops explores the application parameter space, resulting in configurations that both deploy reliably and perform optimally—a nearly impossible task for even the most capable DevOps teams to undertake by hand. We also created Red Sky Ops to learn over time, allowing the platform to become even more efficient over time.
Now, Carbon Relay is collaborating with the IBM Cloud Kubernetes Service, a complete managed container service, to tackle the Kubernetes complexity challenge head-on. I’ve worked with IBM’s Chris Rosen, program director, offering management of the IBM Kubernetes Service, to describe in detail the work we’re doing together to help deliver on the vision of Kubernetes.
In Turning a Glimpse of Kubernetes’ Future into Reality, Chris and I describe the collaboration between IBM and Carbon Relay, and how we’re providing enterprises with new and effective ways to use Kubernetes to achieve their business goals—reliably, efficiently, and flexibly.
We’re excited to announce that from November 18-21, we’ll be at KubeCon + CloudNativeCon 2019 in San Diego. As proud CNCF members, we’re looking forward to joining the Kubernetes community in advocating for the advancement of cloud-native technologies. If you’re attending as well, stop by our booth to learn more about Carbon Relay and how we’re applying AI and machine learning to Kubernetes configuration management.
We know first-hand the challenges of managing containerized applications, and that’s why we built Red Sky Ops, the first AIOps solution specifically designed to make DevOps pros’ lives easier by automatically identifying and implementing the optimal settings for any containerized application, on-premise or in the cloud.
Over the last few months, we’ve steadily enhanced Red Sky Ops’ AI to learn even faster, to better support applications deployed in complex environments. Visit us at KubeCon + CloudNativeCon to see how Red Sky Ops can address the challenges your DevOps team may be experiencing.
This year, Carbon Relay has seen incredible growth—first with our launch, then with the unveiling of Red Sky Ops and most recently through our integration with Helm to support charts. We’re excited about what comes next. Make sure you visit us in the exhibition hall to learn about all of our new integrations and features planned for the coming year.
See you there!
Our VP of AI & Machine Learning, Ofer Idan, breaks down the multidimensional chess game that is configuration management in Kubernetes. Today DevOps & IT teams lose the game too often, but our new Red Sky Ops AIOps solution can help.
Against the backdrop of cloud technologies going mainstream, the enterprise IT migration to containerization in general, and to Kubernetes in particular, is well underway. Some organizations are making the move in response to competitive pressures and the need for greater business agility. Others are making the switch for economic reasons; they want more cost-effective IT operations and see Kubernetes as a smart way to get there.
This momentum of this push to Kubernetes is understandable. Its benefits are too compelling to ignore. For IT operations, it makes applications more portable and scalable than alternatives, simpler to develop, and easier, faster and cheaper to deploy. Essentially, Kubernetes enables companies to support their growth and change in nimble, efficient and cost-effective ways.
That’s the promise. But the reality is that DevOps and IT teams in many organizations still can’t quite get their Kubernetes-powered operations to “fly right.”
The reason is the system’s complexity. This stems partially from the flexibility of Kubernetes, which gives teams seemingly endless options and choices. However, that flexibility morphs into complexity as teams initially work to get their clusters up and running. With their clusters up but applications not performing to their liking, teams then try to tune their apps. That’s when they really hit the complexity wall with Kubernetes.
For organizations that are early in their Kubernetes journey, this complexity makes it difficult for their teams to get applications to deploy reliably and have consistently high performance. For enterprises that are further along in their Kubernetes migrations, complexity is what’s preventing them from realizing their anticipated cost savings.
As for software products that help teams get over their Kubernetes complexity hurdles, the options have been limited. There’s no shortage of services for deploying Kubernetes clusters, and products for monitoring application performance. But to date, there have been no available solutions specifically designed for optimizing how applications run in Kubernetes environments.
Without software-driven options, DevOps and IT teams have tackled it old fashioned way — manually using trial and error. They change one or two variables, then nervously wait to see the impact. Often it’s unclear why changing “A” caused “B” to break, so they keep on tinkering. For businesses where application performance is paramount, such as with SaaS companies or MSPs, their teams often default to costly overprovisioning.
Hence, the complexity-related problems cited above. Some of these occur at the cluster level, like having to decide how large to make nodes and how many of them to create. But many more problems crop up at the application level.
As an example, let’s look at a web app such as an e-commerce site. Minimizing latency is critical for a smooth user experience, so that is a key consideration. To achieve that goal consistently, the app needs to be tuned properly.
When the app is deployed in Kubernetes, it’s up to DevOps or IT team member to select the number of instances, and choose how much CPU, memory, and other types of resources to allocate to each instance. Allocate too few resources, and the app can slow down or even crash. Allocate too many resources, and suddenly the cloud costs skyrocket. Figuring out the “just right” configuration settings, and doing so quickly, accurately and consistently for a growing roster of apps, is a tall order.
The fact is, configuration management in Kubernetes is a multidimensional chess game, and one that DevOps and IT teams are losing too often. To win, and do so consistently, they need a better way forward.
There’s good news for DevOps and IT teams that are presently wrestling with Kubernetes’ complexity. A new, software-driven approach for handling the basics of application configuration in Kubernetes environments has emerged. Powered by advanced machine learning, this new approach eliminates most of this complexity by automatically determining optimal application configuration parameters.
These technologies, which build upon established methods in data science, allow DevOps teams to automate the process of parameter tuning, thereby freeing them to focus on other mission-critical tasks. Using machine learning-powered experimentation, these platforms allow for efficient exploration of the application parameter space, resulting in configurations that are guaranteed to both deploy reliably and perform optimally. As with all powerful ML techniques, the ability to learn over time plays a crucial role in making the process scalable and more efficient. With the help of these technologies, teams can rest assured that development and scaling of their applications will fit naturally into the optimization process, which will become more intelligent with time.
In short, ML-powered approaches for deploying, optimizing, scaling and managing containerized applications in Kubernetes environments are coming into the spotlight. They are proving themselves by intelligently analyzing and managing hundreds of interrelated variables with millions of potential combinations to automatically select the optimal settings for each application.
With our web app example, rather the DevOps team struggling to determine the best parameter values for their app, with these new basics of config optimization, the team gets optimized parameters delivered to them automatically. In addition, the organization and its customers both benefit from a more reliable, high-quality user experience.
It’s all about high performance and reliability with cost-efficiency. By enabling easier and more effective deployment of applications, and ensuring that they are properly resourced and optimally configured, the new, ML-based approach will be a catalyst that creates even more Kubernetes adoption and success. And that’s a very good thing.
Article featured on The NewStack: https://thenewstack.io/the-new-basics-of-configuration-management-in-kubernetes/
Ready to take your Kubernetes environment to the next level? Schedule a demo with our team.