Ready to champion your DevOps project? Hosted by Mike Kavis, Enterprise Initiatives provides lively discussions with some of the top thinkers and practitioners in the industry, focusing on hot topics including DevOps, PaaS, the Internet of Things and more.
Our guest on the podcast this week is Don Duet, President and COO at Vapor IO. We discuss the many definitions floating around for what edge computing is. Some call it fog computing or MEC (mobile edge computing). It is simple. It is a perspective that as the shift goes from wireless networks and person-to-person interaction to machine-to-machine interaction, underlying architectures must change along with that shift. For things like IoT and distributed data, they require reliability and speed at the source. Increased embeddedness from IoT to healthcare, finance to automotives requires lower latency and need solutions that are mass-scaled to meet distribution requirements. There are many use-cases for edge computing. Many often think first of Fitbits or Amazon Echos. However, it is transforming enterprises and changing entire business models based on the access to real-time data on the edge. It’s used in places like industrial IoT where sensors and sound collection takes place in factories. Managing these at the source helps solve scalability issues. Hospitals are transforming into mobile computing labs. As this technology enters the operating room, it needs to be 100% reliable and fast because lives are on the line. Last, autonomous vehicles and drones use edge computing. Decision-making is brought closer to the device, reducing steps of communication that can result in errors or latency and could cause accidents. Clearly, edge computing is the future in many budding industries.
Our guest on the podcast this week is Shannon Williams, Co-Founder and VP, Sales at Rancher Labs. We discuss how Kubernetes has won the war as a leader in orchestration. However, it is still not easy to use or maintain. We explore what organizations need to consider to build operational efficiencies around the technology. Kubernetes, Docker, and containers are very different from something like VMware and Amazon in terms of adoption. When new technology comes into an organization, they usually would quickly become an IT-led project. With containers, they are more like Puppet, Chef, and Ansible. They pop up in clusters around the organization, started directly by DevOps teams, developers, and users. A lot of times they already exist scattered around an organization and need a framework or logic to manage it all. That is what Rancher does, they act as a services platform to help manage many clusters. Key Considerations for Kubernetes We talk about the key considerations organizations need to think about when trying to deploy Kubernetes across multiple clusters across multi-cloud. Some key points are: Understanding expectations of scale Where will you be running it How to make it highly available Understanding your organization’s tolerance for failure What many don’t realize is that often organizations are running Kubernetes in conjunction with legacy technologies. If you keep up with cloud news, it can sometimes seem like everyone is using the latest technology. In reality, many organizations still use legacy technology like VMs and only make small incremental changes. The two biggest place Rancher runs containers is on VMware and on Amazon. In summary, Kubernetes may prolong the life of legacy technologies like VMware.
Our guest on the podcast this week is Brian Gracely, Director of Product Strategy at Red Hat. Brian discusses a small bank in Ohio and how they have a business use-case for Kubernetes. Because Kubernetes was built to manage Google-sized technology, it is surprising that there is a reason to apply it to a small brick and mortar bank just starting with web and mobile. What they noticed in the switch is that because customers get paid on Fridays and are more likely to check if they got paid on their personal mobile device, as soon as they launched mobile their Friday traffic soared. For just one fifth of business days, Fridays received ten times the amount of traffic as any other day. This unexpected spiky traffic pattern was a great use-case for Kubernetes. Kubernetes was built to deal with problems like that, even at small businesses. We also look at the state of the big three orchestration engines: Kubernetes, Mesos, and Swarm. Kubernetes and Mesos began as internal projects from larger companies. Mesos began in 2014 as a container scheduler Twitter was using to manage its own containers. They released it as open-source, so a community began to form around that. It focused on big data elements because of their applications to Twitter. To this day, Mesos is still preferred to run big data applications compared to the other two. Kubernetes began as Google’s Borg technology, used internally then released open-source. It was focused on the 80/20 type of use-cases such as batch use-cases and container use-cases. What happens when open-source solutions are released is that the community flocks to one and Kubernetes won more of the community than any solution. With a strong community, Kubernetes is better suited to work with many different types of applications. Last, Docker came out with Swarm to compete in the market. They keep things as simple as possible to get a few containers clustered together. Swarm has evolved to work mostly around Docker’s data center products. Tracking this industry over time reveals how containers have evolved to having use-cases in enterprises of all sizes.
We discuss RightScale’s State of the Cloud report analyzing trends in the cloud. RightScale helps customers adopt cloud by helping them with a cloud management and optimization. This is the sixth year of the report so we can start to see trends over time now and there were a few interesting takeaways this year. In the report, RightScale asks two big questions for enterprises. First is about cloud strategy and what their intention is on cloud – to use private, public, or combinations of those. Second, they are asked about what they use today for private or public clouds. From a strategy point of view, people are still focused on multi-cloud with a special focus on hybrid cloud. In strategy, there was a shift away from private-only strategies. Fewer people were saying they plan to use only private cloud or multiple private clouds as their strategies. On adoption, there was a slight drop in people who are already using private cloud from 77% last year down to 72% this year. This may indicate companies who had tried to build their own private cloud with Openstack, and are now backing off from that strategy. The survey found that the average company leverages about four different cloud vendors. This a result of a combination of acquisitions of companies that use different cloud providers and a strategy to leverage different cloud providers. Rightscale asked people for a list of public and private clouds they are running applications on (focused on IaaS and PaaS, not SaaS) and whether they’re experimenting with particular public and private clouds. They found that among people that are using at least one public cloud, they’re running applications in 1.8 public clouds. They are typically experimenting with another 1.8 clouds. Even if they are not using one of the big cloud vendors, they are often at least experimenting with it. The ones that are adopting private cloud are reporting about 2.3 different private clouds. For top challenges this year there was a three-way tie between security, spend and skills (access to skilled resources). Last year skills was highest and it has dropped a bit this year. The people in IT who are concerned about security has been declining each year. Among enterprises, in 2017 over 35% rated cloud security as a significant challenge, and six years ago that number was about 10% higher. We have now reached a tipping point where people realize that when done right, cloud can be as secure if not more secure than a traditional data center. As people adopt public cloud, the cost has been increasing and companies are starting to realize they are inefficient with their spend. On average companies believe they are wasting 30% of their cloud spend. RightScale has found that 30-45% or more is typically what companies are wasting on their cloud spend. The survey found that the more mature a cloud instance is, the more important spend becomes. This year Docker has moved into first place in the list of tools RightScale researches, and while all tools had an increase in usage, Chef and Puppet had a decrease in usage. The survey specifically focuses on configuration management tools and container tools. Docker usage moved from 13% in 2015 to 27% in 2016 to 35% this year, while Chef and Puppet each dropped about 4% this year. The other big increase seen this year was in Kubernetes, which doubled from 7% last year to 14% this year and seems to be in the lead for scheduling and orchestration tools. There is an early trend RightScale noticed that people are starting to use Docker to take advantage of the temporary instances from the cloud providers such as AWS Spot or Google Preemptibles. For people looking to use those, which can mean 70-90% savings on demand, they need the ability to be very portable when they lose their temporary instances, so using Docker along with a container as a service can be helpful in saving those costs. We look at predictions for next year’s State of the Cloud report. Private cloud will likely continue to be under pressure, though we may see a slight uptick with VMware on AWS. It is likely that Docker will continue to grow and that the cost of the cloud will continue to be an ongoing challenge for enterprises.
Our guest on the podcast this week is Gene Kim, DevOps expert and author at "The Phoenix Project" and "The DevOps Handbook." We discuss “The DevOps Handbook”, which was started over five and a half years ago and released in October 2016. Gene co-wrote the book with Jez Humble, Patrick Debois, and John Willis. The book includes over 48 case studies that range from unicorns like Google, Amazon, Facebook, but also horses like Nordstrom, Target, and Capital One. Many of the case studies came from the DevOps Enterprise Summit. The book has discussions of both greenfield and brownfield deployments, even touching on mainframes. The most common question enterprises ask about DevOps is: Where do we start? “The DevOps Handbook” starts with a chapter on starting with DevOps by picking the right value stream. The research started by looking at where successful unicorns and horses started with their DevOps initiatives. They also look at the failed DevOps initiatives to learn what to avoid. Failures can be categorized in two ways: starting too small or starting too big. Initiatives that start too small often start with a simple Chef or Puppet project end up looking like like more of a hobby. When they finish the project, they haven’t actually proven anything and the project gets easily dismissed. Initiatives that start too large often choose something too critical to the operations of the organization, and is unforgiving of mistakes. The most successful journeys start with something that creates a material contribution to the organization, but is small enough that it does not get shut down early for small mistakes. Their studies found that one out of three leaders who were starting these transformations were being promoted for their contributions. There is no one answer for how to change an organization’s culture for DevOps, but there is a prescriptive set of guidelines. The technical practices do not change such as version control, continuous testing, continuous integration, automated deployments, proactive monitoring of the production environment, security integrated into every step. What is different is where these initiatives start from. Some come from the Director ofr Operations, a Chief Architect, or even Director of Development. This transformation starts from different people and teams at different organizations. There are many different ways to reach a great DevOps practice. The three guiding principles of DevOps are: Flow : Maximizing the flow of work and minimizing the lead-time Feedback: Creating check-ins and the equivalent of being able to stop the assembly line Culture: Fostering a culture of continuous experimentation and learning There is a myth that with DevOps you can’t have any central control. We discuss differences in self-service teams like Netflix and Amazon to function-based teams used at Google and Disney. We look at Etsy’s liason model assigning ops engineers to various service and product teams. Many companies do well with the technical aspects of DevOps, but struggle with the culture changes it requires. Ten years from now, it will be about creating learning organizations and the command and control model will not be effective. It won’t be about who caused a problem or who to blame, but will be about creating a culture of learning for a successful team.
Our guest on the podcast this week is Jim Ford, Chief Strategic Architect at ADP. We discuss Docker adoption, DevOps, and security for a company as big as ADP with over 630,000 customers, 35 million users in 100 countries, and 55 million stored social security numbers. We see how DevOps processes and security is never a completed initiative and can always be improved. We also see how DevOps at ADP varies significantly across projects and seems to mold to the team using it.
Our guest on the podcast this week is Rob Hirschfeld, Founder and CEO at RackN. We discuss the differences between Docker Swarm and Kubernetes as well as the debate about how many companies use containers the right way. We also look at hybrid and how it can be easy to lift a container from Amazon and move it to Google, but there will still be key differences in networking and storage between the two that make the shift more complex. Last, we talk about the pros and cons of going all in on one vendor.
Our guests on the podcast this week are Luke Kanies, CEO, andAlanna Brown, Senior Product Marketing Manager at Puppet. We discuss key findings from Puppet's 2016 State of DevOps Report. The report separates high performers by key metrics: deploy frequency, lead time for changes, mean to to recover, and change failure rate. Puppet found that high performers had 200x more frequent deployments than low performers, and 2,555x faster lead times. We look at how IT can take notes from the lean manufacturing model, especially in moving security and quality forward so that they can be closely involved in the initial design of features to avoid future issues.
Our guest on the podcast this week is Kim Weins, VP of Marketing at RightScale. We discuss RightScale’s recent State of the Cloud Report, taking a deep dive into DevOps and Docker adoption. Adoption is increasing across the board for configuration management tools such as Chef, Puppet, and Ansible, but Docker has seen the greatest growth since last year. Last year 13% of the teams surveyed were using Docker, while this year 27% are using it and 35% reported they plan to use it in the future. At this rate of growth, Docker could take the lead on widespread adoption as soon as next year. We also see a trend in large enterprises embracing DevOps, sometimes with more enthusiasm that SMBs. RightScale has taken their own steps to migrate to Docker to save costs and we see how they were successfully able to make the switch.
Our guest on the podcast this week is Michael Heflin, CEO at Sensuron. We discuss Sensuron’s research on fiber optic sensing, which is sure to change the game for internet of things in multiple industries from aerospace to energy to medicine. These sensors may be expensive now, but in less than ten years they are predicted to be offered at a consumer-level price. Small data, or data directly from sensors, is becoming increasingly important as internet of things spreads more widely.
Our guest on the podcast this week is Heather Andrus, Managing Director at Radius. We discuss hardware development for internet of things, overcoming the challenges of getting a new device to work reliably and scaling production. Cloud data storage will be a key factor in the future of hardware and connected devices.
Our guest on the podcast this week is John Dulchinos, VP, Global Automation and 3D Printing at Jabil. Up until 2016, 3D printing was only used for creating models, however on the immediate horizon is a new generation of printers that will allow you to print parts with the quality for real end use. We discuss how designs can be optimized in a number of hours, accelerating the product introduction process. We also discuss how 3D printing will impact traditional manufacturing and design for consumer goods from athletic shoes to internal bone structure and what this means for its corresponding industries
Our guest on the podcast this week is Damon Edwards, Founder atDTO Solutions and SimplifyOps and one of the early thought leaders in the DevOps movement. We discuss why Bimodal IT is the problem, not the solution. It is true that if technology moves too quickly an organization can lose control, but that does not mean the opposite is true and that an organization is completely safe if it moves very slowly. Large enterprises who have been successful at implementing DevOps practices include Ticketmaster, Nordstrom, and unicorn companies like Amazon who have been able to build DevOps processes from scratch as they grew.
Our guest on the podcast this week is Solomon Hykes, Founder and CTO at Docker. We discuss Docker’s upcoming third birthday and how they have been able to grow an entrepreneurial and engineering-focused team that moves quickly. We look at innovations in internet of things and the importance of Docker’s recent acquisition of Unikernel Systems.
Our guest on the podcast this week is Sean Hull, a cloud solutions architect specializing in data & DevOps at scale. We discuss the five tech challenges on Sean’s mind from issues with Docker, Redshift and Hadoop, to DevOps automation, to too much hardware and Amazon’s disruption of VC’s. Sean blogs weekly on his Scalable Startups blog. Find him on twitter [email protected] and LinkedIn.
Our guest on the podcast this week is Rob Hirschfeld, Founder and CEO at RackN. We discuss hybrid devops and why the hybrid cloud is here to stay. We talk about the importance of abstraction layers in allowing each team to use whichever vendor works best for them while continuing to be supportable. We debate the security of containers versus private cloud and how the discussion is changing with the emergence of new companies that monitor security overall and will not run a container with a vulnerability.
We discuss how the industrial and consumer IoT worlds are beginning to collide with technologies such as the Apple Watch and Microsoft Kinect. We also look at the growth of IoT in 2015, specifically focusing on the increase of acquisitions and mergers in the semiconductor space. Finally, we consider the transcendence of drones, robots and augmented reality into mainstream movements in 2016 and beyond.
Our guest on the podcast this week is Matthew Lodge, Chief Operating Officer at Weaveworks. We discuss how speed triumphs over size when competing as a disruptor, which is why microservices and containers are increasingly important to allow companies to keep up. A focus on small, dispersed teams changes the landscape of cloud computing. With many new companies entering the microservices market, we anticipate future trends for the next few years from mergers and acquisitions to IPOs.
Our guest on the podcast this week is Jason Parsons, Senior Architect and Big Data expert at Cloud Technology Partners. We discuss the benefits of real time, big data processing in the cloud and why some companies are hesitant to migrate from their traditional data centers. Jason explains why planning is key when moving from on-premise centers and why all companies should begin their cloud initiatives with a minimum viable product before going into full production.
Our guest on the podcast this week is Preeti Somal, VP of Engineering at Yahoo. We discuss the cloud challenges that come with supporting over a billion consumers per day at Yahoo. One way the Yahoo team simplifies this is through automation and smarter monitoring. Yahoo also leverages open source projects such as Hadoop and OpenStack, which helps onboard employees through their many mergers and acquisitions.