Book Image

Docker Orchestration

By : Randall Smith
Book Image

Docker Orchestration

By: Randall Smith

Overview of this book

Docker orchestration is what you need when transitioning from deploying containers individually on a single host to deploying complex multi-container apps on many machines. This book covers the new orchestration features of Docker 1.12 and helps you efficiently build, test, and deploy your application using Docker. You will be shown how to build multi-container applications using Docker Compose. You will also be introduced to the building blocks for multi-host Docker clusters such as registry, overlay networks, and shared storage using practical examples. This book gives an overview of core tools such as Docker Machine, Swarm, and Compose which will enhance your orchestration skills. You’ll learn how to set up a swarm using the decentralized building block. Next, you’ll be shown how to make the most out of the in-built orchestration feature of Docker engine and you’ll use third-party tools such as Kubernetes, Mesosphere, and CoreOS to orchestrate your existing process. Finally, you will learn to deploy cluster hosts on cloud services and automate your infrastructure.
Table of Contents (17 chapters)
Docker Orchestration
Credits
About the Author
About the Reviewer
www.PacktPub.com
Customer Feedback
Preface

Providing persistent storage


Running ephemeral containers like these is all well and good, but more complicated workloads will require some form of additional storage. DC/OS and Marathon provide two options. The first is local storage, which is only available on the node that the task starts on. The second is external volumes, which are available anywhere in the cluster.

Note

As of this writing, volumes are considered experimental. The functionality may change.

Using local volumes

Local persistent volumes are blocks of data allocated on a node that persists across application upgrades and restarts. Because the volume is located on a single node, the application is pinned to that node and will always start there. This means that the data will always be available for the application, but it also means that if that node goes away, the application will not be able to start. The following example shows how to add a local volume:

{ 
  "id" : "nginx-volume", 
  "instances" : 1, 
  "cpus...