Dispatching containers to cloud VMs » History » Version 27
Tom Clegg, 10/22/2018 06:17 PM
1 | 1 | Tom Clegg | h1. Dispatching containers to cloud VMs |
---|---|---|---|
2 | |||
3 | 6 | Tom Clegg | (Draft) |
4 | 1 | Tom Clegg | |
5 | 11 | Tom Clegg | {{>toc}} |
6 | |||
7 | 7 | Tom Clegg | h2. Component name / purpose |
8 | 1 | Tom Clegg | |
9 | 7 | Tom Clegg | crunch-dispatch-cloud runs Arvados user containers on generic public cloud infrastructure by automatically creating and destroying VMs of various sizes according to demand, preparing the VMs' runtime environments, and running containers on them. |
10 | 1 | Tom Clegg | |
11 | 12 | Tom Clegg | h2. Deployment |
12 | |||
13 | 21 | Tom Clegg | *Where to install:* The crunch-dispatch-cloud process can run anywhere, as long as it has network access to the Arvados controller, the cloud provider's API, and the worker VMs. Each Arvados cluster should run only one crunch-dispatch-cloud process (future versions will support multiple dispatchers). |
14 | |||
15 | *SSH key:* The installer must generate an SSH private key, with no passphrase, for the dispatcher to use when connecting to cloud VMs. The private key is stored in the cluster configuration file. It does not need to be saved in @~/.ssh/@. |
||
16 | |||
17 | *Cloud VM image:* The installer must provide a VM image with an SSH server on port 22, reachable by the dispatcher. The dispatcher's SSH public key must be listed in @/root/.ssh/authorized_keys@. The image should also include suitable versions of docker and crunch-run. _(Alternatively, it is possible to install these using a custom boot probe command, but pre-installing is more efficient.)_ |
||
18 | 22 | Tom Clegg | * (Future) Configurable SSH port number. |
19 | * (Future) Automatically sync the crunch-run binary from the dispatcher host to each worker node. |
||
20 | 12 | Tom Clegg | |
21 | 23 | Tom Clegg | *Cloud provider account:* The dispatcher uses cloud provider credentials to create and delete VMs and other cloud resources. An Arvados user can create an arbitrary number of long-running containers, and the dispatcher will try to run all of them. Currently the dispatcher does not enforce any resource limits of its own, so the installer must ensure the cloud provider itself is enforcing a suitable quota. |
22 | |||
23 | 24 | Tom Clegg | *Migrating from nodemanager/SLURM:* When VM images, SSH keys, and configuration files are ready, disable nodemanager and crunch-dispatch-slurm. Install crunch-dispatch-cloud deb/rpm package. Confirm success with @systemctl status crunch-dispatch-cloud@ and @journalctl -fu crunch-dispatch-cloud@. |
24 | |||
25 | 9 | Tom Clegg | h2. Overview of operation |
26 | 1 | Tom Clegg | |
27 | 9 | Tom Clegg | The dispatcher waits for containers to appear in the queue, and runs them on appropriately sized cloud VMs. When there are no idle cloud VMs with the desired size, the dispatcher brings up more VMs using the cloud provider's API. The dispatcher also shuts down idle VMs that exceed the configured idle timer -- and sooner if the provider starts refusing to create new VMs. |
28 | 1 | Tom Clegg | |
29 | 6 | Tom Clegg | h2. Interaction with other components |
30 | 1 | Tom Clegg | |
31 | 9 | Tom Clegg | Controller (backed by RailsAPI and PostgreSQL) supplies the container queue: which containers the system should be trying to execute (or cancel) at any given time. |
32 | 1 | Tom Clegg | |
33 | 6 | Tom Clegg | The cloud provider's API supplies a list of VMs that exist (or are being created) at a given time and their network addresses, accepts orders to create new VMs, updates instance tags, and (optionally, depending on the driver) obtains the VMs' SSH server public keys. |
34 | 1 | Tom Clegg | |
35 | 6 | Tom Clegg | The SSH server on each cloud VM allows the dispatcher to authenticate with a private key and execute shell commands as root. |
36 | 1 | Tom Clegg | |
37 | 6 | Tom Clegg | h2. Configuration |
38 | 1 | Tom Clegg | |
39 | 6 | Tom Clegg | Arvados configuration (currently a file in /etc) supplies cloud provider credentials, allowed node types, spending limits/policies, etc. |
40 | 1 | Tom Clegg | |
41 | 6 | Tom Clegg | <pre><code class="yaml"> |
42 | CloudVMs: |
||
43 | 8 | Tom Clegg | BootProbeCommand: "docker ps -q" |
44 | 27 | Tom Clegg | SyncInterval: 1m # how often to get list of active instances from cloud provider |
45 | 8 | Tom Clegg | TimeoutIdle: 1m # shutdown if idle longer than this |
46 | TimeoutBooting: 10m # shutdown if exists longer than this without running BootProbeCommand successfully |
||
47 | TimeoutProbe: 2m # shutdown if (after booting) communication fails longer than this, even if ctrs are running |
||
48 | TimeoutShutdown: 1m # shutdown again if node still exists this long after shutdown |
||
49 | 6 | Tom Clegg | Driver: Amazon |
50 | 8 | Tom Clegg | DriverParameters: # following configs are driver dependent |
51 | 6 | Tom Clegg | Region: us-east-1 |
52 | APITimeout: 20s |
||
53 | EC2Key: abcdef |
||
54 | EC2Secret: abcdefghijklmnopqrstuvwxyz |
||
55 | StorageKey: abcdef |
||
56 | StorageSecret: abcdefghijklmnopqrstuvwxyz |
||
57 | ImageID: ami-0123456789abcdef0 |
||
58 | 1 | Tom Clegg | SubnetID: subnet-01234567 |
59 | SecurityGroups: sg-01234567 |
||
60 | 8 | Tom Clegg | Dispatch: |
61 | StaleLockTimeout: 1m # after restart, time to wait for workers to come up before abandoning locks from previous run |
||
62 | PollInterval: 1m # how often to get latest queue from arvados controller |
||
63 | ProbeInterval: 10s # how often to probe each instance for current status/vital signs |
||
64 | MaxProbesPerSecond: 1000 # limit total probe rate for dispatch process (across all instances) |
||
65 | PrivateKey: | # SSH key able to log in as root@ worker VMs |
||
66 | -----BEGIN RSA PRIVATE KEY----- |
||
67 | MIIEowIBAAKCAQEAqYm4XsQHm8sBSZFwUX5VeW1OkGsfoNzcGPG2nzzYRhNhClYZ |
||
68 | 0ABHhUk82HkaC/8l6d/jpYTf42HrK42nNQ0r0Yzs7qw8yZMQioK4Yk+kFyVLF78E |
||
69 | GRG4pGAWXFs6pUchs/lm8fo9zcda4R3XeqgI+NO+nEERXmdRJa1FhI+Za3/S/+CV |
||
70 | mg+6O00wZz2+vKmDPptGN4MCKmQOCKsMJts7wSZGyVcTtdNv7jjfr6yPAIOIL8X7 |
||
71 | LtarBCFaK/pD7uWll/Uj7h7D8K48nIZUrvBJJjXL8Sm4LxCNoz3Z83k8J5ZzuDRD |
||
72 | gRiQe/C085mhO6VL+2fypDLwcKt1tOL8fI81MwIDAQABAoIBACR3tEnmHsDbNOav |
||
73 | Oxq8cwRQh9K2yDHg8BMJgz/TZa4FIx2HEbxVIw0/iLADtJ+Z/XzGJQCIiWQuvtg6 |
||
74 | exoFQESt7JUWRWkSkj9JCQJUoTY9Vl7APtBpqG7rIEQzd3TvzQcagZNRQZQO6rR7 |
||
75 | p8sBdBSZ72lK8cJ9tM3G7Kor/VNK7KgRZFNhEWnmvEa3qMd4hzDcQ4faOn7C9NZK |
||
76 | dwJAuJVVfwOLlOORYcyEkvksLaDOK2DsB/p0AaCpfSmThRbBKN5fPXYaKgUdfp3w |
||
77 | 70Hpp27WWymb1cgjyqSH3DY+V/kvid+5QxgxCBRq865jPLn3FFT9bWEVS/0wvJRj |
||
78 | iMIRrjECgYEA4Ffv9rBJXqVXonNQbbstd2PaprJDXMUy9/UmfHL6pkq1xdBeuM7v |
||
79 | yf2ocXheA8AahHtIOhtgKqwv/aRhVK0ErYtiSvIk+tXG+dAtj/1ZAKbKiFyxjkZV |
||
80 | X72BH7cTlR6As5SRRfWM/HaBGEgED391gKsI5PyMdqWWdczT5KfxAksCgYEAwXYE |
||
81 | ewPmV1GaR5fbh2RupoPnUJPMj36gJCnwls7sGaXDQIpdlq56zfKgrLocGXGgj+8f |
||
82 | QH7FHTJQO15YCYebtsXWwB3++iG43gVlJlecPAydsap2CCshqNWC5JU5pan0QzsP |
||
83 | exzNzWqfUPSbTkR2SRaN+MenZo2Y/WqScOAth7kCgYBgVoLujW9EXH5QfXJpXLq+ |
||
84 | jTvE38I7oVcs0bJwOLPYGzcJtlwmwn6IYAwohgbhV2pLv+EZSs42JPEK278MLKxY |
||
85 | lgVkp60npgunFTWroqDIvdc1TZDVxvA8h9VeODEJlSqxczgbMcIUXBM9yRctTI+5 |
||
86 | 7DiKlMUA4kTFW2sWwuOlFwKBgGXvrYS0FVbFJKm8lmvMu5D5x5RpjEu/yNnFT4Pn |
||
87 | G/iXoz4Kqi2PWh3STl804UF24cd1k94D7hDoReZCW9kJnz67F+C67XMW+bXi2d1O |
||
88 | JIBvlVfcHb1IHMA9YG7ZQjrMRmx2Xj3ce4RVPgUGHh8ra7gvLjd72/Tpf0doNClN |
||
89 | ti/hAoGBAMW5D3LhU05LXWmOqpeT4VDgqk4MrTBcstVe7KdVjwzHrVHCAmI927vI |
||
90 | 1 | Tom Clegg | pjpphWzpC9m3x4OsTNf8m+g6H7f3IiQS0aiFNtduXYlcuT5FHS2fSATTzg5PBon9 |
91 | 1E6BudOve+WyFyBs7hFWAqWFBdWujAl4Qk5Ek09U2ilFEPE7RTgJ |
||
92 | -----END RSA PRIVATE KEY----- |
||
93 | 9 | Tom Clegg | InstanceTypes: |
94 | - Name: m4.large |
||
95 | VCPUs: 2 |
||
96 | RAM: 7782000000 |
||
97 | Scratch: 32000000000 |
||
98 | Price: 0.1 |
||
99 | - Name: m4.large.spot |
||
100 | Preemptible: true |
||
101 | VCPUs: 2 |
||
102 | RAM: 7782000000 |
||
103 | Scratch: 32000000000 |
||
104 | Price: 0.1 |
||
105 | - Name: m4.xlarge |
||
106 | VCPUs: 4 |
||
107 | RAM: 15564000000 |
||
108 | Scratch: 80000000000 |
||
109 | Price: 0.2 |
||
110 | - Name: m4.xlarge.spot |
||
111 | Preemptible: true |
||
112 | VCPUs: 4 |
||
113 | RAM: 15564000000 |
||
114 | Scratch: 80000000000 |
||
115 | Price: 0.2 |
||
116 | - Name: m4.2xlarge |
||
117 | VCPUs: 8 |
||
118 | RAM: 31129000000 |
||
119 | Scratch: 160000000000 |
||
120 | Price: 0.4 |
||
121 | - Name: m4.2xlarge.spot |
||
122 | Preemptible: true |
||
123 | VCPUs: 8 |
||
124 | RAM: 31129000000 |
||
125 | Scratch: 160000000000 |
||
126 | Price: 0.4 |
||
127 | 6 | Tom Clegg | </code></pre> |
128 | 1 | Tom Clegg | |
129 | 10 | Tom Clegg | h2. Management API |
130 | 1 | Tom Clegg | |
131 | 10 | Tom Clegg | APIs for monitoring/diagnostics/control are available via HTTP on a configurable address/port. Request headers must include "Authorization: Bearer {management token}". |
132 | |||
133 | Responses are JSON-encoded and resemble other Arvados APIs: |
||
134 | <pre><code class="json"> |
||
135 | { |
||
136 | "Items": [ |
||
137 | { |
||
138 | "Name": "...", |
||
139 | ... |
||
140 | }, |
||
141 | ... |
||
142 | ] |
||
143 | } |
||
144 | </code></pre> |
||
145 | |||
146 | @GET /arvados/v1/dispatch/instances@ lists cloud VMs. Each returned item includes: |
||
147 | * provider's instance ID |
||
148 | * hourly price (from configuration file) |
||
149 | * instance type (from configuration file) |
||
150 | * instance type (from provider's menu) |
||
151 | * UUID of the current / most recent container attempted (if known) |
||
152 | * time last container finished (or boot time, if nothing run yet) |
||
153 | |||
154 | @GET /arvados/v1/dispatch/containers@ lists queued/locked/running containers. Each returned item includes: |
||
155 | * container UUID |
||
156 | * container state (Queued/Locked/Running/Complete/Cancelled) |
||
157 | * desired instance type |
||
158 | * time appeared in queue |
||
159 | * time started (if started) |
||
160 | |||
161 | @POST /arvados/v1/dispatch/instances/:instance_id/drain@ puts an instance in "drain" state. |
||
162 | * if the instance is currently running a container, it is allowed to continue |
||
163 | * no further containers will be scheduled on the instance |
||
164 | * (TBD) the instance will not be shut down automatically |
||
165 | |||
166 | @POST /arvados/v1/dispatch/instances/:instance_id/shutdown@ puts an instance in "shutdown" state. |
||
167 | * if the instance is currently running a container, the instance is shut down when the container finishes |
||
168 | * otherwise, the instance is shut down immediately |
||
169 | |||
170 | h2. Metrics |
||
171 | |||
172 | 13 | Tom Clegg | Metrics are available via HTTP on a configurable address/port (conventionally :9005). Request headers must include "Authorization: Bearer {management token}". |
173 | 10 | Tom Clegg | |
174 | Metrics include: |
||
175 | 13 | Tom Clegg | * [future] (summary) time elapsed between VM creation and first successful SSH connection to that VM |
176 | * [future] (summary) time elapsed between first successful SSH connection on a VM and ready to run a container on that VM |
||
177 | 10 | Tom Clegg | * (gauge) total hourly price of all existing VMs |
178 | * (gauge) total VCPUs and memory allocated to containers |
||
179 | 1 | Tom Clegg | * (gauge) number of containers running |
180 | 10 | Tom Clegg | * (gauge) number of containers allocated to VMs but not started yet (because VMs are pending/booting) |
181 | * (gauge) number of containers not allocated to VMs (because provider quota is reached) |
||
182 | 13 | Tom Clegg | |
183 | 14 | Tom Clegg | h2. Logs |
184 | |||
185 | 20 | Tom Clegg | For purposes of troubleshooting, a JSON-formatted log entry is printed on stderr when... |
186 | 16 | Tom Clegg | |
187 | 20 | Tom Clegg | | |...including timestamp and...| |
188 | 16 | Tom Clegg | |a new instance is created/ordered |instance type name| |
189 | |an instance appears on the provider's list of instances |instance ID| |
||
190 | |an instance's boot probe succeeds |instance ID| |
||
191 | |an instance is shut down after boot timeout |instance ID, stdout/stderr/error from last boot probe attempt| |
||
192 | |an instance shutdown is requested |instance ID| |
||
193 | |an instance disappears from the provider's list of instances |instance ID and previous state (booting/idle/shutdown)| |
||
194 | |a cloud provider API or driver error occurs |provider/driver's error message| |
||
195 | |a new container appears in the Arvados queue |container UUID, desired instance type name| |
||
196 | |a container is locked by the dispatcher |container UUID| |
||
197 | |a crunch-run process is started on an instance |container UUID, instance ID, crunch-run PID| |
||
198 | |a crunch-run process fails to start on an instance |container UUID, instance ID, stdout/stderr/exitcode| |
||
199 | |a crunch-run process ends |container UUID, instance ID| |
||
200 | |an active container's state changes to Complete or Cancelled |container UUID, new state| |
||
201 | |an active container is requeued after being locked |container UUID| |
||
202 | |an Arvados API error occurs |error message| |
||
203 | |||
204 | 14 | Tom Clegg | |
205 | (Example log entries should be shown here) |
||
206 | |||
207 | If the dispatcher starts with a non-empty ARVADOS_DEBUG environment variable, it also prints more detailed logs about other internal state changes, using level=debug. |
||
208 | 10 | Tom Clegg | |
209 | h2. Internal details |
||
210 | |||
211 | h3. Scheduling policy |
||
212 | |||
213 | 6 | Tom Clegg | The container priority field determines the order in which resources are allocated. |
214 | * If container C1 has priority P1, |
||
215 | * ...and C2 has higher priority P2, |
||
216 | * ...and there is no pending/booting/idle VM suitable for running C2, |
||
217 | * ...then C1 will not be started. |
||
218 | |||
219 | However, containers that run on different VM types don't necessarily start in priority order. |
||
220 | 1 | Tom Clegg | * If container C1 has priority P1, |
221 | * ...and C2 has higher priority P2, |
||
222 | 5 | Peter Amstutz | * ...and there is no idle VM suitable for running C2, |
223 | 6 | Tom Clegg | * ...and there is a pending/booting VM that will be suitable for running C2 when it comes up, |
224 | * ...and there is an idle VM suitable for running C1, |
||
225 | 1 | Tom Clegg | * ...then C1 will start before C2. |
226 | 6 | Tom Clegg | |
227 | 10 | Tom Clegg | h3. Special cases / synchronizing state |
228 | 1 | Tom Clegg | |
229 | 6 | Tom Clegg | When first starting up, dispatcher inspects API server’s container queue and the cloud provider’s list of dispatcher-tagged cloud nodes, and restores internal state accordingly. |
230 | |||
231 | 10 | Tom Clegg | Some containers might have state=Locked at startup. The dispatcher can't be sure these have no corresponding crunch-run process anywhere until it establishes communication with all running instances. To avoid breaking priority order by guessing wrong, the dispatcher avoids scheduling any new containers until all such "stale-locked" containers are matched up with crunch-run processes on existing VMs (typically preparing a docker image) or all of the existing VMs have been probed successfully (meaning the locked containers aren't running anywhere and need to be rescheduled). |
232 | 6 | Tom Clegg | |
233 | 1 | Tom Clegg | When a user cancels a container request with state=Locked or Running, the container priority changes to 0. On its next poll, the dispatcher notices this and kills any corresponding crunch-run processes (or, if there is no such process, just unlocks the container). |
234 | 4 | Peter Amstutz | |
235 | 6 | Tom Clegg | When a crunch-run process ends without finalizing its container's state, the dispatcher notices this and sets state to Cancelled. |
236 | |||
237 | 4 | Peter Amstutz | h3. Probes |
238 | 5 | Peter Amstutz | |
239 | Sometimes (on the happy path) the dispatcher knows the state of each worker, whether it's idle, and which container it's running. In general, it's necessary to probe the worker node itself. |
||
240 | |||
241 | Probe: |
||
242 | * Check whether the SSH connection is alive; reopen if needed. |
||
243 | * Run the configured "ready?" command (e.g., "grep /encrypted-tmp /etc/mtab"); if this fails, conclude the node is still booting. |
||
244 | * Run "crunch-run --list" to get a list of crunch-run supervisors (pid + container UUID) |
||
245 | |||
246 | 6 | Tom Clegg | h3. Detecting dead/lame nodes |
247 | 5 | Peter Amstutz | |
248 | 10 | Tom Clegg | If a node has been up for N seconds without a successful probe, despite at least M attempts, it is shut down, even if it was running a container last time it was contacted successfully. |