/src/systemd/src/core/unit.c
Line | Count | Source (jump to first uncovered line) |
1 | | /* SPDX-License-Identifier: LGPL-2.1+ */ |
2 | | |
3 | | #include <errno.h> |
4 | | #include <stdlib.h> |
5 | | #include <string.h> |
6 | | #include <sys/prctl.h> |
7 | | #include <sys/stat.h> |
8 | | #include <unistd.h> |
9 | | |
10 | | #include "sd-id128.h" |
11 | | #include "sd-messages.h" |
12 | | |
13 | | #include "all-units.h" |
14 | | #include "alloc-util.h" |
15 | | #include "bus-common-errors.h" |
16 | | #include "bus-util.h" |
17 | | #include "cgroup-util.h" |
18 | | #include "dbus-unit.h" |
19 | | #include "dbus.h" |
20 | | #include "dropin.h" |
21 | | #include "escape.h" |
22 | | #include "execute.h" |
23 | | #include "fd-util.h" |
24 | | #include "fileio-label.h" |
25 | | #include "fileio.h" |
26 | | #include "format-util.h" |
27 | | #include "fs-util.h" |
28 | | #include "id128-util.h" |
29 | | #include "io-util.h" |
30 | | #include "load-dropin.h" |
31 | | #include "load-fragment.h" |
32 | | #include "log.h" |
33 | | #include "macro.h" |
34 | | #include "missing.h" |
35 | | #include "mkdir.h" |
36 | | #include "parse-util.h" |
37 | | #include "path-util.h" |
38 | | #include "process-util.h" |
39 | | #include "serialize.h" |
40 | | #include "set.h" |
41 | | #include "signal-util.h" |
42 | | #include "sparse-endian.h" |
43 | | #include "special.h" |
44 | | #include "specifier.h" |
45 | | #include "stat-util.h" |
46 | | #include "stdio-util.h" |
47 | | #include "string-table.h" |
48 | | #include "string-util.h" |
49 | | #include "strv.h" |
50 | | #include "terminal-util.h" |
51 | | #include "tmpfile-util.h" |
52 | | #include "umask-util.h" |
53 | | #include "unit-name.h" |
54 | | #include "unit.h" |
55 | | #include "user-util.h" |
56 | | #include "virt.h" |
57 | | |
58 | | const UnitVTable * const unit_vtable[_UNIT_TYPE_MAX] = { |
59 | | [UNIT_SERVICE] = &service_vtable, |
60 | | [UNIT_SOCKET] = &socket_vtable, |
61 | | [UNIT_TARGET] = &target_vtable, |
62 | | [UNIT_DEVICE] = &device_vtable, |
63 | | [UNIT_MOUNT] = &mount_vtable, |
64 | | [UNIT_AUTOMOUNT] = &automount_vtable, |
65 | | [UNIT_SWAP] = &swap_vtable, |
66 | | [UNIT_TIMER] = &timer_vtable, |
67 | | [UNIT_PATH] = &path_vtable, |
68 | | [UNIT_SLICE] = &slice_vtable, |
69 | | [UNIT_SCOPE] = &scope_vtable, |
70 | | }; |
71 | | |
72 | | static void maybe_warn_about_dependency(Unit *u, const char *other, UnitDependency dependency); |
73 | | |
74 | 36.8k | Unit *unit_new(Manager *m, size_t size) { |
75 | 36.8k | Unit *u; |
76 | 36.8k | |
77 | 36.8k | assert(m); |
78 | 36.8k | assert(size >= sizeof(Unit)); |
79 | 36.8k | |
80 | 36.8k | u = malloc0(size); |
81 | 36.8k | if (!u) |
82 | 0 | return NULL; |
83 | 36.8k | |
84 | 36.8k | u->names = set_new(&string_hash_ops); |
85 | 36.8k | if (!u->names) |
86 | 0 | return mfree(u); |
87 | 36.8k | |
88 | 36.8k | u->manager = m; |
89 | 36.8k | u->type = _UNIT_TYPE_INVALID; |
90 | 36.8k | u->default_dependencies = true; |
91 | 36.8k | u->unit_file_state = _UNIT_FILE_STATE_INVALID; |
92 | 36.8k | u->unit_file_preset = -1; |
93 | 36.8k | u->on_failure_job_mode = JOB_REPLACE; |
94 | 36.8k | u->cgroup_control_inotify_wd = -1; |
95 | 36.8k | u->cgroup_memory_inotify_wd = -1; |
96 | 36.8k | u->job_timeout = USEC_INFINITY; |
97 | 36.8k | u->job_running_timeout = USEC_INFINITY; |
98 | 36.8k | u->ref_uid = UID_INVALID; |
99 | 36.8k | u->ref_gid = GID_INVALID; |
100 | 36.8k | u->cpu_usage_last = NSEC_INFINITY; |
101 | 36.8k | u->cgroup_invalidated_mask |= CGROUP_MASK_BPF_FIREWALL; |
102 | 36.8k | u->failure_action_exit_status = u->success_action_exit_status = -1; |
103 | 36.8k | |
104 | 36.8k | u->ip_accounting_ingress_map_fd = -1; |
105 | 36.8k | u->ip_accounting_egress_map_fd = -1; |
106 | 36.8k | u->ipv4_allow_map_fd = -1; |
107 | 36.8k | u->ipv6_allow_map_fd = -1; |
108 | 36.8k | u->ipv4_deny_map_fd = -1; |
109 | 36.8k | u->ipv6_deny_map_fd = -1; |
110 | 36.8k | |
111 | 36.8k | u->last_section_private = -1; |
112 | 36.8k | |
113 | 36.8k | RATELIMIT_INIT(u->start_limit, m->default_start_limit_interval, m->default_start_limit_burst); |
114 | 36.8k | RATELIMIT_INIT(u->auto_stop_ratelimit, 10 * USEC_PER_SEC, 16); |
115 | 36.8k | |
116 | 184k | for (CGroupIOAccountingMetric i = 0; i < _CGROUP_IO_ACCOUNTING_METRIC_MAX; i++) |
117 | 147k | u->io_accounting_last[i] = UINT64_MAX; |
118 | 36.8k | |
119 | 36.8k | return u; |
120 | 36.8k | } |
121 | | |
122 | 18.0k | int unit_new_for_name(Manager *m, size_t size, const char *name, Unit **ret) { |
123 | 18.0k | _cleanup_(unit_freep) Unit *u = NULL; |
124 | 18.0k | int r; |
125 | 18.0k | |
126 | 18.0k | u = unit_new(m, size); |
127 | 18.0k | if (!u) |
128 | 0 | return -ENOMEM; |
129 | 18.0k | |
130 | 18.0k | r = unit_add_name(u, name); |
131 | 18.0k | if (r < 0) |
132 | 0 | return r; |
133 | 18.0k | |
134 | 18.0k | *ret = TAKE_PTR(u); |
135 | 18.0k | |
136 | 18.0k | return r; |
137 | 18.0k | } |
138 | | |
139 | 1.27M | bool unit_has_name(const Unit *u, const char *name) { |
140 | 1.27M | assert(u); |
141 | 1.27M | assert(name); |
142 | 1.27M | |
143 | 1.27M | return set_contains(u->names, (char*) name); |
144 | 1.27M | } |
145 | | |
146 | 35.2k | static void unit_init(Unit *u) { |
147 | 35.2k | CGroupContext *cc; |
148 | 35.2k | ExecContext *ec; |
149 | 35.2k | KillContext *kc; |
150 | 35.2k | |
151 | 35.2k | assert(u); |
152 | 35.2k | assert(u->manager); |
153 | 35.2k | assert(u->type >= 0); |
154 | 35.2k | |
155 | 35.2k | cc = unit_get_cgroup_context(u); |
156 | 35.2k | if (cc) { |
157 | 27.5k | cgroup_context_init(cc); |
158 | 27.5k | |
159 | 27.5k | /* Copy in the manager defaults into the cgroup |
160 | 27.5k | * context, _before_ the rest of the settings have |
161 | 27.5k | * been initialized */ |
162 | 27.5k | |
163 | 27.5k | cc->cpu_accounting = u->manager->default_cpu_accounting; |
164 | 27.5k | cc->io_accounting = u->manager->default_io_accounting; |
165 | 27.5k | cc->blockio_accounting = u->manager->default_blockio_accounting; |
166 | 27.5k | cc->memory_accounting = u->manager->default_memory_accounting; |
167 | 27.5k | cc->tasks_accounting = u->manager->default_tasks_accounting; |
168 | 27.5k | cc->ip_accounting = u->manager->default_ip_accounting; |
169 | 27.5k | |
170 | 27.5k | if (u->type != UNIT_SLICE) |
171 | 21.3k | cc->tasks_max = u->manager->default_tasks_max; |
172 | 27.5k | } |
173 | 35.2k | |
174 | 35.2k | ec = unit_get_exec_context(u); |
175 | 35.2k | if (ec) { |
176 | 20.7k | exec_context_init(ec); |
177 | 20.7k | |
178 | 20.7k | ec->keyring_mode = MANAGER_IS_SYSTEM(u->manager) ? |
179 | 20.7k | EXEC_KEYRING_SHARED : EXEC_KEYRING_INHERIT; |
180 | 20.7k | } |
181 | 35.2k | |
182 | 35.2k | kc = unit_get_kill_context(u); |
183 | 35.2k | if (kc) |
184 | 21.3k | kill_context_init(kc); |
185 | 35.2k | |
186 | 35.2k | if (UNIT_VTABLE(u)->init) |
187 | 33.9k | UNIT_VTABLE(u)->init(u); |
188 | 35.2k | } |
189 | | |
190 | 36.8k | int unit_add_name(Unit *u, const char *text) { |
191 | 36.8k | _cleanup_free_ char *s = NULL, *i = NULL; |
192 | 36.8k | UnitType t; |
193 | 36.8k | int r; |
194 | 36.8k | |
195 | 36.8k | assert(u); |
196 | 36.8k | assert(text); |
197 | 36.8k | |
198 | 36.8k | if (unit_name_is_valid(text, UNIT_NAME_TEMPLATE)) { |
199 | 0 |
|
200 | 0 | if (!u->instance) |
201 | 0 | return -EINVAL; |
202 | 0 | |
203 | 0 | r = unit_name_replace_instance(text, u->instance, &s); |
204 | 0 | if (r < 0) |
205 | 0 | return r; |
206 | 36.8k | } else { |
207 | 36.8k | s = strdup(text); |
208 | 36.8k | if (!s) |
209 | 0 | return -ENOMEM; |
210 | 36.8k | } |
211 | 36.8k | |
212 | 36.8k | if (set_contains(u->names, s)) |
213 | 0 | return 0; |
214 | 36.8k | if (hashmap_contains(u->manager->units, s)) |
215 | 0 | return -EEXIST; |
216 | 36.8k | |
217 | 36.8k | if (!unit_name_is_valid(s, UNIT_NAME_PLAIN|UNIT_NAME_INSTANCE)) |
218 | 0 | return -EINVAL; |
219 | 36.8k | |
220 | 36.8k | t = unit_name_to_type(s); |
221 | 36.8k | if (t < 0) |
222 | 0 | return -EINVAL; |
223 | 36.8k | |
224 | 36.8k | if (u->type != _UNIT_TYPE_INVALID && t != u->type) |
225 | 0 | return -EINVAL; |
226 | 36.8k | |
227 | 36.8k | r = unit_name_to_instance(s, &i); |
228 | 36.8k | if (r < 0) |
229 | 0 | return r; |
230 | 36.8k | |
231 | 36.8k | if (i && !unit_type_may_template(t)) |
232 | 1.53k | return -EINVAL; |
233 | 35.2k | |
234 | 35.2k | /* Ensure that this unit is either instanced or not instanced, |
235 | 35.2k | * but not both. Note that we do allow names with different |
236 | 35.2k | * instance names however! */ |
237 | 35.2k | if (u->type != _UNIT_TYPE_INVALID && !u->instance != !i) |
238 | 0 | return -EINVAL; |
239 | 35.2k | |
240 | 35.2k | if (!unit_type_may_alias(t) && !set_isempty(u->names)) |
241 | 0 | return -EEXIST; |
242 | 35.2k | |
243 | 35.2k | if (hashmap_size(u->manager->units) >= MANAGER_MAX_NAMES) |
244 | 35.2k | return -E2BIG; |
245 | 35.2k | |
246 | 35.2k | r = set_put(u->names, s); |
247 | 35.2k | if (r < 0) |
248 | 0 | return r; |
249 | 35.2k | assert(r > 0); |
250 | 35.2k | |
251 | 35.2k | r = hashmap_put(u->manager->units, s, u); |
252 | 35.2k | if (r < 0) { |
253 | 0 | (void) set_remove(u->names, s); |
254 | 0 | return r; |
255 | 0 | } |
256 | 35.2k | |
257 | 35.2k | if (u->type == _UNIT_TYPE_INVALID) { |
258 | 35.2k | u->type = t; |
259 | 35.2k | u->id = s; |
260 | 35.2k | u->instance = TAKE_PTR(i); |
261 | 35.2k | |
262 | 35.2k | LIST_PREPEND(units_by_type, u->manager->units_by_type[t], u); |
263 | 35.2k | |
264 | 35.2k | unit_init(u); |
265 | 35.2k | } |
266 | 35.2k | |
267 | 35.2k | s = NULL; |
268 | 35.2k | |
269 | 35.2k | unit_add_to_dbus_queue(u); |
270 | 35.2k | return 0; |
271 | 35.2k | } |
272 | | |
273 | 0 | int unit_choose_id(Unit *u, const char *name) { |
274 | 0 | _cleanup_free_ char *t = NULL; |
275 | 0 | char *s, *i; |
276 | 0 | int r; |
277 | 0 |
|
278 | 0 | assert(u); |
279 | 0 | assert(name); |
280 | 0 |
|
281 | 0 | if (unit_name_is_valid(name, UNIT_NAME_TEMPLATE)) { |
282 | 0 |
|
283 | 0 | if (!u->instance) |
284 | 0 | return -EINVAL; |
285 | 0 | |
286 | 0 | r = unit_name_replace_instance(name, u->instance, &t); |
287 | 0 | if (r < 0) |
288 | 0 | return r; |
289 | 0 | |
290 | 0 | name = t; |
291 | 0 | } |
292 | 0 |
|
293 | 0 | /* Selects one of the names of this unit as the id */ |
294 | 0 | s = set_get(u->names, (char*) name); |
295 | 0 | if (!s) |
296 | 0 | return -ENOENT; |
297 | 0 | |
298 | 0 | /* Determine the new instance from the new id */ |
299 | 0 | r = unit_name_to_instance(s, &i); |
300 | 0 | if (r < 0) |
301 | 0 | return r; |
302 | 0 | |
303 | 0 | u->id = s; |
304 | 0 |
|
305 | 0 | free(u->instance); |
306 | 0 | u->instance = i; |
307 | 0 |
|
308 | 0 | unit_add_to_dbus_queue(u); |
309 | 0 |
|
310 | 0 | return 0; |
311 | 0 | } |
312 | | |
313 | 0 | int unit_set_description(Unit *u, const char *description) { |
314 | 0 | int r; |
315 | 0 |
|
316 | 0 | assert(u); |
317 | 0 |
|
318 | 0 | r = free_and_strdup(&u->description, empty_to_null(description)); |
319 | 0 | if (r < 0) |
320 | 0 | return r; |
321 | 0 | if (r > 0) |
322 | 0 | unit_add_to_dbus_queue(u); |
323 | 0 |
|
324 | 0 | return 0; |
325 | 0 | } |
326 | | |
327 | 71.1k | bool unit_may_gc(Unit *u) { |
328 | 71.1k | UnitActiveState state; |
329 | 71.1k | int r; |
330 | 71.1k | |
331 | 71.1k | assert(u); |
332 | 71.1k | |
333 | 71.1k | /* Checks whether the unit is ready to be unloaded for garbage collection. |
334 | 71.1k | * Returns true when the unit may be collected, and false if there's some |
335 | 71.1k | * reason to keep it loaded. |
336 | 71.1k | * |
337 | 71.1k | * References from other units are *not* checked here. Instead, this is done |
338 | 71.1k | * in unit_gc_sweep(), but using markers to properly collect dependency loops. |
339 | 71.1k | */ |
340 | 71.1k | |
341 | 71.1k | if (u->job) |
342 | 0 | return false; |
343 | 71.1k | |
344 | 71.1k | if (u->nop_job) |
345 | 0 | return false; |
346 | 71.1k | |
347 | 71.1k | state = unit_active_state(u); |
348 | 71.1k | |
349 | 71.1k | /* If the unit is inactive and failed and no job is queued for it, then release its runtime resources */ |
350 | 71.1k | if (UNIT_IS_INACTIVE_OR_FAILED(state) && |
351 | 71.1k | UNIT_VTABLE(u)->release_resources) |
352 | 14.1k | UNIT_VTABLE(u)->release_resources(u); |
353 | 71.1k | |
354 | 71.1k | if (u->perpetual) |
355 | 778 | return false; |
356 | 70.3k | |
357 | 70.3k | if (sd_bus_track_count(u->bus_track) > 0) |
358 | 0 | return false; |
359 | 70.3k | |
360 | 70.3k | /* But we keep the unit object around for longer when it is referenced or configured to not be gc'ed */ |
361 | 70.3k | switch (u->collect_mode) { |
362 | 70.3k | |
363 | 70.3k | case COLLECT_INACTIVE: |
364 | 70.3k | if (state != UNIT_INACTIVE) |
365 | 0 | return false; |
366 | 70.3k | |
367 | 70.3k | break; |
368 | 70.3k | |
369 | 70.3k | case COLLECT_INACTIVE_OR_FAILED: |
370 | 22 | if (!IN_SET(state, UNIT_INACTIVE, UNIT_FAILED)) |
371 | 22 | return false; |
372 | 22 | |
373 | 22 | break; |
374 | 22 | |
375 | 22 | default: |
376 | 0 | assert_not_reached("Unknown garbage collection mode"); |
377 | 70.3k | } |
378 | 70.3k | |
379 | 70.3k | if (u->cgroup_path) { |
380 | 0 | /* If the unit has a cgroup, then check whether there's anything in it. If so, we should stay |
381 | 0 | * around. Units with active processes should never be collected. */ |
382 | 0 |
|
383 | 0 | r = cg_is_empty_recursive(SYSTEMD_CGROUP_CONTROLLER, u->cgroup_path); |
384 | 0 | if (r < 0) |
385 | 0 | log_unit_debug_errno(u, r, "Failed to determine whether cgroup %s is empty: %m", u->cgroup_path); |
386 | 0 | if (r <= 0) |
387 | 0 | return false; |
388 | 70.3k | } |
389 | 70.3k | |
390 | 70.3k | if (UNIT_VTABLE(u)->may_gc && !UNIT_VTABLE(u)->may_gc(u)) |
391 | 0 | return false; |
392 | 70.3k | |
393 | 70.3k | return true; |
394 | 70.3k | } |
395 | | |
396 | 17.2k | void unit_add_to_load_queue(Unit *u) { |
397 | 17.2k | assert(u); |
398 | 17.2k | assert(u->type != _UNIT_TYPE_INVALID); |
399 | 17.2k | |
400 | 17.2k | if (u->load_state != UNIT_STUB || u->in_load_queue) |
401 | 0 | return; |
402 | 17.2k | |
403 | 17.2k | LIST_PREPEND(load_queue, u->manager->load_queue, u); |
404 | 17.2k | u->in_load_queue = true; |
405 | 17.2k | } |
406 | | |
407 | 0 | void unit_add_to_cleanup_queue(Unit *u) { |
408 | 0 | assert(u); |
409 | 0 |
|
410 | 0 | if (u->in_cleanup_queue) |
411 | 0 | return; |
412 | 0 | |
413 | 0 | LIST_PREPEND(cleanup_queue, u->manager->cleanup_queue, u); |
414 | 0 | u->in_cleanup_queue = true; |
415 | 0 | } |
416 | | |
417 | 89.6k | void unit_add_to_gc_queue(Unit *u) { |
418 | 89.6k | assert(u); |
419 | 89.6k | |
420 | 89.6k | if (u->in_gc_queue || u->in_cleanup_queue) |
421 | 71.8k | return; |
422 | 17.7k | |
423 | 17.7k | if (!unit_may_gc(u)) |
424 | 0 | return; |
425 | 17.7k | |
426 | 17.7k | LIST_PREPEND(gc_queue, u->manager->gc_unit_queue, u); |
427 | 17.7k | u->in_gc_queue = true; |
428 | 17.7k | } |
429 | | |
430 | 104k | void unit_add_to_dbus_queue(Unit *u) { |
431 | 104k | assert(u); |
432 | 104k | assert(u->type != _UNIT_TYPE_INVALID); |
433 | 104k | |
434 | 104k | if (u->load_state == UNIT_STUB || u->in_dbus_queue) |
435 | 68.6k | return; |
436 | 35.3k | |
437 | 35.3k | /* Shortcut things if nobody cares */ |
438 | 35.3k | if (sd_bus_track_count(u->manager->subscribed) <= 0 && |
439 | 35.3k | sd_bus_track_count(u->bus_track) <= 0 && |
440 | 35.3k | set_isempty(u->manager->private_buses)) { |
441 | 35.3k | u->sent_dbus_new_signal = true; |
442 | 35.3k | return; |
443 | 35.3k | } |
444 | 0 | |
445 | 0 | LIST_PREPEND(dbus_queue, u->manager->dbus_unit_queue, u); |
446 | 0 | u->in_dbus_queue = true; |
447 | 0 | } |
448 | | |
449 | 0 | void unit_submit_to_stop_when_unneeded_queue(Unit *u) { |
450 | 0 | assert(u); |
451 | 0 |
|
452 | 0 | if (u->in_stop_when_unneeded_queue) |
453 | 0 | return; |
454 | 0 | |
455 | 0 | if (!u->stop_when_unneeded) |
456 | 0 | return; |
457 | 0 | |
458 | 0 | if (!UNIT_IS_ACTIVE_OR_RELOADING(unit_active_state(u))) |
459 | 0 | return; |
460 | 0 | |
461 | 0 | LIST_PREPEND(stop_when_unneeded_queue, u->manager->stop_when_unneeded_queue, u); |
462 | 0 | u->in_stop_when_unneeded_queue = true; |
463 | 0 | } |
464 | | |
465 | 810k | static void bidi_set_free(Unit *u, Hashmap *h) { |
466 | 810k | Unit *other; |
467 | 810k | Iterator i; |
468 | 810k | void *v; |
469 | 810k | |
470 | 810k | assert(u); |
471 | 810k | |
472 | 810k | /* Frees the hashmap and makes sure we are dropped from the inverse pointers */ |
473 | 810k | |
474 | 810k | HASHMAP_FOREACH_KEY(v, other, h, i) { |
475 | 49.6k | UnitDependency d; |
476 | 49.6k | |
477 | 1.14M | for (d = 0; d < _UNIT_DEPENDENCY_MAX; d++) |
478 | 1.09M | hashmap_remove(other->dependencies[d], u); |
479 | 49.6k | |
480 | 49.6k | unit_add_to_gc_queue(other); |
481 | 49.6k | } |
482 | 810k | |
483 | 810k | hashmap_free(h); |
484 | 810k | } |
485 | | |
486 | 36.8k | static void unit_remove_transient(Unit *u) { |
487 | 36.8k | char **i; |
488 | 36.8k | |
489 | 36.8k | assert(u); |
490 | 36.8k | |
491 | 36.8k | if (!u->transient) |
492 | 36.8k | return; |
493 | 0 | |
494 | 0 | if (u->fragment_path) |
495 | 0 | (void) unlink(u->fragment_path); |
496 | 0 |
|
497 | 0 | STRV_FOREACH(i, u->dropin_paths) { |
498 | 0 | _cleanup_free_ char *p = NULL, *pp = NULL; |
499 | 0 |
|
500 | 0 | p = dirname_malloc(*i); /* Get the drop-in directory from the drop-in file */ |
501 | 0 | if (!p) |
502 | 0 | continue; |
503 | 0 | |
504 | 0 | pp = dirname_malloc(p); /* Get the config directory from the drop-in directory */ |
505 | 0 | if (!pp) |
506 | 0 | continue; |
507 | 0 | |
508 | 0 | /* Only drop transient drop-ins */ |
509 | 0 | if (!path_equal(u->manager->lookup_paths.transient, pp)) |
510 | 0 | continue; |
511 | 0 | |
512 | 0 | (void) unlink(*i); |
513 | 0 | (void) rmdir(p); |
514 | 0 | } |
515 | 0 | } |
516 | | |
517 | 36.8k | static void unit_free_requires_mounts_for(Unit *u) { |
518 | 36.8k | assert(u); |
519 | 36.8k | |
520 | 38.6k | for (;;) { |
521 | 38.6k | _cleanup_free_ char *path; |
522 | 38.6k | |
523 | 38.6k | path = hashmap_steal_first_key(u->requires_mounts_for); |
524 | 38.6k | if (!path) |
525 | 36.8k | break; |
526 | 1.80k | else { |
527 | 1.80k | char s[strlen(path) + 1]; |
528 | 1.80k | |
529 | 176k | PATH_FOREACH_PREFIX_MORE(s, path) { |
530 | 176k | char *y; |
531 | 176k | Set *x; |
532 | 176k | |
533 | 176k | x = hashmap_get2(u->manager->units_requiring_mounts_for, s, (void**) &y); |
534 | 176k | if (!x) |
535 | 14.9k | continue; |
536 | 161k | |
537 | 161k | (void) set_remove(x, u); |
538 | 161k | |
539 | 161k | if (set_isempty(x)) { |
540 | 161k | (void) hashmap_remove(u->manager->units_requiring_mounts_for, y); |
541 | 161k | free(y); |
542 | 161k | set_free(x); |
543 | 161k | } |
544 | 161k | } |
545 | 1.80k | } |
546 | 38.6k | } |
547 | 36.8k | |
548 | 36.8k | u->requires_mounts_for = hashmap_free(u->requires_mounts_for); |
549 | 36.8k | } |
550 | | |
551 | 36.8k | static void unit_done(Unit *u) { |
552 | 36.8k | ExecContext *ec; |
553 | 36.8k | CGroupContext *cc; |
554 | 36.8k | |
555 | 36.8k | assert(u); |
556 | 36.8k | |
557 | 36.8k | if (u->type < 0) |
558 | 1.53k | return; |
559 | 35.2k | |
560 | 35.2k | if (UNIT_VTABLE(u)->done) |
561 | 27.6k | UNIT_VTABLE(u)->done(u); |
562 | 35.2k | |
563 | 35.2k | ec = unit_get_exec_context(u); |
564 | 35.2k | if (ec) |
565 | 20.7k | exec_context_done(ec); |
566 | 35.2k | |
567 | 35.2k | cc = unit_get_cgroup_context(u); |
568 | 35.2k | if (cc) |
569 | 27.5k | cgroup_context_done(cc); |
570 | 35.2k | } |
571 | | |
572 | 36.8k | void unit_free(Unit *u) { |
573 | 36.8k | UnitDependency d; |
574 | 36.8k | Iterator i; |
575 | 36.8k | char *t; |
576 | 36.8k | |
577 | 36.8k | if (!u) |
578 | 0 | return; |
579 | 36.8k | |
580 | 36.8k | if (UNIT_ISSET(u->slice)) { |
581 | 2.60k | /* A unit is being dropped from the tree, make sure our parent slice recalculates the member mask */ |
582 | 2.60k | unit_invalidate_cgroup_members_masks(UNIT_DEREF(u->slice)); |
583 | 2.60k | |
584 | 2.60k | /* And make sure the parent is realized again, updating cgroup memberships */ |
585 | 2.60k | unit_add_to_cgroup_realize_queue(UNIT_DEREF(u->slice)); |
586 | 2.60k | } |
587 | 36.8k | |
588 | 36.8k | u->transient_file = safe_fclose(u->transient_file); |
589 | 36.8k | |
590 | 36.8k | if (!MANAGER_IS_RELOADING(u->manager)) |
591 | 36.8k | unit_remove_transient(u); |
592 | 36.8k | |
593 | 36.8k | bus_unit_send_removed_signal(u); |
594 | 36.8k | |
595 | 36.8k | unit_done(u); |
596 | 36.8k | |
597 | 36.8k | unit_dequeue_rewatch_pids(u); |
598 | 36.8k | |
599 | 36.8k | sd_bus_slot_unref(u->match_bus_slot); |
600 | 36.8k | sd_bus_track_unref(u->bus_track); |
601 | 36.8k | u->deserialized_refs = strv_free(u->deserialized_refs); |
602 | 36.8k | |
603 | 36.8k | unit_free_requires_mounts_for(u); |
604 | 36.8k | |
605 | 36.8k | SET_FOREACH(t, u->names, i) |
606 | 36.8k | hashmap_remove_value(u->manager->units, t, u); |
607 | 36.8k | |
608 | 36.8k | if (!sd_id128_is_null(u->invocation_id)) |
609 | 0 | hashmap_remove_value(u->manager->units_by_invocation_id, &u->invocation_id, u); |
610 | 36.8k | |
611 | 36.8k | if (u->job) { |
612 | 0 | Job *j = u->job; |
613 | 0 | job_uninstall(j); |
614 | 0 | job_free(j); |
615 | 0 | } |
616 | 36.8k | |
617 | 36.8k | if (u->nop_job) { |
618 | 0 | Job *j = u->nop_job; |
619 | 0 | job_uninstall(j); |
620 | 0 | job_free(j); |
621 | 0 | } |
622 | 36.8k | |
623 | 846k | for (d = 0; d < _UNIT_DEPENDENCY_MAX; d++) |
624 | 810k | bidi_set_free(u, u->dependencies[d]); |
625 | 36.8k | |
626 | 36.8k | if (u->on_console) |
627 | 0 | manager_unref_console(u->manager); |
628 | 36.8k | |
629 | 36.8k | unit_release_cgroup(u); |
630 | 36.8k | |
631 | 36.8k | if (!MANAGER_IS_RELOADING(u->manager)) |
632 | 36.8k | unit_unlink_state_files(u); |
633 | 36.8k | |
634 | 36.8k | unit_unref_uid_gid(u, false); |
635 | 36.8k | |
636 | 36.8k | (void) manager_update_failed_units(u->manager, u, false); |
637 | 36.8k | set_remove(u->manager->startup_units, u); |
638 | 36.8k | |
639 | 36.8k | unit_unwatch_all_pids(u); |
640 | 36.8k | |
641 | 36.8k | unit_ref_unset(&u->slice); |
642 | 39.2k | while (u->refs_by_target) |
643 | 2.39k | unit_ref_unset(u->refs_by_target); |
644 | 36.8k | |
645 | 36.8k | if (u->type != _UNIT_TYPE_INVALID) |
646 | 36.8k | LIST_REMOVE(units_by_type, u->manager->units_by_type[u->type], u); |
647 | 36.8k | |
648 | 36.8k | if (u->in_load_queue) |
649 | 36.8k | LIST_REMOVE(load_queue, u->manager->load_queue, u); |
650 | 36.8k | |
651 | 36.8k | if (u->in_dbus_queue) |
652 | 36.8k | LIST_REMOVE(dbus_queue, u->manager->dbus_unit_queue, u); |
653 | 36.8k | |
654 | 36.8k | if (u->in_gc_queue) |
655 | 36.8k | LIST_REMOVE(gc_queue, u->manager->gc_unit_queue, u); |
656 | 36.8k | |
657 | 36.8k | if (u->in_cgroup_realize_queue) |
658 | 36.8k | LIST_REMOVE(cgroup_realize_queue, u->manager->cgroup_realize_queue, u); |
659 | 36.8k | |
660 | 36.8k | if (u->in_cgroup_empty_queue) |
661 | 36.8k | LIST_REMOVE(cgroup_empty_queue, u->manager->cgroup_empty_queue, u); |
662 | 36.8k | |
663 | 36.8k | if (u->in_cleanup_queue) |
664 | 36.8k | LIST_REMOVE(cleanup_queue, u->manager->cleanup_queue, u); |
665 | 36.8k | |
666 | 36.8k | if (u->in_target_deps_queue) |
667 | 36.8k | LIST_REMOVE(target_deps_queue, u->manager->target_deps_queue, u); |
668 | 36.8k | |
669 | 36.8k | if (u->in_stop_when_unneeded_queue) |
670 | 36.8k | LIST_REMOVE(stop_when_unneeded_queue, u->manager->stop_when_unneeded_queue, u); |
671 | 36.8k | |
672 | 36.8k | safe_close(u->ip_accounting_ingress_map_fd); |
673 | 36.8k | safe_close(u->ip_accounting_egress_map_fd); |
674 | 36.8k | |
675 | 36.8k | safe_close(u->ipv4_allow_map_fd); |
676 | 36.8k | safe_close(u->ipv6_allow_map_fd); |
677 | 36.8k | safe_close(u->ipv4_deny_map_fd); |
678 | 36.8k | safe_close(u->ipv6_deny_map_fd); |
679 | 36.8k | |
680 | 36.8k | bpf_program_unref(u->ip_bpf_ingress); |
681 | 36.8k | bpf_program_unref(u->ip_bpf_ingress_installed); |
682 | 36.8k | bpf_program_unref(u->ip_bpf_egress); |
683 | 36.8k | bpf_program_unref(u->ip_bpf_egress_installed); |
684 | 36.8k | |
685 | 36.8k | bpf_program_unref(u->bpf_device_control_installed); |
686 | 36.8k | |
687 | 36.8k | condition_free_list(u->conditions); |
688 | 36.8k | condition_free_list(u->asserts); |
689 | 36.8k | |
690 | 36.8k | free(u->description); |
691 | 36.8k | strv_free(u->documentation); |
692 | 36.8k | free(u->fragment_path); |
693 | 36.8k | free(u->source_path); |
694 | 36.8k | strv_free(u->dropin_paths); |
695 | 36.8k | free(u->instance); |
696 | 36.8k | |
697 | 36.8k | free(u->job_timeout_reboot_arg); |
698 | 36.8k | |
699 | 36.8k | set_free_free(u->names); |
700 | 36.8k | |
701 | 36.8k | free(u->reboot_arg); |
702 | 36.8k | |
703 | 36.8k | free(u); |
704 | 36.8k | } |
705 | | |
706 | 126k | UnitActiveState unit_active_state(Unit *u) { |
707 | 126k | assert(u); |
708 | 126k | |
709 | 126k | if (u->load_state == UNIT_MERGED) |
710 | 0 | return unit_active_state(unit_follow_merge(u)); |
711 | 126k | |
712 | 126k | /* After a reload it might happen that a unit is not correctly |
713 | 126k | * loaded but still has a process around. That's why we won't |
714 | 126k | * shortcut failed loading to UNIT_INACTIVE_FAILED. */ |
715 | 126k | |
716 | 126k | return UNIT_VTABLE(u)->active_state(u); |
717 | 126k | } |
718 | | |
719 | 0 | const char* unit_sub_state_to_string(Unit *u) { |
720 | 0 | assert(u); |
721 | 0 |
|
722 | 0 | return UNIT_VTABLE(u)->sub_state_to_string(u); |
723 | 0 | } |
724 | | |
725 | 0 | static int set_complete_move(Set **s, Set **other) { |
726 | 0 | assert(s); |
727 | 0 | assert(other); |
728 | 0 |
|
729 | 0 | if (!other) |
730 | 0 | return 0; |
731 | 0 | |
732 | 0 | if (*s) |
733 | 0 | return set_move(*s, *other); |
734 | 0 | else |
735 | 0 | *s = TAKE_PTR(*other); |
736 | 0 |
|
737 | 0 | return 0; |
738 | 0 | } |
739 | | |
740 | 0 | static int hashmap_complete_move(Hashmap **s, Hashmap **other) { |
741 | 0 | assert(s); |
742 | 0 | assert(other); |
743 | 0 |
|
744 | 0 | if (!*other) |
745 | 0 | return 0; |
746 | 0 | |
747 | 0 | if (*s) |
748 | 0 | return hashmap_move(*s, *other); |
749 | 0 | else |
750 | 0 | *s = TAKE_PTR(*other); |
751 | 0 |
|
752 | 0 | return 0; |
753 | 0 | } |
754 | | |
755 | 0 | static int merge_names(Unit *u, Unit *other) { |
756 | 0 | char *t; |
757 | 0 | Iterator i; |
758 | 0 | int r; |
759 | 0 |
|
760 | 0 | assert(u); |
761 | 0 | assert(other); |
762 | 0 |
|
763 | 0 | r = set_complete_move(&u->names, &other->names); |
764 | 0 | if (r < 0) |
765 | 0 | return r; |
766 | 0 | |
767 | 0 | set_free_free(other->names); |
768 | 0 | other->names = NULL; |
769 | 0 | other->id = NULL; |
770 | 0 |
|
771 | 0 | SET_FOREACH(t, u->names, i) |
772 | 0 | assert_se(hashmap_replace(u->manager->units, t, u) == 0); |
773 | 0 |
|
774 | 0 | return 0; |
775 | 0 | } |
776 | | |
777 | 0 | static int reserve_dependencies(Unit *u, Unit *other, UnitDependency d) { |
778 | 0 | unsigned n_reserve; |
779 | 0 |
|
780 | 0 | assert(u); |
781 | 0 | assert(other); |
782 | 0 | assert(d < _UNIT_DEPENDENCY_MAX); |
783 | 0 |
|
784 | 0 | /* |
785 | 0 | * If u does not have this dependency set allocated, there is no need |
786 | 0 | * to reserve anything. In that case other's set will be transferred |
787 | 0 | * as a whole to u by complete_move(). |
788 | 0 | */ |
789 | 0 | if (!u->dependencies[d]) |
790 | 0 | return 0; |
791 | 0 | |
792 | 0 | /* merge_dependencies() will skip a u-on-u dependency */ |
793 | 0 | n_reserve = hashmap_size(other->dependencies[d]) - !!hashmap_get(other->dependencies[d], u); |
794 | 0 |
|
795 | 0 | return hashmap_reserve(u->dependencies[d], n_reserve); |
796 | 0 | } |
797 | | |
798 | 0 | static void merge_dependencies(Unit *u, Unit *other, const char *other_id, UnitDependency d) { |
799 | 0 | Iterator i; |
800 | 0 | Unit *back; |
801 | 0 | void *v; |
802 | 0 | int r; |
803 | 0 |
|
804 | 0 | /* Merges all dependencies of type 'd' of the unit 'other' into the deps of the unit 'u' */ |
805 | 0 |
|
806 | 0 | assert(u); |
807 | 0 | assert(other); |
808 | 0 | assert(d < _UNIT_DEPENDENCY_MAX); |
809 | 0 |
|
810 | 0 | /* Fix backwards pointers. Let's iterate through all dependent units of the other unit. */ |
811 | 0 | HASHMAP_FOREACH_KEY(v, back, other->dependencies[d], i) { |
812 | 0 | UnitDependency k; |
813 | 0 |
|
814 | 0 | /* Let's now iterate through the dependencies of that dependencies of the other units, looking for |
815 | 0 | * pointers back, and let's fix them up, to instead point to 'u'. */ |
816 | 0 |
|
817 | 0 | for (k = 0; k < _UNIT_DEPENDENCY_MAX; k++) { |
818 | 0 | if (back == u) { |
819 | 0 | /* Do not add dependencies between u and itself. */ |
820 | 0 | if (hashmap_remove(back->dependencies[k], other)) |
821 | 0 | maybe_warn_about_dependency(u, other_id, k); |
822 | 0 | } else { |
823 | 0 | UnitDependencyInfo di_u, di_other, di_merged; |
824 | 0 |
|
825 | 0 | /* Let's drop this dependency between "back" and "other", and let's create it between |
826 | 0 | * "back" and "u" instead. Let's merge the bit masks of the dependency we are moving, |
827 | 0 | * and any such dependency which might already exist */ |
828 | 0 |
|
829 | 0 | di_other.data = hashmap_get(back->dependencies[k], other); |
830 | 0 | if (!di_other.data) |
831 | 0 | continue; /* dependency isn't set, let's try the next one */ |
832 | 0 | |
833 | 0 | di_u.data = hashmap_get(back->dependencies[k], u); |
834 | 0 |
|
835 | 0 | di_merged = (UnitDependencyInfo) { |
836 | 0 | .origin_mask = di_u.origin_mask | di_other.origin_mask, |
837 | 0 | .destination_mask = di_u.destination_mask | di_other.destination_mask, |
838 | 0 | }; |
839 | 0 |
|
840 | 0 | r = hashmap_remove_and_replace(back->dependencies[k], other, u, di_merged.data); |
841 | 0 | if (r < 0) |
842 | 0 | log_warning_errno(r, "Failed to remove/replace: back=%s other=%s u=%s: %m", back->id, other_id, u->id); |
843 | 0 | assert(r >= 0); |
844 | 0 |
|
845 | 0 | /* assert_se(hashmap_remove_and_replace(back->dependencies[k], other, u, di_merged.data) >= 0); */ |
846 | 0 | } |
847 | 0 | } |
848 | 0 |
|
849 | 0 | } |
850 | 0 |
|
851 | 0 | /* Also do not move dependencies on u to itself */ |
852 | 0 | back = hashmap_remove(other->dependencies[d], u); |
853 | 0 | if (back) |
854 | 0 | maybe_warn_about_dependency(u, other_id, d); |
855 | 0 |
|
856 | 0 | /* The move cannot fail. The caller must have performed a reservation. */ |
857 | 0 | assert_se(hashmap_complete_move(&u->dependencies[d], &other->dependencies[d]) == 0); |
858 | 0 |
|
859 | 0 | other->dependencies[d] = hashmap_free(other->dependencies[d]); |
860 | 0 | } |
861 | | |
862 | 0 | int unit_merge(Unit *u, Unit *other) { |
863 | 0 | UnitDependency d; |
864 | 0 | const char *other_id = NULL; |
865 | 0 | int r; |
866 | 0 |
|
867 | 0 | assert(u); |
868 | 0 | assert(other); |
869 | 0 | assert(u->manager == other->manager); |
870 | 0 | assert(u->type != _UNIT_TYPE_INVALID); |
871 | 0 |
|
872 | 0 | other = unit_follow_merge(other); |
873 | 0 |
|
874 | 0 | if (other == u) |
875 | 0 | return 0; |
876 | 0 | |
877 | 0 | if (u->type != other->type) |
878 | 0 | return -EINVAL; |
879 | 0 | |
880 | 0 | if (!u->instance != !other->instance) |
881 | 0 | return -EINVAL; |
882 | 0 | |
883 | 0 | if (!unit_type_may_alias(u->type)) /* Merging only applies to unit names that support aliases */ |
884 | 0 | return -EEXIST; |
885 | 0 | |
886 | 0 | if (!IN_SET(other->load_state, UNIT_STUB, UNIT_NOT_FOUND)) |
887 | 0 | return -EEXIST; |
888 | 0 | |
889 | 0 | if (other->job) |
890 | 0 | return -EEXIST; |
891 | 0 | |
892 | 0 | if (other->nop_job) |
893 | 0 | return -EEXIST; |
894 | 0 | |
895 | 0 | if (!UNIT_IS_INACTIVE_OR_FAILED(unit_active_state(other))) |
896 | 0 | return -EEXIST; |
897 | 0 | |
898 | 0 | if (other->id) |
899 | 0 | other_id = strdupa(other->id); |
900 | 0 |
|
901 | 0 | /* Make reservations to ensure merge_dependencies() won't fail */ |
902 | 0 | for (d = 0; d < _UNIT_DEPENDENCY_MAX; d++) { |
903 | 0 | r = reserve_dependencies(u, other, d); |
904 | 0 | /* |
905 | 0 | * We don't rollback reservations if we fail. We don't have |
906 | 0 | * a way to undo reservations. A reservation is not a leak. |
907 | 0 | */ |
908 | 0 | if (r < 0) |
909 | 0 | return r; |
910 | 0 | } |
911 | 0 |
|
912 | 0 | /* Merge names */ |
913 | 0 | r = merge_names(u, other); |
914 | 0 | if (r < 0) |
915 | 0 | return r; |
916 | 0 | |
917 | 0 | /* Redirect all references */ |
918 | 0 | while (other->refs_by_target) |
919 | 0 | unit_ref_set(other->refs_by_target, other->refs_by_target->source, u); |
920 | 0 |
|
921 | 0 | /* Merge dependencies */ |
922 | 0 | for (d = 0; d < _UNIT_DEPENDENCY_MAX; d++) |
923 | 0 | merge_dependencies(u, other, other_id, d); |
924 | 0 |
|
925 | 0 | other->load_state = UNIT_MERGED; |
926 | 0 | other->merged_into = u; |
927 | 0 |
|
928 | 0 | /* If there is still some data attached to the other node, we |
929 | 0 | * don't need it anymore, and can free it. */ |
930 | 0 | if (other->load_state != UNIT_STUB) |
931 | 0 | if (UNIT_VTABLE(other)->done) |
932 | 0 | UNIT_VTABLE(other)->done(other); |
933 | 0 |
|
934 | 0 | unit_add_to_dbus_queue(u); |
935 | 0 | unit_add_to_cleanup_queue(other); |
936 | 0 |
|
937 | 0 | return 0; |
938 | 0 | } |
939 | | |
940 | 0 | int unit_merge_by_name(Unit *u, const char *name) { |
941 | 0 | _cleanup_free_ char *s = NULL; |
942 | 0 | Unit *other; |
943 | 0 | int r; |
944 | 0 |
|
945 | 0 | assert(u); |
946 | 0 | assert(name); |
947 | 0 |
|
948 | 0 | if (unit_name_is_valid(name, UNIT_NAME_TEMPLATE)) { |
949 | 0 | if (!u->instance) |
950 | 0 | return -EINVAL; |
951 | 0 | |
952 | 0 | r = unit_name_replace_instance(name, u->instance, &s); |
953 | 0 | if (r < 0) |
954 | 0 | return r; |
955 | 0 | |
956 | 0 | name = s; |
957 | 0 | } |
958 | 0 |
|
959 | 0 | other = manager_get_unit(u->manager, name); |
960 | 0 | if (other) |
961 | 0 | return unit_merge(u, other); |
962 | 0 | |
963 | 0 | return unit_add_name(u, name); |
964 | 0 | } |
965 | | |
966 | 105k | Unit* unit_follow_merge(Unit *u) { |
967 | 105k | assert(u); |
968 | 105k | |
969 | 105k | while (u->load_state == UNIT_MERGED) |
970 | 105k | assert_se(u = u->merged_into); |
971 | 105k | |
972 | 105k | return u; |
973 | 105k | } |
974 | | |
975 | 8 | int unit_add_exec_dependencies(Unit *u, ExecContext *c) { |
976 | 8 | ExecDirectoryType dt; |
977 | 8 | char **dp; |
978 | 8 | int r; |
979 | 8 | |
980 | 8 | assert(u); |
981 | 8 | assert(c); |
982 | 8 | |
983 | 8 | if (c->working_directory && !c->working_directory_missing_ok) { |
984 | 0 | r = unit_require_mounts_for(u, c->working_directory, UNIT_DEPENDENCY_FILE); |
985 | 0 | if (r < 0) |
986 | 0 | return r; |
987 | 8 | } |
988 | 8 | |
989 | 8 | if (c->root_directory) { |
990 | 0 | r = unit_require_mounts_for(u, c->root_directory, UNIT_DEPENDENCY_FILE); |
991 | 0 | if (r < 0) |
992 | 0 | return r; |
993 | 8 | } |
994 | 8 | |
995 | 8 | if (c->root_image) { |
996 | 0 | r = unit_require_mounts_for(u, c->root_image, UNIT_DEPENDENCY_FILE); |
997 | 0 | if (r < 0) |
998 | 0 | return r; |
999 | 8 | } |
1000 | 8 | |
1001 | 48 | for (dt = 0; dt < _EXEC_DIRECTORY_TYPE_MAX; dt++) { |
1002 | 40 | if (!u->manager->prefix[dt]) |
1003 | 0 | continue; |
1004 | 40 | |
1005 | 40 | STRV_FOREACH(dp, c->directories[dt].paths) { |
1006 | 0 | _cleanup_free_ char *p; |
1007 | 0 |
|
1008 | 0 | p = strjoin(u->manager->prefix[dt], "/", *dp); |
1009 | 0 | if (!p) |
1010 | 0 | return -ENOMEM; |
1011 | 0 | |
1012 | 0 | r = unit_require_mounts_for(u, p, UNIT_DEPENDENCY_FILE); |
1013 | 0 | if (r < 0) |
1014 | 0 | return r; |
1015 | 0 | } |
1016 | 40 | } |
1017 | 8 | |
1018 | 8 | if (!MANAGER_IS_SYSTEM(u->manager)) |
1019 | 8 | return 0; |
1020 | 8 | |
1021 | 8 | if (c->private_tmp) { |
1022 | 0 | const char *p; |
1023 | 0 |
|
1024 | 0 | FOREACH_STRING(p, "/tmp", "/var/tmp") { |
1025 | 0 | r = unit_require_mounts_for(u, p, UNIT_DEPENDENCY_FILE); |
1026 | 0 | if (r < 0) |
1027 | 0 | return r; |
1028 | 0 | } |
1029 | 0 |
|
1030 | 0 | r = unit_add_dependency_by_name(u, UNIT_AFTER, SPECIAL_TMPFILES_SETUP_SERVICE, true, UNIT_DEPENDENCY_FILE); |
1031 | 0 | if (r < 0) |
1032 | 0 | return r; |
1033 | 8 | } |
1034 | 8 | |
1035 | 8 | if (!IN_SET(c->std_output, |
1036 | 8 | EXEC_OUTPUT_JOURNAL, EXEC_OUTPUT_JOURNAL_AND_CONSOLE, |
1037 | 8 | EXEC_OUTPUT_KMSG, EXEC_OUTPUT_KMSG_AND_CONSOLE, |
1038 | 8 | EXEC_OUTPUT_SYSLOG, EXEC_OUTPUT_SYSLOG_AND_CONSOLE) && |
1039 | 8 | !IN_SET(c->std_error, |
1040 | 8 | EXEC_OUTPUT_JOURNAL, EXEC_OUTPUT_JOURNAL_AND_CONSOLE, |
1041 | 8 | EXEC_OUTPUT_KMSG, EXEC_OUTPUT_KMSG_AND_CONSOLE, |
1042 | 8 | EXEC_OUTPUT_SYSLOG, EXEC_OUTPUT_SYSLOG_AND_CONSOLE)) |
1043 | 8 | return 0; |
1044 | 0 | |
1045 | 0 | /* If syslog or kernel logging is requested, make sure our own |
1046 | 0 | * logging daemon is run first. */ |
1047 | 0 | |
1048 | 0 | r = unit_add_dependency_by_name(u, UNIT_AFTER, SPECIAL_JOURNALD_SOCKET, true, UNIT_DEPENDENCY_FILE); |
1049 | 0 | if (r < 0) |
1050 | 0 | return r; |
1051 | 0 | |
1052 | 0 | return 0; |
1053 | 0 | } |
1054 | | |
1055 | 53.3k | const char *unit_description(Unit *u) { |
1056 | 53.3k | assert(u); |
1057 | 53.3k | |
1058 | 53.3k | if (u->description) |
1059 | 1.82k | return u->description; |
1060 | 51.5k | |
1061 | 51.5k | return strna(u->id); |
1062 | 51.5k | } |
1063 | | |
1064 | 255k | static void print_unit_dependency_mask(FILE *f, const char *kind, UnitDependencyMask mask, bool *space) { |
1065 | 255k | const struct { |
1066 | 255k | UnitDependencyMask mask; |
1067 | 255k | const char *name; |
1068 | 255k | } table[] = { |
1069 | 255k | { UNIT_DEPENDENCY_FILE, "file" }, |
1070 | 255k | { UNIT_DEPENDENCY_IMPLICIT, "implicit" }, |
1071 | 255k | { UNIT_DEPENDENCY_DEFAULT, "default" }, |
1072 | 255k | { UNIT_DEPENDENCY_UDEV, "udev" }, |
1073 | 255k | { UNIT_DEPENDENCY_PATH, "path" }, |
1074 | 255k | { UNIT_DEPENDENCY_MOUNTINFO_IMPLICIT, "mountinfo-implicit" }, |
1075 | 255k | { UNIT_DEPENDENCY_MOUNTINFO_DEFAULT, "mountinfo-default" }, |
1076 | 255k | { UNIT_DEPENDENCY_PROC_SWAP, "proc-swap" }, |
1077 | 255k | }; |
1078 | 255k | size_t i; |
1079 | 255k | |
1080 | 255k | assert(f); |
1081 | 255k | assert(kind); |
1082 | 255k | assert(space); |
1083 | 255k | |
1084 | 466k | for (i = 0; i < ELEMENTSOF(table); i++) { |
1085 | 466k | |
1086 | 466k | if (mask == 0) |
1087 | 255k | break; |
1088 | 210k | |
1089 | 210k | if (FLAGS_SET(mask, table[i].mask)) { |
1090 | 128k | if (*space) |
1091 | 570 | fputc(' ', f); |
1092 | 127k | else |
1093 | 127k | *space = true; |
1094 | 128k | |
1095 | 128k | fputs(kind, f); |
1096 | 128k | fputs("-", f); |
1097 | 128k | fputs(table[i].name, f); |
1098 | 128k | |
1099 | 128k | mask &= ~table[i].mask; |
1100 | 128k | } |
1101 | 210k | } |
1102 | 255k | |
1103 | 255k | assert(mask == 0); |
1104 | 255k | } |
1105 | | |
1106 | 53.3k | void unit_dump(Unit *u, FILE *f, const char *prefix) { |
1107 | 53.3k | char *t, **j; |
1108 | 53.3k | UnitDependency d; |
1109 | 53.3k | Iterator i; |
1110 | 53.3k | const char *prefix2; |
1111 | 53.3k | char |
1112 | 53.3k | timestamp0[FORMAT_TIMESTAMP_MAX], |
1113 | 53.3k | timestamp1[FORMAT_TIMESTAMP_MAX], |
1114 | 53.3k | timestamp2[FORMAT_TIMESTAMP_MAX], |
1115 | 53.3k | timestamp3[FORMAT_TIMESTAMP_MAX], |
1116 | 53.3k | timestamp4[FORMAT_TIMESTAMP_MAX], |
1117 | 53.3k | timespan[FORMAT_TIMESPAN_MAX]; |
1118 | 53.3k | Unit *following; |
1119 | 53.3k | _cleanup_set_free_ Set *following_set = NULL; |
1120 | 53.3k | const char *n; |
1121 | 53.3k | CGroupMask m; |
1122 | 53.3k | int r; |
1123 | 53.3k | |
1124 | 53.3k | assert(u); |
1125 | 53.3k | assert(u->type >= 0); |
1126 | 53.3k | |
1127 | 53.3k | prefix = strempty(prefix); |
1128 | 106k | prefix2 = strjoina(prefix, "\t"); |
1129 | 106k | |
1130 | 106k | fprintf(f, |
1131 | 106k | "%s-> Unit %s:\n" |
1132 | 106k | "%s\tDescription: %s\n" |
1133 | 106k | "%s\tInstance: %s\n" |
1134 | 106k | "%s\tUnit Load State: %s\n" |
1135 | 106k | "%s\tUnit Active State: %s\n" |
1136 | 106k | "%s\tState Change Timestamp: %s\n" |
1137 | 106k | "%s\tInactive Exit Timestamp: %s\n" |
1138 | 106k | "%s\tActive Enter Timestamp: %s\n" |
1139 | 106k | "%s\tActive Exit Timestamp: %s\n" |
1140 | 106k | "%s\tInactive Enter Timestamp: %s\n" |
1141 | 106k | "%s\tMay GC: %s\n" |
1142 | 106k | "%s\tNeed Daemon Reload: %s\n" |
1143 | 106k | "%s\tTransient: %s\n" |
1144 | 106k | "%s\tPerpetual: %s\n" |
1145 | 106k | "%s\tGarbage Collection Mode: %s\n" |
1146 | 106k | "%s\tSlice: %s\n" |
1147 | 106k | "%s\tCGroup: %s\n" |
1148 | 106k | "%s\tCGroup realized: %s\n", |
1149 | 106k | prefix, u->id, |
1150 | 106k | prefix, unit_description(u), |
1151 | 106k | prefix, strna(u->instance), |
1152 | 106k | prefix, unit_load_state_to_string(u->load_state), |
1153 | 106k | prefix, unit_active_state_to_string(unit_active_state(u)), |
1154 | 106k | prefix, strna(format_timestamp(timestamp0, sizeof(timestamp0), u->state_change_timestamp.realtime)), |
1155 | 106k | prefix, strna(format_timestamp(timestamp1, sizeof(timestamp1), u->inactive_exit_timestamp.realtime)), |
1156 | 106k | prefix, strna(format_timestamp(timestamp2, sizeof(timestamp2), u->active_enter_timestamp.realtime)), |
1157 | 106k | prefix, strna(format_timestamp(timestamp3, sizeof(timestamp3), u->active_exit_timestamp.realtime)), |
1158 | 106k | prefix, strna(format_timestamp(timestamp4, sizeof(timestamp4), u->inactive_enter_timestamp.realtime)), |
1159 | 106k | prefix, yes_no(unit_may_gc(u)), |
1160 | 106k | prefix, yes_no(unit_need_daemon_reload(u)), |
1161 | 106k | prefix, yes_no(u->transient), |
1162 | 106k | prefix, yes_no(u->perpetual), |
1163 | 106k | prefix, collect_mode_to_string(u->collect_mode), |
1164 | 106k | prefix, strna(unit_slice_name(u)), |
1165 | 106k | prefix, strna(u->cgroup_path), |
1166 | 106k | prefix, yes_no(u->cgroup_realized)); |
1167 | 106k | |
1168 | 106k | if (u->cgroup_realized_mask != 0) { |
1169 | 0 | _cleanup_free_ char *s = NULL; |
1170 | 0 | (void) cg_mask_to_string(u->cgroup_realized_mask, &s); |
1171 | 0 | fprintf(f, "%s\tCGroup realized mask: %s\n", prefix, strnull(s)); |
1172 | 0 | } |
1173 | 106k | |
1174 | 106k | if (u->cgroup_enabled_mask != 0) { |
1175 | 0 | _cleanup_free_ char *s = NULL; |
1176 | 0 | (void) cg_mask_to_string(u->cgroup_enabled_mask, &s); |
1177 | 0 | fprintf(f, "%s\tCGroup enabled mask: %s\n", prefix, strnull(s)); |
1178 | 0 | } |
1179 | 106k | |
1180 | 106k | m = unit_get_own_mask(u); |
1181 | 106k | if (m != 0) { |
1182 | 5.30k | _cleanup_free_ char *s = NULL; |
1183 | 5.30k | (void) cg_mask_to_string(m, &s); |
1184 | 5.30k | fprintf(f, "%s\tCGroup own mask: %s\n", prefix, strnull(s)); |
1185 | 5.30k | } |
1186 | 106k | |
1187 | 106k | m = unit_get_members_mask(u); |
1188 | 106k | if (m != 0) { |
1189 | 2.68k | _cleanup_free_ char *s = NULL; |
1190 | 2.68k | (void) cg_mask_to_string(m, &s); |
1191 | 2.68k | fprintf(f, "%s\tCGroup members mask: %s\n", prefix, strnull(s)); |
1192 | 2.68k | } |
1193 | 106k | |
1194 | 106k | m = unit_get_delegate_mask(u); |
1195 | 106k | if (m != 0) { |
1196 | 36 | _cleanup_free_ char *s = NULL; |
1197 | 36 | (void) cg_mask_to_string(m, &s); |
1198 | 36 | fprintf(f, "%s\tCGroup delegate mask: %s\n", prefix, strnull(s)); |
1199 | 36 | } |
1200 | 106k | |
1201 | 106k | SET_FOREACH(t, u->names, i) |
1202 | 106k | fprintf(f, "%s\tName: %s\n", prefix, t); |
1203 | 106k | |
1204 | 106k | if (!sd_id128_is_null(u->invocation_id)) |
1205 | 0 | fprintf(f, "%s\tInvocation ID: " SD_ID128_FORMAT_STR "\n", |
1206 | 0 | prefix, SD_ID128_FORMAT_VAL(u->invocation_id)); |
1207 | 106k | |
1208 | 106k | STRV_FOREACH(j, u->documentation) |
1209 | 106k | fprintf(f, "%s\tDocumentation: %s\n", prefix, *j); |
1210 | 106k | |
1211 | 106k | following = unit_following(u); |
1212 | 106k | if (following) |
1213 | 0 | fprintf(f, "%s\tFollowing: %s\n", prefix, following->id); |
1214 | 106k | |
1215 | 106k | r = unit_following_set(u, &following_set); |
1216 | 106k | if (r >= 0) { |
1217 | 53.3k | Unit *other; |
1218 | 53.3k | |
1219 | 53.3k | SET_FOREACH(other, following_set, i) |
1220 | 53.3k | fprintf(f, "%s\tFollowing Set Member: %s\n", prefix, other->id); |
1221 | 53.3k | } |
1222 | 106k | |
1223 | 106k | if (u->fragment_path) |
1224 | 0 | fprintf(f, "%s\tFragment Path: %s\n", prefix, u->fragment_path); |
1225 | 106k | |
1226 | 106k | if (u->source_path) |
1227 | 40 | fprintf(f, "%s\tSource Path: %s\n", prefix, u->source_path); |
1228 | 106k | |
1229 | 106k | STRV_FOREACH(j, u->dropin_paths) |
1230 | 106k | fprintf(f, "%s\tDropIn Path: %s\n", prefix, *j); |
1231 | 106k | |
1232 | 106k | if (u->failure_action != EMERGENCY_ACTION_NONE) |
1233 | 0 | fprintf(f, "%s\tFailure Action: %s\n", prefix, emergency_action_to_string(u->failure_action)); |
1234 | 106k | if (u->failure_action_exit_status >= 0) |
1235 | 26 | fprintf(f, "%s\tFailure Action Exit Status: %i\n", prefix, u->failure_action_exit_status); |
1236 | 106k | if (u->success_action != EMERGENCY_ACTION_NONE) |
1237 | 36 | fprintf(f, "%s\tSuccess Action: %s\n", prefix, emergency_action_to_string(u->success_action)); |
1238 | 106k | if (u->success_action_exit_status >= 0) |
1239 | 24 | fprintf(f, "%s\tSuccess Action Exit Status: %i\n", prefix, u->success_action_exit_status); |
1240 | 106k | |
1241 | 106k | if (u->job_timeout != USEC_INFINITY) |
1242 | 106k | fprintf(f, "%s\tJob Timeout: %s\n", prefix, format_timespan(timespan, sizeof(timespan), u->job_timeout, 0)); |
1243 | 106k | |
1244 | 106k | if (u->job_timeout_action != EMERGENCY_ACTION_NONE) |
1245 | 6 | fprintf(f, "%s\tJob Timeout Action: %s\n", prefix, emergency_action_to_string(u->job_timeout_action)); |
1246 | 106k | |
1247 | 106k | if (u->job_timeout_reboot_arg) |
1248 | 4 | fprintf(f, "%s\tJob Timeout Reboot Argument: %s\n", prefix, u->job_timeout_reboot_arg); |
1249 | 106k | |
1250 | 106k | condition_dump_list(u->conditions, f, prefix, condition_type_to_string); |
1251 | 106k | condition_dump_list(u->asserts, f, prefix, assert_type_to_string); |
1252 | 106k | |
1253 | 106k | if (dual_timestamp_is_set(&u->condition_timestamp)) |
1254 | 0 | fprintf(f, |
1255 | 0 | "%s\tCondition Timestamp: %s\n" |
1256 | 0 | "%s\tCondition Result: %s\n", |
1257 | 0 | prefix, strna(format_timestamp(timestamp1, sizeof(timestamp1), u->condition_timestamp.realtime)), |
1258 | 0 | prefix, yes_no(u->condition_result)); |
1259 | 106k | |
1260 | 106k | if (dual_timestamp_is_set(&u->assert_timestamp)) |
1261 | 0 | fprintf(f, |
1262 | 0 | "%s\tAssert Timestamp: %s\n" |
1263 | 0 | "%s\tAssert Result: %s\n", |
1264 | 0 | prefix, strna(format_timestamp(timestamp1, sizeof(timestamp1), u->assert_timestamp.realtime)), |
1265 | 0 | prefix, yes_no(u->assert_result)); |
1266 | 106k | |
1267 | 1.22M | for (d = 0; d < _UNIT_DEPENDENCY_MAX; d++) { |
1268 | 1.17M | UnitDependencyInfo di; |
1269 | 1.17M | Unit *other; |
1270 | 1.17M | |
1271 | 1.17M | HASHMAP_FOREACH_KEY(di.data, other, u->dependencies[d], i) { |
1272 | 124k | bool space = false; |
1273 | 124k | |
1274 | 124k | fprintf(f, "%s\t%s: %s (", prefix, unit_dependency_to_string(d), other->id); |
1275 | 124k | |
1276 | 124k | print_unit_dependency_mask(f, "origin", di.origin_mask, &space); |
1277 | 124k | print_unit_dependency_mask(f, "destination", di.destination_mask, &space); |
1278 | 124k | |
1279 | 124k | fputs(")\n", f); |
1280 | 124k | } |
1281 | 1.17M | } |
1282 | 106k | |
1283 | 106k | if (!hashmap_isempty(u->requires_mounts_for)) { |
1284 | 448 | UnitDependencyInfo di; |
1285 | 448 | const char *path; |
1286 | 448 | |
1287 | 3.60k | HASHMAP_FOREACH_KEY(di.data, path, u->requires_mounts_for, i) { |
1288 | 3.60k | bool space = false; |
1289 | 3.60k | |
1290 | 3.60k | fprintf(f, "%s\tRequiresMountsFor: %s (", prefix, path); |
1291 | 3.60k | |
1292 | 3.60k | print_unit_dependency_mask(f, "origin", di.origin_mask, &space); |
1293 | 3.60k | print_unit_dependency_mask(f, "destination", di.destination_mask, &space); |
1294 | 3.60k | |
1295 | 3.60k | fputs(")\n", f); |
1296 | 3.60k | } |
1297 | 448 | } |
1298 | 106k | |
1299 | 106k | if (u->load_state == UNIT_LOADED) { |
1300 | 6.94k | |
1301 | 6.94k | fprintf(f, |
1302 | 6.94k | "%s\tStopWhenUnneeded: %s\n" |
1303 | 6.94k | "%s\tRefuseManualStart: %s\n" |
1304 | 6.94k | "%s\tRefuseManualStop: %s\n" |
1305 | 6.94k | "%s\tDefaultDependencies: %s\n" |
1306 | 6.94k | "%s\tOnFailureJobMode: %s\n" |
1307 | 6.94k | "%s\tIgnoreOnIsolate: %s\n", |
1308 | 6.94k | prefix, yes_no(u->stop_when_unneeded), |
1309 | 6.94k | prefix, yes_no(u->refuse_manual_start), |
1310 | 6.94k | prefix, yes_no(u->refuse_manual_stop), |
1311 | 6.94k | prefix, yes_no(u->default_dependencies), |
1312 | 6.94k | prefix, job_mode_to_string(u->on_failure_job_mode), |
1313 | 6.94k | prefix, yes_no(u->ignore_on_isolate)); |
1314 | 6.94k | |
1315 | 6.94k | if (UNIT_VTABLE(u)->dump) |
1316 | 6.94k | UNIT_VTABLE(u)->dump(u, f, prefix2); |
1317 | 6.94k | |
1318 | 46.3k | } else if (u->load_state == UNIT_MERGED) |
1319 | 0 | fprintf(f, |
1320 | 0 | "%s\tMerged into: %s\n", |
1321 | 0 | prefix, u->merged_into->id); |
1322 | 46.3k | else if (u->load_state == UNIT_ERROR) |
1323 | 422 | fprintf(f, "%s\tLoad Error Code: %s\n", prefix, strerror(-u->load_error)); |
1324 | 106k | |
1325 | 106k | for (n = sd_bus_track_first(u->bus_track); n; n = sd_bus_track_next(u->bus_track)) |
1326 | 0 | fprintf(f, "%s\tBus Ref: %s\n", prefix, n); |
1327 | 106k | |
1328 | 106k | if (u->job) |
1329 | 0 | job_dump(u->job, f, prefix2); |
1330 | 106k | |
1331 | 106k | if (u->nop_job) |
1332 | 0 | job_dump(u->nop_job, f, prefix2); |
1333 | 106k | } |
1334 | | |
1335 | | /* Common implementation for multiple backends */ |
1336 | 6.19k | int unit_load_fragment_and_dropin(Unit *u) { |
1337 | 6.19k | int r; |
1338 | 6.19k | |
1339 | 6.19k | assert(u); |
1340 | 6.19k | |
1341 | 6.19k | /* Load a .{service,socket,...} file */ |
1342 | 6.19k | r = unit_load_fragment(u); |
1343 | 6.19k | if (r < 0) |
1344 | 0 | return r; |
1345 | 6.19k | |
1346 | 6.19k | if (u->load_state == UNIT_STUB) |
1347 | 6.19k | return -ENOENT; |
1348 | 0 | |
1349 | 0 | /* Load drop-in directory data. If u is an alias, we might be reloading the |
1350 | 0 | * target unit needlessly. But we cannot be sure which drops-ins have already |
1351 | 0 | * been loaded and which not, at least without doing complicated book-keeping, |
1352 | 0 | * so let's always reread all drop-ins. */ |
1353 | 0 | return unit_load_dropin(unit_follow_merge(u)); |
1354 | 0 | } |
1355 | | |
1356 | | /* Common implementation for multiple backends */ |
1357 | 7.36k | int unit_load_fragment_and_dropin_optional(Unit *u) { |
1358 | 7.36k | int r; |
1359 | 7.36k | |
1360 | 7.36k | assert(u); |
1361 | 7.36k | |
1362 | 7.36k | /* Same as unit_load_fragment_and_dropin(), but whether |
1363 | 7.36k | * something can be loaded or not doesn't matter. */ |
1364 | 7.36k | |
1365 | 7.36k | /* Load a .service/.socket/.slice/… file */ |
1366 | 7.36k | r = unit_load_fragment(u); |
1367 | 7.36k | if (r < 0) |
1368 | 0 | return r; |
1369 | 7.36k | |
1370 | 7.36k | if (u->load_state == UNIT_STUB) |
1371 | 7.36k | u->load_state = UNIT_LOADED; |
1372 | 7.36k | |
1373 | 7.36k | /* Load drop-in directory data */ |
1374 | 7.36k | return unit_load_dropin(unit_follow_merge(u)); |
1375 | 7.36k | } |
1376 | | |
1377 | 6.94k | void unit_add_to_target_deps_queue(Unit *u) { |
1378 | 6.94k | Manager *m = u->manager; |
1379 | 6.94k | |
1380 | 6.94k | assert(u); |
1381 | 6.94k | |
1382 | 6.94k | if (u->in_target_deps_queue) |
1383 | 0 | return; |
1384 | 6.94k | |
1385 | 6.94k | LIST_PREPEND(target_deps_queue, m->target_deps_queue, u); |
1386 | 6.94k | u->in_target_deps_queue = true; |
1387 | 6.94k | } |
1388 | | |
1389 | 2.56k | int unit_add_default_target_dependency(Unit *u, Unit *target) { |
1390 | 2.56k | assert(u); |
1391 | 2.56k | assert(target); |
1392 | 2.56k | |
1393 | 2.56k | if (target->type != UNIT_TARGET) |
1394 | 2.56k | return 0; |
1395 | 0 | |
1396 | 0 | /* Only add the dependency if both units are loaded, so that |
1397 | 0 | * that loop check below is reliable */ |
1398 | 0 | if (u->load_state != UNIT_LOADED || |
1399 | 0 | target->load_state != UNIT_LOADED) |
1400 | 0 | return 0; |
1401 | 0 | |
1402 | 0 | /* If either side wants no automatic dependencies, then let's |
1403 | 0 | * skip this */ |
1404 | 0 | if (!u->default_dependencies || |
1405 | 0 | !target->default_dependencies) |
1406 | 0 | return 0; |
1407 | 0 | |
1408 | 0 | /* Don't create loops */ |
1409 | 0 | if (hashmap_get(target->dependencies[UNIT_BEFORE], u)) |
1410 | 0 | return 0; |
1411 | 0 | |
1412 | 0 | return unit_add_dependency(target, UNIT_AFTER, u, true, UNIT_DEPENDENCY_DEFAULT); |
1413 | 0 | } |
1414 | | |
1415 | 6.94k | static int unit_add_slice_dependencies(Unit *u) { |
1416 | 6.94k | UnitDependencyMask mask; |
1417 | 6.94k | assert(u); |
1418 | 6.94k | |
1419 | 6.94k | if (!UNIT_HAS_CGROUP_CONTEXT(u)) |
1420 | 6.94k | return 0; |
1421 | 5.30k | |
1422 | 5.30k | /* Slice units are implicitly ordered against their parent slices (as this relationship is encoded in the |
1423 | 5.30k | name), while all other units are ordered based on configuration (as in their case Slice= configures the |
1424 | 5.30k | relationship). */ |
1425 | 5.30k | mask = u->type == UNIT_SLICE ? UNIT_DEPENDENCY_IMPLICIT : UNIT_DEPENDENCY_FILE; |
1426 | 5.30k | |
1427 | 5.30k | if (UNIT_ISSET(u->slice)) |
1428 | 5.30k | return unit_add_two_dependencies(u, UNIT_AFTER, UNIT_REQUIRES, UNIT_DEREF(u->slice), true, mask); |
1429 | 760 | |
1430 | 760 | if (unit_has_name(u, SPECIAL_ROOT_SLICE)) |
1431 | 760 | return 0; |
1432 | 0 | |
1433 | 0 | return unit_add_two_dependencies_by_name(u, UNIT_AFTER, UNIT_REQUIRES, SPECIAL_ROOT_SLICE, true, mask); |
1434 | 0 | } |
1435 | | |
1436 | 6.94k | static int unit_add_mount_dependencies(Unit *u) { |
1437 | 6.94k | UnitDependencyInfo di; |
1438 | 6.94k | const char *path; |
1439 | 6.94k | Iterator i; |
1440 | 6.94k | int r; |
1441 | 6.94k | |
1442 | 6.94k | assert(u); |
1443 | 6.94k | |
1444 | 6.94k | HASHMAP_FOREACH_KEY(di.data, path, u->requires_mounts_for, i) { |
1445 | 0 | char prefix[strlen(path) + 1]; |
1446 | 0 |
|
1447 | 0 | PATH_FOREACH_PREFIX_MORE(prefix, path) { |
1448 | 0 | _cleanup_free_ char *p = NULL; |
1449 | 0 | Unit *m; |
1450 | 0 |
|
1451 | 0 | r = unit_name_from_path(prefix, ".mount", &p); |
1452 | 0 | if (r < 0) |
1453 | 0 | return r; |
1454 | 0 | |
1455 | 0 | m = manager_get_unit(u->manager, p); |
1456 | 0 | if (!m) { |
1457 | 0 | /* Make sure to load the mount unit if |
1458 | 0 | * it exists. If so the dependencies |
1459 | 0 | * on this unit will be added later |
1460 | 0 | * during the loading of the mount |
1461 | 0 | * unit. */ |
1462 | 0 | (void) manager_load_unit_prepare(u->manager, p, NULL, NULL, &m); |
1463 | 0 | continue; |
1464 | 0 | } |
1465 | 0 | if (m == u) |
1466 | 0 | continue; |
1467 | 0 | |
1468 | 0 | if (m->load_state != UNIT_LOADED) |
1469 | 0 | continue; |
1470 | 0 | |
1471 | 0 | r = unit_add_dependency(u, UNIT_AFTER, m, true, di.origin_mask); |
1472 | 0 | if (r < 0) |
1473 | 0 | return r; |
1474 | 0 | |
1475 | 0 | if (m->fragment_path) { |
1476 | 0 | r = unit_add_dependency(u, UNIT_REQUIRES, m, true, di.origin_mask); |
1477 | 0 | if (r < 0) |
1478 | 0 | return r; |
1479 | 0 | } |
1480 | 0 | } |
1481 | 0 | } |
1482 | 6.94k | |
1483 | 6.94k | return 0; |
1484 | 6.94k | } |
1485 | | |
1486 | 6.94k | static int unit_add_startup_units(Unit *u) { |
1487 | 6.94k | CGroupContext *c; |
1488 | 6.94k | int r; |
1489 | 6.94k | |
1490 | 6.94k | c = unit_get_cgroup_context(u); |
1491 | 6.94k | if (!c) |
1492 | 1.64k | return 0; |
1493 | 5.30k | |
1494 | 5.30k | if (c->startup_cpu_shares == CGROUP_CPU_SHARES_INVALID && |
1495 | 5.30k | c->startup_io_weight == CGROUP_WEIGHT_INVALID && |
1496 | 5.30k | c->startup_blockio_weight == CGROUP_BLKIO_WEIGHT_INVALID) |
1497 | 5.30k | return 0; |
1498 | 0 | |
1499 | 0 | r = set_ensure_allocated(&u->manager->startup_units, NULL); |
1500 | 0 | if (r < 0) |
1501 | 0 | return r; |
1502 | 0 | |
1503 | 0 | return set_put(u->manager->startup_units, u); |
1504 | 0 | } |
1505 | | |
1506 | 17.2k | int unit_load(Unit *u) { |
1507 | 17.2k | int r; |
1508 | 17.2k | |
1509 | 17.2k | assert(u); |
1510 | 17.2k | |
1511 | 17.2k | if (u->in_load_queue) { |
1512 | 17.2k | LIST_REMOVE(load_queue, u->manager->load_queue, u); |
1513 | 17.2k | u->in_load_queue = false; |
1514 | 17.2k | } |
1515 | 17.2k | |
1516 | 17.2k | if (u->type == _UNIT_TYPE_INVALID) |
1517 | 0 | return -EINVAL; |
1518 | 17.2k | |
1519 | 17.2k | if (u->load_state != UNIT_STUB) |
1520 | 0 | return 0; |
1521 | 17.2k | |
1522 | 17.2k | if (u->transient_file) { |
1523 | 0 | /* Finalize transient file: if this is a transient unit file, as soon as we reach unit_load() the setup |
1524 | 0 | * is complete, hence let's synchronize the unit file we just wrote to disk. */ |
1525 | 0 |
|
1526 | 0 | r = fflush_and_check(u->transient_file); |
1527 | 0 | if (r < 0) |
1528 | 0 | goto fail; |
1529 | 0 | |
1530 | 0 | u->transient_file = safe_fclose(u->transient_file); |
1531 | 0 | u->fragment_mtime = now(CLOCK_REALTIME); |
1532 | 0 | } |
1533 | 17.2k | |
1534 | 17.2k | if (UNIT_VTABLE(u)->load) { |
1535 | 17.2k | r = UNIT_VTABLE(u)->load(u); |
1536 | 17.2k | if (r < 0) |
1537 | 10.3k | goto fail; |
1538 | 6.94k | } |
1539 | 6.94k | |
1540 | 6.94k | if (u->load_state == UNIT_STUB) { |
1541 | 0 | r = -ENOENT; |
1542 | 0 | goto fail; |
1543 | 0 | } |
1544 | 6.94k | |
1545 | 6.94k | if (u->load_state == UNIT_LOADED) { |
1546 | 6.94k | unit_add_to_target_deps_queue(u); |
1547 | 6.94k | |
1548 | 6.94k | r = unit_add_slice_dependencies(u); |
1549 | 6.94k | if (r < 0) |
1550 | 0 | goto fail; |
1551 | 6.94k | |
1552 | 6.94k | r = unit_add_mount_dependencies(u); |
1553 | 6.94k | if (r < 0) |
1554 | 0 | goto fail; |
1555 | 6.94k | |
1556 | 6.94k | r = unit_add_startup_units(u); |
1557 | 6.94k | if (r < 0) |
1558 | 0 | goto fail; |
1559 | 6.94k | |
1560 | 6.94k | if (u->on_failure_job_mode == JOB_ISOLATE && hashmap_size(u->dependencies[UNIT_ON_FAILURE]) > 1) { |
1561 | 0 | log_unit_error(u, "More than one OnFailure= dependencies specified but OnFailureJobMode=isolate set. Refusing."); |
1562 | 0 | r = -ENOEXEC; |
1563 | 0 | goto fail; |
1564 | 0 | } |
1565 | 6.94k | |
1566 | 6.94k | if (u->job_running_timeout != USEC_INFINITY && u->job_running_timeout > u->job_timeout) |
1567 | 6.94k | log_unit_warning(u, "JobRunningTimeoutSec= is greater than JobTimeoutSec=, it has no effect."); |
1568 | 6.94k | |
1569 | 6.94k | /* We finished loading, let's ensure our parents recalculate the members mask */ |
1570 | 6.94k | unit_invalidate_cgroup_members_masks(u); |
1571 | 6.94k | } |
1572 | 6.94k | |
1573 | 6.94k | assert((u->load_state != UNIT_MERGED) == !u->merged_into); |
1574 | 6.94k | |
1575 | 6.94k | unit_add_to_dbus_queue(unit_follow_merge(u)); |
1576 | 6.94k | unit_add_to_gc_queue(u); |
1577 | 6.94k | |
1578 | 6.94k | return 0; |
1579 | 10.3k | |
1580 | 10.3k | fail: |
1581 | 10.3k | /* We convert ENOEXEC errors to the UNIT_BAD_SETTING load state here. Configuration parsing code should hence |
1582 | 10.3k | * return ENOEXEC to ensure units are placed in this state after loading */ |
1583 | 10.3k | |
1584 | 10.3k | u->load_state = u->load_state == UNIT_STUB ? UNIT_NOT_FOUND : |
1585 | 10.3k | r == -ENOEXEC ? UNIT_BAD_SETTING : |
1586 | 422 | UNIT_ERROR; |
1587 | 10.3k | u->load_error = r; |
1588 | 10.3k | |
1589 | 10.3k | unit_add_to_dbus_queue(u); |
1590 | 10.3k | unit_add_to_gc_queue(u); |
1591 | 10.3k | |
1592 | 10.3k | return log_unit_debug_errno(u, r, "Failed to load configuration: %m"); |
1593 | 6.94k | } |
1594 | | |
1595 | | _printf_(7, 8) |
1596 | 0 | static int log_unit_internal(void *userdata, int level, int error, const char *file, int line, const char *func, const char *format, ...) { |
1597 | 0 | Unit *u = userdata; |
1598 | 0 | va_list ap; |
1599 | 0 | int r; |
1600 | 0 |
|
1601 | 0 | va_start(ap, format); |
1602 | 0 | if (u) |
1603 | 0 | r = log_object_internalv(level, error, file, line, func, |
1604 | 0 | u->manager->unit_log_field, |
1605 | 0 | u->id, |
1606 | 0 | u->manager->invocation_log_field, |
1607 | 0 | u->invocation_id_string, |
1608 | 0 | format, ap); |
1609 | 0 | else |
1610 | 0 | r = log_internalv(level, error, file, line, func, format, ap); |
1611 | 0 | va_end(ap); |
1612 | 0 |
|
1613 | 0 | return r; |
1614 | 0 | } |
1615 | | |
1616 | 0 | static bool unit_test_condition(Unit *u) { |
1617 | 0 | assert(u); |
1618 | 0 |
|
1619 | 0 | dual_timestamp_get(&u->condition_timestamp); |
1620 | 0 | u->condition_result = condition_test_list(u->conditions, condition_type_to_string, log_unit_internal, u); |
1621 | 0 |
|
1622 | 0 | unit_add_to_dbus_queue(u); |
1623 | 0 |
|
1624 | 0 | return u->condition_result; |
1625 | 0 | } |
1626 | | |
1627 | 0 | static bool unit_test_assert(Unit *u) { |
1628 | 0 | assert(u); |
1629 | 0 |
|
1630 | 0 | dual_timestamp_get(&u->assert_timestamp); |
1631 | 0 | u->assert_result = condition_test_list(u->asserts, assert_type_to_string, log_unit_internal, u); |
1632 | 0 |
|
1633 | 0 | unit_add_to_dbus_queue(u); |
1634 | 0 |
|
1635 | 0 | return u->assert_result; |
1636 | 0 | } |
1637 | | |
1638 | 0 | void unit_status_printf(Unit *u, const char *status, const char *unit_status_msg_format) { |
1639 | 0 | const char *d; |
1640 | 0 |
|
1641 | 0 | d = unit_description(u); |
1642 | 0 | if (log_get_show_color()) |
1643 | 0 | d = strjoina(ANSI_HIGHLIGHT, d, ANSI_NORMAL); |
1644 | 0 |
|
1645 | 0 | DISABLE_WARNING_FORMAT_NONLITERAL; |
1646 | 0 | manager_status_printf(u->manager, STATUS_TYPE_NORMAL, status, unit_status_msg_format, d); |
1647 | 0 | REENABLE_WARNING; |
1648 | 0 | } |
1649 | | |
1650 | 0 | int unit_test_start_limit(Unit *u) { |
1651 | 0 | const char *reason; |
1652 | 0 |
|
1653 | 0 | assert(u); |
1654 | 0 |
|
1655 | 0 | if (ratelimit_below(&u->start_limit)) { |
1656 | 0 | u->start_limit_hit = false; |
1657 | 0 | return 0; |
1658 | 0 | } |
1659 | 0 | |
1660 | 0 | log_unit_warning(u, "Start request repeated too quickly."); |
1661 | 0 | u->start_limit_hit = true; |
1662 | 0 |
|
1663 | 0 | reason = strjoina("unit ", u->id, " failed"); |
1664 | 0 |
|
1665 | 0 | emergency_action(u->manager, u->start_limit_action, |
1666 | 0 | EMERGENCY_ACTION_IS_WATCHDOG|EMERGENCY_ACTION_WARN, |
1667 | 0 | u->reboot_arg, -1, reason); |
1668 | 0 |
|
1669 | 0 | return -ECANCELED; |
1670 | 0 | } |
1671 | | |
1672 | 0 | bool unit_shall_confirm_spawn(Unit *u) { |
1673 | 0 | assert(u); |
1674 | 0 |
|
1675 | 0 | if (manager_is_confirm_spawn_disabled(u->manager)) |
1676 | 0 | return false; |
1677 | 0 | |
1678 | 0 | /* For some reasons units remaining in the same process group |
1679 | 0 | * as PID 1 fail to acquire the console even if it's not used |
1680 | 0 | * by any process. So skip the confirmation question for them. */ |
1681 | 0 | return !unit_get_exec_context(u)->same_pgrp; |
1682 | 0 | } |
1683 | | |
1684 | 0 | static bool unit_verify_deps(Unit *u) { |
1685 | 0 | Unit *other; |
1686 | 0 | Iterator j; |
1687 | 0 | void *v; |
1688 | 0 |
|
1689 | 0 | assert(u); |
1690 | 0 |
|
1691 | 0 | /* Checks whether all BindsTo= dependencies of this unit are fulfilled — if they are also combined with |
1692 | 0 | * After=. We do not check Requires= or Requisite= here as they only should have an effect on the job |
1693 | 0 | * processing, but do not have any effect afterwards. We don't check BindsTo= dependencies that are not used in |
1694 | 0 | * conjunction with After= as for them any such check would make things entirely racy. */ |
1695 | 0 |
|
1696 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_BINDS_TO], j) { |
1697 | 0 |
|
1698 | 0 | if (!hashmap_contains(u->dependencies[UNIT_AFTER], other)) |
1699 | 0 | continue; |
1700 | 0 | |
1701 | 0 | if (!UNIT_IS_ACTIVE_OR_RELOADING(unit_active_state(other))) { |
1702 | 0 | log_unit_notice(u, "Bound to unit %s, but unit isn't active.", other->id); |
1703 | 0 | return false; |
1704 | 0 | } |
1705 | 0 | } |
1706 | 0 |
|
1707 | 0 | return true; |
1708 | 0 | } |
1709 | | |
1710 | | /* Errors that aren't really errors: |
1711 | | * -EALREADY: Unit is already started. |
1712 | | * -ECOMM: Condition failed |
1713 | | * -EAGAIN: An operation is already in progress. Retry later. |
1714 | | * |
1715 | | * Errors that are real errors: |
1716 | | * -EBADR: This unit type does not support starting. |
1717 | | * -ECANCELED: Start limit hit, too many requests for now |
1718 | | * -EPROTO: Assert failed |
1719 | | * -EINVAL: Unit not loaded |
1720 | | * -EOPNOTSUPP: Unit type not supported |
1721 | | * -ENOLINK: The necessary dependencies are not fulfilled. |
1722 | | * -ESTALE: This unit has been started before and can't be started a second time |
1723 | | * -ENOENT: This is a triggering unit and unit to trigger is not loaded |
1724 | | */ |
1725 | 0 | int unit_start(Unit *u) { |
1726 | 0 | UnitActiveState state; |
1727 | 0 | Unit *following; |
1728 | 0 | int r; |
1729 | 0 |
|
1730 | 0 | assert(u); |
1731 | 0 |
|
1732 | 0 | /* If this is already started, then this will succeed. Note that this will even succeed if this unit |
1733 | 0 | * is not startable by the user. This is relied on to detect when we need to wait for units and when |
1734 | 0 | * waiting is finished. */ |
1735 | 0 | state = unit_active_state(u); |
1736 | 0 | if (UNIT_IS_ACTIVE_OR_RELOADING(state)) |
1737 | 0 | return -EALREADY; |
1738 | 0 | |
1739 | 0 | /* Units that aren't loaded cannot be started */ |
1740 | 0 | if (u->load_state != UNIT_LOADED) |
1741 | 0 | return -EINVAL; |
1742 | 0 | |
1743 | 0 | /* Refuse starting scope units more than once */ |
1744 | 0 | if (UNIT_VTABLE(u)->once_only && dual_timestamp_is_set(&u->inactive_enter_timestamp)) |
1745 | 0 | return -ESTALE; |
1746 | 0 | |
1747 | 0 | /* If the conditions failed, don't do anything at all. If we already are activating this call might |
1748 | 0 | * still be useful to speed up activation in case there is some hold-off time, but we don't want to |
1749 | 0 | * recheck the condition in that case. */ |
1750 | 0 | if (state != UNIT_ACTIVATING && |
1751 | 0 | !unit_test_condition(u)) { |
1752 | 0 |
|
1753 | 0 | /* Let's also check the start limit here. Normally, the start limit is only checked by the |
1754 | 0 | * .start() method of the unit type after it did some additional checks verifying everything |
1755 | 0 | * is in order (so that those other checks can propagate errors properly). However, if a |
1756 | 0 | * condition check doesn't hold we don't get that far but we should still ensure we are not |
1757 | 0 | * called in a tight loop without a rate limit check enforced, hence do the check here. Note |
1758 | 0 | * that ECOMM is generally not a reason for a job to fail, unlike most other errors here, |
1759 | 0 | * hence the chance is big that any triggering unit for us will trigger us again. Note this |
1760 | 0 | * condition check is a bit different from the condition check inside the per-unit .start() |
1761 | 0 | * function, as this one will not change the unit's state in any way (and we shouldn't here, |
1762 | 0 | * after all the condition failed). */ |
1763 | 0 |
|
1764 | 0 | r = unit_test_start_limit(u); |
1765 | 0 | if (r < 0) |
1766 | 0 | return r; |
1767 | 0 | |
1768 | 0 | return log_unit_debug_errno(u, SYNTHETIC_ERRNO(ECOMM), "Starting requested but condition failed. Not starting unit."); |
1769 | 0 | } |
1770 | 0 |
|
1771 | 0 | /* If the asserts failed, fail the entire job */ |
1772 | 0 | if (state != UNIT_ACTIVATING && |
1773 | 0 | !unit_test_assert(u)) |
1774 | 0 | return log_unit_notice_errno(u, SYNTHETIC_ERRNO(EPROTO), "Starting requested but asserts failed."); |
1775 | 0 | |
1776 | 0 | /* Units of types that aren't supported cannot be started. Note that we do this test only after the |
1777 | 0 | * condition checks, so that we rather return condition check errors (which are usually not |
1778 | 0 | * considered a true failure) than "not supported" errors (which are considered a failure). |
1779 | 0 | */ |
1780 | 0 | if (!unit_supported(u)) |
1781 | 0 | return -EOPNOTSUPP; |
1782 | 0 | |
1783 | 0 | /* Let's make sure that the deps really are in order before we start this. Normally the job engine |
1784 | 0 | * should have taken care of this already, but let's check this here again. After all, our |
1785 | 0 | * dependencies might not be in effect anymore, due to a reload or due to a failed condition. */ |
1786 | 0 | if (!unit_verify_deps(u)) |
1787 | 0 | return -ENOLINK; |
1788 | 0 | |
1789 | 0 | /* Forward to the main object, if we aren't it. */ |
1790 | 0 | following = unit_following(u); |
1791 | 0 | if (following) { |
1792 | 0 | log_unit_debug(u, "Redirecting start request from %s to %s.", u->id, following->id); |
1793 | 0 | return unit_start(following); |
1794 | 0 | } |
1795 | 0 |
|
1796 | 0 | /* If it is stopped, but we cannot start it, then fail */ |
1797 | 0 | if (!UNIT_VTABLE(u)->start) |
1798 | 0 | return -EBADR; |
1799 | 0 | |
1800 | 0 | /* We don't suppress calls to ->start() here when we are already starting, to allow this request to |
1801 | 0 | * be used as a "hurry up" call, for example when the unit is in some "auto restart" state where it |
1802 | 0 | * waits for a holdoff timer to elapse before it will start again. */ |
1803 | 0 | |
1804 | 0 | unit_add_to_dbus_queue(u); |
1805 | 0 |
|
1806 | 0 | return UNIT_VTABLE(u)->start(u); |
1807 | 0 | } |
1808 | | |
1809 | 0 | bool unit_can_start(Unit *u) { |
1810 | 0 | assert(u); |
1811 | 0 |
|
1812 | 0 | if (u->load_state != UNIT_LOADED) |
1813 | 0 | return false; |
1814 | 0 | |
1815 | 0 | if (!unit_supported(u)) |
1816 | 0 | return false; |
1817 | 0 | |
1818 | 0 | /* Scope units may be started only once */ |
1819 | 0 | if (UNIT_VTABLE(u)->once_only && dual_timestamp_is_set(&u->inactive_exit_timestamp)) |
1820 | 0 | return false; |
1821 | 0 | |
1822 | 0 | return !!UNIT_VTABLE(u)->start; |
1823 | 0 | } |
1824 | | |
1825 | 0 | bool unit_can_isolate(Unit *u) { |
1826 | 0 | assert(u); |
1827 | 0 |
|
1828 | 0 | return unit_can_start(u) && |
1829 | 0 | u->allow_isolate; |
1830 | 0 | } |
1831 | | |
1832 | | /* Errors: |
1833 | | * -EBADR: This unit type does not support stopping. |
1834 | | * -EALREADY: Unit is already stopped. |
1835 | | * -EAGAIN: An operation is already in progress. Retry later. |
1836 | | */ |
1837 | 0 | int unit_stop(Unit *u) { |
1838 | 0 | UnitActiveState state; |
1839 | 0 | Unit *following; |
1840 | 0 |
|
1841 | 0 | assert(u); |
1842 | 0 |
|
1843 | 0 | state = unit_active_state(u); |
1844 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(state)) |
1845 | 0 | return -EALREADY; |
1846 | 0 | |
1847 | 0 | following = unit_following(u); |
1848 | 0 | if (following) { |
1849 | 0 | log_unit_debug(u, "Redirecting stop request from %s to %s.", u->id, following->id); |
1850 | 0 | return unit_stop(following); |
1851 | 0 | } |
1852 | 0 |
|
1853 | 0 | if (!UNIT_VTABLE(u)->stop) |
1854 | 0 | return -EBADR; |
1855 | 0 | |
1856 | 0 | unit_add_to_dbus_queue(u); |
1857 | 0 |
|
1858 | 0 | return UNIT_VTABLE(u)->stop(u); |
1859 | 0 | } |
1860 | | |
1861 | 0 | bool unit_can_stop(Unit *u) { |
1862 | 0 | assert(u); |
1863 | 0 |
|
1864 | 0 | if (!unit_supported(u)) |
1865 | 0 | return false; |
1866 | 0 | |
1867 | 0 | if (u->perpetual) |
1868 | 0 | return false; |
1869 | 0 | |
1870 | 0 | return !!UNIT_VTABLE(u)->stop; |
1871 | 0 | } |
1872 | | |
1873 | | /* Errors: |
1874 | | * -EBADR: This unit type does not support reloading. |
1875 | | * -ENOEXEC: Unit is not started. |
1876 | | * -EAGAIN: An operation is already in progress. Retry later. |
1877 | | */ |
1878 | 0 | int unit_reload(Unit *u) { |
1879 | 0 | UnitActiveState state; |
1880 | 0 | Unit *following; |
1881 | 0 |
|
1882 | 0 | assert(u); |
1883 | 0 |
|
1884 | 0 | if (u->load_state != UNIT_LOADED) |
1885 | 0 | return -EINVAL; |
1886 | 0 | |
1887 | 0 | if (!unit_can_reload(u)) |
1888 | 0 | return -EBADR; |
1889 | 0 | |
1890 | 0 | state = unit_active_state(u); |
1891 | 0 | if (state == UNIT_RELOADING) |
1892 | 0 | return -EAGAIN; |
1893 | 0 | |
1894 | 0 | if (state != UNIT_ACTIVE) { |
1895 | 0 | log_unit_warning(u, "Unit cannot be reloaded because it is inactive."); |
1896 | 0 | return -ENOEXEC; |
1897 | 0 | } |
1898 | 0 |
|
1899 | 0 | following = unit_following(u); |
1900 | 0 | if (following) { |
1901 | 0 | log_unit_debug(u, "Redirecting reload request from %s to %s.", u->id, following->id); |
1902 | 0 | return unit_reload(following); |
1903 | 0 | } |
1904 | 0 |
|
1905 | 0 | unit_add_to_dbus_queue(u); |
1906 | 0 |
|
1907 | 0 | if (!UNIT_VTABLE(u)->reload) { |
1908 | 0 | /* Unit doesn't have a reload function, but we need to propagate the reload anyway */ |
1909 | 0 | unit_notify(u, unit_active_state(u), unit_active_state(u), 0); |
1910 | 0 | return 0; |
1911 | 0 | } |
1912 | 0 | |
1913 | 0 | return UNIT_VTABLE(u)->reload(u); |
1914 | 0 | } |
1915 | | |
1916 | 0 | bool unit_can_reload(Unit *u) { |
1917 | 0 | assert(u); |
1918 | 0 |
|
1919 | 0 | if (UNIT_VTABLE(u)->can_reload) |
1920 | 0 | return UNIT_VTABLE(u)->can_reload(u); |
1921 | 0 | |
1922 | 0 | if (!hashmap_isempty(u->dependencies[UNIT_PROPAGATES_RELOAD_TO])) |
1923 | 0 | return true; |
1924 | 0 | |
1925 | 0 | return UNIT_VTABLE(u)->reload; |
1926 | 0 | } |
1927 | | |
1928 | 0 | bool unit_is_unneeded(Unit *u) { |
1929 | 0 | static const UnitDependency deps[] = { |
1930 | 0 | UNIT_REQUIRED_BY, |
1931 | 0 | UNIT_REQUISITE_OF, |
1932 | 0 | UNIT_WANTED_BY, |
1933 | 0 | UNIT_BOUND_BY, |
1934 | 0 | }; |
1935 | 0 | size_t j; |
1936 | 0 |
|
1937 | 0 | assert(u); |
1938 | 0 |
|
1939 | 0 | if (!u->stop_when_unneeded) |
1940 | 0 | return false; |
1941 | 0 | |
1942 | 0 | /* Don't clean up while the unit is transitioning or is even inactive. */ |
1943 | 0 | if (!UNIT_IS_ACTIVE_OR_RELOADING(unit_active_state(u))) |
1944 | 0 | return false; |
1945 | 0 | if (u->job) |
1946 | 0 | return false; |
1947 | 0 | |
1948 | 0 | for (j = 0; j < ELEMENTSOF(deps); j++) { |
1949 | 0 | Unit *other; |
1950 | 0 | Iterator i; |
1951 | 0 | void *v; |
1952 | 0 |
|
1953 | 0 | /* If a dependent unit has a job queued, is active or transitioning, or is marked for |
1954 | 0 | * restart, then don't clean this one up. */ |
1955 | 0 |
|
1956 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[deps[j]], i) { |
1957 | 0 | if (other->job) |
1958 | 0 | return false; |
1959 | 0 | |
1960 | 0 | if (!UNIT_IS_INACTIVE_OR_FAILED(unit_active_state(other))) |
1961 | 0 | return false; |
1962 | 0 | |
1963 | 0 | if (unit_will_restart(other)) |
1964 | 0 | return false; |
1965 | 0 | } |
1966 | 0 | } |
1967 | 0 |
|
1968 | 0 | return true; |
1969 | 0 | } |
1970 | | |
1971 | 0 | static void check_unneeded_dependencies(Unit *u) { |
1972 | 0 |
|
1973 | 0 | static const UnitDependency deps[] = { |
1974 | 0 | UNIT_REQUIRES, |
1975 | 0 | UNIT_REQUISITE, |
1976 | 0 | UNIT_WANTS, |
1977 | 0 | UNIT_BINDS_TO, |
1978 | 0 | }; |
1979 | 0 | size_t j; |
1980 | 0 |
|
1981 | 0 | assert(u); |
1982 | 0 |
|
1983 | 0 | /* Add all units this unit depends on to the queue that processes StopWhenUnneeded= behaviour. */ |
1984 | 0 |
|
1985 | 0 | for (j = 0; j < ELEMENTSOF(deps); j++) { |
1986 | 0 | Unit *other; |
1987 | 0 | Iterator i; |
1988 | 0 | void *v; |
1989 | 0 |
|
1990 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[deps[j]], i) |
1991 | 0 | unit_submit_to_stop_when_unneeded_queue(other); |
1992 | 0 | } |
1993 | 0 | } |
1994 | | |
1995 | 0 | static void unit_check_binds_to(Unit *u) { |
1996 | 0 | _cleanup_(sd_bus_error_free) sd_bus_error error = SD_BUS_ERROR_NULL; |
1997 | 0 | bool stop = false; |
1998 | 0 | Unit *other; |
1999 | 0 | Iterator i; |
2000 | 0 | void *v; |
2001 | 0 | int r; |
2002 | 0 |
|
2003 | 0 | assert(u); |
2004 | 0 |
|
2005 | 0 | if (u->job) |
2006 | 0 | return; |
2007 | 0 | |
2008 | 0 | if (unit_active_state(u) != UNIT_ACTIVE) |
2009 | 0 | return; |
2010 | 0 | |
2011 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_BINDS_TO], i) { |
2012 | 0 | if (other->job) |
2013 | 0 | continue; |
2014 | 0 | |
2015 | 0 | if (!other->coldplugged) |
2016 | 0 | /* We might yet create a job for the other unit… */ |
2017 | 0 | continue; |
2018 | 0 | |
2019 | 0 | if (!UNIT_IS_INACTIVE_OR_FAILED(unit_active_state(other))) |
2020 | 0 | continue; |
2021 | 0 | |
2022 | 0 | stop = true; |
2023 | 0 | break; |
2024 | 0 | } |
2025 | 0 |
|
2026 | 0 | if (!stop) |
2027 | 0 | return; |
2028 | 0 | |
2029 | 0 | /* If stopping a unit fails continuously we might enter a stop |
2030 | 0 | * loop here, hence stop acting on the service being |
2031 | 0 | * unnecessary after a while. */ |
2032 | 0 | if (!ratelimit_below(&u->auto_stop_ratelimit)) { |
2033 | 0 | log_unit_warning(u, "Unit is bound to inactive unit %s, but not stopping since we tried this too often recently.", other->id); |
2034 | 0 | return; |
2035 | 0 | } |
2036 | 0 |
|
2037 | 0 | assert(other); |
2038 | 0 | log_unit_info(u, "Unit is bound to inactive unit %s. Stopping, too.", other->id); |
2039 | 0 |
|
2040 | 0 | /* A unit we need to run is gone. Sniff. Let's stop this. */ |
2041 | 0 | r = manager_add_job(u->manager, JOB_STOP, u, JOB_FAIL, NULL, &error, NULL); |
2042 | 0 | if (r < 0) |
2043 | 0 | log_unit_warning_errno(u, r, "Failed to enqueue stop job, ignoring: %s", bus_error_message(&error, r)); |
2044 | 0 | } |
2045 | | |
2046 | 0 | static void retroactively_start_dependencies(Unit *u) { |
2047 | 0 | Iterator i; |
2048 | 0 | Unit *other; |
2049 | 0 | void *v; |
2050 | 0 |
|
2051 | 0 | assert(u); |
2052 | 0 | assert(UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(u))); |
2053 | 0 |
|
2054 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_REQUIRES], i) |
2055 | 0 | if (!hashmap_get(u->dependencies[UNIT_AFTER], other) && |
2056 | 0 | !UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(other))) |
2057 | 0 | manager_add_job(u->manager, JOB_START, other, JOB_REPLACE, NULL, NULL, NULL); |
2058 | 0 |
|
2059 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_BINDS_TO], i) |
2060 | 0 | if (!hashmap_get(u->dependencies[UNIT_AFTER], other) && |
2061 | 0 | !UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(other))) |
2062 | 0 | manager_add_job(u->manager, JOB_START, other, JOB_REPLACE, NULL, NULL, NULL); |
2063 | 0 |
|
2064 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_WANTS], i) |
2065 | 0 | if (!hashmap_get(u->dependencies[UNIT_AFTER], other) && |
2066 | 0 | !UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(other))) |
2067 | 0 | manager_add_job(u->manager, JOB_START, other, JOB_FAIL, NULL, NULL, NULL); |
2068 | 0 |
|
2069 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_CONFLICTS], i) |
2070 | 0 | if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other))) |
2071 | 0 | manager_add_job(u->manager, JOB_STOP, other, JOB_REPLACE, NULL, NULL, NULL); |
2072 | 0 |
|
2073 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_CONFLICTED_BY], i) |
2074 | 0 | if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other))) |
2075 | 0 | manager_add_job(u->manager, JOB_STOP, other, JOB_REPLACE, NULL, NULL, NULL); |
2076 | 0 | } |
2077 | | |
2078 | 0 | static void retroactively_stop_dependencies(Unit *u) { |
2079 | 0 | Unit *other; |
2080 | 0 | Iterator i; |
2081 | 0 | void *v; |
2082 | 0 |
|
2083 | 0 | assert(u); |
2084 | 0 | assert(UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(u))); |
2085 | 0 |
|
2086 | 0 | /* Pull down units which are bound to us recursively if enabled */ |
2087 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_BOUND_BY], i) |
2088 | 0 | if (!UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(other))) |
2089 | 0 | manager_add_job(u->manager, JOB_STOP, other, JOB_REPLACE, NULL, NULL, NULL); |
2090 | 0 | } |
2091 | | |
2092 | 0 | void unit_start_on_failure(Unit *u) { |
2093 | 0 | Unit *other; |
2094 | 0 | Iterator i; |
2095 | 0 | void *v; |
2096 | 0 | int r; |
2097 | 0 |
|
2098 | 0 | assert(u); |
2099 | 0 |
|
2100 | 0 | if (hashmap_size(u->dependencies[UNIT_ON_FAILURE]) <= 0) |
2101 | 0 | return; |
2102 | 0 | |
2103 | 0 | log_unit_info(u, "Triggering OnFailure= dependencies."); |
2104 | 0 |
|
2105 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_ON_FAILURE], i) { |
2106 | 0 | _cleanup_(sd_bus_error_free) sd_bus_error error = SD_BUS_ERROR_NULL; |
2107 | 0 |
|
2108 | 0 | r = manager_add_job(u->manager, JOB_START, other, u->on_failure_job_mode, NULL, &error, NULL); |
2109 | 0 | if (r < 0) |
2110 | 0 | log_unit_warning_errno(u, r, "Failed to enqueue OnFailure= job, ignoring: %s", bus_error_message(&error, r)); |
2111 | 0 | } |
2112 | 0 | } |
2113 | | |
2114 | 0 | void unit_trigger_notify(Unit *u) { |
2115 | 0 | Unit *other; |
2116 | 0 | Iterator i; |
2117 | 0 | void *v; |
2118 | 0 |
|
2119 | 0 | assert(u); |
2120 | 0 |
|
2121 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_TRIGGERED_BY], i) |
2122 | 0 | if (UNIT_VTABLE(other)->trigger_notify) |
2123 | 0 | UNIT_VTABLE(other)->trigger_notify(other, u); |
2124 | 0 | } |
2125 | | |
2126 | 0 | static int unit_log_resources(Unit *u) { |
2127 | 0 | struct iovec iovec[1 + _CGROUP_IP_ACCOUNTING_METRIC_MAX + _CGROUP_IO_ACCOUNTING_METRIC_MAX + 4]; |
2128 | 0 | bool any_traffic = false, have_ip_accounting = false, any_io = false, have_io_accounting = false; |
2129 | 0 | _cleanup_free_ char *igress = NULL, *egress = NULL, *rr = NULL, *wr = NULL; |
2130 | 0 | size_t n_message_parts = 0, n_iovec = 0; |
2131 | 0 | char* message_parts[1 + 2 + 2 + 1], *t; |
2132 | 0 | nsec_t nsec = NSEC_INFINITY; |
2133 | 0 | CGroupIPAccountingMetric m; |
2134 | 0 | size_t i; |
2135 | 0 | int r; |
2136 | 0 | const char* const ip_fields[_CGROUP_IP_ACCOUNTING_METRIC_MAX] = { |
2137 | 0 | [CGROUP_IP_INGRESS_BYTES] = "IP_METRIC_INGRESS_BYTES", |
2138 | 0 | [CGROUP_IP_INGRESS_PACKETS] = "IP_METRIC_INGRESS_PACKETS", |
2139 | 0 | [CGROUP_IP_EGRESS_BYTES] = "IP_METRIC_EGRESS_BYTES", |
2140 | 0 | [CGROUP_IP_EGRESS_PACKETS] = "IP_METRIC_EGRESS_PACKETS", |
2141 | 0 | }; |
2142 | 0 | const char* const io_fields[_CGROUP_IO_ACCOUNTING_METRIC_MAX] = { |
2143 | 0 | [CGROUP_IO_READ_BYTES] = "IO_METRIC_READ_BYTES", |
2144 | 0 | [CGROUP_IO_WRITE_BYTES] = "IO_METRIC_WRITE_BYTES", |
2145 | 0 | [CGROUP_IO_READ_OPERATIONS] = "IO_METRIC_READ_OPERATIONS", |
2146 | 0 | [CGROUP_IO_WRITE_OPERATIONS] = "IO_METRIC_WRITE_OPERATIONS", |
2147 | 0 | }; |
2148 | 0 |
|
2149 | 0 | assert(u); |
2150 | 0 |
|
2151 | 0 | /* Invoked whenever a unit enters failed or dead state. Logs information about consumed resources if resource |
2152 | 0 | * accounting was enabled for a unit. It does this in two ways: a friendly human readable string with reduced |
2153 | 0 | * information and the complete data in structured fields. */ |
2154 | 0 |
|
2155 | 0 | (void) unit_get_cpu_usage(u, &nsec); |
2156 | 0 | if (nsec != NSEC_INFINITY) { |
2157 | 0 | char buf[FORMAT_TIMESPAN_MAX] = ""; |
2158 | 0 |
|
2159 | 0 | /* Format the CPU time for inclusion in the structured log message */ |
2160 | 0 | if (asprintf(&t, "CPU_USAGE_NSEC=%" PRIu64, nsec) < 0) { |
2161 | 0 | r = log_oom(); |
2162 | 0 | goto finish; |
2163 | 0 | } |
2164 | 0 | iovec[n_iovec++] = IOVEC_MAKE_STRING(t); |
2165 | 0 |
|
2166 | 0 | /* Format the CPU time for inclusion in the human language message string */ |
2167 | 0 | format_timespan(buf, sizeof(buf), nsec / NSEC_PER_USEC, USEC_PER_MSEC); |
2168 | 0 | t = strjoin("consumed ", buf, " CPU time"); |
2169 | 0 | if (!t) { |
2170 | 0 | r = log_oom(); |
2171 | 0 | goto finish; |
2172 | 0 | } |
2173 | 0 |
|
2174 | 0 | message_parts[n_message_parts++] = t; |
2175 | 0 | } |
2176 | 0 |
|
2177 | 0 | for (CGroupIOAccountingMetric k = 0; k < _CGROUP_IO_ACCOUNTING_METRIC_MAX; k++) { |
2178 | 0 | char buf[FORMAT_BYTES_MAX] = ""; |
2179 | 0 | uint64_t value = UINT64_MAX; |
2180 | 0 |
|
2181 | 0 | assert(io_fields[k]); |
2182 | 0 |
|
2183 | 0 | (void) unit_get_io_accounting(u, k, k > 0, &value); |
2184 | 0 | if (value == UINT64_MAX) |
2185 | 0 | continue; |
2186 | 0 | |
2187 | 0 | have_io_accounting = true; |
2188 | 0 | if (value > 0) |
2189 | 0 | any_io = true; |
2190 | 0 |
|
2191 | 0 | /* Format IO accounting data for inclusion in the structured log message */ |
2192 | 0 | if (asprintf(&t, "%s=%" PRIu64, io_fields[k], value) < 0) { |
2193 | 0 | r = log_oom(); |
2194 | 0 | goto finish; |
2195 | 0 | } |
2196 | 0 | iovec[n_iovec++] = IOVEC_MAKE_STRING(t); |
2197 | 0 |
|
2198 | 0 | /* Format the IO accounting data for inclusion in the human language message string, but only |
2199 | 0 | * for the bytes counters (and not for the operations counters) */ |
2200 | 0 | if (k == CGROUP_IO_READ_BYTES) { |
2201 | 0 | assert(!rr); |
2202 | 0 | rr = strjoin("read ", format_bytes(buf, sizeof(buf), value), " from disk"); |
2203 | 0 | if (!rr) { |
2204 | 0 | r = log_oom(); |
2205 | 0 | goto finish; |
2206 | 0 | } |
2207 | 0 | } else if (k == CGROUP_IO_WRITE_BYTES) { |
2208 | 0 | assert(!wr); |
2209 | 0 | wr = strjoin("written ", format_bytes(buf, sizeof(buf), value), " to disk"); |
2210 | 0 | if (!wr) { |
2211 | 0 | r = log_oom(); |
2212 | 0 | goto finish; |
2213 | 0 | } |
2214 | 0 | } |
2215 | 0 | } |
2216 | 0 |
|
2217 | 0 | if (have_io_accounting) { |
2218 | 0 | if (any_io) { |
2219 | 0 | if (rr) |
2220 | 0 | message_parts[n_message_parts++] = TAKE_PTR(rr); |
2221 | 0 | if (wr) |
2222 | 0 | message_parts[n_message_parts++] = TAKE_PTR(wr); |
2223 | 0 |
|
2224 | 0 | } else { |
2225 | 0 | char *k; |
2226 | 0 |
|
2227 | 0 | k = strdup("no IO"); |
2228 | 0 | if (!k) { |
2229 | 0 | r = log_oom(); |
2230 | 0 | goto finish; |
2231 | 0 | } |
2232 | 0 |
|
2233 | 0 | message_parts[n_message_parts++] = k; |
2234 | 0 | } |
2235 | 0 | } |
2236 | 0 |
|
2237 | 0 | for (m = 0; m < _CGROUP_IP_ACCOUNTING_METRIC_MAX; m++) { |
2238 | 0 | char buf[FORMAT_BYTES_MAX] = ""; |
2239 | 0 | uint64_t value = UINT64_MAX; |
2240 | 0 |
|
2241 | 0 | assert(ip_fields[m]); |
2242 | 0 |
|
2243 | 0 | (void) unit_get_ip_accounting(u, m, &value); |
2244 | 0 | if (value == UINT64_MAX) |
2245 | 0 | continue; |
2246 | 0 | |
2247 | 0 | have_ip_accounting = true; |
2248 | 0 | if (value > 0) |
2249 | 0 | any_traffic = true; |
2250 | 0 |
|
2251 | 0 | /* Format IP accounting data for inclusion in the structured log message */ |
2252 | 0 | if (asprintf(&t, "%s=%" PRIu64, ip_fields[m], value) < 0) { |
2253 | 0 | r = log_oom(); |
2254 | 0 | goto finish; |
2255 | 0 | } |
2256 | 0 | iovec[n_iovec++] = IOVEC_MAKE_STRING(t); |
2257 | 0 |
|
2258 | 0 | /* Format the IP accounting data for inclusion in the human language message string, but only for the |
2259 | 0 | * bytes counters (and not for the packets counters) */ |
2260 | 0 | if (m == CGROUP_IP_INGRESS_BYTES) { |
2261 | 0 | assert(!igress); |
2262 | 0 | igress = strjoin("received ", format_bytes(buf, sizeof(buf), value), " IP traffic"); |
2263 | 0 | if (!igress) { |
2264 | 0 | r = log_oom(); |
2265 | 0 | goto finish; |
2266 | 0 | } |
2267 | 0 | } else if (m == CGROUP_IP_EGRESS_BYTES) { |
2268 | 0 | assert(!egress); |
2269 | 0 | egress = strjoin("sent ", format_bytes(buf, sizeof(buf), value), " IP traffic"); |
2270 | 0 | if (!egress) { |
2271 | 0 | r = log_oom(); |
2272 | 0 | goto finish; |
2273 | 0 | } |
2274 | 0 | } |
2275 | 0 | } |
2276 | 0 |
|
2277 | 0 | if (have_ip_accounting) { |
2278 | 0 | if (any_traffic) { |
2279 | 0 | if (igress) |
2280 | 0 | message_parts[n_message_parts++] = TAKE_PTR(igress); |
2281 | 0 | if (egress) |
2282 | 0 | message_parts[n_message_parts++] = TAKE_PTR(egress); |
2283 | 0 |
|
2284 | 0 | } else { |
2285 | 0 | char *k; |
2286 | 0 |
|
2287 | 0 | k = strdup("no IP traffic"); |
2288 | 0 | if (!k) { |
2289 | 0 | r = log_oom(); |
2290 | 0 | goto finish; |
2291 | 0 | } |
2292 | 0 |
|
2293 | 0 | message_parts[n_message_parts++] = k; |
2294 | 0 | } |
2295 | 0 | } |
2296 | 0 |
|
2297 | 0 | /* Is there any accounting data available at all? */ |
2298 | 0 | if (n_iovec == 0) { |
2299 | 0 | r = 0; |
2300 | 0 | goto finish; |
2301 | 0 | } |
2302 | 0 | |
2303 | 0 | if (n_message_parts == 0) |
2304 | 0 | t = strjoina("MESSAGE=", u->id, ": Completed."); |
2305 | 0 | else { |
2306 | 0 | _cleanup_free_ char *joined; |
2307 | 0 |
|
2308 | 0 | message_parts[n_message_parts] = NULL; |
2309 | 0 |
|
2310 | 0 | joined = strv_join(message_parts, ", "); |
2311 | 0 | if (!joined) { |
2312 | 0 | r = log_oom(); |
2313 | 0 | goto finish; |
2314 | 0 | } |
2315 | 0 |
|
2316 | 0 | joined[0] = ascii_toupper(joined[0]); |
2317 | 0 | t = strjoina("MESSAGE=", u->id, ": ", joined, "."); |
2318 | 0 | } |
2319 | 0 |
|
2320 | 0 | /* The following four fields we allocate on the stack or are static strings, we hence don't want to free them, |
2321 | 0 | * and hence don't increase n_iovec for them */ |
2322 | 0 | iovec[n_iovec] = IOVEC_MAKE_STRING(t); |
2323 | 0 | iovec[n_iovec + 1] = IOVEC_MAKE_STRING("MESSAGE_ID=" SD_MESSAGE_UNIT_RESOURCES_STR); |
2324 | 0 |
|
2325 | 0 | t = strjoina(u->manager->unit_log_field, u->id); |
2326 | 0 | iovec[n_iovec + 2] = IOVEC_MAKE_STRING(t); |
2327 | 0 |
|
2328 | 0 | t = strjoina(u->manager->invocation_log_field, u->invocation_id_string); |
2329 | 0 | iovec[n_iovec + 3] = IOVEC_MAKE_STRING(t); |
2330 | 0 |
|
2331 | 0 | log_struct_iovec(LOG_INFO, iovec, n_iovec + 4); |
2332 | 0 | r = 0; |
2333 | 0 |
|
2334 | 0 | finish: |
2335 | 0 | for (i = 0; i < n_message_parts; i++) |
2336 | 0 | free(message_parts[i]); |
2337 | 0 |
|
2338 | 0 | for (i = 0; i < n_iovec; i++) |
2339 | 0 | free(iovec[i].iov_base); |
2340 | 0 |
|
2341 | 0 | return r; |
2342 | 0 |
|
2343 | 0 | } |
2344 | | |
2345 | 0 | static void unit_update_on_console(Unit *u) { |
2346 | 0 | bool b; |
2347 | 0 |
|
2348 | 0 | assert(u); |
2349 | 0 |
|
2350 | 0 | b = unit_needs_console(u); |
2351 | 0 | if (u->on_console == b) |
2352 | 0 | return; |
2353 | 0 | |
2354 | 0 | u->on_console = b; |
2355 | 0 | if (b) |
2356 | 0 | manager_ref_console(u->manager); |
2357 | 0 | else |
2358 | 0 | manager_unref_console(u->manager); |
2359 | 0 | } |
2360 | | |
2361 | 0 | static void unit_emit_audit_start(Unit *u) { |
2362 | 0 | assert(u); |
2363 | 0 |
|
2364 | 0 | if (u->type != UNIT_SERVICE) |
2365 | 0 | return; |
2366 | 0 | |
2367 | 0 | /* Write audit record if we have just finished starting up */ |
2368 | 0 | manager_send_unit_audit(u->manager, u, AUDIT_SERVICE_START, true); |
2369 | 0 | u->in_audit = true; |
2370 | 0 | } |
2371 | | |
2372 | 0 | static void unit_emit_audit_stop(Unit *u, UnitActiveState state) { |
2373 | 0 | assert(u); |
2374 | 0 |
|
2375 | 0 | if (u->type != UNIT_SERVICE) |
2376 | 0 | return; |
2377 | 0 | |
2378 | 0 | if (u->in_audit) { |
2379 | 0 | /* Write audit record if we have just finished shutting down */ |
2380 | 0 | manager_send_unit_audit(u->manager, u, AUDIT_SERVICE_STOP, state == UNIT_INACTIVE); |
2381 | 0 | u->in_audit = false; |
2382 | 0 | } else { |
2383 | 0 | /* Hmm, if there was no start record written write it now, so that we always have a nice pair */ |
2384 | 0 | manager_send_unit_audit(u->manager, u, AUDIT_SERVICE_START, state == UNIT_INACTIVE); |
2385 | 0 |
|
2386 | 0 | if (state == UNIT_INACTIVE) |
2387 | 0 | manager_send_unit_audit(u->manager, u, AUDIT_SERVICE_STOP, true); |
2388 | 0 | } |
2389 | 0 | } |
2390 | | |
2391 | 0 | static bool unit_process_job(Job *j, UnitActiveState ns, UnitNotifyFlags flags) { |
2392 | 0 | bool unexpected = false; |
2393 | 0 |
|
2394 | 0 | assert(j); |
2395 | 0 |
|
2396 | 0 | if (j->state == JOB_WAITING) |
2397 | 0 |
|
2398 | 0 | /* So we reached a different state for this job. Let's see if we can run it now if it failed previously |
2399 | 0 | * due to EAGAIN. */ |
2400 | 0 | job_add_to_run_queue(j); |
2401 | 0 |
|
2402 | 0 | /* Let's check whether the unit's new state constitutes a finished job, or maybe contradicts a running job and |
2403 | 0 | * hence needs to invalidate jobs. */ |
2404 | 0 |
|
2405 | 0 | switch (j->type) { |
2406 | 0 |
|
2407 | 0 | case JOB_START: |
2408 | 0 | case JOB_VERIFY_ACTIVE: |
2409 | 0 |
|
2410 | 0 | if (UNIT_IS_ACTIVE_OR_RELOADING(ns)) |
2411 | 0 | job_finish_and_invalidate(j, JOB_DONE, true, false); |
2412 | 0 | else if (j->state == JOB_RUNNING && ns != UNIT_ACTIVATING) { |
2413 | 0 | unexpected = true; |
2414 | 0 |
|
2415 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(ns)) |
2416 | 0 | job_finish_and_invalidate(j, ns == UNIT_FAILED ? JOB_FAILED : JOB_DONE, true, false); |
2417 | 0 | } |
2418 | 0 |
|
2419 | 0 | break; |
2420 | 0 |
|
2421 | 0 | case JOB_RELOAD: |
2422 | 0 | case JOB_RELOAD_OR_START: |
2423 | 0 | case JOB_TRY_RELOAD: |
2424 | 0 |
|
2425 | 0 | if (j->state == JOB_RUNNING) { |
2426 | 0 | if (ns == UNIT_ACTIVE) |
2427 | 0 | job_finish_and_invalidate(j, (flags & UNIT_NOTIFY_RELOAD_FAILURE) ? JOB_FAILED : JOB_DONE, true, false); |
2428 | 0 | else if (!IN_SET(ns, UNIT_ACTIVATING, UNIT_RELOADING)) { |
2429 | 0 | unexpected = true; |
2430 | 0 |
|
2431 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(ns)) |
2432 | 0 | job_finish_and_invalidate(j, ns == UNIT_FAILED ? JOB_FAILED : JOB_DONE, true, false); |
2433 | 0 | } |
2434 | 0 | } |
2435 | 0 |
|
2436 | 0 | break; |
2437 | 0 |
|
2438 | 0 | case JOB_STOP: |
2439 | 0 | case JOB_RESTART: |
2440 | 0 | case JOB_TRY_RESTART: |
2441 | 0 |
|
2442 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(ns)) |
2443 | 0 | job_finish_and_invalidate(j, JOB_DONE, true, false); |
2444 | 0 | else if (j->state == JOB_RUNNING && ns != UNIT_DEACTIVATING) { |
2445 | 0 | unexpected = true; |
2446 | 0 | job_finish_and_invalidate(j, JOB_FAILED, true, false); |
2447 | 0 | } |
2448 | 0 |
|
2449 | 0 | break; |
2450 | 0 |
|
2451 | 0 | default: |
2452 | 0 | assert_not_reached("Job type unknown"); |
2453 | 0 | } |
2454 | 0 |
|
2455 | 0 | return unexpected; |
2456 | 0 | } |
2457 | | |
2458 | 0 | void unit_notify(Unit *u, UnitActiveState os, UnitActiveState ns, UnitNotifyFlags flags) { |
2459 | 0 | const char *reason; |
2460 | 0 | Manager *m; |
2461 | 0 |
|
2462 | 0 | assert(u); |
2463 | 0 | assert(os < _UNIT_ACTIVE_STATE_MAX); |
2464 | 0 | assert(ns < _UNIT_ACTIVE_STATE_MAX); |
2465 | 0 |
|
2466 | 0 | /* Note that this is called for all low-level state changes, even if they might map to the same high-level |
2467 | 0 | * UnitActiveState! That means that ns == os is an expected behavior here. For example: if a mount point is |
2468 | 0 | * remounted this function will be called too! */ |
2469 | 0 |
|
2470 | 0 | m = u->manager; |
2471 | 0 |
|
2472 | 0 | /* Let's enqueue the change signal early. In case this unit has a job associated we want that this unit is in |
2473 | 0 | * the bus queue, so that any job change signal queued will force out the unit change signal first. */ |
2474 | 0 | unit_add_to_dbus_queue(u); |
2475 | 0 |
|
2476 | 0 | /* Update timestamps for state changes */ |
2477 | 0 | if (!MANAGER_IS_RELOADING(m)) { |
2478 | 0 | dual_timestamp_get(&u->state_change_timestamp); |
2479 | 0 |
|
2480 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(os) && !UNIT_IS_INACTIVE_OR_FAILED(ns)) |
2481 | 0 | u->inactive_exit_timestamp = u->state_change_timestamp; |
2482 | 0 | else if (!UNIT_IS_INACTIVE_OR_FAILED(os) && UNIT_IS_INACTIVE_OR_FAILED(ns)) |
2483 | 0 | u->inactive_enter_timestamp = u->state_change_timestamp; |
2484 | 0 |
|
2485 | 0 | if (!UNIT_IS_ACTIVE_OR_RELOADING(os) && UNIT_IS_ACTIVE_OR_RELOADING(ns)) |
2486 | 0 | u->active_enter_timestamp = u->state_change_timestamp; |
2487 | 0 | else if (UNIT_IS_ACTIVE_OR_RELOADING(os) && !UNIT_IS_ACTIVE_OR_RELOADING(ns)) |
2488 | 0 | u->active_exit_timestamp = u->state_change_timestamp; |
2489 | 0 | } |
2490 | 0 |
|
2491 | 0 | /* Keep track of failed units */ |
2492 | 0 | (void) manager_update_failed_units(m, u, ns == UNIT_FAILED); |
2493 | 0 |
|
2494 | 0 | /* Make sure the cgroup and state files are always removed when we become inactive */ |
2495 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(ns)) { |
2496 | 0 | unit_prune_cgroup(u); |
2497 | 0 | unit_unlink_state_files(u); |
2498 | 0 | } |
2499 | 0 |
|
2500 | 0 | unit_update_on_console(u); |
2501 | 0 |
|
2502 | 0 | if (!MANAGER_IS_RELOADING(m)) { |
2503 | 0 | bool unexpected; |
2504 | 0 |
|
2505 | 0 | /* Let's propagate state changes to the job */ |
2506 | 0 | if (u->job) |
2507 | 0 | unexpected = unit_process_job(u->job, ns, flags); |
2508 | 0 | else |
2509 | 0 | unexpected = true; |
2510 | 0 |
|
2511 | 0 | /* If this state change happened without being requested by a job, then let's retroactively start or |
2512 | 0 | * stop dependencies. We skip that step when deserializing, since we don't want to create any |
2513 | 0 | * additional jobs just because something is already activated. */ |
2514 | 0 |
|
2515 | 0 | if (unexpected) { |
2516 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(os) && UNIT_IS_ACTIVE_OR_ACTIVATING(ns)) |
2517 | 0 | retroactively_start_dependencies(u); |
2518 | 0 | else if (UNIT_IS_ACTIVE_OR_ACTIVATING(os) && UNIT_IS_INACTIVE_OR_DEACTIVATING(ns)) |
2519 | 0 | retroactively_stop_dependencies(u); |
2520 | 0 | } |
2521 | 0 |
|
2522 | 0 | /* stop unneeded units regardless if going down was expected or not */ |
2523 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(ns)) |
2524 | 0 | check_unneeded_dependencies(u); |
2525 | 0 |
|
2526 | 0 | if (ns != os && ns == UNIT_FAILED) { |
2527 | 0 | log_unit_debug(u, "Unit entered failed state."); |
2528 | 0 |
|
2529 | 0 | if (!(flags & UNIT_NOTIFY_WILL_AUTO_RESTART)) |
2530 | 0 | unit_start_on_failure(u); |
2531 | 0 | } |
2532 | 0 |
|
2533 | 0 | if (UNIT_IS_ACTIVE_OR_RELOADING(ns) && !UNIT_IS_ACTIVE_OR_RELOADING(os)) { |
2534 | 0 | /* This unit just finished starting up */ |
2535 | 0 |
|
2536 | 0 | unit_emit_audit_start(u); |
2537 | 0 | manager_send_unit_plymouth(m, u); |
2538 | 0 | } |
2539 | 0 |
|
2540 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(ns) && !UNIT_IS_INACTIVE_OR_FAILED(os)) { |
2541 | 0 | /* This unit just stopped/failed. */ |
2542 | 0 |
|
2543 | 0 | unit_emit_audit_stop(u, ns); |
2544 | 0 | unit_log_resources(u); |
2545 | 0 | } |
2546 | 0 | } |
2547 | 0 |
|
2548 | 0 | manager_recheck_journal(m); |
2549 | 0 | manager_recheck_dbus(m); |
2550 | 0 |
|
2551 | 0 | unit_trigger_notify(u); |
2552 | 0 |
|
2553 | 0 | if (!MANAGER_IS_RELOADING(m)) { |
2554 | 0 | /* Maybe we finished startup and are now ready for being stopped because unneeded? */ |
2555 | 0 | unit_submit_to_stop_when_unneeded_queue(u); |
2556 | 0 |
|
2557 | 0 | /* Maybe we finished startup, but something we needed has vanished? Let's die then. (This happens when |
2558 | 0 | * something BindsTo= to a Type=oneshot unit, as these units go directly from starting to inactive, |
2559 | 0 | * without ever entering started.) */ |
2560 | 0 | unit_check_binds_to(u); |
2561 | 0 |
|
2562 | 0 | if (os != UNIT_FAILED && ns == UNIT_FAILED) { |
2563 | 0 | reason = strjoina("unit ", u->id, " failed"); |
2564 | 0 | emergency_action(m, u->failure_action, 0, u->reboot_arg, unit_failure_action_exit_status(u), reason); |
2565 | 0 | } else if (!UNIT_IS_INACTIVE_OR_FAILED(os) && ns == UNIT_INACTIVE) { |
2566 | 0 | reason = strjoina("unit ", u->id, " succeeded"); |
2567 | 0 | emergency_action(m, u->success_action, 0, u->reboot_arg, unit_success_action_exit_status(u), reason); |
2568 | 0 | } |
2569 | 0 | } |
2570 | 0 |
|
2571 | 0 | unit_add_to_gc_queue(u); |
2572 | 0 | } |
2573 | | |
2574 | 0 | int unit_watch_pid(Unit *u, pid_t pid, bool exclusive) { |
2575 | 0 | int r; |
2576 | 0 |
|
2577 | 0 | assert(u); |
2578 | 0 | assert(pid_is_valid(pid)); |
2579 | 0 |
|
2580 | 0 | /* Watch a specific PID */ |
2581 | 0 |
|
2582 | 0 | /* Caller might be sure that this PID belongs to this unit only. Let's take this |
2583 | 0 | * opportunity to remove any stalled references to this PID as they can be created |
2584 | 0 | * easily (when watching a process which is not our direct child). */ |
2585 | 0 | if (exclusive) |
2586 | 0 | manager_unwatch_pid(u->manager, pid); |
2587 | 0 |
|
2588 | 0 | r = set_ensure_allocated(&u->pids, NULL); |
2589 | 0 | if (r < 0) |
2590 | 0 | return r; |
2591 | 0 | |
2592 | 0 | r = hashmap_ensure_allocated(&u->manager->watch_pids, NULL); |
2593 | 0 | if (r < 0) |
2594 | 0 | return r; |
2595 | 0 | |
2596 | 0 | /* First try, let's add the unit keyed by "pid". */ |
2597 | 0 | r = hashmap_put(u->manager->watch_pids, PID_TO_PTR(pid), u); |
2598 | 0 | if (r == -EEXIST) { |
2599 | 0 | Unit **array; |
2600 | 0 | bool found = false; |
2601 | 0 | size_t n = 0; |
2602 | 0 |
|
2603 | 0 | /* OK, the "pid" key is already assigned to a different unit. Let's see if the "-pid" key (which points |
2604 | 0 | * to an array of Units rather than just a Unit), lists us already. */ |
2605 | 0 |
|
2606 | 0 | array = hashmap_get(u->manager->watch_pids, PID_TO_PTR(-pid)); |
2607 | 0 | if (array) |
2608 | 0 | for (; array[n]; n++) |
2609 | 0 | if (array[n] == u) |
2610 | 0 | found = true; |
2611 | 0 |
|
2612 | 0 | if (found) /* Found it already? if so, do nothing */ |
2613 | 0 | r = 0; |
2614 | 0 | else { |
2615 | 0 | Unit **new_array; |
2616 | 0 |
|
2617 | 0 | /* Allocate a new array */ |
2618 | 0 | new_array = new(Unit*, n + 2); |
2619 | 0 | if (!new_array) |
2620 | 0 | return -ENOMEM; |
2621 | 0 | |
2622 | 0 | memcpy_safe(new_array, array, sizeof(Unit*) * n); |
2623 | 0 | new_array[n] = u; |
2624 | 0 | new_array[n+1] = NULL; |
2625 | 0 |
|
2626 | 0 | /* Add or replace the old array */ |
2627 | 0 | r = hashmap_replace(u->manager->watch_pids, PID_TO_PTR(-pid), new_array); |
2628 | 0 | if (r < 0) { |
2629 | 0 | free(new_array); |
2630 | 0 | return r; |
2631 | 0 | } |
2632 | 0 | |
2633 | 0 | free(array); |
2634 | 0 | } |
2635 | 0 | } else if (r < 0) |
2636 | 0 | return r; |
2637 | 0 | |
2638 | 0 | r = set_put(u->pids, PID_TO_PTR(pid)); |
2639 | 0 | if (r < 0) |
2640 | 0 | return r; |
2641 | 0 | |
2642 | 0 | return 0; |
2643 | 0 | } |
2644 | | |
2645 | 0 | void unit_unwatch_pid(Unit *u, pid_t pid) { |
2646 | 0 | Unit **array; |
2647 | 0 |
|
2648 | 0 | assert(u); |
2649 | 0 | assert(pid_is_valid(pid)); |
2650 | 0 |
|
2651 | 0 | /* First let's drop the unit in case it's keyed as "pid". */ |
2652 | 0 | (void) hashmap_remove_value(u->manager->watch_pids, PID_TO_PTR(pid), u); |
2653 | 0 |
|
2654 | 0 | /* Then, let's also drop the unit, in case it's in the array keyed by -pid */ |
2655 | 0 | array = hashmap_get(u->manager->watch_pids, PID_TO_PTR(-pid)); |
2656 | 0 | if (array) { |
2657 | 0 | size_t n, m = 0; |
2658 | 0 |
|
2659 | 0 | /* Let's iterate through the array, dropping our own entry */ |
2660 | 0 | for (n = 0; array[n]; n++) |
2661 | 0 | if (array[n] != u) |
2662 | 0 | array[m++] = array[n]; |
2663 | 0 | array[m] = NULL; |
2664 | 0 |
|
2665 | 0 | if (m == 0) { |
2666 | 0 | /* The array is now empty, remove the entire entry */ |
2667 | 0 | assert(hashmap_remove(u->manager->watch_pids, PID_TO_PTR(-pid)) == array); |
2668 | 0 | free(array); |
2669 | 0 | } |
2670 | 0 | } |
2671 | 0 |
|
2672 | 0 | (void) set_remove(u->pids, PID_TO_PTR(pid)); |
2673 | 0 | } |
2674 | | |
2675 | 36.8k | void unit_unwatch_all_pids(Unit *u) { |
2676 | 36.8k | assert(u); |
2677 | 36.8k | |
2678 | 36.8k | while (!set_isempty(u->pids)) |
2679 | 0 | unit_unwatch_pid(u, PTR_TO_PID(set_first(u->pids))); |
2680 | 36.8k | |
2681 | 36.8k | u->pids = set_free(u->pids); |
2682 | 36.8k | } |
2683 | | |
2684 | 0 | static void unit_tidy_watch_pids(Unit *u) { |
2685 | 0 | pid_t except1, except2; |
2686 | 0 | Iterator i; |
2687 | 0 | void *e; |
2688 | 0 |
|
2689 | 0 | assert(u); |
2690 | 0 |
|
2691 | 0 | /* Cleans dead PIDs from our list */ |
2692 | 0 |
|
2693 | 0 | except1 = unit_main_pid(u); |
2694 | 0 | except2 = unit_control_pid(u); |
2695 | 0 |
|
2696 | 0 | SET_FOREACH(e, u->pids, i) { |
2697 | 0 | pid_t pid = PTR_TO_PID(e); |
2698 | 0 |
|
2699 | 0 | if (pid == except1 || pid == except2) |
2700 | 0 | continue; |
2701 | 0 | |
2702 | 0 | if (!pid_is_unwaited(pid)) |
2703 | 0 | unit_unwatch_pid(u, pid); |
2704 | 0 | } |
2705 | 0 | } |
2706 | | |
2707 | 0 | static int on_rewatch_pids_event(sd_event_source *s, void *userdata) { |
2708 | 0 | Unit *u = userdata; |
2709 | 0 |
|
2710 | 0 | assert(s); |
2711 | 0 | assert(u); |
2712 | 0 |
|
2713 | 0 | unit_tidy_watch_pids(u); |
2714 | 0 | unit_watch_all_pids(u); |
2715 | 0 |
|
2716 | 0 | /* If the PID set is empty now, then let's finish this off. */ |
2717 | 0 | unit_synthesize_cgroup_empty_event(u); |
2718 | 0 |
|
2719 | 0 | return 0; |
2720 | 0 | } |
2721 | | |
2722 | 0 | int unit_enqueue_rewatch_pids(Unit *u) { |
2723 | 0 | int r; |
2724 | 0 |
|
2725 | 0 | assert(u); |
2726 | 0 |
|
2727 | 0 | if (!u->cgroup_path) |
2728 | 0 | return -ENOENT; |
2729 | 0 | |
2730 | 0 | r = cg_unified_controller(SYSTEMD_CGROUP_CONTROLLER); |
2731 | 0 | if (r < 0) |
2732 | 0 | return r; |
2733 | 0 | if (r > 0) /* On unified we can use proper notifications */ |
2734 | 0 | return 0; |
2735 | 0 | |
2736 | 0 | /* Enqueues a low-priority job that will clean up dead PIDs from our list of PIDs to watch and subscribe to new |
2737 | 0 | * PIDs that might have appeared. We do this in a delayed job because the work might be quite slow, as it |
2738 | 0 | * involves issuing kill(pid, 0) on all processes we watch. */ |
2739 | 0 | |
2740 | 0 | if (!u->rewatch_pids_event_source) { |
2741 | 0 | _cleanup_(sd_event_source_unrefp) sd_event_source *s = NULL; |
2742 | 0 |
|
2743 | 0 | r = sd_event_add_defer(u->manager->event, &s, on_rewatch_pids_event, u); |
2744 | 0 | if (r < 0) |
2745 | 0 | return log_error_errno(r, "Failed to allocate event source for tidying watched PIDs: %m"); |
2746 | 0 | |
2747 | 0 | r = sd_event_source_set_priority(s, SD_EVENT_PRIORITY_IDLE); |
2748 | 0 | if (r < 0) |
2749 | 0 | return log_error_errno(r, "Failed to adjust priority of event source for tidying watched PIDs: m"); |
2750 | 0 | |
2751 | 0 | (void) sd_event_source_set_description(s, "tidy-watch-pids"); |
2752 | 0 |
|
2753 | 0 | u->rewatch_pids_event_source = TAKE_PTR(s); |
2754 | 0 | } |
2755 | 0 |
|
2756 | 0 | r = sd_event_source_set_enabled(u->rewatch_pids_event_source, SD_EVENT_ONESHOT); |
2757 | 0 | if (r < 0) |
2758 | 0 | return log_error_errno(r, "Failed to enable event source for tidying watched PIDs: %m"); |
2759 | 0 | |
2760 | 0 | return 0; |
2761 | 0 | } |
2762 | | |
2763 | 36.8k | void unit_dequeue_rewatch_pids(Unit *u) { |
2764 | 36.8k | int r; |
2765 | 36.8k | assert(u); |
2766 | 36.8k | |
2767 | 36.8k | if (!u->rewatch_pids_event_source) |
2768 | 36.8k | return; |
2769 | 0 | |
2770 | 0 | r = sd_event_source_set_enabled(u->rewatch_pids_event_source, SD_EVENT_OFF); |
2771 | 0 | if (r < 0) |
2772 | 0 | log_warning_errno(r, "Failed to disable event source for tidying watched PIDs, ignoring: %m"); |
2773 | 0 |
|
2774 | 0 | u->rewatch_pids_event_source = sd_event_source_unref(u->rewatch_pids_event_source); |
2775 | 0 | } |
2776 | | |
2777 | 0 | bool unit_job_is_applicable(Unit *u, JobType j) { |
2778 | 0 | assert(u); |
2779 | 0 | assert(j >= 0 && j < _JOB_TYPE_MAX); |
2780 | 0 |
|
2781 | 0 | switch (j) { |
2782 | 0 |
|
2783 | 0 | case JOB_VERIFY_ACTIVE: |
2784 | 0 | case JOB_START: |
2785 | 0 | case JOB_NOP: |
2786 | 0 | /* Note that we don't check unit_can_start() here. That's because .device units and suchlike are not |
2787 | 0 | * startable by us but may appear due to external events, and it thus makes sense to permit enqueing |
2788 | 0 | * jobs for it. */ |
2789 | 0 | return true; |
2790 | 0 |
|
2791 | 0 | case JOB_STOP: |
2792 | 0 | /* Similar as above. However, perpetual units can never be stopped (neither explicitly nor due to |
2793 | 0 | * external events), hence it makes no sense to permit enqueing such a request either. */ |
2794 | 0 | return !u->perpetual; |
2795 | 0 |
|
2796 | 0 | case JOB_RESTART: |
2797 | 0 | case JOB_TRY_RESTART: |
2798 | 0 | return unit_can_stop(u) && unit_can_start(u); |
2799 | 0 |
|
2800 | 0 | case JOB_RELOAD: |
2801 | 0 | case JOB_TRY_RELOAD: |
2802 | 0 | return unit_can_reload(u); |
2803 | 0 |
|
2804 | 0 | case JOB_RELOAD_OR_START: |
2805 | 0 | return unit_can_reload(u) && unit_can_start(u); |
2806 | 0 |
|
2807 | 0 | default: |
2808 | 0 | assert_not_reached("Invalid job type"); |
2809 | 0 | } |
2810 | 0 | } |
2811 | | |
2812 | 948 | static void maybe_warn_about_dependency(Unit *u, const char *other, UnitDependency dependency) { |
2813 | 948 | assert(u); |
2814 | 948 | |
2815 | 948 | /* Only warn about some unit types */ |
2816 | 948 | if (!IN_SET(dependency, UNIT_CONFLICTS, UNIT_CONFLICTED_BY, UNIT_BEFORE, UNIT_AFTER, UNIT_ON_FAILURE, UNIT_TRIGGERS, UNIT_TRIGGERED_BY)) |
2817 | 948 | return; |
2818 | 744 | |
2819 | 744 | if (streq_ptr(u->id, other)) |
2820 | 744 | log_unit_warning(u, "Dependency %s=%s dropped", unit_dependency_to_string(dependency), u->id); |
2821 | 744 | else |
2822 | 744 | log_unit_warning(u, "Dependency %s=%s dropped, merged into %s", unit_dependency_to_string(dependency), strna(other), u->id); |
2823 | 744 | } |
2824 | | |
2825 | | static int unit_add_dependency_hashmap( |
2826 | | Hashmap **h, |
2827 | | Unit *other, |
2828 | | UnitDependencyMask origin_mask, |
2829 | 132k | UnitDependencyMask destination_mask) { |
2830 | 132k | |
2831 | 132k | UnitDependencyInfo info; |
2832 | 132k | int r; |
2833 | 132k | |
2834 | 132k | assert(h); |
2835 | 132k | assert(other); |
2836 | 132k | assert(origin_mask < _UNIT_DEPENDENCY_MASK_FULL); |
2837 | 132k | assert(destination_mask < _UNIT_DEPENDENCY_MASK_FULL); |
2838 | 132k | assert(origin_mask > 0 || destination_mask > 0); |
2839 | 132k | |
2840 | 132k | r = hashmap_ensure_allocated(h, NULL); |
2841 | 132k | if (r < 0) |
2842 | 0 | return r; |
2843 | 132k | |
2844 | 132k | assert_cc(sizeof(void*) == sizeof(info)); |
2845 | 132k | |
2846 | 132k | info.data = hashmap_get(*h, other); |
2847 | 132k | if (info.data) { |
2848 | 33.4k | /* Entry already exists. Add in our mask. */ |
2849 | 33.4k | |
2850 | 33.4k | if (FLAGS_SET(origin_mask, info.origin_mask) && |
2851 | 33.4k | FLAGS_SET(destination_mask, info.destination_mask)) |
2852 | 33.4k | return 0; /* NOP */ |
2853 | 1.00k | |
2854 | 1.00k | info.origin_mask |= origin_mask; |
2855 | 1.00k | info.destination_mask |= destination_mask; |
2856 | 1.00k | |
2857 | 1.00k | r = hashmap_update(*h, other, info.data); |
2858 | 99.2k | } else { |
2859 | 99.2k | info = (UnitDependencyInfo) { |
2860 | 99.2k | .origin_mask = origin_mask, |
2861 | 99.2k | .destination_mask = destination_mask, |
2862 | 99.2k | }; |
2863 | 99.2k | |
2864 | 99.2k | r = hashmap_put(*h, other, info.data); |
2865 | 99.2k | } |
2866 | 132k | if (r < 0) |
2867 | 0 | return r; |
2868 | 100k | |
2869 | 100k | return 1; |
2870 | 100k | } |
2871 | | |
2872 | | int unit_add_dependency( |
2873 | | Unit *u, |
2874 | | UnitDependency d, |
2875 | | Unit *other, |
2876 | | bool add_reference, |
2877 | 37.0k | UnitDependencyMask mask) { |
2878 | 37.0k | |
2879 | 37.0k | static const UnitDependency inverse_table[_UNIT_DEPENDENCY_MAX] = { |
2880 | 37.0k | [UNIT_REQUIRES] = UNIT_REQUIRED_BY, |
2881 | 37.0k | [UNIT_WANTS] = UNIT_WANTED_BY, |
2882 | 37.0k | [UNIT_REQUISITE] = UNIT_REQUISITE_OF, |
2883 | 37.0k | [UNIT_BINDS_TO] = UNIT_BOUND_BY, |
2884 | 37.0k | [UNIT_PART_OF] = UNIT_CONSISTS_OF, |
2885 | 37.0k | [UNIT_REQUIRED_BY] = UNIT_REQUIRES, |
2886 | 37.0k | [UNIT_REQUISITE_OF] = UNIT_REQUISITE, |
2887 | 37.0k | [UNIT_WANTED_BY] = UNIT_WANTS, |
2888 | 37.0k | [UNIT_BOUND_BY] = UNIT_BINDS_TO, |
2889 | 37.0k | [UNIT_CONSISTS_OF] = UNIT_PART_OF, |
2890 | 37.0k | [UNIT_CONFLICTS] = UNIT_CONFLICTED_BY, |
2891 | 37.0k | [UNIT_CONFLICTED_BY] = UNIT_CONFLICTS, |
2892 | 37.0k | [UNIT_BEFORE] = UNIT_AFTER, |
2893 | 37.0k | [UNIT_AFTER] = UNIT_BEFORE, |
2894 | 37.0k | [UNIT_ON_FAILURE] = _UNIT_DEPENDENCY_INVALID, |
2895 | 37.0k | [UNIT_REFERENCES] = UNIT_REFERENCED_BY, |
2896 | 37.0k | [UNIT_REFERENCED_BY] = UNIT_REFERENCES, |
2897 | 37.0k | [UNIT_TRIGGERS] = UNIT_TRIGGERED_BY, |
2898 | 37.0k | [UNIT_TRIGGERED_BY] = UNIT_TRIGGERS, |
2899 | 37.0k | [UNIT_PROPAGATES_RELOAD_TO] = UNIT_RELOAD_PROPAGATED_FROM, |
2900 | 37.0k | [UNIT_RELOAD_PROPAGATED_FROM] = UNIT_PROPAGATES_RELOAD_TO, |
2901 | 37.0k | [UNIT_JOINS_NAMESPACE_OF] = UNIT_JOINS_NAMESPACE_OF, |
2902 | 37.0k | }; |
2903 | 37.0k | Unit *original_u = u, *original_other = other; |
2904 | 37.0k | int r; |
2905 | 37.0k | |
2906 | 37.0k | assert(u); |
2907 | 37.0k | assert(d >= 0 && d < _UNIT_DEPENDENCY_MAX); |
2908 | 37.0k | assert(other); |
2909 | 37.0k | |
2910 | 37.0k | u = unit_follow_merge(u); |
2911 | 37.0k | other = unit_follow_merge(other); |
2912 | 37.0k | |
2913 | 37.0k | /* We won't allow dependencies on ourselves. We will not |
2914 | 37.0k | * consider them an error however. */ |
2915 | 37.0k | if (u == other) { |
2916 | 948 | maybe_warn_about_dependency(original_u, original_other->id, d); |
2917 | 948 | return 0; |
2918 | 948 | } |
2919 | 36.1k | |
2920 | 36.1k | if ((d == UNIT_BEFORE && other->type == UNIT_DEVICE) || |
2921 | 36.1k | (d == UNIT_AFTER && u->type == UNIT_DEVICE)) { |
2922 | 1.84k | log_unit_warning(u, "Dependency Before=%s ignored (.device units cannot be delayed)", other->id); |
2923 | 1.84k | return 0; |
2924 | 1.84k | } |
2925 | 34.2k | |
2926 | 34.2k | r = unit_add_dependency_hashmap(u->dependencies + d, other, mask, 0); |
2927 | 34.2k | if (r < 0) |
2928 | 0 | return r; |
2929 | 34.2k | |
2930 | 34.2k | if (inverse_table[d] != _UNIT_DEPENDENCY_INVALID && inverse_table[d] != d) { |
2931 | 29.9k | r = unit_add_dependency_hashmap(other->dependencies + inverse_table[d], u, 0, mask); |
2932 | 29.9k | if (r < 0) |
2933 | 0 | return r; |
2934 | 34.2k | } |
2935 | 34.2k | |
2936 | 34.2k | if (add_reference) { |
2937 | 34.2k | r = unit_add_dependency_hashmap(u->dependencies + UNIT_REFERENCES, other, mask, 0); |
2938 | 34.2k | if (r < 0) |
2939 | 0 | return r; |
2940 | 34.2k | |
2941 | 34.2k | r = unit_add_dependency_hashmap(other->dependencies + UNIT_REFERENCED_BY, u, 0, mask); |
2942 | 34.2k | if (r < 0) |
2943 | 0 | return r; |
2944 | 34.2k | } |
2945 | 34.2k | |
2946 | 34.2k | unit_add_to_dbus_queue(u); |
2947 | 34.2k | return 0; |
2948 | 34.2k | } |
2949 | | |
2950 | 9.59k | int unit_add_two_dependencies(Unit *u, UnitDependency d, UnitDependency e, Unit *other, bool add_reference, UnitDependencyMask mask) { |
2951 | 9.59k | int r; |
2952 | 9.59k | |
2953 | 9.59k | assert(u); |
2954 | 9.59k | |
2955 | 9.59k | r = unit_add_dependency(u, d, other, add_reference, mask); |
2956 | 9.59k | if (r < 0) |
2957 | 0 | return r; |
2958 | 9.59k | |
2959 | 9.59k | return unit_add_dependency(u, e, other, add_reference, mask); |
2960 | 9.59k | } |
2961 | | |
2962 | 32.2k | static int resolve_template(Unit *u, const char *name, char **buf, const char **ret) { |
2963 | 32.2k | int r; |
2964 | 32.2k | |
2965 | 32.2k | assert(u); |
2966 | 32.2k | assert(name); |
2967 | 32.2k | assert(buf); |
2968 | 32.2k | assert(ret); |
2969 | 32.2k | |
2970 | 32.2k | if (!unit_name_is_valid(name, UNIT_NAME_TEMPLATE)) { |
2971 | 31.3k | *buf = NULL; |
2972 | 31.3k | *ret = name; |
2973 | 31.3k | return 0; |
2974 | 31.3k | } |
2975 | 923 | |
2976 | 923 | if (u->instance) |
2977 | 0 | r = unit_name_replace_instance(name, u->instance, buf); |
2978 | 923 | else { |
2979 | 923 | _cleanup_free_ char *i = NULL; |
2980 | 923 | |
2981 | 923 | r = unit_name_to_prefix(u->id, &i); |
2982 | 923 | if (r < 0) |
2983 | 0 | return r; |
2984 | 923 | |
2985 | 923 | r = unit_name_replace_instance(name, i, buf); |
2986 | 923 | } |
2987 | 923 | if (r < 0) |
2988 | 0 | return r; |
2989 | 923 | |
2990 | 923 | *ret = *buf; |
2991 | 923 | return 0; |
2992 | 923 | } |
2993 | | |
2994 | 26.6k | int unit_add_dependency_by_name(Unit *u, UnitDependency d, const char *name, bool add_reference, UnitDependencyMask mask) { |
2995 | 26.6k | _cleanup_free_ char *buf = NULL; |
2996 | 26.6k | Unit *other; |
2997 | 26.6k | int r; |
2998 | 26.6k | |
2999 | 26.6k | assert(u); |
3000 | 26.6k | assert(name); |
3001 | 26.6k | |
3002 | 26.6k | r = resolve_template(u, name, &buf, &name); |
3003 | 26.6k | if (r < 0) |
3004 | 0 | return r; |
3005 | 26.6k | |
3006 | 26.6k | r = manager_load_unit(u->manager, name, NULL, NULL, &other); |
3007 | 26.6k | if (r < 0) |
3008 | 8.74k | return r; |
3009 | 17.8k | |
3010 | 17.8k | return unit_add_dependency(u, d, other, add_reference, mask); |
3011 | 17.8k | } |
3012 | | |
3013 | 5.68k | int unit_add_two_dependencies_by_name(Unit *u, UnitDependency d, UnitDependency e, const char *name, bool add_reference, UnitDependencyMask mask) { |
3014 | 5.68k | _cleanup_free_ char *buf = NULL; |
3015 | 5.68k | Unit *other; |
3016 | 5.68k | int r; |
3017 | 5.68k | |
3018 | 5.68k | assert(u); |
3019 | 5.68k | assert(name); |
3020 | 5.68k | |
3021 | 5.68k | r = resolve_template(u, name, &buf, &name); |
3022 | 5.68k | if (r < 0) |
3023 | 0 | return r; |
3024 | 5.68k | |
3025 | 5.68k | r = manager_load_unit(u->manager, name, NULL, NULL, &other); |
3026 | 5.68k | if (r < 0) |
3027 | 631 | return r; |
3028 | 5.05k | |
3029 | 5.05k | return unit_add_two_dependencies(u, d, e, other, add_reference, mask); |
3030 | 5.05k | } |
3031 | | |
3032 | 0 | int set_unit_path(const char *p) { |
3033 | 0 | /* This is mostly for debug purposes */ |
3034 | 0 | if (setenv("SYSTEMD_UNIT_PATH", p, 1) < 0) |
3035 | 0 | return -errno; |
3036 | 0 | |
3037 | 0 | return 0; |
3038 | 0 | } |
3039 | | |
3040 | 0 | char *unit_dbus_path(Unit *u) { |
3041 | 0 | assert(u); |
3042 | 0 |
|
3043 | 0 | if (!u->id) |
3044 | 0 | return NULL; |
3045 | 0 | |
3046 | 0 | return unit_dbus_path_from_name(u->id); |
3047 | 0 | } |
3048 | | |
3049 | 0 | char *unit_dbus_path_invocation_id(Unit *u) { |
3050 | 0 | assert(u); |
3051 | 0 |
|
3052 | 0 | if (sd_id128_is_null(u->invocation_id)) |
3053 | 0 | return NULL; |
3054 | 0 | |
3055 | 0 | return unit_dbus_path_from_name(u->invocation_id_string); |
3056 | 0 | } |
3057 | | |
3058 | 2.13k | int unit_set_slice(Unit *u, Unit *slice) { |
3059 | 2.13k | assert(u); |
3060 | 2.13k | assert(slice); |
3061 | 2.13k | |
3062 | 2.13k | /* Sets the unit slice if it has not been set before. Is extra |
3063 | 2.13k | * careful, to only allow this for units that actually have a |
3064 | 2.13k | * cgroup context. Also, we don't allow to set this for slices |
3065 | 2.13k | * (since the parent slice is derived from the name). Make |
3066 | 2.13k | * sure the unit we set is actually a slice. */ |
3067 | 2.13k | |
3068 | 2.13k | if (!UNIT_HAS_CGROUP_CONTEXT(u)) |
3069 | 2.13k | return -EOPNOTSUPP; |
3070 | 2.13k | |
3071 | 2.13k | if (u->type == UNIT_SLICE) |
3072 | 0 | return -EINVAL; |
3073 | 2.13k | |
3074 | 2.13k | if (unit_active_state(u) != UNIT_INACTIVE) |
3075 | 0 | return -EBUSY; |
3076 | 2.13k | |
3077 | 2.13k | if (slice->type != UNIT_SLICE) |
3078 | 962 | return -EINVAL; |
3079 | 1.17k | |
3080 | 1.17k | if (unit_has_name(u, SPECIAL_INIT_SCOPE) && |
3081 | 1.17k | !unit_has_name(slice, SPECIAL_ROOT_SLICE)) |
3082 | 0 | return -EPERM; |
3083 | 1.17k | |
3084 | 1.17k | if (UNIT_DEREF(u->slice) == slice) |
3085 | 406 | return 0; |
3086 | 767 | |
3087 | 767 | /* Disallow slice changes if @u is already bound to cgroups */ |
3088 | 767 | if (UNIT_ISSET(u->slice) && u->cgroup_realized) |
3089 | 0 | return -EBUSY; |
3090 | 767 | |
3091 | 767 | unit_ref_set(&u->slice, u, slice); |
3092 | 767 | return 1; |
3093 | 767 | } |
3094 | | |
3095 | 8 | int unit_set_default_slice(Unit *u) { |
3096 | 8 | const char *slice_name; |
3097 | 8 | Unit *slice; |
3098 | 8 | int r; |
3099 | 8 | |
3100 | 8 | assert(u); |
3101 | 8 | |
3102 | 8 | if (UNIT_ISSET(u->slice)) |
3103 | 8 | return 0; |
3104 | 8 | |
3105 | 8 | if (u->instance) { |
3106 | 0 | _cleanup_free_ char *prefix = NULL, *escaped = NULL; |
3107 | 0 |
|
3108 | 0 | /* Implicitly place all instantiated units in their |
3109 | 0 | * own per-template slice */ |
3110 | 0 |
|
3111 | 0 | r = unit_name_to_prefix(u->id, &prefix); |
3112 | 0 | if (r < 0) |
3113 | 0 | return r; |
3114 | 0 | |
3115 | 0 | /* The prefix is already escaped, but it might include |
3116 | 0 | * "-" which has a special meaning for slice units, |
3117 | 0 | * hence escape it here extra. */ |
3118 | 0 | escaped = unit_name_escape(prefix); |
3119 | 0 | if (!escaped) |
3120 | 0 | return -ENOMEM; |
3121 | 0 | |
3122 | 0 | if (MANAGER_IS_SYSTEM(u->manager)) |
3123 | 0 | slice_name = strjoina("system-", escaped, ".slice"); |
3124 | 0 | else |
3125 | 0 | slice_name = strjoina(escaped, ".slice"); |
3126 | 0 | } else |
3127 | 8 | slice_name = |
3128 | 8 | MANAGER_IS_SYSTEM(u->manager) && !unit_has_name(u, SPECIAL_INIT_SCOPE) |
3129 | 8 | ? SPECIAL_SYSTEM_SLICE |
3130 | 8 | : SPECIAL_ROOT_SLICE; |
3131 | 8 | |
3132 | 8 | r = manager_load_unit(u->manager, slice_name, NULL, NULL, &slice); |
3133 | 8 | if (r < 0) |
3134 | 0 | return r; |
3135 | 8 | |
3136 | 8 | return unit_set_slice(u, slice); |
3137 | 8 | } |
3138 | | |
3139 | 53.3k | const char *unit_slice_name(Unit *u) { |
3140 | 53.3k | assert(u); |
3141 | 53.3k | |
3142 | 53.3k | if (!UNIT_ISSET(u->slice)) |
3143 | 53.3k | return NULL; |
3144 | 5.44k | |
3145 | 5.44k | return UNIT_DEREF(u->slice)->id; |
3146 | 5.44k | } |
3147 | | |
3148 | 0 | int unit_load_related_unit(Unit *u, const char *type, Unit **_found) { |
3149 | 0 | _cleanup_free_ char *t = NULL; |
3150 | 0 | int r; |
3151 | 0 |
|
3152 | 0 | assert(u); |
3153 | 0 | assert(type); |
3154 | 0 | assert(_found); |
3155 | 0 |
|
3156 | 0 | r = unit_name_change_suffix(u->id, type, &t); |
3157 | 0 | if (r < 0) |
3158 | 0 | return r; |
3159 | 0 | if (unit_has_name(u, t)) |
3160 | 0 | return -EINVAL; |
3161 | 0 | |
3162 | 0 | r = manager_load_unit(u->manager, t, NULL, NULL, _found); |
3163 | 0 | assert(r < 0 || *_found != u); |
3164 | 0 | return r; |
3165 | 0 | } |
3166 | | |
3167 | 0 | static int signal_name_owner_changed(sd_bus_message *message, void *userdata, sd_bus_error *error) { |
3168 | 0 | const char *name, *old_owner, *new_owner; |
3169 | 0 | Unit *u = userdata; |
3170 | 0 | int r; |
3171 | 0 |
|
3172 | 0 | assert(message); |
3173 | 0 | assert(u); |
3174 | 0 |
|
3175 | 0 | r = sd_bus_message_read(message, "sss", &name, &old_owner, &new_owner); |
3176 | 0 | if (r < 0) { |
3177 | 0 | bus_log_parse_error(r); |
3178 | 0 | return 0; |
3179 | 0 | } |
3180 | 0 | |
3181 | 0 | old_owner = empty_to_null(old_owner); |
3182 | 0 | new_owner = empty_to_null(new_owner); |
3183 | 0 |
|
3184 | 0 | if (UNIT_VTABLE(u)->bus_name_owner_change) |
3185 | 0 | UNIT_VTABLE(u)->bus_name_owner_change(u, name, old_owner, new_owner); |
3186 | 0 |
|
3187 | 0 | return 0; |
3188 | 0 | } |
3189 | | |
3190 | 0 | int unit_install_bus_match(Unit *u, sd_bus *bus, const char *name) { |
3191 | 0 | const char *match; |
3192 | 0 |
|
3193 | 0 | assert(u); |
3194 | 0 | assert(bus); |
3195 | 0 | assert(name); |
3196 | 0 |
|
3197 | 0 | if (u->match_bus_slot) |
3198 | 0 | return -EBUSY; |
3199 | 0 | |
3200 | 0 | match = strjoina("type='signal'," |
3201 | 0 | "sender='org.freedesktop.DBus'," |
3202 | 0 | "path='/org/freedesktop/DBus'," |
3203 | 0 | "interface='org.freedesktop.DBus'," |
3204 | 0 | "member='NameOwnerChanged'," |
3205 | 0 | "arg0='", name, "'"); |
3206 | 0 |
|
3207 | 0 | return sd_bus_add_match_async(bus, &u->match_bus_slot, match, signal_name_owner_changed, NULL, u); |
3208 | 0 | } |
3209 | | |
3210 | 0 | int unit_watch_bus_name(Unit *u, const char *name) { |
3211 | 0 | int r; |
3212 | 0 |
|
3213 | 0 | assert(u); |
3214 | 0 | assert(name); |
3215 | 0 |
|
3216 | 0 | /* Watch a specific name on the bus. We only support one unit |
3217 | 0 | * watching each name for now. */ |
3218 | 0 |
|
3219 | 0 | if (u->manager->api_bus) { |
3220 | 0 | /* If the bus is already available, install the match directly. |
3221 | 0 | * Otherwise, just put the name in the list. bus_setup_api() will take care later. */ |
3222 | 0 | r = unit_install_bus_match(u, u->manager->api_bus, name); |
3223 | 0 | if (r < 0) |
3224 | 0 | return log_warning_errno(r, "Failed to subscribe to NameOwnerChanged signal for '%s': %m", name); |
3225 | 0 | } |
3226 | 0 | |
3227 | 0 | r = hashmap_put(u->manager->watch_bus, name, u); |
3228 | 0 | if (r < 0) { |
3229 | 0 | u->match_bus_slot = sd_bus_slot_unref(u->match_bus_slot); |
3230 | 0 | return log_warning_errno(r, "Failed to put bus name to hashmap: %m"); |
3231 | 0 | } |
3232 | 0 |
|
3233 | 0 | return 0; |
3234 | 0 | } |
3235 | | |
3236 | 43 | void unit_unwatch_bus_name(Unit *u, const char *name) { |
3237 | 43 | assert(u); |
3238 | 43 | assert(name); |
3239 | 43 | |
3240 | 43 | (void) hashmap_remove_value(u->manager->watch_bus, name, u); |
3241 | 43 | u->match_bus_slot = sd_bus_slot_unref(u->match_bus_slot); |
3242 | 43 | } |
3243 | | |
3244 | 0 | bool unit_can_serialize(Unit *u) { |
3245 | 0 | assert(u); |
3246 | 0 |
|
3247 | 0 | return UNIT_VTABLE(u)->serialize && UNIT_VTABLE(u)->deserialize_item; |
3248 | 0 | } |
3249 | | |
3250 | 0 | static int serialize_cgroup_mask(FILE *f, const char *key, CGroupMask mask) { |
3251 | 0 | _cleanup_free_ char *s = NULL; |
3252 | 0 | int r; |
3253 | 0 |
|
3254 | 0 | assert(f); |
3255 | 0 | assert(key); |
3256 | 0 |
|
3257 | 0 | if (mask == 0) |
3258 | 0 | return 0; |
3259 | 0 | |
3260 | 0 | r = cg_mask_to_string(mask, &s); |
3261 | 0 | if (r < 0) |
3262 | 0 | return log_error_errno(r, "Failed to format cgroup mask: %m"); |
3263 | 0 | |
3264 | 0 | return serialize_item(f, key, s); |
3265 | 0 | } |
3266 | | |
3267 | | static const char *const ip_accounting_metric_field[_CGROUP_IP_ACCOUNTING_METRIC_MAX] = { |
3268 | | [CGROUP_IP_INGRESS_BYTES] = "ip-accounting-ingress-bytes", |
3269 | | [CGROUP_IP_INGRESS_PACKETS] = "ip-accounting-ingress-packets", |
3270 | | [CGROUP_IP_EGRESS_BYTES] = "ip-accounting-egress-bytes", |
3271 | | [CGROUP_IP_EGRESS_PACKETS] = "ip-accounting-egress-packets", |
3272 | | }; |
3273 | | |
3274 | | static const char *const io_accounting_metric_field_base[_CGROUP_IO_ACCOUNTING_METRIC_MAX] = { |
3275 | | [CGROUP_IO_READ_BYTES] = "io-accounting-read-bytes-base", |
3276 | | [CGROUP_IO_WRITE_BYTES] = "io-accounting-write-bytes-base", |
3277 | | [CGROUP_IO_READ_OPERATIONS] = "io-accounting-read-operations-base", |
3278 | | [CGROUP_IO_WRITE_OPERATIONS] = "io-accounting-write-operations-base", |
3279 | | }; |
3280 | | |
3281 | | static const char *const io_accounting_metric_field_last[_CGROUP_IO_ACCOUNTING_METRIC_MAX] = { |
3282 | | [CGROUP_IO_READ_BYTES] = "io-accounting-read-bytes-last", |
3283 | | [CGROUP_IO_WRITE_BYTES] = "io-accounting-write-bytes-last", |
3284 | | [CGROUP_IO_READ_OPERATIONS] = "io-accounting-read-operations-last", |
3285 | | [CGROUP_IO_WRITE_OPERATIONS] = "io-accounting-write-operations-last", |
3286 | | }; |
3287 | | |
3288 | 0 | int unit_serialize(Unit *u, FILE *f, FDSet *fds, bool serialize_jobs) { |
3289 | 0 | CGroupIPAccountingMetric m; |
3290 | 0 | int r; |
3291 | 0 |
|
3292 | 0 | assert(u); |
3293 | 0 | assert(f); |
3294 | 0 | assert(fds); |
3295 | 0 |
|
3296 | 0 | if (unit_can_serialize(u)) { |
3297 | 0 | r = UNIT_VTABLE(u)->serialize(u, f, fds); |
3298 | 0 | if (r < 0) |
3299 | 0 | return r; |
3300 | 0 | } |
3301 | 0 | |
3302 | 0 | (void) serialize_dual_timestamp(f, "state-change-timestamp", &u->state_change_timestamp); |
3303 | 0 |
|
3304 | 0 | (void) serialize_dual_timestamp(f, "inactive-exit-timestamp", &u->inactive_exit_timestamp); |
3305 | 0 | (void) serialize_dual_timestamp(f, "active-enter-timestamp", &u->active_enter_timestamp); |
3306 | 0 | (void) serialize_dual_timestamp(f, "active-exit-timestamp", &u->active_exit_timestamp); |
3307 | 0 | (void) serialize_dual_timestamp(f, "inactive-enter-timestamp", &u->inactive_enter_timestamp); |
3308 | 0 |
|
3309 | 0 | (void) serialize_dual_timestamp(f, "condition-timestamp", &u->condition_timestamp); |
3310 | 0 | (void) serialize_dual_timestamp(f, "assert-timestamp", &u->assert_timestamp); |
3311 | 0 |
|
3312 | 0 | if (dual_timestamp_is_set(&u->condition_timestamp)) |
3313 | 0 | (void) serialize_bool(f, "condition-result", u->condition_result); |
3314 | 0 |
|
3315 | 0 | if (dual_timestamp_is_set(&u->assert_timestamp)) |
3316 | 0 | (void) serialize_bool(f, "assert-result", u->assert_result); |
3317 | 0 |
|
3318 | 0 | (void) serialize_bool(f, "transient", u->transient); |
3319 | 0 | (void) serialize_bool(f, "in-audit", u->in_audit); |
3320 | 0 |
|
3321 | 0 | (void) serialize_bool(f, "exported-invocation-id", u->exported_invocation_id); |
3322 | 0 | (void) serialize_bool(f, "exported-log-level-max", u->exported_log_level_max); |
3323 | 0 | (void) serialize_bool(f, "exported-log-extra-fields", u->exported_log_extra_fields); |
3324 | 0 | (void) serialize_bool(f, "exported-log-rate-limit-interval", u->exported_log_rate_limit_interval); |
3325 | 0 | (void) serialize_bool(f, "exported-log-rate-limit-burst", u->exported_log_rate_limit_burst); |
3326 | 0 |
|
3327 | 0 | (void) serialize_item_format(f, "cpu-usage-base", "%" PRIu64, u->cpu_usage_base); |
3328 | 0 | if (u->cpu_usage_last != NSEC_INFINITY) |
3329 | 0 | (void) serialize_item_format(f, "cpu-usage-last", "%" PRIu64, u->cpu_usage_last); |
3330 | 0 |
|
3331 | 0 | if (u->oom_kill_last > 0) |
3332 | 0 | (void) serialize_item_format(f, "oom-kill-last", "%" PRIu64, u->oom_kill_last); |
3333 | 0 |
|
3334 | 0 | for (CGroupIOAccountingMetric im = 0; im < _CGROUP_IO_ACCOUNTING_METRIC_MAX; im++) { |
3335 | 0 | (void) serialize_item_format(f, io_accounting_metric_field_base[im], "%" PRIu64, u->io_accounting_base[im]); |
3336 | 0 |
|
3337 | 0 | if (u->io_accounting_last[im] != UINT64_MAX) |
3338 | 0 | (void) serialize_item_format(f, io_accounting_metric_field_last[im], "%" PRIu64, u->io_accounting_last[im]); |
3339 | 0 | } |
3340 | 0 |
|
3341 | 0 | if (u->cgroup_path) |
3342 | 0 | (void) serialize_item(f, "cgroup", u->cgroup_path); |
3343 | 0 |
|
3344 | 0 | (void) serialize_bool(f, "cgroup-realized", u->cgroup_realized); |
3345 | 0 | (void) serialize_cgroup_mask(f, "cgroup-realized-mask", u->cgroup_realized_mask); |
3346 | 0 | (void) serialize_cgroup_mask(f, "cgroup-enabled-mask", u->cgroup_enabled_mask); |
3347 | 0 | (void) serialize_cgroup_mask(f, "cgroup-invalidated-mask", u->cgroup_invalidated_mask); |
3348 | 0 |
|
3349 | 0 | if (uid_is_valid(u->ref_uid)) |
3350 | 0 | (void) serialize_item_format(f, "ref-uid", UID_FMT, u->ref_uid); |
3351 | 0 | if (gid_is_valid(u->ref_gid)) |
3352 | 0 | (void) serialize_item_format(f, "ref-gid", GID_FMT, u->ref_gid); |
3353 | 0 |
|
3354 | 0 | if (!sd_id128_is_null(u->invocation_id)) |
3355 | 0 | (void) serialize_item_format(f, "invocation-id", SD_ID128_FORMAT_STR, SD_ID128_FORMAT_VAL(u->invocation_id)); |
3356 | 0 |
|
3357 | 0 | bus_track_serialize(u->bus_track, f, "ref"); |
3358 | 0 |
|
3359 | 0 | for (m = 0; m < _CGROUP_IP_ACCOUNTING_METRIC_MAX; m++) { |
3360 | 0 | uint64_t v; |
3361 | 0 |
|
3362 | 0 | r = unit_get_ip_accounting(u, m, &v); |
3363 | 0 | if (r >= 0) |
3364 | 0 | (void) serialize_item_format(f, ip_accounting_metric_field[m], "%" PRIu64, v); |
3365 | 0 | } |
3366 | 0 |
|
3367 | 0 | if (serialize_jobs) { |
3368 | 0 | if (u->job) { |
3369 | 0 | fputs("job\n", f); |
3370 | 0 | job_serialize(u->job, f); |
3371 | 0 | } |
3372 | 0 |
|
3373 | 0 | if (u->nop_job) { |
3374 | 0 | fputs("job\n", f); |
3375 | 0 | job_serialize(u->nop_job, f); |
3376 | 0 | } |
3377 | 0 | } |
3378 | 0 |
|
3379 | 0 | /* End marker */ |
3380 | 0 | fputc('\n', f); |
3381 | 0 | return 0; |
3382 | 0 | } |
3383 | | |
3384 | 0 | static int unit_deserialize_job(Unit *u, FILE *f) { |
3385 | 0 | _cleanup_(job_freep) Job *j = NULL; |
3386 | 0 | int r; |
3387 | 0 |
|
3388 | 0 | assert(u); |
3389 | 0 | assert(f); |
3390 | 0 |
|
3391 | 0 | j = job_new_raw(u); |
3392 | 0 | if (!j) |
3393 | 0 | return log_oom(); |
3394 | 0 | |
3395 | 0 | r = job_deserialize(j, f); |
3396 | 0 | if (r < 0) |
3397 | 0 | return r; |
3398 | 0 | |
3399 | 0 | r = job_install_deserialized(j); |
3400 | 0 | if (r < 0) |
3401 | 0 | return r; |
3402 | 0 | |
3403 | 0 | TAKE_PTR(j); |
3404 | 0 | return 0; |
3405 | 0 | } |
3406 | | |
3407 | | int unit_deserialize(Unit *u, FILE *f, FDSet *fds) { |
3408 | | int r; |
3409 | | |
3410 | | assert(u); |
3411 | | assert(f); |
3412 | | assert(fds); |
3413 | | |
3414 | | for (;;) { |
3415 | | _cleanup_free_ char *line = NULL; |
3416 | | char *l, *v; |
3417 | | ssize_t m; |
3418 | | size_t k; |
3419 | | |
3420 | | r = read_line(f, LONG_LINE_MAX, &line); |
3421 | | if (r < 0) |
3422 | | return log_error_errno(r, "Failed to read serialization line: %m"); |
3423 | | if (r == 0) /* eof */ |
3424 | | break; |
3425 | | |
3426 | | l = strstrip(line); |
3427 | | if (isempty(l)) /* End marker */ |
3428 | | break; |
3429 | | |
3430 | | k = strcspn(l, "="); |
3431 | | |
3432 | | if (l[k] == '=') { |
3433 | | l[k] = 0; |
3434 | | v = l+k+1; |
3435 | | } else |
3436 | | v = l+k; |
3437 | | |
3438 | | if (streq(l, "job")) { |
3439 | | if (v[0] == '\0') { |
3440 | | /* New-style serialized job */ |
3441 | | r = unit_deserialize_job(u, f); |
3442 | | if (r < 0) |
3443 | | return r; |
3444 | | } else /* Legacy for pre-44 */ |
3445 | | log_unit_warning(u, "Update from too old systemd versions are unsupported, cannot deserialize job: %s", v); |
3446 | | continue; |
3447 | | } else if (streq(l, "state-change-timestamp")) { |
3448 | | (void) deserialize_dual_timestamp(v, &u->state_change_timestamp); |
3449 | | continue; |
3450 | | } else if (streq(l, "inactive-exit-timestamp")) { |
3451 | | (void) deserialize_dual_timestamp(v, &u->inactive_exit_timestamp); |
3452 | | continue; |
3453 | | } else if (streq(l, "active-enter-timestamp")) { |
3454 | | (void) deserialize_dual_timestamp(v, &u->active_enter_timestamp); |
3455 | | continue; |
3456 | | } else if (streq(l, "active-exit-timestamp")) { |
3457 | | (void) deserialize_dual_timestamp(v, &u->active_exit_timestamp); |
3458 | | continue; |
3459 | | } else if (streq(l, "inactive-enter-timestamp")) { |
3460 | | (void) deserialize_dual_timestamp(v, &u->inactive_enter_timestamp); |
3461 | | continue; |
3462 | | } else if (streq(l, "condition-timestamp")) { |
3463 | | (void) deserialize_dual_timestamp(v, &u->condition_timestamp); |
3464 | | continue; |
3465 | | } else if (streq(l, "assert-timestamp")) { |
3466 | | (void) deserialize_dual_timestamp(v, &u->assert_timestamp); |
3467 | | continue; |
3468 | | } else if (streq(l, "condition-result")) { |
3469 | | |
3470 | | r = parse_boolean(v); |
3471 | | if (r < 0) |
3472 | | log_unit_debug(u, "Failed to parse condition result value %s, ignoring.", v); |
3473 | | else |
3474 | | u->condition_result = r; |
3475 | | |
3476 | | continue; |
3477 | | |
3478 | | } else if (streq(l, "assert-result")) { |
3479 | | |
3480 | | r = parse_boolean(v); |
3481 | | if (r < 0) |
3482 | | log_unit_debug(u, "Failed to parse assert result value %s, ignoring.", v); |
3483 | | else |
3484 | | u->assert_result = r; |
3485 | | |
3486 | | continue; |
3487 | | |
3488 | | } else if (streq(l, "transient")) { |
3489 | | |
3490 | | r = parse_boolean(v); |
3491 | | if (r < 0) |
3492 | | log_unit_debug(u, "Failed to parse transient bool %s, ignoring.", v); |
3493 | | else |
3494 | | u->transient = r; |
3495 | | |
3496 | | continue; |
3497 | | |
3498 | | } else if (streq(l, "in-audit")) { |
3499 | | |
3500 | | r = parse_boolean(v); |
3501 | | if (r < 0) |
3502 | | log_unit_debug(u, "Failed to parse in-audit bool %s, ignoring.", v); |
3503 | | else |
3504 | | u->in_audit = r; |
3505 | | |
3506 | | continue; |
3507 | | |
3508 | | } else if (streq(l, "exported-invocation-id")) { |
3509 | | |
3510 | | r = parse_boolean(v); |
3511 | | if (r < 0) |
3512 | | log_unit_debug(u, "Failed to parse exported invocation ID bool %s, ignoring.", v); |
3513 | | else |
3514 | | u->exported_invocation_id = r; |
3515 | | |
3516 | | continue; |
3517 | | |
3518 | | } else if (streq(l, "exported-log-level-max")) { |
3519 | | |
3520 | | r = parse_boolean(v); |
3521 | | if (r < 0) |
3522 | | log_unit_debug(u, "Failed to parse exported log level max bool %s, ignoring.", v); |
3523 | | else |
3524 | | u->exported_log_level_max = r; |
3525 | | |
3526 | | continue; |
3527 | | |
3528 | | } else if (streq(l, "exported-log-extra-fields")) { |
3529 | | |
3530 | | r = parse_boolean(v); |
3531 | | if (r < 0) |
3532 | | log_unit_debug(u, "Failed to parse exported log extra fields bool %s, ignoring.", v); |
3533 | | else |
3534 | | u->exported_log_extra_fields = r; |
3535 | | |
3536 | | continue; |
3537 | | |
3538 | | } else if (streq(l, "exported-log-rate-limit-interval")) { |
3539 | | |
3540 | | r = parse_boolean(v); |
3541 | | if (r < 0) |
3542 | | log_unit_debug(u, "Failed to parse exported log rate limit interval %s, ignoring.", v); |
3543 | | else |
3544 | | u->exported_log_rate_limit_interval = r; |
3545 | | |
3546 | | continue; |
3547 | | |
3548 | | } else if (streq(l, "exported-log-rate-limit-burst")) { |
3549 | | |
3550 | | r = parse_boolean(v); |
3551 | | if (r < 0) |
3552 | | log_unit_debug(u, "Failed to parse exported log rate limit burst %s, ignoring.", v); |
3553 | | else |
3554 | | u->exported_log_rate_limit_burst = r; |
3555 | | |
3556 | | continue; |
3557 | | |
3558 | | } else if (STR_IN_SET(l, "cpu-usage-base", "cpuacct-usage-base")) { |
3559 | | |
3560 | | r = safe_atou64(v, &u->cpu_usage_base); |
3561 | | if (r < 0) |
3562 | | log_unit_debug(u, "Failed to parse CPU usage base %s, ignoring.", v); |
3563 | | |
3564 | | continue; |
3565 | | |
3566 | | } else if (streq(l, "cpu-usage-last")) { |
3567 | | |
3568 | | r = safe_atou64(v, &u->cpu_usage_last); |
3569 | | if (r < 0) |
3570 | | log_unit_debug(u, "Failed to read CPU usage last %s, ignoring.", v); |
3571 | | |
3572 | | continue; |
3573 | | |
3574 | | } else if (streq(l, "oom-kill-last")) { |
3575 | | |
3576 | | r = safe_atou64(v, &u->oom_kill_last); |
3577 | | if (r < 0) |
3578 | | log_unit_debug(u, "Failed to read OOM kill last %s, ignoring.", v); |
3579 | | |
3580 | | continue; |
3581 | | |
3582 | | } else if (streq(l, "cgroup")) { |
3583 | | |
3584 | | r = unit_set_cgroup_path(u, v); |
3585 | | if (r < 0) |
3586 | | log_unit_debug_errno(u, r, "Failed to set cgroup path %s, ignoring: %m", v); |
3587 | | |
3588 | | (void) unit_watch_cgroup(u); |
3589 | | (void) unit_watch_cgroup_memory(u); |
3590 | | |
3591 | | continue; |
3592 | | } else if (streq(l, "cgroup-realized")) { |
3593 | | int b; |
3594 | | |
3595 | | b = parse_boolean(v); |
3596 | | if (b < 0) |
3597 | | log_unit_debug(u, "Failed to parse cgroup-realized bool %s, ignoring.", v); |
3598 | | else |
3599 | | u->cgroup_realized = b; |
3600 | | |
3601 | | continue; |
3602 | | |
3603 | | } else if (streq(l, "cgroup-realized-mask")) { |
3604 | | |
3605 | | r = cg_mask_from_string(v, &u->cgroup_realized_mask); |
3606 | | if (r < 0) |
3607 | | log_unit_debug(u, "Failed to parse cgroup-realized-mask %s, ignoring.", v); |
3608 | | continue; |
3609 | | |
3610 | | } else if (streq(l, "cgroup-enabled-mask")) { |
3611 | | |
3612 | | r = cg_mask_from_string(v, &u->cgroup_enabled_mask); |
3613 | | if (r < 0) |
3614 | | log_unit_debug(u, "Failed to parse cgroup-enabled-mask %s, ignoring.", v); |
3615 | | continue; |
3616 | | |
3617 | | } else if (streq(l, "cgroup-invalidated-mask")) { |
3618 | | |
3619 | | r = cg_mask_from_string(v, &u->cgroup_invalidated_mask); |
3620 | | if (r < 0) |
3621 | | log_unit_debug(u, "Failed to parse cgroup-invalidated-mask %s, ignoring.", v); |
3622 | | continue; |
3623 | | |
3624 | | } else if (streq(l, "ref-uid")) { |
3625 | | uid_t uid; |
3626 | | |
3627 | | r = parse_uid(v, &uid); |
3628 | | if (r < 0) |
3629 | | log_unit_debug(u, "Failed to parse referenced UID %s, ignoring.", v); |
3630 | | else |
3631 | | unit_ref_uid_gid(u, uid, GID_INVALID); |
3632 | | |
3633 | | continue; |
3634 | | |
3635 | | } else if (streq(l, "ref-gid")) { |
3636 | | gid_t gid; |
3637 | | |
3638 | | r = parse_gid(v, &gid); |
3639 | | if (r < 0) |
3640 | | log_unit_debug(u, "Failed to parse referenced GID %s, ignoring.", v); |
3641 | | else |
3642 | | unit_ref_uid_gid(u, UID_INVALID, gid); |
3643 | | |
3644 | | continue; |
3645 | | |
3646 | | } else if (streq(l, "ref")) { |
3647 | | |
3648 | | r = strv_extend(&u->deserialized_refs, v); |
3649 | | if (r < 0) |
3650 | | return log_oom(); |
3651 | | |
3652 | | continue; |
3653 | | } else if (streq(l, "invocation-id")) { |
3654 | | sd_id128_t id; |
3655 | | |
3656 | | r = sd_id128_from_string(v, &id); |
3657 | | if (r < 0) |
3658 | | log_unit_debug(u, "Failed to parse invocation id %s, ignoring.", v); |
3659 | | else { |
3660 | | r = unit_set_invocation_id(u, id); |
3661 | | if (r < 0) |
3662 | | log_unit_warning_errno(u, r, "Failed to set invocation ID for unit: %m"); |
3663 | | } |
3664 | | |
3665 | | continue; |
3666 | | } |
3667 | | |
3668 | | /* Check if this is an IP accounting metric serialization field */ |
3669 | | m = string_table_lookup(ip_accounting_metric_field, ELEMENTSOF(ip_accounting_metric_field), l); |
3670 | | if (m >= 0) { |
3671 | | uint64_t c; |
3672 | | |
3673 | | r = safe_atou64(v, &c); |
3674 | | if (r < 0) |
3675 | | log_unit_debug(u, "Failed to parse IP accounting value %s, ignoring.", v); |
3676 | | else |
3677 | | u->ip_accounting_extra[m] = c; |
3678 | | continue; |
3679 | | } |
3680 | | |
3681 | | m = string_table_lookup(io_accounting_metric_field_base, ELEMENTSOF(io_accounting_metric_field_base), l); |
3682 | | if (m >= 0) { |
3683 | | uint64_t c; |
3684 | | |
3685 | | r = safe_atou64(v, &c); |
3686 | | if (r < 0) |
3687 | | log_unit_debug(u, "Failed to parse IO accounting base value %s, ignoring.", v); |
3688 | | else |
3689 | | u->io_accounting_base[m] = c; |
3690 | | continue; |
3691 | | } |
3692 | | |
3693 | | m = string_table_lookup(io_accounting_metric_field_last, ELEMENTSOF(io_accounting_metric_field_last), l); |
3694 | | if (m >= 0) { |
3695 | | uint64_t c; |
3696 | | |
3697 | | r = safe_atou64(v, &c); |
3698 | | if (r < 0) |
3699 | | log_unit_debug(u, "Failed to parse IO accounting last value %s, ignoring.", v); |
3700 | | else |
3701 | | u->io_accounting_last[m] = c; |
3702 | | continue; |
3703 | | } |
3704 | | |
3705 | | if (unit_can_serialize(u)) { |
3706 | | r = exec_runtime_deserialize_compat(u, l, v, fds); |
3707 | | if (r < 0) { |
3708 | | log_unit_warning(u, "Failed to deserialize runtime parameter '%s', ignoring.", l); |
3709 | | continue; |
3710 | | } |
3711 | | |
3712 | | /* Returns positive if key was handled by the call */ |
3713 | | if (r > 0) |
3714 | | continue; |
3715 | | |
3716 | | r = UNIT_VTABLE(u)->deserialize_item(u, l, v, fds); |
3717 | | if (r < 0) |
3718 | | log_unit_warning(u, "Failed to deserialize unit parameter '%s', ignoring.", l); |
3719 | | } |
3720 | | } |
3721 | | |
3722 | | /* Versions before 228 did not carry a state change timestamp. In this case, take the current time. This is |
3723 | | * useful, so that timeouts based on this timestamp don't trigger too early, and is in-line with the logic from |
3724 | | * before 228 where the base for timeouts was not persistent across reboots. */ |
3725 | | |
3726 | | if (!dual_timestamp_is_set(&u->state_change_timestamp)) |
3727 | | dual_timestamp_get(&u->state_change_timestamp); |
3728 | | |
3729 | | /* Let's make sure that everything that is deserialized also gets any potential new cgroup settings applied |
3730 | | * after we are done. For that we invalidate anything already realized, so that we can realize it again. */ |
3731 | | unit_invalidate_cgroup(u, _CGROUP_MASK_ALL); |
3732 | | unit_invalidate_cgroup_bpf(u); |
3733 | | |
3734 | | return 0; |
3735 | | } |
3736 | | |
3737 | 0 | int unit_deserialize_skip(FILE *f) { |
3738 | 0 | int r; |
3739 | 0 | assert(f); |
3740 | 0 |
|
3741 | 0 | /* Skip serialized data for this unit. We don't know what it is. */ |
3742 | 0 |
|
3743 | 0 | for (;;) { |
3744 | 0 | _cleanup_free_ char *line = NULL; |
3745 | 0 | char *l; |
3746 | 0 |
|
3747 | 0 | r = read_line(f, LONG_LINE_MAX, &line); |
3748 | 0 | if (r < 0) |
3749 | 0 | return log_error_errno(r, "Failed to read serialization line: %m"); |
3750 | 0 | if (r == 0) |
3751 | 0 | return 0; |
3752 | 0 | |
3753 | 0 | l = strstrip(line); |
3754 | 0 |
|
3755 | 0 | /* End marker */ |
3756 | 0 | if (isempty(l)) |
3757 | 0 | return 1; |
3758 | 0 | } |
3759 | 0 | } |
3760 | | |
3761 | 0 | int unit_add_node_dependency(Unit *u, const char *what, bool wants, UnitDependency dep, UnitDependencyMask mask) { |
3762 | 0 | Unit *device; |
3763 | 0 | _cleanup_free_ char *e = NULL; |
3764 | 0 | int r; |
3765 | 0 |
|
3766 | 0 | assert(u); |
3767 | 0 |
|
3768 | 0 | /* Adds in links to the device node that this unit is based on */ |
3769 | 0 | if (isempty(what)) |
3770 | 0 | return 0; |
3771 | 0 | |
3772 | 0 | if (!is_device_path(what)) |
3773 | 0 | return 0; |
3774 | 0 | |
3775 | 0 | /* When device units aren't supported (such as in a |
3776 | 0 | * container), don't create dependencies on them. */ |
3777 | 0 | if (!unit_type_supported(UNIT_DEVICE)) |
3778 | 0 | return 0; |
3779 | 0 | |
3780 | 0 | r = unit_name_from_path(what, ".device", &e); |
3781 | 0 | if (r < 0) |
3782 | 0 | return r; |
3783 | 0 | |
3784 | 0 | r = manager_load_unit(u->manager, e, NULL, NULL, &device); |
3785 | 0 | if (r < 0) |
3786 | 0 | return r; |
3787 | 0 | |
3788 | 0 | if (dep == UNIT_REQUIRES && device_shall_be_bound_by(device, u)) |
3789 | 0 | dep = UNIT_BINDS_TO; |
3790 | 0 |
|
3791 | 0 | r = unit_add_two_dependencies(u, UNIT_AFTER, |
3792 | 0 | MANAGER_IS_SYSTEM(u->manager) ? dep : UNIT_WANTS, |
3793 | 0 | device, true, mask); |
3794 | 0 | if (r < 0) |
3795 | 0 | return r; |
3796 | 0 | |
3797 | 0 | if (wants) { |
3798 | 0 | r = unit_add_dependency(device, UNIT_WANTS, u, false, mask); |
3799 | 0 | if (r < 0) |
3800 | 0 | return r; |
3801 | 0 | } |
3802 | 0 | |
3803 | 0 | return 0; |
3804 | 0 | } |
3805 | | |
3806 | 0 | int unit_coldplug(Unit *u) { |
3807 | 0 | int r = 0, q; |
3808 | 0 | char **i; |
3809 | 0 |
|
3810 | 0 | assert(u); |
3811 | 0 |
|
3812 | 0 | /* Make sure we don't enter a loop, when coldplugging recursively. */ |
3813 | 0 | if (u->coldplugged) |
3814 | 0 | return 0; |
3815 | 0 | |
3816 | 0 | u->coldplugged = true; |
3817 | 0 |
|
3818 | 0 | STRV_FOREACH(i, u->deserialized_refs) { |
3819 | 0 | q = bus_unit_track_add_name(u, *i); |
3820 | 0 | if (q < 0 && r >= 0) |
3821 | 0 | r = q; |
3822 | 0 | } |
3823 | 0 | u->deserialized_refs = strv_free(u->deserialized_refs); |
3824 | 0 |
|
3825 | 0 | if (UNIT_VTABLE(u)->coldplug) { |
3826 | 0 | q = UNIT_VTABLE(u)->coldplug(u); |
3827 | 0 | if (q < 0 && r >= 0) |
3828 | 0 | r = q; |
3829 | 0 | } |
3830 | 0 |
|
3831 | 0 | if (u->job) { |
3832 | 0 | q = job_coldplug(u->job); |
3833 | 0 | if (q < 0 && r >= 0) |
3834 | 0 | r = q; |
3835 | 0 | } |
3836 | 0 |
|
3837 | 0 | return r; |
3838 | 0 | } |
3839 | | |
3840 | 0 | void unit_catchup(Unit *u) { |
3841 | 0 | assert(u); |
3842 | 0 |
|
3843 | 0 | if (UNIT_VTABLE(u)->catchup) |
3844 | 0 | UNIT_VTABLE(u)->catchup(u); |
3845 | 0 | } |
3846 | | |
3847 | 106k | static bool fragment_mtime_newer(const char *path, usec_t mtime, bool path_masked) { |
3848 | 106k | struct stat st; |
3849 | 106k | |
3850 | 106k | if (!path) |
3851 | 106k | return false; |
3852 | 40 | |
3853 | 40 | /* If the source is some virtual kernel file system, then we assume we watch it anyway, and hence pretend we |
3854 | 40 | * are never out-of-date. */ |
3855 | 40 | if (PATH_STARTSWITH_SET(path, "/proc", "/sys")) |
3856 | 40 | return false; |
3857 | 38 | |
3858 | 38 | if (stat(path, &st) < 0) |
3859 | 20 | /* What, cannot access this anymore? */ |
3860 | 20 | return true; |
3861 | 18 | |
3862 | 18 | if (path_masked) |
3863 | 0 | /* For masked files check if they are still so */ |
3864 | 0 | return !null_or_empty(&st); |
3865 | 18 | else |
3866 | 18 | /* For non-empty files check the mtime */ |
3867 | 18 | return timespec_load(&st.st_mtim) > mtime; |
3868 | 0 | |
3869 | 0 | return false; |
3870 | 0 | } |
3871 | | |
3872 | 53.3k | bool unit_need_daemon_reload(Unit *u) { |
3873 | 53.3k | _cleanup_strv_free_ char **t = NULL; |
3874 | 53.3k | char **path; |
3875 | 53.3k | |
3876 | 53.3k | assert(u); |
3877 | 53.3k | |
3878 | 53.3k | /* For unit files, we allow masking… */ |
3879 | 53.3k | if (fragment_mtime_newer(u->fragment_path, u->fragment_mtime, |
3880 | 53.3k | u->load_state == UNIT_MASKED)) |
3881 | 0 | return true; |
3882 | 53.3k | |
3883 | 53.3k | /* Source paths should not be masked… */ |
3884 | 53.3k | if (fragment_mtime_newer(u->source_path, u->source_mtime, false)) |
3885 | 38 | return true; |
3886 | 53.3k | |
3887 | 53.3k | if (u->load_state == UNIT_LOADED) |
3888 | 6.94k | (void) unit_find_dropin_paths(u, &t); |
3889 | 53.3k | if (!strv_equal(u->dropin_paths, t)) |
3890 | 0 | return true; |
3891 | 53.3k | |
3892 | 53.3k | /* … any drop-ins that are masked are simply omitted from the list. */ |
3893 | 53.3k | STRV_FOREACH(path, u->dropin_paths) |
3894 | 53.3k | if (fragment_mtime_newer(*path, u->dropin_mtime, false)) |
3895 | 0 | return true; |
3896 | 53.3k | |
3897 | 53.3k | return false; |
3898 | 53.3k | } |
3899 | | |
3900 | 0 | void unit_reset_failed(Unit *u) { |
3901 | 0 | assert(u); |
3902 | 0 |
|
3903 | 0 | if (UNIT_VTABLE(u)->reset_failed) |
3904 | 0 | UNIT_VTABLE(u)->reset_failed(u); |
3905 | 0 |
|
3906 | 0 | RATELIMIT_RESET(u->start_limit); |
3907 | 0 | u->start_limit_hit = false; |
3908 | 0 | } |
3909 | | |
3910 | 53.3k | Unit *unit_following(Unit *u) { |
3911 | 53.3k | assert(u); |
3912 | 53.3k | |
3913 | 53.3k | if (UNIT_VTABLE(u)->following) |
3914 | 9.38k | return UNIT_VTABLE(u)->following(u); |
3915 | 43.9k | |
3916 | 43.9k | return NULL; |
3917 | 43.9k | } |
3918 | | |
3919 | 0 | bool unit_stop_pending(Unit *u) { |
3920 | 0 | assert(u); |
3921 | 0 |
|
3922 | 0 | /* This call does check the current state of the unit. It's |
3923 | 0 | * hence useful to be called from state change calls of the |
3924 | 0 | * unit itself, where the state isn't updated yet. This is |
3925 | 0 | * different from unit_inactive_or_pending() which checks both |
3926 | 0 | * the current state and for a queued job. */ |
3927 | 0 |
|
3928 | 0 | return u->job && u->job->type == JOB_STOP; |
3929 | 0 | } |
3930 | | |
3931 | 0 | bool unit_inactive_or_pending(Unit *u) { |
3932 | 0 | assert(u); |
3933 | 0 |
|
3934 | 0 | /* Returns true if the unit is inactive or going down */ |
3935 | 0 |
|
3936 | 0 | if (UNIT_IS_INACTIVE_OR_DEACTIVATING(unit_active_state(u))) |
3937 | 0 | return true; |
3938 | 0 | |
3939 | 0 | if (unit_stop_pending(u)) |
3940 | 0 | return true; |
3941 | 0 | |
3942 | 0 | return false; |
3943 | 0 | } |
3944 | | |
3945 | 0 | bool unit_active_or_pending(Unit *u) { |
3946 | 0 | assert(u); |
3947 | 0 |
|
3948 | 0 | /* Returns true if the unit is active or going up */ |
3949 | 0 |
|
3950 | 0 | if (UNIT_IS_ACTIVE_OR_ACTIVATING(unit_active_state(u))) |
3951 | 0 | return true; |
3952 | 0 | |
3953 | 0 | if (u->job && |
3954 | 0 | IN_SET(u->job->type, JOB_START, JOB_RELOAD_OR_START, JOB_RESTART)) |
3955 | 0 | return true; |
3956 | 0 | |
3957 | 0 | return false; |
3958 | 0 | } |
3959 | | |
3960 | 0 | bool unit_will_restart(Unit *u) { |
3961 | 0 | assert(u); |
3962 | 0 |
|
3963 | 0 | if (!UNIT_VTABLE(u)->will_restart) |
3964 | 0 | return false; |
3965 | 0 | |
3966 | 0 | return UNIT_VTABLE(u)->will_restart(u); |
3967 | 0 | } |
3968 | | |
3969 | 0 | int unit_kill(Unit *u, KillWho w, int signo, sd_bus_error *error) { |
3970 | 0 | assert(u); |
3971 | 0 | assert(w >= 0 && w < _KILL_WHO_MAX); |
3972 | 0 | assert(SIGNAL_VALID(signo)); |
3973 | 0 |
|
3974 | 0 | if (!UNIT_VTABLE(u)->kill) |
3975 | 0 | return -EOPNOTSUPP; |
3976 | 0 | |
3977 | 0 | return UNIT_VTABLE(u)->kill(u, w, signo, error); |
3978 | 0 | } |
3979 | | |
3980 | 0 | static Set *unit_pid_set(pid_t main_pid, pid_t control_pid) { |
3981 | 0 | _cleanup_set_free_ Set *pid_set = NULL; |
3982 | 0 | int r; |
3983 | 0 |
|
3984 | 0 | pid_set = set_new(NULL); |
3985 | 0 | if (!pid_set) |
3986 | 0 | return NULL; |
3987 | 0 | |
3988 | 0 | /* Exclude the main/control pids from being killed via the cgroup */ |
3989 | 0 | if (main_pid > 0) { |
3990 | 0 | r = set_put(pid_set, PID_TO_PTR(main_pid)); |
3991 | 0 | if (r < 0) |
3992 | 0 | return NULL; |
3993 | 0 | } |
3994 | 0 | |
3995 | 0 | if (control_pid > 0) { |
3996 | 0 | r = set_put(pid_set, PID_TO_PTR(control_pid)); |
3997 | 0 | if (r < 0) |
3998 | 0 | return NULL; |
3999 | 0 | } |
4000 | 0 | |
4001 | 0 | return TAKE_PTR(pid_set); |
4002 | 0 | } |
4003 | | |
4004 | | int unit_kill_common( |
4005 | | Unit *u, |
4006 | | KillWho who, |
4007 | | int signo, |
4008 | | pid_t main_pid, |
4009 | | pid_t control_pid, |
4010 | 0 | sd_bus_error *error) { |
4011 | 0 |
|
4012 | 0 | int r = 0; |
4013 | 0 | bool killed = false; |
4014 | 0 |
|
4015 | 0 | if (IN_SET(who, KILL_MAIN, KILL_MAIN_FAIL)) { |
4016 | 0 | if (main_pid < 0) |
4017 | 0 | return sd_bus_error_setf(error, BUS_ERROR_NO_SUCH_PROCESS, "%s units have no main processes", unit_type_to_string(u->type)); |
4018 | 0 | else if (main_pid == 0) |
4019 | 0 | return sd_bus_error_set_const(error, BUS_ERROR_NO_SUCH_PROCESS, "No main process to kill"); |
4020 | 0 | } |
4021 | 0 |
|
4022 | 0 | if (IN_SET(who, KILL_CONTROL, KILL_CONTROL_FAIL)) { |
4023 | 0 | if (control_pid < 0) |
4024 | 0 | return sd_bus_error_setf(error, BUS_ERROR_NO_SUCH_PROCESS, "%s units have no control processes", unit_type_to_string(u->type)); |
4025 | 0 | else if (control_pid == 0) |
4026 | 0 | return sd_bus_error_set_const(error, BUS_ERROR_NO_SUCH_PROCESS, "No control process to kill"); |
4027 | 0 | } |
4028 | 0 |
|
4029 | 0 | if (IN_SET(who, KILL_CONTROL, KILL_CONTROL_FAIL, KILL_ALL, KILL_ALL_FAIL)) |
4030 | 0 | if (control_pid > 0) { |
4031 | 0 | if (kill(control_pid, signo) < 0) |
4032 | 0 | r = -errno; |
4033 | 0 | else |
4034 | 0 | killed = true; |
4035 | 0 | } |
4036 | 0 |
|
4037 | 0 | if (IN_SET(who, KILL_MAIN, KILL_MAIN_FAIL, KILL_ALL, KILL_ALL_FAIL)) |
4038 | 0 | if (main_pid > 0) { |
4039 | 0 | if (kill(main_pid, signo) < 0) |
4040 | 0 | r = -errno; |
4041 | 0 | else |
4042 | 0 | killed = true; |
4043 | 0 | } |
4044 | 0 |
|
4045 | 0 | if (IN_SET(who, KILL_ALL, KILL_ALL_FAIL) && u->cgroup_path) { |
4046 | 0 | _cleanup_set_free_ Set *pid_set = NULL; |
4047 | 0 | int q; |
4048 | 0 |
|
4049 | 0 | /* Exclude the main/control pids from being killed via the cgroup */ |
4050 | 0 | pid_set = unit_pid_set(main_pid, control_pid); |
4051 | 0 | if (!pid_set) |
4052 | 0 | return -ENOMEM; |
4053 | 0 | |
4054 | 0 | q = cg_kill_recursive(SYSTEMD_CGROUP_CONTROLLER, u->cgroup_path, signo, 0, pid_set, NULL, NULL); |
4055 | 0 | if (q < 0 && !IN_SET(q, -EAGAIN, -ESRCH, -ENOENT)) |
4056 | 0 | r = q; |
4057 | 0 | else |
4058 | 0 | killed = true; |
4059 | 0 | } |
4060 | 0 |
|
4061 | 0 | if (r == 0 && !killed && IN_SET(who, KILL_ALL_FAIL, KILL_CONTROL_FAIL)) |
4062 | 0 | return -ESRCH; |
4063 | 0 | |
4064 | 0 | return r; |
4065 | 0 | } |
4066 | | |
4067 | 53.3k | int unit_following_set(Unit *u, Set **s) { |
4068 | 53.3k | assert(u); |
4069 | 53.3k | assert(s); |
4070 | 53.3k | |
4071 | 53.3k | if (UNIT_VTABLE(u)->following_set) |
4072 | 9.38k | return UNIT_VTABLE(u)->following_set(u, s); |
4073 | 43.9k | |
4074 | 43.9k | *s = NULL; |
4075 | 43.9k | return 0; |
4076 | 43.9k | } |
4077 | | |
4078 | 0 | UnitFileState unit_get_unit_file_state(Unit *u) { |
4079 | 0 | int r; |
4080 | 0 |
|
4081 | 0 | assert(u); |
4082 | 0 |
|
4083 | 0 | if (u->unit_file_state < 0 && u->fragment_path) { |
4084 | 0 | r = unit_file_get_state( |
4085 | 0 | u->manager->unit_file_scope, |
4086 | 0 | NULL, |
4087 | 0 | u->id, |
4088 | 0 | &u->unit_file_state); |
4089 | 0 | if (r < 0) |
4090 | 0 | u->unit_file_state = UNIT_FILE_BAD; |
4091 | 0 | } |
4092 | 0 |
|
4093 | 0 | return u->unit_file_state; |
4094 | 0 | } |
4095 | | |
4096 | 0 | int unit_get_unit_file_preset(Unit *u) { |
4097 | 0 | assert(u); |
4098 | 0 |
|
4099 | 0 | if (u->unit_file_preset < 0 && u->fragment_path) |
4100 | 0 | u->unit_file_preset = unit_file_query_preset( |
4101 | 0 | u->manager->unit_file_scope, |
4102 | 0 | NULL, |
4103 | 0 | basename(u->fragment_path)); |
4104 | 0 |
|
4105 | 0 | return u->unit_file_preset; |
4106 | 0 | } |
4107 | | |
4108 | 5.50k | Unit* unit_ref_set(UnitRef *ref, Unit *source, Unit *target) { |
4109 | 5.50k | assert(ref); |
4110 | 5.50k | assert(source); |
4111 | 5.50k | assert(target); |
4112 | 5.50k | |
4113 | 5.50k | if (ref->target) |
4114 | 497 | unit_ref_unset(ref); |
4115 | 5.50k | |
4116 | 5.50k | ref->source = source; |
4117 | 5.50k | ref->target = target; |
4118 | 5.50k | LIST_PREPEND(refs_by_target, target->refs_by_target, ref); |
4119 | 5.50k | return target; |
4120 | 5.50k | } |
4121 | | |
4122 | 54.8k | void unit_ref_unset(UnitRef *ref) { |
4123 | 54.8k | assert(ref); |
4124 | 54.8k | |
4125 | 54.8k | if (!ref->target) |
4126 | 49.3k | return; |
4127 | 5.50k | |
4128 | 5.50k | /* We are about to drop a reference to the unit, make sure the garbage collection has a look at it as it might |
4129 | 5.50k | * be unreferenced now. */ |
4130 | 5.50k | unit_add_to_gc_queue(ref->target); |
4131 | 5.50k | |
4132 | 5.50k | LIST_REMOVE(refs_by_target, ref->target->refs_by_target, ref); |
4133 | 5.50k | ref->source = ref->target = NULL; |
4134 | 5.50k | } |
4135 | | |
4136 | 0 | static int user_from_unit_name(Unit *u, char **ret) { |
4137 | 0 |
|
4138 | 0 | static const uint8_t hash_key[] = { |
4139 | 0 | 0x58, 0x1a, 0xaf, 0xe6, 0x28, 0x58, 0x4e, 0x96, |
4140 | 0 | 0xb4, 0x4e, 0xf5, 0x3b, 0x8c, 0x92, 0x07, 0xec |
4141 | 0 | }; |
4142 | 0 |
|
4143 | 0 | _cleanup_free_ char *n = NULL; |
4144 | 0 | int r; |
4145 | 0 |
|
4146 | 0 | r = unit_name_to_prefix(u->id, &n); |
4147 | 0 | if (r < 0) |
4148 | 0 | return r; |
4149 | 0 | |
4150 | 0 | if (valid_user_group_name(n)) { |
4151 | 0 | *ret = TAKE_PTR(n); |
4152 | 0 | return 0; |
4153 | 0 | } |
4154 | 0 |
|
4155 | 0 | /* If we can't use the unit name as a user name, then let's hash it and use that */ |
4156 | 0 | if (asprintf(ret, "_du%016" PRIx64, siphash24(n, strlen(n), hash_key)) < 0) |
4157 | 0 | return -ENOMEM; |
4158 | 0 | |
4159 | 0 | return 0; |
4160 | 0 | } |
4161 | | |
4162 | 5.72k | int unit_patch_contexts(Unit *u) { |
4163 | 5.72k | CGroupContext *cc; |
4164 | 5.72k | ExecContext *ec; |
4165 | 5.72k | unsigned i; |
4166 | 5.72k | int r; |
4167 | 5.72k | |
4168 | 5.72k | assert(u); |
4169 | 5.72k | |
4170 | 5.72k | /* Patch in the manager defaults into the exec and cgroup |
4171 | 5.72k | * contexts, _after_ the rest of the settings have been |
4172 | 5.72k | * initialized */ |
4173 | 5.72k | |
4174 | 5.72k | ec = unit_get_exec_context(u); |
4175 | 5.72k | if (ec) { |
4176 | 8 | /* This only copies in the ones that need memory */ |
4177 | 136 | for (i = 0; i < _RLIMIT_MAX; i++) |
4178 | 128 | if (u->manager->rlimit[i] && !ec->rlimit[i]) { |
4179 | 0 | ec->rlimit[i] = newdup(struct rlimit, u->manager->rlimit[i], 1); |
4180 | 0 | if (!ec->rlimit[i]) |
4181 | 0 | return -ENOMEM; |
4182 | 0 | } |
4183 | 8 | |
4184 | 8 | if (MANAGER_IS_USER(u->manager) && |
4185 | 8 | !ec->working_directory) { |
4186 | 0 |
|
4187 | 0 | r = get_home_dir(&ec->working_directory); |
4188 | 0 | if (r < 0) |
4189 | 0 | return r; |
4190 | 0 | |
4191 | 0 | /* Allow user services to run, even if the |
4192 | 0 | * home directory is missing */ |
4193 | 0 | ec->working_directory_missing_ok = true; |
4194 | 0 | } |
4195 | 8 | |
4196 | 8 | if (ec->private_devices) |
4197 | 0 | ec->capability_bounding_set &= ~((UINT64_C(1) << CAP_MKNOD) | (UINT64_C(1) << CAP_SYS_RAWIO)); |
4198 | 8 | |
4199 | 8 | if (ec->protect_kernel_modules) |
4200 | 0 | ec->capability_bounding_set &= ~(UINT64_C(1) << CAP_SYS_MODULE); |
4201 | 8 | |
4202 | 8 | if (ec->dynamic_user) { |
4203 | 0 | if (!ec->user) { |
4204 | 0 | r = user_from_unit_name(u, &ec->user); |
4205 | 0 | if (r < 0) |
4206 | 0 | return r; |
4207 | 0 | } |
4208 | 0 | |
4209 | 0 | if (!ec->group) { |
4210 | 0 | ec->group = strdup(ec->user); |
4211 | 0 | if (!ec->group) |
4212 | 0 | return -ENOMEM; |
4213 | 0 | } |
4214 | 0 | |
4215 | 0 | /* If the dynamic user option is on, let's make sure that the unit can't leave its |
4216 | 0 | * UID/GID around in the file system or on IPC objects. Hence enforce a strict |
4217 | 0 | * sandbox. */ |
4218 | 0 | |
4219 | 0 | ec->private_tmp = true; |
4220 | 0 | ec->remove_ipc = true; |
4221 | 0 | ec->protect_system = PROTECT_SYSTEM_STRICT; |
4222 | 0 | if (ec->protect_home == PROTECT_HOME_NO) |
4223 | 0 | ec->protect_home = PROTECT_HOME_READ_ONLY; |
4224 | 0 |
|
4225 | 0 | /* Make sure this service can neither benefit from SUID/SGID binaries nor create |
4226 | 0 | * them. */ |
4227 | 0 | ec->no_new_privileges = true; |
4228 | 0 | ec->restrict_suid_sgid = true; |
4229 | 0 | } |
4230 | 8 | } |
4231 | 5.72k | |
4232 | 5.72k | cc = unit_get_cgroup_context(u); |
4233 | 5.72k | if (cc && ec) { |
4234 | 8 | |
4235 | 8 | if (ec->private_devices && |
4236 | 8 | cc->device_policy == CGROUP_AUTO) |
4237 | 0 | cc->device_policy = CGROUP_CLOSED; |
4238 | 8 | |
4239 | 8 | if (ec->root_image && |
4240 | 8 | (cc->device_policy != CGROUP_AUTO || cc->device_allow)) { |
4241 | 0 |
|
4242 | 0 | /* When RootImage= is specified, the following devices are touched. */ |
4243 | 0 | r = cgroup_add_device_allow(cc, "/dev/loop-control", "rw"); |
4244 | 0 | if (r < 0) |
4245 | 0 | return r; |
4246 | 0 | |
4247 | 0 | r = cgroup_add_device_allow(cc, "block-loop", "rwm"); |
4248 | 0 | if (r < 0) |
4249 | 0 | return r; |
4250 | 0 | |
4251 | 0 | r = cgroup_add_device_allow(cc, "block-blkext", "rwm"); |
4252 | 0 | if (r < 0) |
4253 | 0 | return r; |
4254 | 5.72k | } |
4255 | 8 | } |
4256 | 5.72k | |
4257 | 5.72k | return 0; |
4258 | 5.72k | } |
4259 | | |
4260 | 76.4k | ExecContext *unit_get_exec_context(Unit *u) { |
4261 | 76.4k | size_t offset; |
4262 | 76.4k | assert(u); |
4263 | 76.4k | |
4264 | 76.4k | if (u->type < 0) |
4265 | 0 | return NULL; |
4266 | 76.4k | |
4267 | 76.4k | offset = UNIT_VTABLE(u)->exec_context_offset; |
4268 | 76.4k | if (offset <= 0) |
4269 | 34.8k | return NULL; |
4270 | 41.6k | |
4271 | 41.6k | return (ExecContext*) ((uint8_t*) u + offset); |
4272 | 41.6k | } |
4273 | | |
4274 | 35.2k | KillContext *unit_get_kill_context(Unit *u) { |
4275 | 35.2k | size_t offset; |
4276 | 35.2k | assert(u); |
4277 | 35.2k | |
4278 | 35.2k | if (u->type < 0) |
4279 | 0 | return NULL; |
4280 | 35.2k | |
4281 | 35.2k | offset = UNIT_VTABLE(u)->kill_context_offset; |
4282 | 35.2k | if (offset <= 0) |
4283 | 13.9k | return NULL; |
4284 | 21.3k | |
4285 | 21.3k | return (KillContext*) ((uint8_t*) u + offset); |
4286 | 21.3k | } |
4287 | | |
4288 | 167k | CGroupContext *unit_get_cgroup_context(Unit *u) { |
4289 | 167k | size_t offset; |
4290 | 167k | |
4291 | 167k | if (u->type < 0) |
4292 | 0 | return NULL; |
4293 | 167k | |
4294 | 167k | offset = UNIT_VTABLE(u)->cgroup_context_offset; |
4295 | 167k | if (offset <= 0) |
4296 | 18.7k | return NULL; |
4297 | 148k | |
4298 | 148k | return (CGroupContext*) ((uint8_t*) u + offset); |
4299 | 148k | } |
4300 | | |
4301 | 0 | ExecRuntime *unit_get_exec_runtime(Unit *u) { |
4302 | 0 | size_t offset; |
4303 | 0 |
|
4304 | 0 | if (u->type < 0) |
4305 | 0 | return NULL; |
4306 | 0 | |
4307 | 0 | offset = UNIT_VTABLE(u)->exec_runtime_offset; |
4308 | 0 | if (offset <= 0) |
4309 | 0 | return NULL; |
4310 | 0 | |
4311 | 0 | return *(ExecRuntime**) ((uint8_t*) u + offset); |
4312 | 0 | } |
4313 | | |
4314 | 0 | static const char* unit_drop_in_dir(Unit *u, UnitWriteFlags flags) { |
4315 | 0 | assert(u); |
4316 | 0 |
|
4317 | 0 | if (UNIT_WRITE_FLAGS_NOOP(flags)) |
4318 | 0 | return NULL; |
4319 | 0 | |
4320 | 0 | if (u->transient) /* Redirect drop-ins for transient units always into the transient directory. */ |
4321 | 0 | return u->manager->lookup_paths.transient; |
4322 | 0 | |
4323 | 0 | if (flags & UNIT_PERSISTENT) |
4324 | 0 | return u->manager->lookup_paths.persistent_control; |
4325 | 0 | |
4326 | 0 | if (flags & UNIT_RUNTIME) |
4327 | 0 | return u->manager->lookup_paths.runtime_control; |
4328 | 0 | |
4329 | 0 | return NULL; |
4330 | 0 | } |
4331 | | |
4332 | 0 | char* unit_escape_setting(const char *s, UnitWriteFlags flags, char **buf) { |
4333 | 0 | char *ret = NULL; |
4334 | 0 |
|
4335 | 0 | if (!s) |
4336 | 0 | return NULL; |
4337 | 0 | |
4338 | 0 | /* Escapes the input string as requested. Returns the escaped string. If 'buf' is specified then the allocated |
4339 | 0 | * return buffer pointer is also written to *buf, except if no escaping was necessary, in which case *buf is |
4340 | 0 | * set to NULL, and the input pointer is returned as-is. This means the return value always contains a properly |
4341 | 0 | * escaped version, but *buf when passed only contains a pointer if an allocation was necessary. If *buf is |
4342 | 0 | * not specified, then the return value always needs to be freed. Callers can use this to optimize memory |
4343 | 0 | * allocations. */ |
4344 | 0 | |
4345 | 0 | if (flags & UNIT_ESCAPE_SPECIFIERS) { |
4346 | 0 | ret = specifier_escape(s); |
4347 | 0 | if (!ret) |
4348 | 0 | return NULL; |
4349 | 0 | |
4350 | 0 | s = ret; |
4351 | 0 | } |
4352 | 0 |
|
4353 | 0 | if (flags & UNIT_ESCAPE_C) { |
4354 | 0 | char *a; |
4355 | 0 |
|
4356 | 0 | a = cescape(s); |
4357 | 0 | free(ret); |
4358 | 0 | if (!a) |
4359 | 0 | return NULL; |
4360 | 0 | |
4361 | 0 | ret = a; |
4362 | 0 | } |
4363 | 0 |
|
4364 | 0 | if (buf) { |
4365 | 0 | *buf = ret; |
4366 | 0 | return ret ?: (char*) s; |
4367 | 0 | } |
4368 | 0 |
|
4369 | 0 | return ret ?: strdup(s); |
4370 | 0 | } |
4371 | | |
4372 | 0 | char* unit_concat_strv(char **l, UnitWriteFlags flags) { |
4373 | 0 | _cleanup_free_ char *result = NULL; |
4374 | 0 | size_t n = 0, allocated = 0; |
4375 | 0 | char **i; |
4376 | 0 |
|
4377 | 0 | /* Takes a list of strings, escapes them, and concatenates them. This may be used to format command lines in a |
4378 | 0 | * way suitable for ExecStart= stanzas */ |
4379 | 0 |
|
4380 | 0 | STRV_FOREACH(i, l) { |
4381 | 0 | _cleanup_free_ char *buf = NULL; |
4382 | 0 | const char *p; |
4383 | 0 | size_t a; |
4384 | 0 | char *q; |
4385 | 0 |
|
4386 | 0 | p = unit_escape_setting(*i, flags, &buf); |
4387 | 0 | if (!p) |
4388 | 0 | return NULL; |
4389 | 0 | |
4390 | 0 | a = (n > 0) + 1 + strlen(p) + 1; /* separating space + " + entry + " */ |
4391 | 0 | if (!GREEDY_REALLOC(result, allocated, n + a + 1)) |
4392 | 0 | return NULL; |
4393 | 0 | |
4394 | 0 | q = result + n; |
4395 | 0 | if (n > 0) |
4396 | 0 | *(q++) = ' '; |
4397 | 0 |
|
4398 | 0 | *(q++) = '"'; |
4399 | 0 | q = stpcpy(q, p); |
4400 | 0 | *(q++) = '"'; |
4401 | 0 |
|
4402 | 0 | n += a; |
4403 | 0 | } |
4404 | 0 |
|
4405 | 0 | if (!GREEDY_REALLOC(result, allocated, n + 1)) |
4406 | 0 | return NULL; |
4407 | 0 | |
4408 | 0 | result[n] = 0; |
4409 | 0 |
|
4410 | 0 | return TAKE_PTR(result); |
4411 | 0 | } |
4412 | | |
4413 | 0 | int unit_write_setting(Unit *u, UnitWriteFlags flags, const char *name, const char *data) { |
4414 | 0 | _cleanup_free_ char *p = NULL, *q = NULL, *escaped = NULL; |
4415 | 0 | const char *dir, *wrapped; |
4416 | 0 | int r; |
4417 | 0 |
|
4418 | 0 | assert(u); |
4419 | 0 | assert(name); |
4420 | 0 | assert(data); |
4421 | 0 |
|
4422 | 0 | if (UNIT_WRITE_FLAGS_NOOP(flags)) |
4423 | 0 | return 0; |
4424 | 0 | |
4425 | 0 | data = unit_escape_setting(data, flags, &escaped); |
4426 | 0 | if (!data) |
4427 | 0 | return -ENOMEM; |
4428 | 0 | |
4429 | 0 | /* Prefix the section header. If we are writing this out as transient file, then let's suppress this if the |
4430 | 0 | * previous section header is the same */ |
4431 | 0 | |
4432 | 0 | if (flags & UNIT_PRIVATE) { |
4433 | 0 | if (!UNIT_VTABLE(u)->private_section) |
4434 | 0 | return -EINVAL; |
4435 | 0 | |
4436 | 0 | if (!u->transient_file || u->last_section_private < 0) |
4437 | 0 | data = strjoina("[", UNIT_VTABLE(u)->private_section, "]\n", data); |
4438 | 0 | else if (u->last_section_private == 0) |
4439 | 0 | data = strjoina("\n[", UNIT_VTABLE(u)->private_section, "]\n", data); |
4440 | 0 | } else { |
4441 | 0 | if (!u->transient_file || u->last_section_private < 0) |
4442 | 0 | data = strjoina("[Unit]\n", data); |
4443 | 0 | else if (u->last_section_private > 0) |
4444 | 0 | data = strjoina("\n[Unit]\n", data); |
4445 | 0 | } |
4446 | 0 |
|
4447 | 0 | if (u->transient_file) { |
4448 | 0 | /* When this is a transient unit file in creation, then let's not create a new drop-in but instead |
4449 | 0 | * write to the transient unit file. */ |
4450 | 0 | fputs(data, u->transient_file); |
4451 | 0 |
|
4452 | 0 | if (!endswith(data, "\n")) |
4453 | 0 | fputc('\n', u->transient_file); |
4454 | 0 |
|
4455 | 0 | /* Remember which section we wrote this entry to */ |
4456 | 0 | u->last_section_private = !!(flags & UNIT_PRIVATE); |
4457 | 0 | return 0; |
4458 | 0 | } |
4459 | 0 |
|
4460 | 0 | dir = unit_drop_in_dir(u, flags); |
4461 | 0 | if (!dir) |
4462 | 0 | return -EINVAL; |
4463 | 0 | |
4464 | 0 | wrapped = strjoina("# This is a drop-in unit file extension, created via \"systemctl set-property\"\n" |
4465 | 0 | "# or an equivalent operation. Do not edit.\n", |
4466 | 0 | data, |
4467 | 0 | "\n"); |
4468 | 0 |
|
4469 | 0 | r = drop_in_file(dir, u->id, 50, name, &p, &q); |
4470 | 0 | if (r < 0) |
4471 | 0 | return r; |
4472 | 0 | |
4473 | 0 | (void) mkdir_p_label(p, 0755); |
4474 | 0 | r = write_string_file_atomic_label(q, wrapped); |
4475 | 0 | if (r < 0) |
4476 | 0 | return r; |
4477 | 0 | |
4478 | 0 | r = strv_push(&u->dropin_paths, q); |
4479 | 0 | if (r < 0) |
4480 | 0 | return r; |
4481 | 0 | q = NULL; |
4482 | 0 |
|
4483 | 0 | strv_uniq(u->dropin_paths); |
4484 | 0 |
|
4485 | 0 | u->dropin_mtime = now(CLOCK_REALTIME); |
4486 | 0 |
|
4487 | 0 | return 0; |
4488 | 0 | } |
4489 | | |
4490 | 0 | int unit_write_settingf(Unit *u, UnitWriteFlags flags, const char *name, const char *format, ...) { |
4491 | 0 | _cleanup_free_ char *p = NULL; |
4492 | 0 | va_list ap; |
4493 | 0 | int r; |
4494 | 0 |
|
4495 | 0 | assert(u); |
4496 | 0 | assert(name); |
4497 | 0 | assert(format); |
4498 | 0 |
|
4499 | 0 | if (UNIT_WRITE_FLAGS_NOOP(flags)) |
4500 | 0 | return 0; |
4501 | 0 | |
4502 | 0 | va_start(ap, format); |
4503 | 0 | r = vasprintf(&p, format, ap); |
4504 | 0 | va_end(ap); |
4505 | 0 |
|
4506 | 0 | if (r < 0) |
4507 | 0 | return -ENOMEM; |
4508 | 0 | |
4509 | 0 | return unit_write_setting(u, flags, name, p); |
4510 | 0 | } |
4511 | | |
4512 | 0 | int unit_make_transient(Unit *u) { |
4513 | 0 | _cleanup_free_ char *path = NULL; |
4514 | 0 | FILE *f; |
4515 | 0 |
|
4516 | 0 | assert(u); |
4517 | 0 |
|
4518 | 0 | if (!UNIT_VTABLE(u)->can_transient) |
4519 | 0 | return -EOPNOTSUPP; |
4520 | 0 | |
4521 | 0 | (void) mkdir_p_label(u->manager->lookup_paths.transient, 0755); |
4522 | 0 |
|
4523 | 0 | path = strjoin(u->manager->lookup_paths.transient, "/", u->id); |
4524 | 0 | if (!path) |
4525 | 0 | return -ENOMEM; |
4526 | 0 | |
4527 | 0 | /* Let's open the file we'll write the transient settings into. This file is kept open as long as we are |
4528 | 0 | * creating the transient, and is closed in unit_load(), as soon as we start loading the file. */ |
4529 | 0 | |
4530 | 0 | RUN_WITH_UMASK(0022) { |
4531 | 0 | f = fopen(path, "we"); |
4532 | 0 | if (!f) |
4533 | 0 | return -errno; |
4534 | 0 | } |
4535 | 0 |
|
4536 | 0 | safe_fclose(u->transient_file); |
4537 | 0 | u->transient_file = f; |
4538 | 0 |
|
4539 | 0 | free_and_replace(u->fragment_path, path); |
4540 | 0 |
|
4541 | 0 | u->source_path = mfree(u->source_path); |
4542 | 0 | u->dropin_paths = strv_free(u->dropin_paths); |
4543 | 0 | u->fragment_mtime = u->source_mtime = u->dropin_mtime = 0; |
4544 | 0 |
|
4545 | 0 | u->load_state = UNIT_STUB; |
4546 | 0 | u->load_error = 0; |
4547 | 0 | u->transient = true; |
4548 | 0 |
|
4549 | 0 | unit_add_to_dbus_queue(u); |
4550 | 0 | unit_add_to_gc_queue(u); |
4551 | 0 |
|
4552 | 0 | fputs("# This is a transient unit file, created programmatically via the systemd API. Do not edit.\n", |
4553 | 0 | u->transient_file); |
4554 | 0 |
|
4555 | 0 | return 0; |
4556 | 0 | } |
4557 | | |
4558 | 0 | static int log_kill(pid_t pid, int sig, void *userdata) { |
4559 | 0 | _cleanup_free_ char *comm = NULL; |
4560 | 0 |
|
4561 | 0 | (void) get_process_comm(pid, &comm); |
4562 | 0 |
|
4563 | 0 | /* Don't log about processes marked with brackets, under the assumption that these are temporary processes |
4564 | 0 | only, like for example systemd's own PAM stub process. */ |
4565 | 0 | if (comm && comm[0] == '(') |
4566 | 0 | return 0; |
4567 | 0 | |
4568 | 0 | log_unit_notice(userdata, |
4569 | 0 | "Killing process " PID_FMT " (%s) with signal SIG%s.", |
4570 | 0 | pid, |
4571 | 0 | strna(comm), |
4572 | 0 | signal_to_string(sig)); |
4573 | 0 |
|
4574 | 0 | return 1; |
4575 | 0 | } |
4576 | | |
4577 | 0 | static int operation_to_signal(KillContext *c, KillOperation k) { |
4578 | 0 | assert(c); |
4579 | 0 |
|
4580 | 0 | switch (k) { |
4581 | 0 |
|
4582 | 0 | case KILL_TERMINATE: |
4583 | 0 | case KILL_TERMINATE_AND_LOG: |
4584 | 0 | return c->kill_signal; |
4585 | 0 |
|
4586 | 0 | case KILL_KILL: |
4587 | 0 | return c->final_kill_signal; |
4588 | 0 |
|
4589 | 0 | case KILL_WATCHDOG: |
4590 | 0 | return c->watchdog_signal; |
4591 | 0 |
|
4592 | 0 | default: |
4593 | 0 | assert_not_reached("KillOperation unknown"); |
4594 | 0 | } |
4595 | 0 | } |
4596 | | |
4597 | | int unit_kill_context( |
4598 | | Unit *u, |
4599 | | KillContext *c, |
4600 | | KillOperation k, |
4601 | | pid_t main_pid, |
4602 | | pid_t control_pid, |
4603 | 0 | bool main_pid_alien) { |
4604 | 0 |
|
4605 | 0 | bool wait_for_exit = false, send_sighup; |
4606 | 0 | cg_kill_log_func_t log_func = NULL; |
4607 | 0 | int sig, r; |
4608 | 0 |
|
4609 | 0 | assert(u); |
4610 | 0 | assert(c); |
4611 | 0 |
|
4612 | 0 | /* Kill the processes belonging to this unit, in preparation for shutting the unit down. |
4613 | 0 | * Returns > 0 if we killed something worth waiting for, 0 otherwise. */ |
4614 | 0 |
|
4615 | 0 | if (c->kill_mode == KILL_NONE) |
4616 | 0 | return 0; |
4617 | 0 | |
4618 | 0 | sig = operation_to_signal(c, k); |
4619 | 0 |
|
4620 | 0 | send_sighup = |
4621 | 0 | c->send_sighup && |
4622 | 0 | IN_SET(k, KILL_TERMINATE, KILL_TERMINATE_AND_LOG) && |
4623 | 0 | sig != SIGHUP; |
4624 | 0 |
|
4625 | 0 | if (k != KILL_TERMINATE || IN_SET(sig, SIGKILL, SIGABRT)) |
4626 | 0 | log_func = log_kill; |
4627 | 0 |
|
4628 | 0 | if (main_pid > 0) { |
4629 | 0 | if (log_func) |
4630 | 0 | log_func(main_pid, sig, u); |
4631 | 0 |
|
4632 | 0 | r = kill_and_sigcont(main_pid, sig); |
4633 | 0 | if (r < 0 && r != -ESRCH) { |
4634 | 0 | _cleanup_free_ char *comm = NULL; |
4635 | 0 | (void) get_process_comm(main_pid, &comm); |
4636 | 0 |
|
4637 | 0 | log_unit_warning_errno(u, r, "Failed to kill main process " PID_FMT " (%s), ignoring: %m", main_pid, strna(comm)); |
4638 | 0 | } else { |
4639 | 0 | if (!main_pid_alien) |
4640 | 0 | wait_for_exit = true; |
4641 | 0 |
|
4642 | 0 | if (r != -ESRCH && send_sighup) |
4643 | 0 | (void) kill(main_pid, SIGHUP); |
4644 | 0 | } |
4645 | 0 | } |
4646 | 0 |
|
4647 | 0 | if (control_pid > 0) { |
4648 | 0 | if (log_func) |
4649 | 0 | log_func(control_pid, sig, u); |
4650 | 0 |
|
4651 | 0 | r = kill_and_sigcont(control_pid, sig); |
4652 | 0 | if (r < 0 && r != -ESRCH) { |
4653 | 0 | _cleanup_free_ char *comm = NULL; |
4654 | 0 | (void) get_process_comm(control_pid, &comm); |
4655 | 0 |
|
4656 | 0 | log_unit_warning_errno(u, r, "Failed to kill control process " PID_FMT " (%s), ignoring: %m", control_pid, strna(comm)); |
4657 | 0 | } else { |
4658 | 0 | wait_for_exit = true; |
4659 | 0 |
|
4660 | 0 | if (r != -ESRCH && send_sighup) |
4661 | 0 | (void) kill(control_pid, SIGHUP); |
4662 | 0 | } |
4663 | 0 | } |
4664 | 0 |
|
4665 | 0 | if (u->cgroup_path && |
4666 | 0 | (c->kill_mode == KILL_CONTROL_GROUP || (c->kill_mode == KILL_MIXED && k == KILL_KILL))) { |
4667 | 0 | _cleanup_set_free_ Set *pid_set = NULL; |
4668 | 0 |
|
4669 | 0 | /* Exclude the main/control pids from being killed via the cgroup */ |
4670 | 0 | pid_set = unit_pid_set(main_pid, control_pid); |
4671 | 0 | if (!pid_set) |
4672 | 0 | return -ENOMEM; |
4673 | 0 | |
4674 | 0 | r = cg_kill_recursive(SYSTEMD_CGROUP_CONTROLLER, u->cgroup_path, |
4675 | 0 | sig, |
4676 | 0 | CGROUP_SIGCONT|CGROUP_IGNORE_SELF, |
4677 | 0 | pid_set, |
4678 | 0 | log_func, u); |
4679 | 0 | if (r < 0) { |
4680 | 0 | if (!IN_SET(r, -EAGAIN, -ESRCH, -ENOENT)) |
4681 | 0 | log_unit_warning_errno(u, r, "Failed to kill control group %s, ignoring: %m", u->cgroup_path); |
4682 | 0 |
|
4683 | 0 | } else if (r > 0) { |
4684 | 0 |
|
4685 | 0 | /* FIXME: For now, on the legacy hierarchy, we will not wait for the cgroup members to die if |
4686 | 0 | * we are running in a container or if this is a delegation unit, simply because cgroup |
4687 | 0 | * notification is unreliable in these cases. It doesn't work at all in containers, and outside |
4688 | 0 | * of containers it can be confused easily by left-over directories in the cgroup — which |
4689 | 0 | * however should not exist in non-delegated units. On the unified hierarchy that's different, |
4690 | 0 | * there we get proper events. Hence rely on them. */ |
4691 | 0 |
|
4692 | 0 | if (cg_unified_controller(SYSTEMD_CGROUP_CONTROLLER) > 0 || |
4693 | 0 | (detect_container() == 0 && !unit_cgroup_delegate(u))) |
4694 | 0 | wait_for_exit = true; |
4695 | 0 |
|
4696 | 0 | if (send_sighup) { |
4697 | 0 | set_free(pid_set); |
4698 | 0 |
|
4699 | 0 | pid_set = unit_pid_set(main_pid, control_pid); |
4700 | 0 | if (!pid_set) |
4701 | 0 | return -ENOMEM; |
4702 | 0 | |
4703 | 0 | cg_kill_recursive(SYSTEMD_CGROUP_CONTROLLER, u->cgroup_path, |
4704 | 0 | SIGHUP, |
4705 | 0 | CGROUP_IGNORE_SELF, |
4706 | 0 | pid_set, |
4707 | 0 | NULL, NULL); |
4708 | 0 | } |
4709 | 0 | } |
4710 | 0 | } |
4711 | 0 |
|
4712 | 0 | return wait_for_exit; |
4713 | 0 | } |
4714 | | |
4715 | 5.34k | int unit_require_mounts_for(Unit *u, const char *path, UnitDependencyMask mask) { |
4716 | 5.34k | _cleanup_free_ char *p = NULL; |
4717 | 5.34k | UnitDependencyInfo di; |
4718 | 5.34k | int r; |
4719 | 5.34k | |
4720 | 5.34k | assert(u); |
4721 | 5.34k | assert(path); |
4722 | 5.34k | |
4723 | 5.34k | /* Registers a unit for requiring a certain path and all its prefixes. We keep a hashtable of these paths in |
4724 | 5.34k | * the unit (from the path to the UnitDependencyInfo structure indicating how to the dependency came to |
4725 | 5.34k | * be). However, we build a prefix table for all possible prefixes so that new appearing mount units can easily |
4726 | 5.34k | * determine which units to make themselves a dependency of. */ |
4727 | 5.34k | |
4728 | 5.34k | if (!path_is_absolute(path)) |
4729 | 0 | return -EINVAL; |
4730 | 5.34k | |
4731 | 5.34k | r = hashmap_ensure_allocated(&u->requires_mounts_for, &path_hash_ops); |
4732 | 5.34k | if (r < 0) |
4733 | 0 | return r; |
4734 | 5.34k | |
4735 | 5.34k | p = strdup(path); |
4736 | 5.34k | if (!p) |
4737 | 0 | return -ENOMEM; |
4738 | 5.34k | |
4739 | 5.34k | path = path_simplify(p, true); |
4740 | 5.34k | |
4741 | 5.34k | if (!path_is_normalized(path)) |
4742 | 0 | return -EPERM; |
4743 | 5.34k | |
4744 | 5.34k | if (hashmap_contains(u->requires_mounts_for, path)) |
4745 | 3.54k | return 0; |
4746 | 1.80k | |
4747 | 1.80k | di = (UnitDependencyInfo) { |
4748 | 1.80k | .origin_mask = mask |
4749 | 1.80k | }; |
4750 | 1.80k | |
4751 | 1.80k | r = hashmap_put(u->requires_mounts_for, path, di.data); |
4752 | 1.80k | if (r < 0) |
4753 | 0 | return r; |
4754 | 1.80k | p = NULL; |
4755 | 1.80k | |
4756 | 1.80k | char prefix[strlen(path) + 1]; |
4757 | 176k | PATH_FOREACH_PREFIX_MORE(prefix, path) { |
4758 | 176k | Set *x; |
4759 | 176k | |
4760 | 176k | x = hashmap_get(u->manager->units_requiring_mounts_for, prefix); |
4761 | 176k | if (!x) { |
4762 | 161k | _cleanup_free_ char *q = NULL; |
4763 | 161k | |
4764 | 161k | r = hashmap_ensure_allocated(&u->manager->units_requiring_mounts_for, &path_hash_ops); |
4765 | 161k | if (r < 0) |
4766 | 0 | return r; |
4767 | 161k | |
4768 | 161k | q = strdup(prefix); |
4769 | 161k | if (!q) |
4770 | 0 | return -ENOMEM; |
4771 | 161k | |
4772 | 161k | x = set_new(NULL); |
4773 | 161k | if (!x) |
4774 | 0 | return -ENOMEM; |
4775 | 161k | |
4776 | 161k | r = hashmap_put(u->manager->units_requiring_mounts_for, q, x); |
4777 | 161k | if (r < 0) { |
4778 | 0 | set_free(x); |
4779 | 0 | return r; |
4780 | 0 | } |
4781 | 161k | q = NULL; |
4782 | 161k | } |
4783 | 176k | |
4784 | 176k | r = set_put(x, u); |
4785 | 176k | if (r < 0) |
4786 | 0 | return r; |
4787 | 176k | } |
4788 | 1.80k | |
4789 | 1.80k | return 0; |
4790 | 1.80k | } |
4791 | | |
4792 | 0 | int unit_setup_exec_runtime(Unit *u) { |
4793 | 0 | ExecRuntime **rt; |
4794 | 0 | size_t offset; |
4795 | 0 | Unit *other; |
4796 | 0 | Iterator i; |
4797 | 0 | void *v; |
4798 | 0 | int r; |
4799 | 0 |
|
4800 | 0 | offset = UNIT_VTABLE(u)->exec_runtime_offset; |
4801 | 0 | assert(offset > 0); |
4802 | 0 |
|
4803 | 0 | /* Check if there already is an ExecRuntime for this unit? */ |
4804 | 0 | rt = (ExecRuntime**) ((uint8_t*) u + offset); |
4805 | 0 | if (*rt) |
4806 | 0 | return 0; |
4807 | 0 | |
4808 | 0 | /* Try to get it from somebody else */ |
4809 | 0 | HASHMAP_FOREACH_KEY(v, other, u->dependencies[UNIT_JOINS_NAMESPACE_OF], i) { |
4810 | 0 | r = exec_runtime_acquire(u->manager, NULL, other->id, false, rt); |
4811 | 0 | if (r == 1) |
4812 | 0 | return 1; |
4813 | 0 | } |
4814 | 0 |
|
4815 | 0 | return exec_runtime_acquire(u->manager, unit_get_exec_context(u), u->id, true, rt); |
4816 | 0 | } |
4817 | | |
4818 | 0 | int unit_setup_dynamic_creds(Unit *u) { |
4819 | 0 | ExecContext *ec; |
4820 | 0 | DynamicCreds *dcreds; |
4821 | 0 | size_t offset; |
4822 | 0 |
|
4823 | 0 | assert(u); |
4824 | 0 |
|
4825 | 0 | offset = UNIT_VTABLE(u)->dynamic_creds_offset; |
4826 | 0 | assert(offset > 0); |
4827 | 0 | dcreds = (DynamicCreds*) ((uint8_t*) u + offset); |
4828 | 0 |
|
4829 | 0 | ec = unit_get_exec_context(u); |
4830 | 0 | assert(ec); |
4831 | 0 |
|
4832 | 0 | if (!ec->dynamic_user) |
4833 | 0 | return 0; |
4834 | 0 | |
4835 | 0 | return dynamic_creds_acquire(dcreds, u->manager, ec->user, ec->group); |
4836 | 0 | } |
4837 | | |
4838 | 0 | bool unit_type_supported(UnitType t) { |
4839 | 0 | if (_unlikely_(t < 0)) |
4840 | 0 | return false; |
4841 | 0 | if (_unlikely_(t >= _UNIT_TYPE_MAX)) |
4842 | 0 | return false; |
4843 | 0 | |
4844 | 0 | if (!unit_vtable[t]->supported) |
4845 | 0 | return true; |
4846 | 0 | |
4847 | 0 | return unit_vtable[t]->supported(); |
4848 | 0 | } |
4849 | | |
4850 | 0 | void unit_warn_if_dir_nonempty(Unit *u, const char* where) { |
4851 | 0 | int r; |
4852 | 0 |
|
4853 | 0 | assert(u); |
4854 | 0 | assert(where); |
4855 | 0 |
|
4856 | 0 | r = dir_is_empty(where); |
4857 | 0 | if (r > 0 || r == -ENOTDIR) |
4858 | 0 | return; |
4859 | 0 | if (r < 0) { |
4860 | 0 | log_unit_warning_errno(u, r, "Failed to check directory %s: %m", where); |
4861 | 0 | return; |
4862 | 0 | } |
4863 | 0 |
|
4864 | 0 | log_struct(LOG_NOTICE, |
4865 | 0 | "MESSAGE_ID=" SD_MESSAGE_OVERMOUNTING_STR, |
4866 | 0 | LOG_UNIT_ID(u), |
4867 | 0 | LOG_UNIT_INVOCATION_ID(u), |
4868 | 0 | LOG_UNIT_MESSAGE(u, "Directory %s to mount over is not empty, mounting anyway.", where), |
4869 | 0 | "WHERE=%s", where); |
4870 | 0 | } |
4871 | | |
4872 | 0 | int unit_fail_if_noncanonical(Unit *u, const char* where) { |
4873 | 0 | _cleanup_free_ char *canonical_where = NULL; |
4874 | 0 | int r; |
4875 | 0 |
|
4876 | 0 | assert(u); |
4877 | 0 | assert(where); |
4878 | 0 |
|
4879 | 0 | r = chase_symlinks(where, NULL, CHASE_NONEXISTENT, &canonical_where); |
4880 | 0 | if (r < 0) { |
4881 | 0 | log_unit_debug_errno(u, r, "Failed to check %s for symlinks, ignoring: %m", where); |
4882 | 0 | return 0; |
4883 | 0 | } |
4884 | 0 |
|
4885 | 0 | /* We will happily ignore a trailing slash (or any redundant slashes) */ |
4886 | 0 | if (path_equal(where, canonical_where)) |
4887 | 0 | return 0; |
4888 | 0 | |
4889 | 0 | /* No need to mention "." or "..", they would already have been rejected by unit_name_from_path() */ |
4890 | 0 | log_struct(LOG_ERR, |
4891 | 0 | "MESSAGE_ID=" SD_MESSAGE_OVERMOUNTING_STR, |
4892 | 0 | LOG_UNIT_ID(u), |
4893 | 0 | LOG_UNIT_INVOCATION_ID(u), |
4894 | 0 | LOG_UNIT_MESSAGE(u, "Mount path %s is not canonical (contains a symlink).", where), |
4895 | 0 | "WHERE=%s", where); |
4896 | 0 |
|
4897 | 0 | return -ELOOP; |
4898 | 0 | } |
4899 | | |
4900 | 0 | bool unit_is_pristine(Unit *u) { |
4901 | 0 | assert(u); |
4902 | 0 |
|
4903 | 0 | /* Check if the unit already exists or is already around, |
4904 | 0 | * in a number of different ways. Note that to cater for unit |
4905 | 0 | * types such as slice, we are generally fine with units that |
4906 | 0 | * are marked UNIT_LOADED even though nothing was actually |
4907 | 0 | * loaded, as those unit types don't require a file on disk. */ |
4908 | 0 |
|
4909 | 0 | return !(!IN_SET(u->load_state, UNIT_NOT_FOUND, UNIT_LOADED) || |
4910 | 0 | u->fragment_path || |
4911 | 0 | u->source_path || |
4912 | 0 | !strv_isempty(u->dropin_paths) || |
4913 | 0 | u->job || |
4914 | 0 | u->merged_into); |
4915 | 0 | } |
4916 | | |
4917 | 0 | pid_t unit_control_pid(Unit *u) { |
4918 | 0 | assert(u); |
4919 | 0 |
|
4920 | 0 | if (UNIT_VTABLE(u)->control_pid) |
4921 | 0 | return UNIT_VTABLE(u)->control_pid(u); |
4922 | 0 | |
4923 | 0 | return 0; |
4924 | 0 | } |
4925 | | |
4926 | 0 | pid_t unit_main_pid(Unit *u) { |
4927 | 0 | assert(u); |
4928 | 0 |
|
4929 | 0 | if (UNIT_VTABLE(u)->main_pid) |
4930 | 0 | return UNIT_VTABLE(u)->main_pid(u); |
4931 | 0 | |
4932 | 0 | return 0; |
4933 | 0 | } |
4934 | | |
4935 | | static void unit_unref_uid_internal( |
4936 | | Unit *u, |
4937 | | uid_t *ref_uid, |
4938 | | bool destroy_now, |
4939 | 73.6k | void (*_manager_unref_uid)(Manager *m, uid_t uid, bool destroy_now)) { |
4940 | 73.6k | |
4941 | 73.6k | assert(u); |
4942 | 73.6k | assert(ref_uid); |
4943 | 73.6k | assert(_manager_unref_uid); |
4944 | 73.6k | |
4945 | 73.6k | /* Generic implementation of both unit_unref_uid() and unit_unref_gid(), under the assumption that uid_t and |
4946 | 73.6k | * gid_t are actually the same time, with the same validity rules. |
4947 | 73.6k | * |
4948 | 73.6k | * Drops a reference to UID/GID from a unit. */ |
4949 | 73.6k | |
4950 | 73.6k | assert_cc(sizeof(uid_t) == sizeof(gid_t)); |
4951 | 73.6k | assert_cc(UID_INVALID == (uid_t) GID_INVALID); |
4952 | 73.6k | |
4953 | 73.6k | if (!uid_is_valid(*ref_uid)) |
4954 | 73.6k | return; |
4955 | 0 | |
4956 | 0 | _manager_unref_uid(u->manager, *ref_uid, destroy_now); |
4957 | 0 | *ref_uid = UID_INVALID; |
4958 | 0 | } |
4959 | | |
4960 | 36.8k | void unit_unref_uid(Unit *u, bool destroy_now) { |
4961 | 36.8k | unit_unref_uid_internal(u, &u->ref_uid, destroy_now, manager_unref_uid); |
4962 | 36.8k | } |
4963 | | |
4964 | 36.8k | void unit_unref_gid(Unit *u, bool destroy_now) { |
4965 | 36.8k | unit_unref_uid_internal(u, (uid_t*) &u->ref_gid, destroy_now, manager_unref_gid); |
4966 | 36.8k | } |
4967 | | |
4968 | | static int unit_ref_uid_internal( |
4969 | | Unit *u, |
4970 | | uid_t *ref_uid, |
4971 | | uid_t uid, |
4972 | | bool clean_ipc, |
4973 | 0 | int (*_manager_ref_uid)(Manager *m, uid_t uid, bool clean_ipc)) { |
4974 | 0 |
|
4975 | 0 | int r; |
4976 | 0 |
|
4977 | 0 | assert(u); |
4978 | 0 | assert(ref_uid); |
4979 | 0 | assert(uid_is_valid(uid)); |
4980 | 0 | assert(_manager_ref_uid); |
4981 | 0 |
|
4982 | 0 | /* Generic implementation of both unit_ref_uid() and unit_ref_guid(), under the assumption that uid_t and gid_t |
4983 | 0 | * are actually the same type, and have the same validity rules. |
4984 | 0 | * |
4985 | 0 | * Adds a reference on a specific UID/GID to this unit. Each unit referencing the same UID/GID maintains a |
4986 | 0 | * reference so that we can destroy the UID/GID's IPC resources as soon as this is requested and the counter |
4987 | 0 | * drops to zero. */ |
4988 | 0 |
|
4989 | 0 | assert_cc(sizeof(uid_t) == sizeof(gid_t)); |
4990 | 0 | assert_cc(UID_INVALID == (uid_t) GID_INVALID); |
4991 | 0 |
|
4992 | 0 | if (*ref_uid == uid) |
4993 | 0 | return 0; |
4994 | 0 | |
4995 | 0 | if (uid_is_valid(*ref_uid)) /* Already set? */ |
4996 | 0 | return -EBUSY; |
4997 | 0 | |
4998 | 0 | r = _manager_ref_uid(u->manager, uid, clean_ipc); |
4999 | 0 | if (r < 0) |
5000 | 0 | return r; |
5001 | 0 | |
5002 | 0 | *ref_uid = uid; |
5003 | 0 | return 1; |
5004 | 0 | } |
5005 | | |
5006 | 0 | int unit_ref_uid(Unit *u, uid_t uid, bool clean_ipc) { |
5007 | 0 | return unit_ref_uid_internal(u, &u->ref_uid, uid, clean_ipc, manager_ref_uid); |
5008 | 0 | } |
5009 | | |
5010 | 0 | int unit_ref_gid(Unit *u, gid_t gid, bool clean_ipc) { |
5011 | 0 | return unit_ref_uid_internal(u, (uid_t*) &u->ref_gid, (uid_t) gid, clean_ipc, manager_ref_gid); |
5012 | 0 | } |
5013 | | |
5014 | 0 | static int unit_ref_uid_gid_internal(Unit *u, uid_t uid, gid_t gid, bool clean_ipc) { |
5015 | 0 | int r = 0, q = 0; |
5016 | 0 |
|
5017 | 0 | assert(u); |
5018 | 0 |
|
5019 | 0 | /* Reference both a UID and a GID in one go. Either references both, or neither. */ |
5020 | 0 |
|
5021 | 0 | if (uid_is_valid(uid)) { |
5022 | 0 | r = unit_ref_uid(u, uid, clean_ipc); |
5023 | 0 | if (r < 0) |
5024 | 0 | return r; |
5025 | 0 | } |
5026 | 0 | |
5027 | 0 | if (gid_is_valid(gid)) { |
5028 | 0 | q = unit_ref_gid(u, gid, clean_ipc); |
5029 | 0 | if (q < 0) { |
5030 | 0 | if (r > 0) |
5031 | 0 | unit_unref_uid(u, false); |
5032 | 0 |
|
5033 | 0 | return q; |
5034 | 0 | } |
5035 | 0 | } |
5036 | 0 |
|
5037 | 0 | return r > 0 || q > 0; |
5038 | 0 | } |
5039 | | |
5040 | 0 | int unit_ref_uid_gid(Unit *u, uid_t uid, gid_t gid) { |
5041 | 0 | ExecContext *c; |
5042 | 0 | int r; |
5043 | 0 |
|
5044 | 0 | assert(u); |
5045 | 0 |
|
5046 | 0 | c = unit_get_exec_context(u); |
5047 | 0 |
|
5048 | 0 | r = unit_ref_uid_gid_internal(u, uid, gid, c ? c->remove_ipc : false); |
5049 | 0 | if (r < 0) |
5050 | 0 | return log_unit_warning_errno(u, r, "Couldn't add UID/GID reference to unit, proceeding without: %m"); |
5051 | 0 | |
5052 | 0 | return r; |
5053 | 0 | } |
5054 | | |
5055 | 36.8k | void unit_unref_uid_gid(Unit *u, bool destroy_now) { |
5056 | 36.8k | assert(u); |
5057 | 36.8k | |
5058 | 36.8k | unit_unref_uid(u, destroy_now); |
5059 | 36.8k | unit_unref_gid(u, destroy_now); |
5060 | 36.8k | } |
5061 | | |
5062 | 0 | void unit_notify_user_lookup(Unit *u, uid_t uid, gid_t gid) { |
5063 | 0 | int r; |
5064 | 0 |
|
5065 | 0 | assert(u); |
5066 | 0 |
|
5067 | 0 | /* This is invoked whenever one of the forked off processes let's us know the UID/GID its user name/group names |
5068 | 0 | * resolved to. We keep track of which UID/GID is currently assigned in order to be able to destroy its IPC |
5069 | 0 | * objects when no service references the UID/GID anymore. */ |
5070 | 0 |
|
5071 | 0 | r = unit_ref_uid_gid(u, uid, gid); |
5072 | 0 | if (r > 0) |
5073 | 0 | unit_add_to_dbus_queue(u); |
5074 | 0 | } |
5075 | | |
5076 | 0 | int unit_set_invocation_id(Unit *u, sd_id128_t id) { |
5077 | 0 | int r; |
5078 | 0 |
|
5079 | 0 | assert(u); |
5080 | 0 |
|
5081 | 0 | /* Set the invocation ID for this unit. If we cannot, this will not roll back, but reset the whole thing. */ |
5082 | 0 |
|
5083 | 0 | if (sd_id128_equal(u->invocation_id, id)) |
5084 | 0 | return 0; |
5085 | 0 | |
5086 | 0 | if (!sd_id128_is_null(u->invocation_id)) |
5087 | 0 | (void) hashmap_remove_value(u->manager->units_by_invocation_id, &u->invocation_id, u); |
5088 | 0 |
|
5089 | 0 | if (sd_id128_is_null(id)) { |
5090 | 0 | r = 0; |
5091 | 0 | goto reset; |
5092 | 0 | } |
5093 | 0 | |
5094 | 0 | r = hashmap_ensure_allocated(&u->manager->units_by_invocation_id, &id128_hash_ops); |
5095 | 0 | if (r < 0) |
5096 | 0 | goto reset; |
5097 | 0 | |
5098 | 0 | u->invocation_id = id; |
5099 | 0 | sd_id128_to_string(id, u->invocation_id_string); |
5100 | 0 |
|
5101 | 0 | r = hashmap_put(u->manager->units_by_invocation_id, &u->invocation_id, u); |
5102 | 0 | if (r < 0) |
5103 | 0 | goto reset; |
5104 | 0 | |
5105 | 0 | return 0; |
5106 | 0 | |
5107 | 0 | reset: |
5108 | 0 | u->invocation_id = SD_ID128_NULL; |
5109 | 0 | u->invocation_id_string[0] = 0; |
5110 | 0 | return r; |
5111 | 0 | } |
5112 | | |
5113 | 0 | int unit_acquire_invocation_id(Unit *u) { |
5114 | 0 | sd_id128_t id; |
5115 | 0 | int r; |
5116 | 0 |
|
5117 | 0 | assert(u); |
5118 | 0 |
|
5119 | 0 | r = sd_id128_randomize(&id); |
5120 | 0 | if (r < 0) |
5121 | 0 | return log_unit_error_errno(u, r, "Failed to generate invocation ID for unit: %m"); |
5122 | 0 | |
5123 | 0 | r = unit_set_invocation_id(u, id); |
5124 | 0 | if (r < 0) |
5125 | 0 | return log_unit_error_errno(u, r, "Failed to set invocation ID for unit: %m"); |
5126 | 0 | |
5127 | 0 | unit_add_to_dbus_queue(u); |
5128 | 0 | return 0; |
5129 | 0 | } |
5130 | | |
5131 | 0 | int unit_set_exec_params(Unit *u, ExecParameters *p) { |
5132 | 0 | int r; |
5133 | 0 |
|
5134 | 0 | assert(u); |
5135 | 0 | assert(p); |
5136 | 0 |
|
5137 | 0 | /* Copy parameters from manager */ |
5138 | 0 | r = manager_get_effective_environment(u->manager, &p->environment); |
5139 | 0 | if (r < 0) |
5140 | 0 | return r; |
5141 | 0 | |
5142 | 0 | p->confirm_spawn = manager_get_confirm_spawn(u->manager); |
5143 | 0 | p->cgroup_supported = u->manager->cgroup_supported; |
5144 | 0 | p->prefix = u->manager->prefix; |
5145 | 0 | SET_FLAG(p->flags, EXEC_PASS_LOG_UNIT|EXEC_CHOWN_DIRECTORIES, MANAGER_IS_SYSTEM(u->manager)); |
5146 | 0 |
|
5147 | 0 | /* Copy parameters from unit */ |
5148 | 0 | p->cgroup_path = u->cgroup_path; |
5149 | 0 | SET_FLAG(p->flags, EXEC_CGROUP_DELEGATE, unit_cgroup_delegate(u)); |
5150 | 0 |
|
5151 | 0 | return 0; |
5152 | 0 | } |
5153 | | |
5154 | 0 | int unit_fork_helper_process(Unit *u, const char *name, pid_t *ret) { |
5155 | 0 | int r; |
5156 | 0 |
|
5157 | 0 | assert(u); |
5158 | 0 | assert(ret); |
5159 | 0 |
|
5160 | 0 | /* Forks off a helper process and makes sure it is a member of the unit's cgroup. Returns == 0 in the child, |
5161 | 0 | * and > 0 in the parent. The pid parameter is always filled in with the child's PID. */ |
5162 | 0 |
|
5163 | 0 | (void) unit_realize_cgroup(u); |
5164 | 0 |
|
5165 | 0 | r = safe_fork(name, FORK_REOPEN_LOG, ret); |
5166 | 0 | if (r != 0) |
5167 | 0 | return r; |
5168 | 0 | |
5169 | 0 | (void) default_signals(SIGNALS_CRASH_HANDLER, SIGNALS_IGNORE, -1); |
5170 | 0 | (void) ignore_signals(SIGPIPE, -1); |
5171 | 0 |
|
5172 | 0 | (void) prctl(PR_SET_PDEATHSIG, SIGTERM); |
5173 | 0 |
|
5174 | 0 | if (u->cgroup_path) { |
5175 | 0 | r = cg_attach_everywhere(u->manager->cgroup_supported, u->cgroup_path, 0, NULL, NULL); |
5176 | 0 | if (r < 0) { |
5177 | 0 | log_unit_error_errno(u, r, "Failed to join unit cgroup %s: %m", u->cgroup_path); |
5178 | 0 | _exit(EXIT_CGROUP); |
5179 | 0 | } |
5180 | 0 | } |
5181 | 0 |
|
5182 | 0 | return 0; |
5183 | 0 | } |
5184 | | |
5185 | 0 | static void unit_update_dependency_mask(Unit *u, UnitDependency d, Unit *other, UnitDependencyInfo di) { |
5186 | 0 | assert(u); |
5187 | 0 | assert(d >= 0); |
5188 | 0 | assert(d < _UNIT_DEPENDENCY_MAX); |
5189 | 0 | assert(other); |
5190 | 0 |
|
5191 | 0 | if (di.origin_mask == 0 && di.destination_mask == 0) { |
5192 | 0 | /* No bit set anymore, let's drop the whole entry */ |
5193 | 0 | assert_se(hashmap_remove(u->dependencies[d], other)); |
5194 | 0 | log_unit_debug(u, "%s lost dependency %s=%s", u->id, unit_dependency_to_string(d), other->id); |
5195 | 0 | } else |
5196 | 0 | /* Mask was reduced, let's update the entry */ |
5197 | 0 | assert_se(hashmap_update(u->dependencies[d], other, di.data) == 0); |
5198 | 0 | } |
5199 | | |
5200 | 0 | void unit_remove_dependencies(Unit *u, UnitDependencyMask mask) { |
5201 | 0 | UnitDependency d; |
5202 | 0 |
|
5203 | 0 | assert(u); |
5204 | 0 |
|
5205 | 0 | /* Removes all dependencies u has on other units marked for ownership by 'mask'. */ |
5206 | 0 |
|
5207 | 0 | if (mask == 0) |
5208 | 0 | return; |
5209 | 0 | |
5210 | 0 | for (d = 0; d < _UNIT_DEPENDENCY_MAX; d++) { |
5211 | 0 | bool done; |
5212 | 0 |
|
5213 | 0 | do { |
5214 | 0 | UnitDependencyInfo di; |
5215 | 0 | Unit *other; |
5216 | 0 | Iterator i; |
5217 | 0 |
|
5218 | 0 | done = true; |
5219 | 0 |
|
5220 | 0 | HASHMAP_FOREACH_KEY(di.data, other, u->dependencies[d], i) { |
5221 | 0 | UnitDependency q; |
5222 | 0 |
|
5223 | 0 | if ((di.origin_mask & ~mask) == di.origin_mask) |
5224 | 0 | continue; |
5225 | 0 | di.origin_mask &= ~mask; |
5226 | 0 | unit_update_dependency_mask(u, d, other, di); |
5227 | 0 |
|
5228 | 0 | /* We updated the dependency from our unit to the other unit now. But most dependencies |
5229 | 0 | * imply a reverse dependency. Hence, let's delete that one too. For that we go through |
5230 | 0 | * all dependency types on the other unit and delete all those which point to us and |
5231 | 0 | * have the right mask set. */ |
5232 | 0 |
|
5233 | 0 | for (q = 0; q < _UNIT_DEPENDENCY_MAX; q++) { |
5234 | 0 | UnitDependencyInfo dj; |
5235 | 0 |
|
5236 | 0 | dj.data = hashmap_get(other->dependencies[q], u); |
5237 | 0 | if ((dj.destination_mask & ~mask) == dj.destination_mask) |
5238 | 0 | continue; |
5239 | 0 | dj.destination_mask &= ~mask; |
5240 | 0 |
|
5241 | 0 | unit_update_dependency_mask(other, q, u, dj); |
5242 | 0 | } |
5243 | 0 |
|
5244 | 0 | unit_add_to_gc_queue(other); |
5245 | 0 |
|
5246 | 0 | done = false; |
5247 | 0 | break; |
5248 | 0 | } |
5249 | 0 |
|
5250 | 0 | } while (!done); |
5251 | 0 | } |
5252 | 0 | } |
5253 | | |
5254 | 0 | static int unit_export_invocation_id(Unit *u) { |
5255 | 0 | const char *p; |
5256 | 0 | int r; |
5257 | 0 |
|
5258 | 0 | assert(u); |
5259 | 0 |
|
5260 | 0 | if (u->exported_invocation_id) |
5261 | 0 | return 0; |
5262 | 0 | |
5263 | 0 | if (sd_id128_is_null(u->invocation_id)) |
5264 | 0 | return 0; |
5265 | 0 | |
5266 | 0 | p = strjoina("/run/systemd/units/invocation:", u->id); |
5267 | 0 | r = symlink_atomic(u->invocation_id_string, p); |
5268 | 0 | if (r < 0) |
5269 | 0 | return log_unit_debug_errno(u, r, "Failed to create invocation ID symlink %s: %m", p); |
5270 | 0 | |
5271 | 0 | u->exported_invocation_id = true; |
5272 | 0 | return 0; |
5273 | 0 | } |
5274 | | |
5275 | 0 | static int unit_export_log_level_max(Unit *u, const ExecContext *c) { |
5276 | 0 | const char *p; |
5277 | 0 | char buf[2]; |
5278 | 0 | int r; |
5279 | 0 |
|
5280 | 0 | assert(u); |
5281 | 0 | assert(c); |
5282 | 0 |
|
5283 | 0 | if (u->exported_log_level_max) |
5284 | 0 | return 0; |
5285 | 0 | |
5286 | 0 | if (c->log_level_max < 0) |
5287 | 0 | return 0; |
5288 | 0 | |
5289 | 0 | assert(c->log_level_max <= 7); |
5290 | 0 |
|
5291 | 0 | buf[0] = '0' + c->log_level_max; |
5292 | 0 | buf[1] = 0; |
5293 | 0 |
|
5294 | 0 | p = strjoina("/run/systemd/units/log-level-max:", u->id); |
5295 | 0 | r = symlink_atomic(buf, p); |
5296 | 0 | if (r < 0) |
5297 | 0 | return log_unit_debug_errno(u, r, "Failed to create maximum log level symlink %s: %m", p); |
5298 | 0 | |
5299 | 0 | u->exported_log_level_max = true; |
5300 | 0 | return 0; |
5301 | 0 | } |
5302 | | |
5303 | 0 | static int unit_export_log_extra_fields(Unit *u, const ExecContext *c) { |
5304 | 0 | _cleanup_close_ int fd = -1; |
5305 | 0 | struct iovec *iovec; |
5306 | 0 | const char *p; |
5307 | 0 | char *pattern; |
5308 | 0 | le64_t *sizes; |
5309 | 0 | ssize_t n; |
5310 | 0 | size_t i; |
5311 | 0 | int r; |
5312 | 0 |
|
5313 | 0 | if (u->exported_log_extra_fields) |
5314 | 0 | return 0; |
5315 | 0 | |
5316 | 0 | if (c->n_log_extra_fields <= 0) |
5317 | 0 | return 0; |
5318 | 0 | |
5319 | 0 | sizes = newa(le64_t, c->n_log_extra_fields); |
5320 | 0 | iovec = newa(struct iovec, c->n_log_extra_fields * 2); |
5321 | 0 |
|
5322 | 0 | for (i = 0; i < c->n_log_extra_fields; i++) { |
5323 | 0 | sizes[i] = htole64(c->log_extra_fields[i].iov_len); |
5324 | 0 |
|
5325 | 0 | iovec[i*2] = IOVEC_MAKE(sizes + i, sizeof(le64_t)); |
5326 | 0 | iovec[i*2+1] = c->log_extra_fields[i]; |
5327 | 0 | } |
5328 | 0 |
|
5329 | 0 | p = strjoina("/run/systemd/units/log-extra-fields:", u->id); |
5330 | 0 | pattern = strjoina(p, ".XXXXXX"); |
5331 | 0 |
|
5332 | 0 | fd = mkostemp_safe(pattern); |
5333 | 0 | if (fd < 0) |
5334 | 0 | return log_unit_debug_errno(u, fd, "Failed to create extra fields file %s: %m", p); |
5335 | 0 | |
5336 | 0 | n = writev(fd, iovec, c->n_log_extra_fields*2); |
5337 | 0 | if (n < 0) { |
5338 | 0 | r = log_unit_debug_errno(u, errno, "Failed to write extra fields: %m"); |
5339 | 0 | goto fail; |
5340 | 0 | } |
5341 | 0 |
|
5342 | 0 | (void) fchmod(fd, 0644); |
5343 | 0 |
|
5344 | 0 | if (rename(pattern, p) < 0) { |
5345 | 0 | r = log_unit_debug_errno(u, errno, "Failed to rename extra fields file: %m"); |
5346 | 0 | goto fail; |
5347 | 0 | } |
5348 | 0 |
|
5349 | 0 | u->exported_log_extra_fields = true; |
5350 | 0 | return 0; |
5351 | 0 | |
5352 | 0 | fail: |
5353 | 0 | (void) unlink(pattern); |
5354 | 0 | return r; |
5355 | 0 | } |
5356 | | |
5357 | 0 | static int unit_export_log_rate_limit_interval(Unit *u, const ExecContext *c) { |
5358 | 0 | _cleanup_free_ char *buf = NULL; |
5359 | 0 | const char *p; |
5360 | 0 | int r; |
5361 | 0 |
|
5362 | 0 | assert(u); |
5363 | 0 | assert(c); |
5364 | 0 |
|
5365 | 0 | if (u->exported_log_rate_limit_interval) |
5366 | 0 | return 0; |
5367 | 0 | |
5368 | 0 | if (c->log_rate_limit_interval_usec == 0) |
5369 | 0 | return 0; |
5370 | 0 | |
5371 | 0 | p = strjoina("/run/systemd/units/log-rate-limit-interval:", u->id); |
5372 | 0 |
|
5373 | 0 | if (asprintf(&buf, "%" PRIu64, c->log_rate_limit_interval_usec) < 0) |
5374 | 0 | return log_oom(); |
5375 | 0 | |
5376 | 0 | r = symlink_atomic(buf, p); |
5377 | 0 | if (r < 0) |
5378 | 0 | return log_unit_debug_errno(u, r, "Failed to create log rate limit interval symlink %s: %m", p); |
5379 | 0 | |
5380 | 0 | u->exported_log_rate_limit_interval = true; |
5381 | 0 | return 0; |
5382 | 0 | } |
5383 | | |
5384 | 0 | static int unit_export_log_rate_limit_burst(Unit *u, const ExecContext *c) { |
5385 | 0 | _cleanup_free_ char *buf = NULL; |
5386 | 0 | const char *p; |
5387 | 0 | int r; |
5388 | 0 |
|
5389 | 0 | assert(u); |
5390 | 0 | assert(c); |
5391 | 0 |
|
5392 | 0 | if (u->exported_log_rate_limit_burst) |
5393 | 0 | return 0; |
5394 | 0 | |
5395 | 0 | if (c->log_rate_limit_burst == 0) |
5396 | 0 | return 0; |
5397 | 0 | |
5398 | 0 | p = strjoina("/run/systemd/units/log-rate-limit-burst:", u->id); |
5399 | 0 |
|
5400 | 0 | if (asprintf(&buf, "%u", c->log_rate_limit_burst) < 0) |
5401 | 0 | return log_oom(); |
5402 | 0 | |
5403 | 0 | r = symlink_atomic(buf, p); |
5404 | 0 | if (r < 0) |
5405 | 0 | return log_unit_debug_errno(u, r, "Failed to create log rate limit burst symlink %s: %m", p); |
5406 | 0 | |
5407 | 0 | u->exported_log_rate_limit_burst = true; |
5408 | 0 | return 0; |
5409 | 0 | } |
5410 | | |
5411 | 0 | void unit_export_state_files(Unit *u) { |
5412 | 0 | const ExecContext *c; |
5413 | 0 |
|
5414 | 0 | assert(u); |
5415 | 0 |
|
5416 | 0 | if (!u->id) |
5417 | 0 | return; |
5418 | 0 | |
5419 | 0 | if (!MANAGER_IS_SYSTEM(u->manager)) |
5420 | 0 | return; |
5421 | 0 | |
5422 | 0 | if (MANAGER_IS_TEST_RUN(u->manager)) |
5423 | 0 | return; |
5424 | 0 | |
5425 | 0 | /* Exports a couple of unit properties to /run/systemd/units/, so that journald can quickly query this data |
5426 | 0 | * from there. Ideally, journald would use IPC to query this, like everybody else, but that's hard, as long as |
5427 | 0 | * the IPC system itself and PID 1 also log to the journal. |
5428 | 0 | * |
5429 | 0 | * Note that these files really shouldn't be considered API for anyone else, as use a runtime file system as |
5430 | 0 | * IPC replacement is not compatible with today's world of file system namespaces. However, this doesn't really |
5431 | 0 | * apply to communication between the journal and systemd, as we assume that these two daemons live in the same |
5432 | 0 | * namespace at least. |
5433 | 0 | * |
5434 | 0 | * Note that some of the "files" exported here are actually symlinks and not regular files. Symlinks work |
5435 | 0 | * better for storing small bits of data, in particular as we can write them with two system calls, and read |
5436 | 0 | * them with one. */ |
5437 | 0 | |
5438 | 0 | (void) unit_export_invocation_id(u); |
5439 | 0 |
|
5440 | 0 | c = unit_get_exec_context(u); |
5441 | 0 | if (c) { |
5442 | 0 | (void) unit_export_log_level_max(u, c); |
5443 | 0 | (void) unit_export_log_extra_fields(u, c); |
5444 | 0 | (void) unit_export_log_rate_limit_interval(u, c); |
5445 | 0 | (void) unit_export_log_rate_limit_burst(u, c); |
5446 | 0 | } |
5447 | 0 | } |
5448 | | |
5449 | 36.8k | void unit_unlink_state_files(Unit *u) { |
5450 | 36.8k | const char *p; |
5451 | 36.8k | |
5452 | 36.8k | assert(u); |
5453 | 36.8k | |
5454 | 36.8k | if (!u->id) |
5455 | 1.53k | return; |
5456 | 35.2k | |
5457 | 35.2k | if (!MANAGER_IS_SYSTEM(u->manager)) |
5458 | 35.2k | return; |
5459 | 35.2k | |
5460 | 35.2k | /* Undoes the effect of unit_export_state() */ |
5461 | 35.2k | |
5462 | 35.2k | if (u->exported_invocation_id) { |
5463 | 0 | p = strjoina("/run/systemd/units/invocation:", u->id); |
5464 | 0 | (void) unlink(p); |
5465 | 0 |
|
5466 | 0 | u->exported_invocation_id = false; |
5467 | 0 | } |
5468 | 35.2k | |
5469 | 35.2k | if (u->exported_log_level_max) { |
5470 | 0 | p = strjoina("/run/systemd/units/log-level-max:", u->id); |
5471 | 0 | (void) unlink(p); |
5472 | 0 |
|
5473 | 0 | u->exported_log_level_max = false; |
5474 | 0 | } |
5475 | 35.2k | |
5476 | 35.2k | if (u->exported_log_extra_fields) { |
5477 | 0 | p = strjoina("/run/systemd/units/extra-fields:", u->id); |
5478 | 0 | (void) unlink(p); |
5479 | 0 |
|
5480 | 0 | u->exported_log_extra_fields = false; |
5481 | 0 | } |
5482 | 35.2k | |
5483 | 35.2k | if (u->exported_log_rate_limit_interval) { |
5484 | 0 | p = strjoina("/run/systemd/units/log-rate-limit-interval:", u->id); |
5485 | 0 | (void) unlink(p); |
5486 | 0 |
|
5487 | 0 | u->exported_log_rate_limit_interval = false; |
5488 | 0 | } |
5489 | 35.2k | |
5490 | 35.2k | if (u->exported_log_rate_limit_burst) { |
5491 | 0 | p = strjoina("/run/systemd/units/log-rate-limit-burst:", u->id); |
5492 | 0 | (void) unlink(p); |
5493 | 0 |
|
5494 | 0 | u->exported_log_rate_limit_burst = false; |
5495 | 0 | } |
5496 | 35.2k | } |
5497 | | |
5498 | 0 | int unit_prepare_exec(Unit *u) { |
5499 | 0 | int r; |
5500 | 0 |
|
5501 | 0 | assert(u); |
5502 | 0 |
|
5503 | 0 | /* Prepares everything so that we can fork of a process for this unit */ |
5504 | 0 |
|
5505 | 0 | (void) unit_realize_cgroup(u); |
5506 | 0 |
|
5507 | 0 | if (u->reset_accounting) { |
5508 | 0 | (void) unit_reset_accounting(u); |
5509 | 0 | u->reset_accounting = false; |
5510 | 0 | } |
5511 | 0 |
|
5512 | 0 | unit_export_state_files(u); |
5513 | 0 |
|
5514 | 0 | r = unit_setup_exec_runtime(u); |
5515 | 0 | if (r < 0) |
5516 | 0 | return r; |
5517 | 0 | |
5518 | 0 | r = unit_setup_dynamic_creds(u); |
5519 | 0 | if (r < 0) |
5520 | 0 | return r; |
5521 | 0 | |
5522 | 0 | return 0; |
5523 | 0 | } |
5524 | | |
5525 | 0 | static int log_leftover(pid_t pid, int sig, void *userdata) { |
5526 | 0 | _cleanup_free_ char *comm = NULL; |
5527 | 0 |
|
5528 | 0 | (void) get_process_comm(pid, &comm); |
5529 | 0 |
|
5530 | 0 | if (comm && comm[0] == '(') /* Most likely our own helper process (PAM?), ignore */ |
5531 | 0 | return 0; |
5532 | 0 | |
5533 | 0 | log_unit_warning(userdata, |
5534 | 0 | "Found left-over process " PID_FMT " (%s) in control group while starting unit. Ignoring.\n" |
5535 | 0 | "This usually indicates unclean termination of a previous run, or service implementation deficiencies.", |
5536 | 0 | pid, strna(comm)); |
5537 | 0 |
|
5538 | 0 | return 1; |
5539 | 0 | } |
5540 | | |
5541 | 0 | int unit_warn_leftover_processes(Unit *u) { |
5542 | 0 | assert(u); |
5543 | 0 |
|
5544 | 0 | (void) unit_pick_cgroup_path(u); |
5545 | 0 |
|
5546 | 0 | if (!u->cgroup_path) |
5547 | 0 | return 0; |
5548 | 0 | |
5549 | 0 | return cg_kill_recursive(SYSTEMD_CGROUP_CONTROLLER, u->cgroup_path, 0, 0, NULL, log_leftover, u); |
5550 | 0 | } |
5551 | | |
5552 | 0 | bool unit_needs_console(Unit *u) { |
5553 | 0 | ExecContext *ec; |
5554 | 0 | UnitActiveState state; |
5555 | 0 |
|
5556 | 0 | assert(u); |
5557 | 0 |
|
5558 | 0 | state = unit_active_state(u); |
5559 | 0 |
|
5560 | 0 | if (UNIT_IS_INACTIVE_OR_FAILED(state)) |
5561 | 0 | return false; |
5562 | 0 | |
5563 | 0 | if (UNIT_VTABLE(u)->needs_console) |
5564 | 0 | return UNIT_VTABLE(u)->needs_console(u); |
5565 | 0 | |
5566 | 0 | /* If this unit type doesn't implement this call, let's use a generic fallback implementation: */ |
5567 | 0 | ec = unit_get_exec_context(u); |
5568 | 0 | if (!ec) |
5569 | 0 | return false; |
5570 | 0 | |
5571 | 0 | return exec_context_may_touch_console(ec); |
5572 | 0 | } |
5573 | | |
5574 | 0 | const char *unit_label_path(Unit *u) { |
5575 | 0 | const char *p; |
5576 | 0 |
|
5577 | 0 | /* Returns the file system path to use for MAC access decisions, i.e. the file to read the SELinux label off |
5578 | 0 | * when validating access checks. */ |
5579 | 0 |
|
5580 | 0 | p = u->source_path ?: u->fragment_path; |
5581 | 0 | if (!p) |
5582 | 0 | return NULL; |
5583 | 0 | |
5584 | 0 | /* If a unit is masked, then don't read the SELinux label of /dev/null, as that really makes no sense */ |
5585 | 0 | if (path_equal(p, "/dev/null")) |
5586 | 0 | return NULL; |
5587 | 0 | |
5588 | 0 | return p; |
5589 | 0 | } |
5590 | | |
5591 | 0 | int unit_pid_attachable(Unit *u, pid_t pid, sd_bus_error *error) { |
5592 | 0 | int r; |
5593 | 0 |
|
5594 | 0 | assert(u); |
5595 | 0 |
|
5596 | 0 | /* Checks whether the specified PID is generally good for attaching, i.e. a valid PID, not our manager itself, |
5597 | 0 | * and not a kernel thread either */ |
5598 | 0 |
|
5599 | 0 | /* First, a simple range check */ |
5600 | 0 | if (!pid_is_valid(pid)) |
5601 | 0 | return sd_bus_error_setf(error, SD_BUS_ERROR_INVALID_ARGS, "Process identifier " PID_FMT " is not valid.", pid); |
5602 | 0 |
|
5603 | 0 | /* Some extra safety check */ |
5604 | 0 | if (pid == 1 || pid == getpid_cached()) |
5605 | 0 | return sd_bus_error_setf(error, SD_BUS_ERROR_INVALID_ARGS, "Process " PID_FMT " is a manager process, refusing.", pid); |
5606 | 0 |
|
5607 | 0 | /* Don't even begin to bother with kernel threads */ |
5608 | 0 | r = is_kernel_thread(pid); |
5609 | 0 | if (r == -ESRCH) |
5610 | 0 | return sd_bus_error_setf(error, SD_BUS_ERROR_UNIX_PROCESS_ID_UNKNOWN, "Process with ID " PID_FMT " does not exist.", pid); |
5611 | 0 | if (r < 0) |
5612 | 0 | return sd_bus_error_set_errnof(error, r, "Failed to determine whether process " PID_FMT " is a kernel thread: %m", pid); |
5613 | 0 | if (r > 0) |
5614 | 0 | return sd_bus_error_setf(error, SD_BUS_ERROR_INVALID_ARGS, "Process " PID_FMT " is a kernel thread, refusing.", pid); |
5615 | 0 |
|
5616 | 0 | return 0; |
5617 | 0 | } |
5618 | | |
5619 | 0 | void unit_log_success(Unit *u) { |
5620 | 0 | assert(u); |
5621 | 0 |
|
5622 | 0 | log_struct(LOG_INFO, |
5623 | 0 | "MESSAGE_ID=" SD_MESSAGE_UNIT_SUCCESS_STR, |
5624 | 0 | LOG_UNIT_ID(u), |
5625 | 0 | LOG_UNIT_INVOCATION_ID(u), |
5626 | 0 | LOG_UNIT_MESSAGE(u, "Succeeded.")); |
5627 | 0 | } |
5628 | | |
5629 | 0 | void unit_log_failure(Unit *u, const char *result) { |
5630 | 0 | assert(u); |
5631 | 0 | assert(result); |
5632 | 0 |
|
5633 | 0 | log_struct(LOG_WARNING, |
5634 | 0 | "MESSAGE_ID=" SD_MESSAGE_UNIT_FAILURE_RESULT_STR, |
5635 | 0 | LOG_UNIT_ID(u), |
5636 | 0 | LOG_UNIT_INVOCATION_ID(u), |
5637 | 0 | LOG_UNIT_MESSAGE(u, "Failed with result '%s'.", result), |
5638 | 0 | "UNIT_RESULT=%s", result); |
5639 | 0 | } |
5640 | | |
5641 | | void unit_log_process_exit( |
5642 | | Unit *u, |
5643 | | int level, |
5644 | | const char *kind, |
5645 | | const char *command, |
5646 | | int code, |
5647 | 0 | int status) { |
5648 | 0 |
|
5649 | 0 | assert(u); |
5650 | 0 | assert(kind); |
5651 | 0 |
|
5652 | 0 | if (code != CLD_EXITED) |
5653 | 0 | level = LOG_WARNING; |
5654 | 0 |
|
5655 | 0 | log_struct(level, |
5656 | 0 | "MESSAGE_ID=" SD_MESSAGE_UNIT_PROCESS_EXIT_STR, |
5657 | 0 | LOG_UNIT_MESSAGE(u, "%s exited, code=%s, status=%i/%s", |
5658 | 0 | kind, |
5659 | 0 | sigchld_code_to_string(code), status, |
5660 | 0 | strna(code == CLD_EXITED |
5661 | 0 | ? exit_status_to_string(status, EXIT_STATUS_FULL) |
5662 | 0 | : signal_to_string(status))), |
5663 | 0 | "EXIT_CODE=%s", sigchld_code_to_string(code), |
5664 | 0 | "EXIT_STATUS=%i", status, |
5665 | 0 | "COMMAND=%s", strna(command), |
5666 | 0 | LOG_UNIT_ID(u), |
5667 | 0 | LOG_UNIT_INVOCATION_ID(u)); |
5668 | 0 | } |
5669 | | |
5670 | 0 | int unit_exit_status(Unit *u) { |
5671 | 0 | assert(u); |
5672 | 0 |
|
5673 | 0 | /* Returns the exit status to propagate for the most recent cycle of this unit. Returns a value in the range |
5674 | 0 | * 0…255 if there's something to propagate. EOPNOTSUPP if the concept does not apply to this unit type, ENODATA |
5675 | 0 | * if no data is currently known (for example because the unit hasn't deactivated yet) and EBADE if the main |
5676 | 0 | * service process has exited abnormally (signal/coredump). */ |
5677 | 0 |
|
5678 | 0 | if (!UNIT_VTABLE(u)->exit_status) |
5679 | 0 | return -EOPNOTSUPP; |
5680 | 0 | |
5681 | 0 | return UNIT_VTABLE(u)->exit_status(u); |
5682 | 0 | } |
5683 | | |
5684 | 0 | int unit_failure_action_exit_status(Unit *u) { |
5685 | 0 | int r; |
5686 | 0 |
|
5687 | 0 | assert(u); |
5688 | 0 |
|
5689 | 0 | /* Returns the exit status to propagate on failure, or an error if there's nothing to propagate */ |
5690 | 0 |
|
5691 | 0 | if (u->failure_action_exit_status >= 0) |
5692 | 0 | return u->failure_action_exit_status; |
5693 | 0 | |
5694 | 0 | r = unit_exit_status(u); |
5695 | 0 | if (r == -EBADE) /* Exited, but not cleanly (i.e. by signal or such) */ |
5696 | 0 | return 255; |
5697 | 0 | |
5698 | 0 | return r; |
5699 | 0 | } |
5700 | | |
5701 | 0 | int unit_success_action_exit_status(Unit *u) { |
5702 | 0 | int r; |
5703 | 0 |
|
5704 | 0 | assert(u); |
5705 | 0 |
|
5706 | 0 | /* Returns the exit status to propagate on success, or an error if there's nothing to propagate */ |
5707 | 0 |
|
5708 | 0 | if (u->success_action_exit_status >= 0) |
5709 | 0 | return u->success_action_exit_status; |
5710 | 0 | |
5711 | 0 | r = unit_exit_status(u); |
5712 | 0 | if (r == -EBADE) /* Exited, but not cleanly (i.e. by signal or such) */ |
5713 | 0 | return 255; |
5714 | 0 | |
5715 | 0 | return r; |
5716 | 0 | } |
5717 | | |
5718 | 0 | int unit_test_trigger_loaded(Unit *u) { |
5719 | 0 | Unit *trigger; |
5720 | 0 |
|
5721 | 0 | /* Tests whether the unit to trigger is loaded */ |
5722 | 0 |
|
5723 | 0 | trigger = UNIT_TRIGGER(u); |
5724 | 0 | if (!trigger) |
5725 | 0 | return log_unit_error_errno(u, SYNTHETIC_ERRNO(ENOENT), "Refusing to start, unit to trigger not loaded."); |
5726 | 0 | if (trigger->load_state != UNIT_LOADED) |
5727 | 0 | return log_unit_error_errno(u, SYNTHETIC_ERRNO(ENOENT), "Refusing to start, unit %s to trigger not loaded.", u->id); |
5728 | 0 | |
5729 | 0 | return 0; |
5730 | 0 | } |
5731 | | |
5732 | | static const char* const collect_mode_table[_COLLECT_MODE_MAX] = { |
5733 | | [COLLECT_INACTIVE] = "inactive", |
5734 | | [COLLECT_INACTIVE_OR_FAILED] = "inactive-or-failed", |
5735 | | }; |
5736 | | |
5737 | | DEFINE_STRING_TABLE_LOOKUP(collect_mode, CollectMode); |