LXD is a container "hypervisor". This formula provides several states to help manage it and its containers.
This formula will allow you to:
- Initialize LXD with storage, authentication, and network settings.
- Create some default settings for containers (profiles).
- Pull an image from various sources.
- Create a container with an image.
- Bootstrap the given container (run salt-minion in it or whatever you want).
- Start/Stop/Restart/Freeze/Unfreeze/Migrate a container.
- And finally undo all of the above.
Before we forget it, LXD and this formula allows you to migrate unprivileged containers from one host to another!
- There are currently only LXD packages for Ubuntu GNU/Linux so for the daemon you need Ubuntu.
- This has been tested with Saltstack 2017.7.4, we don't know if it works with other versions.
- PyLXD version 2.2.5 from PIP
- Put/symlink the contents of _modules into salt/base/_modules/.
- Put/symlink the contents of _states into salt/base/_states/.
- Put/symlink the directory lxd into salt/base/
gitfs_remotes:
- https://github.com/saltstack-formulas/saltstack-lxd-formula.git
- Master and Minions
salt \* saltutil.sync_modules
salt \* saltutil.sync_states
- Masterless Minion
salt-call --local saltutil.sync_modules
salt-call --local saltutil.sync_states
Does everthing below.
Installs lxd, manages its settings.
To not listen on the network and use the default storage engine:
lxd:
lxd:
run_init: True
To listen on the network:
lxd:
lxd:
run_init: True
init:
trust_password: "PaSsW0rD"
network_address: "[::]"
network_port: "8443"
lxd:
lxd:
run_init: True
init:
trust_password: "PaSsW0rD"
network_address: "[::]"
network_port: "8443"
# Lets say you configured the password wrong on init, or want to change it:
config:
password:
key: core.trust_password
value: "VerySecure!337"
force_password: True # Currently this will be executed every time
# you execute this state.
# Now lets say somewhere else you want to change the ip LXD is listening on
network:
key: core.https_address
value: "[fd57:1:see:bad:c0de::14]:8443"
Installs the lxd client - its a simple package installer for lxd-client (on Debian at least).
Installs pylxd, this requires the pip-formula if you enable "use_pip_formula".
Manages pylxd server connections, this is usefull when you want to create profiles/images/containers on remote LXD instances.
Attention!
Migrations and image copies don't work with provided "local" endpoint, overwrite it if you want to migrate from/to local.
Migrations and image copies don't work with provided "local" endpoint, overwrite it.
lxd:
remotes:
local:
type: lxd
remote_addr : "https://srv02:8443"
cert : "/root/.config/lxc/client.crt"
key : "/root/.config/lxc/client.key"
verify_cert : False
password" : "PaSsW0rD"
This is just here for other states to get its values.
lxd:
remotes:
srv01:
type: lxd
remote_addr : "https://srv01:8443"
cert : "/root/.config/lxc/client.crt"
key : "/root/.config/lxc/client.key"
verify_cert : False
lxd:
remotes:
srv02:
type: lxd
remote_addr : "https://srv02:8443"
cert : "/root/.config/lxc/client.crt"
key" : "/root/.config/lxc/client.key"
verify_cert : False
password" : "PaSsW0rD"
Manages LXD profiles, profiles are something like defaults for a container, you can add multiple profiles to a single container.
It's general a good idea to look how profiles look on the wire:
Also:
salt-call lxd.profile_list --out=json [ --local ]
salt-call lxd.container_list --out=json [ --local ]
gives nice informations about profile config keys and devices.
lxd:
profiles:
local: # local is special, it means local unix socket, no authentication needed.
autostart:
config:
# Enable autostart
boot.autostart: 1
# Delay between containers in seconds.
boot.autostart.delay: 2
# The lesser the later it gets started on autostart.
boot.autostart.priority: 1
lxd:
profiles:
srv01: # Notice the change from "local" to "srv01"
autostart:
config:
# Enable autostart
boot.autostart: 1
# Delay between containers in seconds.
boot.autostart.delay: 2
# The lesser the later it gets started on autostart.
boot.autostart.priority: 1
lxd:
profiles:
local:
add_eth1:
devices:
eth1:
type: "nic"
nictype": "bridged"
parent": "br1"
lxd:
profiles:
local:
shared_mount:
devices:
shared_mount:
type: "disk"
# Source on the host
source: "/home/shared"
# Path in the container
path: "home/shared"
See stgraber's blog
lxd:
profiles:
local:
small:
config:
limits.cpu: 1
limits.memory: 512MB
device:
root:
limits.read: 20Iops
limits.write: 10Iops
If you use the MongoDB _ext_pillar_, you will notice that it doesn't like dots in field names, this is why we added a special case for that:
lxd:
profiles:
local:
autostart:
config:
# Notice the key/value style here
- key: boot.autostart
value: 1
- key: boot.autostart.delay
value: 2
- key: boot.autostart.priority
value: 1
lxd:
profiles:
local:
autostart:
absent: True
Manages LXD images.
lxd:
images:
local:
busybox:
name: busybox # Its alias
source:
type: file
filename: salt://lxd/files/busybox.tar.xz
saltenv: base
On images.linuxcontainers.org you see a list of available images.
And with lxc image list images:
you get a list of aliases.
lxd:
images:
local:
xenial_amd64:
name: xenial/amd64 # Its alias
source:
name: ubuntu/xenial/amd64
remote: images_linuxcontainers_org # See map.jinja for it
aliases: ['x', 'xa64'] # More aliases
public: False
auto_update: True
We also implemented a way to copy images from simplestreams, to do so:
lxd:
images:
local:
trusty_amd64:
source:
name: trusty/amd64
remote: ubuntu # See map.jinja for it
aliases: ['t', 'ta64'] # More aliases
public: False
auto_update: True
Those simplestreams images have cloud-init integrated! Use
$ lxc image alias list ubuntu:
to get a list of available aliases.
lxd:
images:
local:
trusty_amd64:
source:
type: url
url: https://dl.stgraber.org/lxd
aliases: ['busybox-amd64'] # More aliases
public: False
auto_update: True
Manages LXD containers, this includes lxd.images, lxd.profiles, and lxd.remotes.
From the image alias "xenial/amd64"
lxd:
containers:
local:
ubuntu-xenial:
running: True
source: xenial/amd64
We also add a higher start priority and a device eth1
lxd:
containers:
local:
ubuntu-xenial2:
running: True
source: xenial/amd64
profiles:
- default
- autostart
config:
boot.autostart.priority: 1000
devices:
eth1:
type: "nic"
nictype": "bridged"
parent": "br1"
opts:
require:
- lxd_profile: lxd_profile_local_autostart
lxd:
containers:
local:
xenial3:
profiles: [default, autostart]
running: true
source: xenial/amd64
bootstrap_scripts:
- cmd: [ '/bin/sleep', '2' ]
- src: salt://lxd/scripts/bootstrap.sh
dst: /root/bootstrap.sh
cmd: [ '/root/bootstrap.sh', 'xenial3', 'pcdummy.lan', 'salt.pcdummy.lan', 'true' ]
- cmd: [ '/usr/bin/salt-call', 'state.apply' ]
lxd:
containers:
srv01:
ubuntu-xenial:
migrated: True
stop_and_start: True # No live-migration but start/stop.
source: local # Note that we've overwritten "local",
# else this wont work!
lxd:
containers:
srv01:
ubuntu-xenial:
absent: True
stop: True
Please see execution_module doc for it, or better directly the well documented sourcecode of the LXD Module.
René Jochum <[email protected]>
Kees Bos <[email protected]>
Apache Version 2.0