LXD is a next generation system container and virtual machine manager.
It offers a unified user experience around full Linux systems running inside containers or virtual machines.
It's image based with pre-made images available for a wide number of Linux distributions
and is built around a very powerful, yet pretty simple, REST API.
To get a better idea of what LXD is and what it does, you can try it online!
Then if you want to run it locally, take a look at our getting started guide.
Release announcements can be found here: https://linuxcontainers.org/lxd/news/
And the release tarballs here: https://linuxcontainers.org/lxd/downloads/
The LXD daemon only works on Linux but the client tool (lxc
) is available on most platforms.
OS | Format | Command |
---|---|---|
Linux | Snap | snap install lxd |
Windows | Chocolatey | choco install lxc |
MacOS | Homebrew | brew install lxc |
More instructions on installing LXD for a wide variety of Linux distributions and operating systems can be found on our website.
We recommend having the latest versions of liblxc (>= 2.0.0 required) available for LXD development. Additionally, LXD requires Golang 1.10 or later to work. On ubuntu, you can get those with:
sudo apt update
sudo apt install acl autoconf dnsmasq-base git golang libacl1-dev libcap-dev liblxc1 liblxc-dev libtool libudev-dev libuv1-dev make pkg-config rsync squashfs-tools tar tcl xz-utils ebtables
Note that when building LXC yourself, ensure to build it with the appropriate security related libraries installed which our testsuite tests. Again, on ubuntu, you can get those with:
sudo apt install libapparmor-dev libseccomp-dev libcap-dev
There are a few storage backends for LXD besides the default "directory" backend. Installing these tools adds a bit to initramfs and may slow down your host boot, but are needed if you'd like to use a particular backend:
sudo apt install lvm2 thin-provisioning-tools
sudo apt install btrfs-tools
To run the testsuite, you'll also need:
sudo apt install curl gettext jq sqlite3 uuid-runtime bzr socat
These instructions for building from source are suitable for individual developers who want to build the latest version of LXD, or build a specific release of LXD which may not be offered by their Linux distribution. Source builds for integration into Linux distributions are not covered here and may be covered in detail in a separate document in the future.
When building from source, it is customary to configure a GOPATH
which contains the to-be-built source code. When
the sources are done building, the lxc
and lxd
binaries will be available at $GOPATH/bin
, and with a little
LD_LIBRARY_PATH
magic (described later), these binaries can be run directly from the built source tree.
The following lines demonstrate how to configure a GOPATH
with the most recent LXD sources from GitHub:
mkdir -p ~/go
export GOPATH=~/go
go get -d -v github.com/lxc/lxd/lxd
cd $GOPATH/src/github.com/lxc/lxd
When the build process starts, the Makefile will use go get
and git clone
to grab all necessary dependencies
needed for building.
To build an official release of LXD, download and extract a release tarball, and then set up GOPATH to point to the
_dist
directory inside it, which is configured to be used as a GOPATH and contains snapshots of all necessary sources. LXD
will then build using these snapshots rather than grabbing 'live' sources using go get
and git clone
. Once the release
tarball is downloaded and extracted, set the GOPATH
as follows:
cd lxd-3.18
export GOPATH=$(pwd)/_dist
Once the GOPATH
is configured, either to build the latest GitHub version or an official release, the following steps
can be used to build LXD.
The actual building is done by two separate invocations of the Makefile: make deps
-- which builds libraries required
by LXD -- and make
, which builds LXD itself. At the end of make deps
, a message will be displayed which will specify environment variables that should be set prior to invoking make
. As new versions of LXD are released, these environment
variable settings may change, so be sure to use the ones displayed at the end of the make deps
process, as the ones
below (shown for example purposes) may not exactly match what your version of LXD requires:
make deps
# Use the export statements printed in the output of 'make deps' -- these are examples:
export CGO_CFLAGS="${CGO_CFLAGS} -I${GOPATH}/deps/sqlite/ -I${GOPATH}/deps/dqlite/include/ -I${GOPATH}/deps/raft/include/ -I${GOPATH}/deps/libco/"
export CGO_LDFLAGS="${CGO_LDFLAGS} -L${GOPATH}/deps/sqlite/.libs/ -L${GOPATH}/deps/dqlite/.libs/ -L${GOPATH}/deps/raft/.libs -L${GOPATH}/deps/libco/"
export LD_LIBRARY_PATH="${GOPATH}/deps/sqlite/.libs/:${GOPATH}/deps/dqlite/.libs/:${GOPATH}/deps/raft/.libs:${GOPATH}/deps/libco/:${LD_LIBRARY_PATH}"
export CGO_LDFLAGS_ALLOW="-Wl,-wrap,pthread_create"
make
Once the build completes, you simply keep the source tree, add the directory referenced by $GOPATH/bin
to
your shell path, and set the LD_LIBRARY_PATH
variable printed by make deps
to your environment. This might look
something like this for a ~/.bashrc
file:
# No need to export GOPATH:
GOPATH=~/go
# But we need to export these:
export PATH="$PATH:$GOPATH/bin"
export LD_LIBRARY_PATH="${GOPATH}/deps/sqlite/.libs/:${GOPATH}/deps/dqlite/.libs/:${GOPATH}/deps/raft/.libs:${GOPATH}/deps/libco/:${LD_LIBRARY_PATH}"
Now, the lxd
and lxc
binaries will be available to you and can be used to set up LXD. The binaries will automatically find and use the dependencies built in $GOPATH/deps
thanks to the LD_LIBRARY_PATH
environment variable.
You'll need sub{u,g}ids for root, so that LXD can create the unprivileged containers:
echo "root:1000000:65536" | sudo tee -a /etc/subuid /etc/subgid
Now you can run the daemon (the --group sudo
bit allows everyone in the sudo
group to talk to LXD; you can create your own group if you want):
sudo -E LD_LIBRARY_PATH=$LD_LIBRARY_PATH $GOPATH/bin/lxd --group sudo
LXD, similar to other container and VM managers provides a UNIX socket for local communication.
WARNING: Anyone with access to that socket can fully control LXD, which includes the ability to attach host devices and filesystems, this should therefore only be given to users who would be trusted with root access to the host.
When listening on the network, the same API is available on a TLS socket (HTTPS), specific access on the remote API can be restricted through Canonical RBAC.
More details are available here.
Now that you have LXD running on your system you can read the getting started guide or go through more examples and configurations in our documentation.
Bug reports can be filed at: https://github.com/lxc/lxd/issues/new
Fixes and new features are greatly appreciated but please read our contributing guidelines first.
A discussion forum is available at: https://discuss.linuxcontainers.org
We use the LXC mailing-lists for developer and user discussions, you can find and subscribe to those at: https://lists.linuxcontainers.org
If you prefer live discussions, some of us also hang out in #lxcontainers on irc.freenode.net.
By default LXD server is not accessible from the networks as it only listens
on a local unix socket. You can make LXD available from the network by specifying
additional addresses to listen to. This is done with the core.https_address
config variable.
To see the current server configuration, run:
lxc config show
To set the address to listen to, find out what addresses are available and use
the config set
command on the server:
ip addr
lxc config set core.https_address 192.168.1.15
By default, LXD has no password for security reasons, so you can't do a remote add this way. In order to set a password, do:
lxc config set core.trust_password SECRET
on the host LXD is running on. This will set the remote password that you can
then use to do lxc remote add
.
You can also access the server without setting a password by copying the client
certificate from .config/lxc/client.crt
to the server and adding it with:
lxc config trust add client.crt
LXD supports btrfs, ceph, directory, lvm and zfs based storage.
First make sure you have the relevant tools for your filesystem of choice installed on the machine (btrfs-progs, lvm2 or zfsutils-linux).
By default, LXD comes with no configured network or storage. You can get a basic configuration done with:
lxd init
lxd init
supports both directory based storage and ZFS.
If you want something else, you'll need to use the lxc storage
command:
lxc storage create default BACKEND [OPTIONS...]
lxc profile device add default root disk path=/ pool=default
BACKEND is one of btrfs
, ceph
, dir
, lvm
or zfs
.
Unless specified otherwise, LXD will setup loop based storage with a sane default size.
For production environments, you should be using block backed storage instead both for performance and reliability reasons.
Live migration requires a tool installed on both hosts called CRIU, which is available in Ubuntu via:
sudo apt install criu
Then, launch your container with the following,
lxc launch ubuntu $somename
sleep 5s # let the container get to an interesting state
lxc move host1:$somename host2:$somename
And with luck you'll have migrated the container :). Migration is still in experimental stages and may not work for all workloads. Please report bugs on lxc-devel, and we can escalate to CRIU lists as necessary.
Yes. The easiest way to do that is using a privileged container to avoid file ownership issues:
1.a) create a container.
lxc launch ubuntu privilegedContainerName -c security.privileged=true
1.b) or, if your container already exists.
lxc config set privilegedContainerName security.privileged true
- then.
lxc config device add privilegedContainerName shareName disk source=/home/$USER path=/home/ubuntu
In order to run Docker inside a LXD container the security.nesting
property of the container should be set to true
.
lxc config set <container> security.nesting true
Note that LXD containers cannot load kernel modules, so depending on your Docker configuration you may need to have the needed extra kernel modules loaded by the host.
You can do so by setting a comma separate list of kernel modules that your container needs with:
lxc config set <container> linux.kernel_modules <modules>
We have also received some reports that creating a /.dockerenv
file in your
container can help Docker ignore some errors it's getting due to running in a
nested environment.
The LXD REST API can be used locally via unauthenticated Unix socket or remotely via SSL encapsulated TCP.
curl --unix-socket /var/lib/lxd/unix.socket \
-H "Content-Type: application/json" \
-X POST \
-d @hello-ubuntu.json \
lxd/1.0/containers
TCP requires some additional configuration and is not enabled by default.
lxc config set core.https_address "[::]:8443"
curl -k -L \
--cert ~/.config/lxc/client.crt \
--key ~/.config/lxc/client.key \
-H "Content-Type: application/json" \
-X POST \
-d @hello-ubuntu.json \
"https://127.0.0.1:8443/1.0/containers"
The hello-ubuntu.json
file referenced above could contain something like:
{
"name":"some-ubuntu",
"ephemeral":true,
"config":{
"limits.cpu":"2"
},
"source": {
"type":"image",
"mode":"pull",
"protocol":"simplestreams",
"server":"https://cloud-images.ubuntu.com/releases",
"alias":"18.04"
}
}