ROS Development with LXD

Ted Kern

on 6 March 2020

Tags: LXC , LXD , robotics , ROS , ROS2

This article was last updated 1 year ago.


Linux containers allow for easy isolation of developer environments. If you’re often working with a bunch of different ROS versions, it’s a lot easier to do your work entirely in containers.

Getting Started

You’ll first need to install LXD using snap.

ubuntu@lxhost:~$ sudo snap install lxd
lxd 3.21 from Canonical✓ installed

Note

Throughout this guide I will be using the hostname to distinguish which machine I’m running commands on.
lxhost represents the bare metal machine you’ll be creating containers in.
ros1-live is the container we’ll be creating later
remote-pc is a different machine on the same LAN as lxhost
Pay attention to the value after the @ in the shell prompts to make sure you run commands on the right machine.

If you haven’t added /snap/bin to your PATH yet, you’ll want to in order to call the programs snap installs.

ubuntu@lxhost:~$ export PATH=/snap/bin:$PATH

You’ll need to initialize LXD before your first run. I’ll be using default settings for this example, but you should look into the settings to make sure of what you’re getting into. The exact printout may vary depending on your system, as well as what version of LXD you’re using

ubuntu@lxhost:~$ lxd init
Would you like to use LXD clustering? (yes/no) [default=no]:
Do you want to configure a new storage pool? (yes/no) [default=yes]:
Name of the new storage pool [default=default]:
Would you like to connect to a MAAS server? (yes/no) [default=no]:
Would you like to create a new local network bridge? (yes/no) [default=yes]:
What should the new bridge be called? [default=lxdbr0]:
What IPv4 address should be used? (CIDR subnet notation, “auto” or “none”) [default=auto]:
What IPv6 address should be used? (CIDR subnet notation, “auto” or “none”) [default=auto]:
Would you like LXD to be available over the network? (yes/no) [default=no]:
Would you like stale cached images to be updated automatically? (yes/no) [default=yes]
Would you like a YAML "lxd init" preseed to be printed? (yes/no) [default=no]:
ubuntu@lxhost:~$

This handled setting a number of configuration options for us, as well as creating a storage pool for any containers we make and a network bridge (named lxdbr0 above) that will connect our containers to the network.

Heads Up

LXD runs as a daemon on your system that is accessible to root and the lxd group on your system. Users with access to LXD can attach host devices and filesystems, presenting a security risk. Only add users you’d trust with root access to lxd.

You can read more about LXD security in its documentation.

The LXC command

LXD provides an lxc program you’ll use to actually manage the containers. You can invoke this with lxd.lxc, but lxc works just fine. Make sure your PATH resolves to the correct lxc.

ubuntu@lxhost:~$ which lxc
/snap/bin/lxc

If you installed lxc via apt before, and your /usr/bin directory comes before /snap/bin in PATH, you’ll run the wrong lxc command. You probably should go ahead and uninstall that lxc.

LXC provides a number of verbs:

ubuntu@lxhost:~$ lxc
alias      copy       file       info       manpage    operation  publish    restart    start
cluster    delete     help       init       monitor    pause      query      restore    stop
config     exec       image      launch     move       profile    remote     shell      storage
console    export     import     list       network    project    rename     snapshot   version

(Output from tab completion)

We’ll go through some of these, but you can always use lxc <verb> --help to get more details (or man lxc.<verb>, but those are currently just autogenerated from the help text).

Creating a container

We’ll start with a ROS Melodic container. I’m currently running Ubuntu 20.04 Focal preview, so this is actually the easiest way in my case to work with an older ROS.

ubuntu@lxhost:~$ lxc launch ubuntu:18.04 ros1-live
Creating ros1-live
Starting ros1-live
...

Just like that, we’ve spun up a container. An Ubuntu 18.04 machine named ros1-live is now available. lxc list will show it.

ubuntu@lxhost:~$ lxc list
+---------------+---------+-----------------------+-----------------+-----------+-----------+
|     NAME      |  STATE  |         IPV4          |      IPV6       |   TYPE    | SNAPSHOTS |
+---------------+---------+-----------------------+-----------------+-----------+-----------+
| ros1-live     | RUNNING | 10.141.221.65 (eth0)  |                 | CONTAINER | 0         |
+---------------+---------+-----------------------+-----------------+-----------+-----------+

To work in it, we’ll use lxc exec to run a login shell as the default ubuntu user:

ubuntu@lxhost:~$ lxc exec ros1-live -- sudo -iu ubuntu bash
ubuntu@ros1-live:~$

SSH Access

Depending on how we like to develop, we may prefer to ssh in (for instance, to use vscode’s remote tools). We’ll need to add our keys to the container to do so.

If our public key is stored at ~/.ssh/id_rsa.pub, we can add it in a number of ways.

Here we’ll simply create an authorized_keys file with our public key in it, and push it

ubuntu@lxhost:~$ cp ~/.ssh/id_rsa.pub /tmp/authorized_keys
ubuntu@lxhost:~$ lxc file push /tmp/authorized_keys ros1-live/home/ubuntu/.ssh/authorized_keys -p

The -p argument creates parents, as per mkdir -p.

We can now ssh into our container from the host and other containers. You can get the ip address using lxc list

ubuntu@lxhost:~$ lxc list
+---------------+---------+-----------------------+-----------------+-----------+-----------+
|     NAME      |  STATE  |         IPV4          |      IPV6       |   TYPE    | SNAPSHOTS |
+---------------+---------+-----------------------+-----------------+-----------+-----------+
| ros1-live     | RUNNING | 10.141.221.65 (eth0)  |                 | CONTAINER | 0         |
+---------------+---------+-----------------------+-----------------+-----------+-----------+
ubuntu@lxhost:~$ ssh [email protected]
ubuntu@ros1-live:~$

You may want to add this address to your .ssh/config file

Host ros1-live
    HostName 10.141.221.65
    User ubuntu
    IdentityFile ~/.ssh/id_rsa
    ForwardAgent Yes
    ForwardX11 Yes

Remote Access

If you’d like to access the container from a remote pc, the default bridged network setup makes things tricky, as the containers are behind a NAT on the host. The easiest way to ssh in is to use the host as a jumphost.

ubuntu@remote-pc:~$ ssh -J <host-address> ros1-live

or in the config file

ubuntu@remote-pc:~$ cat ~/.ssh/config
Host ros1-live
    HostName 10.141.221.65
    User ubuntu
    ProxyJump <host-address>
    IdentityFile ~/.ssh/id_rsa
    ForwardAgent Yes
    ForwardX11 Yes
ubuntu@remote-pc:~$ ssh ros1-live
ubuntu@ros1-live:~$

If your remote PC is running Windows, you’re likely using a version of ssh without the ProxyJump command. I recommend downloading the latest release of OpenSSH and putting it at the front of your Windows PATH variable, so VSCode and other tools can use it.

If you need more direct access, you can add an lxc proxy device to the container to forward a port into it

ubuntu@lxhost:~$ lxc config device add rosl-live proxy22 proxy connect=tcp:127.0.0.1:22 listen=tcp:0.0.0.0:2222
ubuntu@lxhost:~$ lxc config device show ros1-live
proxy22:
  connect: tcp:127.0.0.1:22
  listen: tcp:0.0.0.0:2222
  type: proxy

This will listen on the host port 2222 and forward connections to the container’s port 22. You can learn more about proxy devices here.

If you don’t want the container behind a NAT on the host, you can specify a different bridge configuration or add an IPVLAN or MACVLAN nic network device to the container. You can read more about nic devices here.

Sharing Files Transparently

You can neatly share directories between the host and the container using disk devices. Disk devices can be a bind-mount of an existing file or directory, a regular mount of a block device, or one of several other source types. You can read about disk devices here.

Lets create a directory and share it with the container:

ubuntu@lxhost:~$ mkdir share && cd share
ubuntu@lxhost:~/share$ mkdir ros1-live
ubuntu@lxhost:~/share$ lxc config device add ros1-live share disk source=~/share/ros1-live path=/home/ubuntu/share

In order to truly share access, we’ll want to use the raw.idmap config option for the container to map your UID and GID. Assuming your UID and GID on the host is 1000 (the default for a single user Ubuntu installation), you’ll use the following command to set the option:

ubuntu@lxhost:~$ lxc config set ros1-live raw.idmap "both 1000 1000"

Now any files you make and modify in the host and the container will be indistinguishable, permission-wise.

To verify the directory is shared:

ubuntu@lxhost:~$ ssh ros1-live
ubuntu@ros1-live:~$ ls
share

Set up ROS

For once, setting up ROS will be the easy part. Just follow the instructions at the ROS wiki like normal.

Developing in the Container

Create a workspace directory in the share directory, and use it as normal to develop.

Devices

If you intend to interface with actual hardware, you’ll need to attach devices to your container. I’ve linked to the device configuration option several times before, but you’ll want to look at it in depth for configuring your hardware.

For example, if you’re controlling an OpenManipulatorX via a U2D2, you’ll need to communicate via serial. This is done via a Unix character device at /dev/ttyUSB0 or similar. To add it to the container as a unix-char device, use

ubuntu@lxhost:~$ lxc config device add ros1-live ttyusb0 unix-char source=/dev/ttyUSB0 path=/dev/ttyUSB0

You can also forward the entire usb device using the vendorid and productid of the device as you would for a udev rule. See the entry on usb devices.

GUIs

You may notice GUI programs don’t work, either well or at all, in containers. Simos Xenitellis wrote a wonderful guide on how to fix this and you’ll be able to run Gazebo, RViz, et al in your containers.

Create a new profile for your containers as follows: (Note: this will use vim, if you do not know vim, use the alternative further below):

ubuntu@lxhost:~$ lxc profile create gui
ubuntu@lxhost:~$ lxc profile edit gui

and add the following to the config file that opened in vim

config:
  environment.DISPLAY: :0
  raw.idmap: both 1000 1000
  user.user-data: |
    #cloud-config
    runcmd:
      - 'sed -i "s/; enable-shm = yes/enable-shm = no/g" /etc/pulse/client.conf'
      - 'echo export PULSE_SERVER=unix:/tmp/.pulse-native | tee --append /home/ubuntu/.profile'
    packages:
      - x11-apps
      - mesa-utils
      - pulseaudio
description: GUI LXD profile
devices:
  PASocket:
    path: /tmp/.pulse-native
    source: /run/user/1000/pulse/native
    type: disk
  X0:
    path: /tmp/.X11-unix/X0
    source: /tmp/.X11-unix/X0
    type: disk
  mygpu:
    type: gpu
name: gui
used_by:

Alternatively, if you do not know or like touching vim, save the above to a temporary file, say, /tmp/foo.yaml, and do the following instead:

ubuntu@lxhost:~$ lxc profile create gui
ubuntu@lxhost:~$ lxc profile edit gui < /tmp/foo.yaml

(see documentation on profiles here).

To summarize the above, we tell lxc that every device with a gui profile on it should

  • Do the raw.idmap config from earlier.
  • Set DISPLAY=:0
  • On cloud-init, disable shm in /etc/pulse/client.conf
  • Set the pulse audio server to a socket /tmp/.pulse_native
  • Install x11-apps, mesa-utils, and pulseaudio
  • Mount the pulseaudio socket from the host to to /tmp/.pulse_native
  • Mount the X11 socket
  • Mount your gpu

We apply this profile to the existing container using:

ubuntu@lxhost:~$ lxc profile add ros1-live gui

Which adds the gui profile alongside the already applied default profile to the container.

Restart the container using:

ubuntu@lxhost:~$ lxc restart ros1-live

And after ~30 seconds of downloading and installing the packages, the container should be up and ready to use. You can get GUI apps over ssh with X11 forwarding, or just from the lxc exec method of getting a shell.

This article originally appeared on Ted Kern’s blog.

Internet of Things

From home control to drones, robots and industrial systems, Ubuntu Core and Snaps provide robust security, app stores and reliable updates for all your IoT devices.

Newsletter signup

Get the latest Ubuntu news and updates in your inbox.

By submitting this form, I confirm that I have read and agree to Canonical's Privacy Policy.

Are you building a robot on top of Ubuntu and looking for a partner? Talk to us!

Contact Us

Related posts

Optimise your ROS snap – Part 2

Welcome to Part 2 of the “optimise your ROS snap” blog series. Make sure to check Part 1 before reading this blog post. This second part is going to present...

Optimise your ROS snap – Part 1

Do you want to optimise the performance of your ROS snap? We reduced the size of the installed Gazebo snap by 95%! This is how you can do it for your snap....

Deploying scalable AI and real-time robots at ROSCon 24

Another year, another ROSCon! This year we’re setting off to Odense, Denmark. At Canonical, we are excited to once again sponsor this event for a community...