For my latest research, I am looking into visual SLAM (e. g. ORB-SLAM2). Since VSLAM libraries are designated for running efficiently on embedded systems, they are generally programmed in C/C++ and designed with just Linux in mind (even though, for example, in version 2, ROS also aims for compatibility with MacOS).
As a MacBook user, this becomes “interesting”. Of course, Docker makes it easy to run libraries for Linux. However, once the software also comprises graphical UIs, e. g. for visualization, things get a bit more complicated.
Of course, anything that involves GPUs on macOS is a nightmare. Not only does the GPU cause overheating if you do crazy things such as connecting a second display, and, hence, heat-throttling, but also developing against it is also very not fun. The solution for CUDA-based ML is to simply not use the Mac and instead do it on remote Linuxes (see my post on remote docker hosts for ML. Analogously, it might be interesting to compute the SLAM visualizations in real-time on a remote GPU (e. g. using a GPU-accelerated VM at AWS) and simply tether the rendering results to the developer machine using VNC.
An alternative to VNC is to render the hardware-accelerated GUI on the developer machine’s GPU using the X Window System. This is what I’ll show in this post. The X Window System follows a client-server approach where the X Server is connected to the user I/O devices (screen, keyboard, mouse) and an X client can communicate to the X Server through the X protocol, either on the same host or over the network. In the context of this post, the main idea is to run the actual SLAM algorithms with all their dependencies in a container and only run the visualization outside of it. Communication between the code running inside the container and the GUI is done using X in the omnipresent version from 2012, X11.
Apart from latency issues, the main problem of the alternative described in this post, is that the GPU driver software must be correctly installed on macOS and it must be compatible with the code running inside the container. This, of course, defeats the idea of containerizing the code in the first place and defining all dependencies through software. E. g. ORB-SLAM2 and ROS2 RViz2 depend on the OpenGL API for 3D rendering which is is run in Mac land and not in container land. However, Apple does not support newer versions of OpenGL anymore and has instead promoted its own Metal 3D graphics library. But surprisingly, at least for research purposes, libraries often do not require the late versions of OpenGL - e. g. ORB-SLAM2’s visualization module is happy with OpenGL 1.4 from 2002, so it’s still worth a try.
Setup up the Mac
Assumming a macOS Catalina (10.15.4) and brew, install XQuartz using
brew cask install xquartz. This will install a macOS-compatible X Server and a variety of tools and libraries (such as
xeyes, OpenGL 1.4, and the
glxgears test app). After installing XQuartz and opening up the macOS-native Terminal, running
glxgears should automatically launch the X server and spin up windows on the local machine. If it doesn’t, open up XQuartz beforehand manually using the Spotlight search.
xeyes renders without OpenGL-provided GPU acceleration,
glxgears will use GPU acceleration through direct rendering - that is use X server for window management only, but bypass it for computing graphics operations directly on the GPU. Since code in the Docker container won’t have access to the GPU directly, we need to use OpenGL’s feature for indirect rendering so that the X server takes the role of a proxy to the GPU:
defaults write org.macosforge.xquartz.X11 enable_iglx -bool true
In the above,
glxgears take the role of X clients. ORB-SLAM2 is the X client I actually want to run in a container.
Running a container as an X client on a remote machine and the XQuartz X Server on the MacBook
The setup I’ll describe in this section can be structured as follows:
Setting up SSH
In order to make sure, an X client can connect to the X server over the internet, we use SSH’s capability for X11 forwarding. The SSH daemon will inject a
DISPLAY environment variable that indicates a virtual X server display address and
- forward all X requests to the virtual display that are taking place on the remote host
- to the X server running on the developer machine through the SSH tunnel.
Make sure the SSH daemon on the remote host allows X11Forwarding:
ubuntu@mo:~$ cat /etc/ssh/sshd_config | grep -i X11Forwarding X11Forwarding yes
If you needed to adapt this, remember to run
sudo service ssh reload.
If you use docker-machine to spin up your remote servers, just using
docker-machine ssh doesn’t work because we cannot pass the
-X argument to enable X11Forwarding for this session. Instead use:
cd ~/.docker/machine/machines/mo ssh -X -i id_rsa email@example.com
Otherwise, just use the SSH command you normally use to connect to your remote server,
but make sure to use the
On my MacBook, running
ssh -X will automatically launch XQuartz. If it doesn’t launch automatically, simply launch it manually.
Running an X and a GLX client on the bare VM
If you don’t mind interfering with your OS installation, e. g. Ubuntu, (because you’re using a throw-away VM anyway) install the X and GLX test apps:
apt update && apt install -y x11-apps mesa-utils
glxgears in the SSH terminal should already work and display windows on your MacBook.
Running an X and a GLX client on a remote Docker host
Let’s now containerize the
glxgears clients. It turns out, for both of them, we don’t need a custom Docker image but can just use a plain Ubuntu image and install two packages using apt.
SSH into the remote host and create the following
docker-compose.yml file (e. g. using vim).
# docker-compose.yml version: "3.7" services: glx-test: image: ubuntu:16.04 environment: - DISPLAY volumes: - /tmp/.X11-unix:/tmp/.X11-unix - /home/ubuntu/.Xauthority:/root/.Xauthority:rw network_mode: "host"
Then, run the
docker-compose run --entrypoint bash glx-test
In the container, run
apt update && apt install -y x11-apps mesa-utils xeyes
xeyes should display on your MacBook. Close it and run
glxgears. It also should now show up on your MacBook.
You can find the Dockerfile I created for ORB-SLAM2 in my ORB-SLAM2 fork.
Running both the container as an X client and the XQuartz X Server on the MacBook
After finishing the above solution and taking a step back, it dawned on me that it doesn’t make much sense. While I was able to achieve the desired result of running the V-SLAM code and visualizing it on the MacBook display, the heavy computation on the GPU is also running on the Mac hardware, thus creating dependencies on the macOS installation.
Originally, I assumed that I would be able to put all dependencies including OpenGL and Graphics drivers into the container. However, what I ended up with, is a container that only runs the CPU-based code on the remote machine, and even at the cost of rendering over the network. So, even though it doesn’t solve the dependency problem, it would be at least better to simply run the container locally on my developer machine. So let’s have a look at how this works.
After starting XQuartzm the DISPLAY environment variable (at least when starting the Terminal through the XQuartz menu) contains the display address:
echo $DISPLAY /private/tmp/com.apple.launchd.ffq9tVZfPf/org.macosforge.xquartz:0
While we could rely on SSH X11Forwarding in the previous setup in order to route X requests from the X client on the remote machine to the XQuartz server, now we must employ a different technique - namely
socat1 - to route requests from inside the container to the XQuartz server.
socat (for socket cat) allows establishing a mapping between two sockets.
brew install socat socat TCP-LISTEN:6000,reuseaddr,fork UNIX-CLIENT:\"$DISPLAY\"
Then, we can run a container locally on the MacBook using
docker run -it -e DISPLAY=host.docker.internal:0 ubuntu:16.04
Inside the container,
glxgears will be able to spin up windows and run OpenGL computations through the
In retrospect of the last three days, even though I haven’t tried the maybe superior VNC alternative, I guess the single best alternative is switching to a local Linux machine for OpenGL-based visualizations.