This is a cache of https://www.96boards.org/blog/k8s_autoware_auto_multiboard/. It is a snapshot of the page at 2024-12-02T00:55:56.397+0000.
<strong>96Boards</strong>: Autoware everywhere | Multi-board Autoware.Auto 3D Perception Stack using k8s - <strong>96Boards</strong>

96Boards: Autoware everywhere | Multi-board Autoware.Auto 3D Perception Stack using k8s

Servando German Serrano
|

Introduction

Following our previous work to show how to distribute Autoware.Auto’s modules to complete the 3D Perception demo it comes as a natural question to ask about the benefits of using k8s as opposed to plain Docker or the ade-cli tool that is used within the Autoware.Auto official documentation to deploy the software.

Well, if you have been following the blog series we showed how to reproduce the 3D perception demo on a Hikey970 using plain docker, which is very similar to the ade-based steps in the documentation, where the level of manual involvement was quite high with multiple terminals to ssh into the board, launch nodes, etc. While it is true that we could put together a ROS2 launch file to kick-off all the nodes, we would still need to start the containers manually prior to using the launch file.

On the other hand, if we go down the k8s route, we still need to ssh into the board to add it to the k8s cluster but once that’s done everything is managed through the master node which provides an easier way to deploy and manage the different components as services within the vehicle. These benefits come closer into play for “stable” software, hence both ways (k8s vs docker/ade) live together as they are meant to target different use cases, development vs closer to production.

In addition, using the k8s infrastructure helps us keep the services running by means of re-spawning if a services dies or while performing an update to a particular module (which we will explore in the near future). Furthermore, while the above is considered for a single board, when we take scalability into account we can introduce redundancies using a k8s cluster with multiple boards or specify nodes attending to board capabilities or interfaces, for example if we had a LIDAR plugged into a k8s node to perform some preprocessing on the data before feeding into the rest of the system so we could use multiple less capable hardware as opposed to a single higher end approach.

All these can be easily achieved thanks to the Eclipse Cyclone DDS middleware for ROS2 as the masterless discovery capabilities mean that we do not need to manually set up the different IP addresses for all the individual containers within each k8s pod.

In this blog we look at this last point. As a proof of concept instead of running the 3D Perception demo modules just on the PCU we will assume that the LIDAR data comes into a separate board (Qualcomm® Robotics (RB3) Dragonboard-845c Development Platform), where the LIDAR drivers are run, with the 3D pointcloud data in ROS2 topics being fed next into the PCU for rest of modules.

This post is organized as follows:


Setting up the RB3

To set up the RB3 board we just need to follow the steps here and then install Docker as we showed in the past.

We can then follow the default steps to install k8s as we did for the PCU.

NOTE we need to create the following symlink for kubelet to run nicely with its default settings.

$ ln -s /run/resolvconf /run/systemd/resolve

k8s cluster bring up

Now that both boards are ready we just need to create the k8s cluster as usual:

  • Kicking off the master node in the laptop first.

  • Then joining the PCU and RB3 as worker nodes.

  • Upon completion we can verify that the 3 nodes are up and running. RB3 named as linaro-alip and PCU as localhost.

Assigning deployments to specific nodes

As a recap for the 3D Perception demo we created 3 different deployment files and associated Docker images:

  • udpreplay: replays the Velodyne pcap data.
  • sensing: 2-pod deployment for the front and rear Velodyne driver nodes.
  • perception3d: multi-pod deployment for the robot state publisher, point cloud filter transform, ray ground classifier and euclidean cluster nodes.

Since we will run the LIDAR drivers on the RB3 we will assing the udpreplay and sensing deployments to that node and the perception3d to the PCU. To do this we just need to modify the deployment files we had and add the keyword nodeName and the associated node name to each file as shown below.

Once modified we can just create the deployments as usual, which leads to each deployment being created in the right node.

For convenience the modified yaml files can be found here and can be applied directly as:

$ kubectl apply -f https://people.linaro.org/~servando.german.serrano/pcu/k8s/autoware.auto-3dperception-demo/multi_board/udpreplay.yaml
$ kubectl apply -f https://people.linaro.org/~servando.german.serrano/pcu/k8s/autoware.auto-3dperception-demo/multi_board/sensing.yaml
$ kubectl apply -f https://people.linaro.org/~servando.german.serrano/pcu/k8s/autoware.auto-3dperception-demo/multi_board/perception3D-demo.yaml

Visualization

Finally we can check that the system is working by using the laptop for visualization purposes as we did before.


Conclusion

In this post we taken the 3D Perception demo further by combining the PCU with the RB3 in a k8s cluster. We have also seen that it is quite straighforward to deploy different Autoware.Auto modules in multiple boards by using the k8s infrastructure as opposed to manually doing the same. We will continue exploring what we can do with k8s in the near future so keep an eye to this space.

This article is Part 14 in a 15-Part Series.

comments powered by Disqus