Development Setup

Requirements

You need to have the following tools available in order to effectively contribute to Cilium:

DependencyVersion / Commit IDDownload Command
gitlatestN/A (OS-specific)
clang>= 10.0 (latest recommended)N/A (OS-specific)
llvm>= 10.0 (latest recommended)N/A (OS-specific)
libelf-devellatestN/A (OS-specific)
go1.14.15N/A (OS-specific)
ginkgo>= 1.4.0go get -u github.com/onsi/ginkgo/ginkgo
gomega>= 1.2.0go get -u github.com/onsi/gomega
ineffassign>= 1003c8bgo get -u github.com/gordonklaus/ineffassign
DockerOS-DependentN/A (OS-specific)
Docker-ComposeOS-DependentN/A (OS-specific)
python3-piplatestN/A (OS-specific)

For Unit Testing, you will need to run docker without privileges. You can usually achieve this by adding your current user to the docker group.

Finally, in order to run Cilium locally on VMs, you need:

DependencyVersion / Commit IDDownload Command
Vagrant>= 2.0Vagrant Install Instructions
VirtualBox (if not using libvirt)>= 5.2N/A (OS-specific)

You should start with the Getting Started Guides, which walks you through the set-up, such as installing Vagrant, getting the Cilium sources, and going through some Cilium basics.

Vagrant Setup

While the Getting Started Guides uses a Vagrantfile tuned for the basic walk through, the setup for the Vagrantfile in the root of the Cilium tree depends on a number of environment variables and network setup that are managed via contrib/vagrant/start.sh.

Using the provided Vagrantfile

To bring up a Vagrant VM with Cilium plus dependencies installed, run:

  1. $ contrib/vagrant/start.sh

This will create and run a vagrant VM based on the base box cilium/ubuntu. The box is currently available for the following providers:

  • virtualbox

Options

The following environment variables can be set to customize the VMs brought up by vagrant:

  • NWORKERS=n: Number of child nodes you want to start with the master, default 0.
  • RELOAD=1: Issue a vagrant reload instead of vagrant up, useful to resume halted VMs.
  • NO_PROVISION=1: Avoid provisioning Cilium inside the VM. Supports quick restart without recompiling all of Cilium.
  • NFS=1: Use NFS for vagrant shared directories instead of rsync.
  • K8S=1: Build & install kubernetes on the nodes. k8s1 is the master node, which contains both master components: etcd, kube-controller-manager, kube-scheduler, kube-apiserver, and node components: kubelet, kube-proxy, kubectl and Cilium. When used in combination with NWORKERS=1 a second node is created, where k8s2 will be a kubernetes node, which contains: kubelet, kube-proxy, kubectl and cilium.
  • NETNEXT=1: Run with net-next kernel.
  • IPV4=1: Run Cilium with IPv4 enabled.
  • RUNTIME=x: Sets up the container runtime to be used inside a kubernetes cluster. Valid options are: docker, containerd and crio. If not set, it defaults to docker.
  • VAGRANT_DEFAULT_PROVIDER={virtualbox \| libvirt \| ...}
  • VM_SET_PROXY=https://127.0.0.1:80/ Sets up VM’s https_proxy.
  • INSTALL=1: Restarts the installation of Cilium, Kubernetes, etc. Only useful when the installation was interrupted.
  • MAKECLEAN=1: Execute make clean before building cilium in the VM.

If you want to start the VM with cilium enabled with containerd, with kubernetes installed and plus a worker, run:

  1. $ RUNTIME=containerd K8S=1 NWORKERS=1 contrib/vagrant/start.sh

If you want to get VM status, run:

  1. $ RUNTIME=containerd K8S=1 NWORKERS=1 vagrant status

If you want to connect to the Kubernetes cluster running inside the developer VM via kubectl from your host machine, set KUBECONFIG environment variable to include new kubeconfig file:

  1. $ export KUBECONFIG=$KUBECONFIG:$GOPATH/src/github.com/cilium/cilium/vagrant.kubeconfig

and add 127.0.0.1 k8s1 to your hosts file.

If you have any issue with the provided vagrant box cilium/ubuntu or need a different box format, you may build the box yourself using the packer scripts

Manual Installation

Alternatively you can import the vagrant box cilium/ubuntu directly and manually install Cilium:

  1. $ vagrant init cilium/ubuntu
  2. $ vagrant up
  3. $ vagrant ssh [...]
  4. $ cd go/src/github.com/cilium/cilium/
  5. $ make
  6. $ sudo make install
  7. $ sudo mkdir -p /etc/sysconfig/
  8. $ sudo cp contrib/systemd/cilium.service /etc/systemd/system/
  9. $ sudo cp contrib/systemd/cilium /etc/sysconfig/cilium
  10. $ sudo usermod -a -G cilium vagrant
  11. $ sudo systemctl enable cilium
  12. $ sudo systemctl restart cilium

Notes

Your Cilium tree is mapped to the VM so that you do not need to keep manually copying files between your host and the VM. Folders are by default synced automatically using VirtualBox Shared Folders . You can also use NFS to access your Cilium tree from the VM by setting the environment variable NFS (mentioned above) before running the startup script (export NFS=1). Note that your host firewall must have a variety of ports open. The Vagrantfile will inform you of the configuration of these addresses and ports to enable NFS.

Note

OSX file system is by default case insensitive, which can confuse git. At the writing of this Cilium repo has no file names that would be considered referring to the same file on a case insensitive file system. Regardless, it may be useful to create a disk image with a case sensitive file system for holding your git repos.

Note

VirtualBox for OSX currently (version 5.1.22) always reports host-only networks’ prefix length as 64. Cilium needs this prefix to be 16, and the startup script will check for this. This check always fails when using VirtualBox on OSX, but it is safe to let the startup script to reset the prefix length to 16.

Note

Make sure your host NFS configuration is setup to use tcp:

  1. # cat /etc/nfs.conf
  2. ...
  3. [nfsd]
  4. # grace-time=90
  5. tcp=y
  6. # vers2=n
  7. # vers3=y
  8. ...

If for some reason, running of the provisioning script fails, you should bring the VM down before trying again:

  1. $ vagrant halt

Local Development in Vagrant Box

See Development Setup for information on how to setup the development environment.

When the development VM is provisioned, it builds and installs Cilium. After the initial build and install you can do further building and testing incrementally inside the VM. vagrant ssh takes you to the Cilium source tree directory (/home/vagrant/go/src/github.com/cilium/cilium) by default, and the following commands assume that you are working within that directory.

Build Cilium

Assuming you have synced (rsync) the source tree after you have made changes, or the tree is automatically in sync via NFS or guest additions folder sharing, you can issue a build as follows:

  1. $ make

Install to dev environment

After a successful build and test you can re-install Cilium by:

  1. $ sudo -E make install

Restart Cilium service

To run the newly installed version of Cilium, restart the service:

  1. $ sudo systemctl restart cilium

You can verify the service and cilium-agent status by the following commands, respectively:

  1. $ sudo systemctl status cilium
  2. $ cilium status

Making Changes

  1. Create a topic branch: git checkout -b myBranch master
  2. Make the changes you want
  3. Separate the changes into logical commits.
    1. Describe the changes in the commit messages. Focus on answering the question why the change is required and document anything that might be unexpected.
    2. If any description is required to understand your code changes, then those instructions should be code comments instead of statements in the commit description.
  4. Make sure your changes meet the following criteria:
    1. New code is covered by Unit Testing.
    2. End to end integration / runtime tests have been extended or added. If not required, mention in the commit message what existing test covers the new code.
    3. Follow-up commits are squashed together nicely. Commits should separate logical chunks of code and not represent a chronological list of changes.
  5. Run git diff --check to catch obvious white space violations
  6. Run make to build your changes. This will also run go fmt and error out on any golang formatting errors.
  7. See Unit Testing on how to run unit tests.
  8. See End-To-End Testing Framework for information how to run the end to end integration tests
  9. If you are making documentation changes, you can generate documentation files and serve them locally on http://localhost:9081 by running make render-docs. This make target works both inside and outside the Vagrant VM, assuming that docker is running in the environment.

Add/update a golang dependency

Lets assume we want to add github.com/containernetworking/cni version v0.5.2:

  1. $ go get github.com/containernetworking/cni@v0.5.2
  2. $ go mod tidy
  3. $ go mod vendor
  4. $ git add go.mod go.sum vendor/

For a first run, it can take a while as it will download all dependencies to your local cache but the remaining runs will be faster.

Updating k8s is a special case, for that one needs to do:

  1. $ # get the tag we are updating (for example ``v0.17.3`` corresponds to k8s ``v1.17.3``)
  2. $ # open go.mod and search and replace all ``v0.17.3`` with the version
  3. $ # that we are trying to upgrade with, for example: ``v0.17.4``.
  4. $ # Close the file and run:
  5. $ go mod tidy
  6. $ go mod vendor
  7. $ make generate-k8s-api
  8. $ git add go.mod go.sum vendor/

Optional: Docker and IPv6

Note that these instructions are useful to you if you care about having IPv6 addresses for your Docker containers.

If you’d like IPv6 addresses, you will need to follow these steps:

  1. Edit /etc/docker/daemon.json and set the ipv6 key to true.
  1. {
  2. "ipv6": true
  3. }

If that doesn’t work alone, try assigning a fixed range. Many people have reported trouble with IPv6 and Docker. Source here.

  1. {
  2. "ipv6": true,
  3. "fixed-cidr-v6": "2001:db8:1::/64"
  4. }

And then:

  1. ip -6 route add 2001:db8:1::/64 dev docker0
  2. sysctl net.ipv6.conf.default.forwarding=1
  3. sysctl net.ipv6.conf.all.forwarding=1
  1. Restart the docker daemon to pick up the new configuration.
  2. The new command for creating a network managed by Cilium:
  1. $ docker network create --ipv6 --driver cilium --ipam-driver cilium cilium-net

Now new containers will have an IPv6 address assigned to them.

Debugging

Datapath code

The tool cilium monitor can also be used to retrieve debugging information from the BPF based datapath. Debugging messages are sent if either the cilium-agent itself or the respective endpoint is in debug mode. The debug mode of the agent can be enabled by starting cilium-agent with the option --debug enabled or by running cilium config debug=true for an already running agent. Debugging of an individual endpoint can be enabled by running cilium endpoint config ID debug=true. Running cilium monitor -v will print the normal form of monitor output along with debug messages:

  1. $ cilium endpoint config 731 debug=true
  2. Endpoint 731 configuration updated successfully
  3. $ cilium monitor -v
  4. Press Ctrl-C to quit
  5. level=info msg="Initializing dissection cache..." subsys=monitor
  6. <- endpoint 745 flow 0x6851276 identity 4->0 state new ifindex 0 orig-ip 0.0.0.0: 8e:3c:a3:67:cc:1e -> 16:f9:cd:dc:87:e5 ARP
  7. -> lxc_health: 16:f9:cd:dc:87:e5 -> 8e:3c:a3:67:cc:1e ARP
  8. CPU 00: MARK 0xbbe3d555 FROM 0 DEBUG: Inheriting identity=1 from stack
  9. <- host flow 0xbbe3d555 identity 1->0 state new ifindex 0 orig-ip 0.0.0.0: 10.11.251.76:57896 -> 10.11.166.21:4240 tcp ACK
  10. CPU 00: MARK 0xbbe3d555 FROM 0 DEBUG: Successfully mapped addr=10.11.251.76 to identity=1
  11. CPU 00: MARK 0xbbe3d555 FROM 0 DEBUG: Attempting local delivery for container id 745 from seclabel 1
  12. CPU 00: MARK 0xbbe3d555 FROM 745 DEBUG: Conntrack lookup 1/2: src=10.11.251.76:57896 dst=10.11.166.21:4240
  13. CPU 00: MARK 0xbbe3d555 FROM 745 DEBUG: Conntrack lookup 2/2: nexthdr=6 flags=0
  14. CPU 00: MARK 0xbbe3d555 FROM 745 DEBUG: CT entry found lifetime=21925, revnat=0
  15. CPU 00: MARK 0xbbe3d555 FROM 745 DEBUG: CT verdict: Established, revnat=0
  16. -> endpoint 745 flow 0xbbe3d555 identity 1->4 state established ifindex lxc_health orig-ip 10.11.251.76: 10.11.251.76:57896 -> 10.11.166.21:4240 tcp ACK

Passing -v -v supports deeper detail, for example:

  1. $ cilium endpoint config 3978 debug=true
  2. Endpoint 3978 configuration updated successfully
  3. $ cilium monitor -v -v --hex
  4. Listening for events on 2 CPUs with 64x4096 of shared memory
  5. Press Ctrl-C to quit
  6. ------------------------------------------------------------------------------
  7. CPU 00: MARK 0x1c56d86c FROM 3978 DEBUG: 70 bytes Incoming packet from container ifindex 85
  8. 00000000 33 33 00 00 00 02 ae 45 75 73 11 04 86 dd 60 00 |33.....Eus....`.|
  9. 00000010 00 00 00 10 3a ff fe 80 00 00 00 00 00 00 ac 45 |....:..........E|
  10. 00000020 75 ff fe 73 11 04 ff 02 00 00 00 00 00 00 00 00 |u..s............|
  11. 00000030 00 00 00 00 00 02 85 00 15 b4 00 00 00 00 01 01 |................|
  12. 00000040 ae 45 75 73 11 04 00 00 00 00 00 00 |.Eus........|
  13. CPU 00: MARK 0x1c56d86c FROM 3978 DEBUG: Handling ICMPv6 type=133
  14. ------------------------------------------------------------------------------
  15. CPU 00: MARK 0x1c56d86c FROM 3978 Packet dropped 131 (Invalid destination mac) 70 bytes ifindex=0 284->0
  16. 00000000 33 33 00 00 00 02 ae 45 75 73 11 04 86 dd 60 00 |33.....Eus....`.|
  17. 00000010 00 00 00 10 3a ff fe 80 00 00 00 00 00 00 ac 45 |....:..........E|
  18. 00000020 75 ff fe 73 11 04 ff 02 00 00 00 00 00 00 00 00 |u..s............|
  19. 00000030 00 00 00 00 00 02 85 00 15 b4 00 00 00 00 01 01 |................|
  20. 00000040 00 00 00 00 |....|
  21. ------------------------------------------------------------------------------
  22. CPU 00: MARK 0x7dc2b704 FROM 3978 DEBUG: 86 bytes Incoming packet from container ifindex 85
  23. 00000000 33 33 ff 00 8a d6 ae 45 75 73 11 04 86 dd 60 00 |33.....Eus....`.|
  24. 00000010 00 00 00 20 3a ff fe 80 00 00 00 00 00 00 ac 45 |... :..........E|
  25. 00000020 75 ff fe 73 11 04 ff 02 00 00 00 00 00 00 00 00 |u..s............|
  26. 00000030 00 01 ff 00 8a d6 87 00 20 40 00 00 00 00 fd 02 |........ @......|
  27. 00000040 00 00 00 00 00 00 c0 a8 21 0b 00 00 8a d6 01 01 |........!.......|
  28. 00000050 ae 45 75 73 11 04 00 00 00 00 00 00 |.Eus........|
  29. CPU 00: MARK 0x7dc2b704 FROM 3978 DEBUG: Handling ICMPv6 type=135
  30. CPU 00: MARK 0x7dc2b704 FROM 3978 DEBUG: ICMPv6 neighbour soliciation for address b21a8c0:d68a0000

One of the most common issues when developing datapath code is that the BPF code cannot be loaded into the kernel. This frequently manifests as the endpoints appearing in the “not-ready” state and never switching out of it:

  1. $ cilium endpoint list
  2. ENDPOINT POLICY IDENTITY LABELS (source:key[=value]) IPv6 IPv4 STATUS
  3. ENFORCEMENT
  4. 48896 Disabled 266 container:id.server fd02::c0a8:210b:0:bf00 10.11.13.37 not-ready
  5. 60670 Disabled 267 container:id.client fd02::c0a8:210b:0:ecfe 10.11.167.158 not-ready

Running cilium endpoint get for one of the endpoints will provide a description of known state about it, which includes BPF verification logs.

The files under /var/run/cilium/state provide context about how the BPF datapath is managed and set up. The .h files describe specific configurations used for BPF program compilation. The numbered directories describe endpoint-specific state, including header configuration files and BPF binaries.

Current BPF map state for particular programs is held under /sys/fs/bpf/, and the bpf-map utility can be useful for debugging what is going on inside them, for example:

  1. # ls /sys/fs/bpf/tc/globals/
  2. cilium_calls_15124 cilium_calls_48896 cilium_ct4_global cilium_lb4_rr_seq cilium_lb6_services cilium_policy_25729 cilium_policy_60670 cilium_proxy6
  3. cilium_calls_25729 cilium_calls_60670 cilium_ct6_global cilium_lb4_services cilium_lxc cilium_policy_3978 cilium_policy_reserved_1 cilium_reserved_policy
  4. cilium_calls_3978 cilium_calls_netdev_ns_1 cilium_events cilium_lb6_reverse_nat cilium_policy cilium_policy_4314 cilium_policy_reserved_2 cilium_tunnel_map
  5. cilium_calls_4314 cilium_calls_overlay_2 cilium_lb4_reverse_nat cilium_lb6_rr_seq cilium_policy_15124 cilium_policy_48896 cilium_proxy4
  6. # bpf-map info /sys/fs/bpf/tc/globals/cilium_policy_15124
  7. Type: Hash
  8. Key size: 8
  9. Value size: 24
  10. Max entries: 1024
  11. Flags: 0x0
  12. # bpf-map dump /sys/fs/bpf/tc/globals/cilium_policy_15124
  13. Key:
  14. 00000000 6a 01 00 00 82 23 06 00 |j....#..|
  15. Value:
  16. 00000000 01 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 |................|
  17. 00000010 00 00 00 00 00 00 00 00 |........|