A playbook that sets up an internal WireGuard network using innernet (as declaratively as possible)
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
Max Mehl 1806110965
bump inventory
2 months ago
group_vars add manually added peers 2 months ago
innernet-src@4715cd0c87 bump innernet-src to 1.5.3 2 months ago
inventory@cf31dd4f63 bump inventory 2 months ago
roles do not report port as changed if not actually changed 2 months ago
.dockerignore reduce docker context 2 months ago
.drone.yml add seaborg IP to testing 3 months ago
.gitignore add .deb files for innernet 1.5.1 in repo 2 months ago
.gitmodules add correct upstream as `innernet-src` 6 months ago
Dockerfile change the way in which the .deb packages are built 5 months ago
README.md enhance update of listen-port on clients 2 months ago
ansible.cfg adding ansible scaffolding 6 months ago
build-debs.sh fix typo in build script 5 months ago
deploy.yml limit global tasks to innernet server and clients 2 months ago
fsfe-innernet.png change picture background 6 months ago
open_the_vault.sh adding ansible scaffolding 6 months ago
shell.nix added shell.nix 4 months ago
vault_passphrase.gpg adding ansible scaffolding 6 months ago
vault_passphrase.gpg.license adding ansible scaffolding 6 months ago

README.md

Table of Contents

Motivation

There is a need for some of our servers to connect to other IPv6-only hosts. Since this is not always possible without introducing major painpoints elsewhere, we simply create an internal WireGuard network so that the machines in question can communicate securely using IPv4.

An overview

You can learn more about innernet by looking at its source code or reading this informative blog post of its creator.

Preparation

Requirements

  • A somewhat recent version of ansible
  • git

Clone the repo

git clone --recurse-submodules git@git.fsfe.org:fsfe-system-hackers/innernet-playbook.git
cd innernet-playbook

Deployment

In the scope of this playbook and its roles, we have three different categories of computers:

  1. The innernet server, being the central connector of all innernet peers
  2. Automatically managed machines that are innernet peers, mostly VMs
  3. Manually managed peers, for example admins and other humans

Configure server and all clients

Run the whole playbook to configure everything. For the innernet server and automatically managed machines, all will be handled. For the manually managed peers, you will be given an invitation file.

ansible-playbook deploy.yml

Add a new machine

In order to add e.g. a virtual machine to the networks, run these steps:

  1. In the inventory, add the innernet_client group to the host
  2. Run the playbook with ansible-playbook -l newserver.org deploy.yml

This will configure both the necessary parts on the server and the new machine.

Add a new manually managed peer

In order to add a new human or otherwise manually managed innernet peer, run these steps:

  1. In all.yml, add a new entry for manual_peers
  2. Run the playbook with ansible-playbook -l innernet_server deploy.yml
  3. Install innernet and import the invitation file on the new peer's computer (see below). They are in roles/client/files/ then.

Distribute the invitation files

Some invitation files are for humans, so you need to send these files to them securely. We suggest using someting like wormohle.

sudo apt install magic-wormhole
cd roles/client/files
wormhole send <name_of_peer>.toml

Update

Since innernet is new software, it is not yet included in the Debian repositories. Thus, before running the playbook we need to build the innernet and innernet-server binaries.

In order to switch to a newer version of innernet, run the following steps:

  1. Check out the desired tag in the innernet-src submodule
  2. Run the build script: ./build-debs.sh
  3. Run the playbook with ansible-playbook -t update deploy.yml

Associations

The different CIDRs can have associations, e.g. so that admins can access machines, although they are not in the same subnet.

These have to be configure by an admin!

Currently, the admins CIDR is associated with all other CIDRs (i.e. humans > others and machines).

Ansible tags

Some tags allow you to specify just certain operations. Here are the currently available ones:

  • cidr: configure CIDRs
  • update: update the innernet binaries
  • listen_port: edit/set the listen port between server and clients
  • uninstall: delete innernet configuration and packages from systems