💾 Archived View for gmi.runtimeterror.dev › tailscale-serve-docker-compose-sidecar › index.gmi captured on 2024-08-25 at 00:17:21. Gemini links have been rewritten to link to archived content

View Raw

More Information

⬅️ Previous capture (2024-08-18)

🚧 View Differences

-=-=-=-=-=-=-

💻 [runtimeterror $]

2023-12-30 ~ 2024-02-07

Tailscale Serve in a Docker Compose Sidecar

Hi, and welcome back to what has become my Tailscale blog [1].

[1] Tailscale blog

I have a few servers that I use for running multiple container workloads. My approach in the past had been to use Caddy webserver [2] on the host to proxy the various containers. With this setup, each app would have its own DNS record, and Caddy would be configured to route traffic to the appropriate internal port based on that. For instance:

[2] Caddy webserver

cyberchef.runtimeterror.dev {
  reverse_proxy localhost:8000
}
ntfy.runtimeterror.dev, http://ntfy.runtimeterror.dev {
  reverse_proxy localhost:8080
  @httpget {
    protocol http
    method GET
    path_regexp ^/([-_a-z0-9]{0,64}$|docs/|static/)
  }
  redir @httpget https://{host}{uri}
}
uptime.runtimeterror.dev {
  reverse_proxy localhost:3001
}
miniflux.runtimeterror.dev {
  reverse_proxy localhost:8080
}

And I don't really need all of these services to be public. Not because they're particularly sensitive, but I just don't really have a reason to share my personal Miniflux [3] or CyberChef [4] instances with the world at large. Those would be great candidates to proxy with Tailscale Serve [5] so they'd only be available on my tailnet. Of course, with that setup I'd then have to differentiate the services based on external port numbers since they'd all be served with the same hostname. That's not ideal either.

[3] Miniflux

[4] CyberChef

[5] Tailscale Serve

sudo tailscale serve --bg --https 8443 8180 
Available within your tailnet: 
https://tsdemo.tailnet-name.ts.net/
|-- proxy http://127.0.0.1:8000
https://tsdemo.tailnet-name.ts.net:8443/
|-- proxy http://127.0.0.1:8080

It would be really great if I could directly attach each container to my tailnet and then access the apps with addresses like `https://miniflux.tailnet-name.ts.net` or `https://cyber.tailnet-name.ts.net`. Tailscale does have an official Docker image [6], and at first glance it seems like that would solve my needs pretty directly. Unfortunately, it looks like trying to leverage that container image directly would still require me to configure Tailscale Serve interactively..

[6] official Docker image

<-- note -->

Tailscale just published a blog post [7] which shares some details about how to configure Funnel and Serve within the official image. The short version is that the `TS_SERVE_CONFIG` variable should point to a `serve-config.json` file. The name of the file doesn't actually matter, but the contents do - and you can generate a config by running `tailscale serve status -json` on a functioning system... or just copy-pasta'ing this example I just made for the Cyberchef [8] setup I describe later in this post:

[7] just published a blog post

[8] Cyberchef

{
  "TCP": {
    "443": {
      "HTTPS": true
    }
  },
  "Web": {
    "cyber.tailnet-name.ts.net:443": {
      "Handlers": {
        "/": {
          "Proxy": "http://127.0.0.1:8000"
        }
      }
    }
  }//, uncomment to enable funnel
  // "AllowFunnel": {
  //   "cyber.tailnet-name.ts.net:443": true
  // }
}

Replace the ports and protocols and hostnames and such, and you'll be good to go.

A compose config using this setup might look something like this:

services:
  tailscale:
    image: tailscale/tailscale:latest 
    container_name: cyberchef-tailscale
    restart: unless-stopped
    environment:
      TS_AUTHKEY: ${TS_AUTHKEY:?err}
      TS_HOSTNAME: ${TS_HOSTNAME:-ts-docker}
      TS_EXTRA_ARGS: ${TS_EXTRA_ARGS:-}
      TS_STATE_DIR: /var/lib/tailscale/
      TS_SERVE_CONFIG: /config/serve-config.json 
    volumes:
      - ./ts_data:/var/lib/tailscale/
      - ./serve-config.json:/config/serve-config.json 
  cyberchef:
    container_name: cyberchef
    image: mpepping/cyberchef:latest
    restart: unless-stopped
    network_mode: service:tailscale

That's a bit cleaner than the workaround I'd put together, but you're totally welcome to keep on reading if you want to see how it compares.

<-- /note -->

And then I came across Louis-Philippe Asselin's post [9] about how he set up Tailscale in Docker Compose. When he wrote his post, there was even less documentation on how to do this stuff, so he used a modified Tailscale docker image [10] which loads a startup script [11] to handle some of the configuration steps. His repo also includes a helpful docker-compose example [12] of how to connect it together.

[9] Louis-Philippe Asselin's post

[10] modified Tailscale docker image

[11] startup script

[12] helpful docker-compose example

I quickly realized I could modify his startup script to take care of my Tailscale Serve need. So here's how I did it.

Docker Image

My image starts out basically the same as Louis-Philippe's, with just pulling in the official image and then adding the customized script:

FROM tailscale/tailscale:v1.56.1
COPY start.sh /usr/bin/start.sh
RUN chmod +x /usr/bin/start.sh
CMD ["/usr/bin/start.sh"]

My `start.sh` script has a few tweaks for brevity/clarity, and also adds a block for conditionally enabling a basic Tailscale Serve (or Funnel) configuration:

#!/bin/ash
trap 'kill -TERM $PID' TERM INT
echo "Starting Tailscale daemon"
tailscaled --tun=userspace-networking --statedir="${TS_STATE_DIR}" ${TS_TAILSCALED_EXTRA_ARGS} &
PID=$!
until tailscale up --authkey="${TS_AUTHKEY}" --hostname="${TS_HOSTNAME}" ${TS_EXTRA_ARGS}; do
  sleep 0.1
done
tailscale status
if [ -n "${TS_SERVE_PORT}" ]; then 
  if [ -n "${TS_FUNNEL}" ]; then
    if ! tailscale funnel status | grep -q -A1 '(Funnel on)' | grep -q "${TS_SERVE_PORT}"; then
      tailscale funnel --bg "${TS_SERVE_PORT}"
    fi
  else
    if ! tailscale serve status | grep -q "${TS_SERVE_PORT}"; then
      tailscale serve --bg "${TS_SERVE_PORT}"
    fi
  fi
fi
wait ${PID}

This script starts the `tailscaled` daemon in userspace mode, and it tells the daemon to store its state in a user-defined location. It then uses a supplied pre-auth key [13] to bring up the new Tailscale node and set the hostname.

[13] pre-auth key

If both `TS_SERVE_PORT` and `TS_FUNNEL` are set, the script will publicly proxy the designated port with Tailscale Funnel. If only `TS_SERVE_PORT` is set, it will just proxy it internal to the tailnet with Tailscale Serve.

I'm using this git repo [14] to track my work on this, and it automatically builds my tailscale-docker [15] image. So now I can can simply reference `ghcr.io/jbowdre/tailscale-docker` in my Docker configurations.

[14] this git repo

[15] tailscale-docker

On that note...

Compose Configuration

There's also a sample `docker-compose.yml` [16] in the repo to show how to use the image:

[16] sample `docker-compose.yml`

services:
  tailscale:
    image: ghcr.io/jbowdre/tailscale-docker:latest
    restart: unless-stopped
    container_name: tailscale
    environment:
      TS_AUTHKEY: ${TS_AUTHKEY:?err} # from https://login.tailscale.com/admin/settings/authkeys
      TS_HOSTNAME: ${TS_HOSTNAME:-ts-docker} # optional hostname to use for this node
      TS_STATE_DIR: "/var/lib/tailscale/" # store ts state in a local volume
      TS_TAILSCALED_EXTRA_ARGS: ${TS_TAILSCALED_EXTRA_ARGS:-} # optional extra args to pass to tailscaled
      TS_EXTRA_ARGS: ${TS_EXTRA_ARGS:-} # optional extra flags to pass to tailscale up
      TS_SERVE_PORT: ${TS_SERVE_PORT:-} # optional port to proxy with tailscale serve (ex: '80')
      TS_FUNNEL: ${TS_FUNNEL:-} # if set, serve publicly with tailscale funnel
    volumes:
      - ./ts_data:/var/lib/tailscale/   # the mount point should match TS_STATE_DIR
  myservice:
    image: nginxdemos/hello
    restart: unless-stopped
    network_mode: "service:tailscale" # use the tailscale network service's network

You'll note that most of those environment variables aren't actually defined in this YAML. Instead, they'll be inherited from the environment used for spawning the containers. This provides a few benefits. First, it lets the `tailscale` service definition block function as a template to allow copying it into other Compose files without having to modify. Second, it avoids holding sensitive data in the YAML itself. And third, it allows us to set default values for undefined variables (if `TS_HOSTNAME` is empty it will be automatically replaced with `ts-docker`) or throw an error if a required value isn't set (an empty `TS_AUTHKEY` will throw an error and abort).

You can create the required variables by exporting them at the command line (`export TS_HOSTNAME=ts-docker`) - but that runs the risk of having sensitive values like an authkey stored in your shell history. It's not a great habit.

Perhaps a better approach is to set the variables in a `.env` file stored alongside the `docker-compose.yaml` but with stricter permissions. This file can be owned and only readable by root (or the defined Docker user), while the Compose file can be owned by your own user or the `docker` group.

Here's how the `.env` for this setup might look:

TS_AUTHKEY=tskey-auth-somestring-somelongerstring
TS_HOSTNAME=tsdemo
TS_TAILSCALED_EXTRA_ARGS=--verbose=1
TS_EXTRA_ARGS=--ssh
TS_SERVE_PORT=8080
TS_FUNNEL=1
| Variable Name              | Example                                  | Description                                                                                             |
|----------------------------|------------------------------------------|---------------------------------------------------------------------------------------------------------|
| `TS_AUTHKEY`               | `tskey-auth-somestring-somelongerstring` | used for unattended auth of the new node, get one here [17]                                             |
| `TS_HOSTNAME`              | `tsdemo`                                 | optional Tailscale hostname for the new node                                                            |
| `TS_STATE_DIR`             | `/var/lib/tailscale/`                    | required directory for storing Tailscale state, this should be mounted to the container for persistence |
| `TS_TAILSCALED_EXTRA_ARGS` | `--verbose=1`                            | optional additional flags [18] for `tailscaled`                                                         |
| `TS_EXTRA_ARGS`            | `--ssh`                                  | optional additional flags [19] for `tailscale up`                                                       |
| `TS_SERVE_PORT`            | `8080`                                   | optional application port to expose with Tailscale Serve [20]                                           |
| `TS_FUNNEL`                | `1`                                      | if set (to anything), will proxy `TS_SERVE_PORT` **publicly** with Tailscale Funnel [21]                |

[17] here

[18] flags

[19] flags

[20] Tailscale Serve

[21] Tailscale Funnel

A few implementation notes:

[22] Funnel ACL policy

[23] here

[24] pre-auth key

[25] the magic

Usage Examples

To tie this all together, I'm going to quickly run through the steps I took to create and publish two container-based services without having to do any interactive configuration.

CyberChef

I'll start with my CyberChef [26] instance.

[26] CyberChef

*CyberChef is a simple, intuitive web app for carrying out all manner of "cyber" operations within a web browser. These operations include simple encoding like XOR and Base64, more complex encryption like AES, DES and Blowfish, creating binary and hexdumps, compression and decompression of data, calculating hashes and checksums, IPv6 and X.509 parsing, changing character encodings, and much more.*

This will be served publicly with Funnel so that my friends can use this instance if they need it.

I'll need a pre-auth key so that the Tailscale container can authenticate to my Tailnet. I can get that by going to the Tailscale Admin Portal [27] and generating a new auth key. I gave it a description, ticked the option to pre-approve whatever device authenticates with this key (since I have Device Approval [28] enabled on my tailnet). I also used the option to auto-apply the `tag:internal` tag I used for grouping my on-prem systems as well as the `tag:funnel` tag I use for approving Funnel devices in the ACL.

[27] Tailscale Admin Portal

[28] Device Approval

Image: authkey creation

That gives me a new single-use authkey:

Image: new authkey

I'll use that new key as well as the knowledge that CyberChef is served by default on port `8000` to create an appropriate `.env` file:

TS_AUTHKEY=tskey-auth-somestring-somelongerstring
TS_HOSTNAME=cyber
TS_EXTRA_ARGS=--ssh
TS_SERVE_PORT=8000
TS_FUNNEL=true

And I can add the corresponding `docker-compose.yml` to go with it:

services:
  tailscale: 
    image: ghcr.io/jbowdre/tailscale-docker:latest
    restart: unless-stopped
    container_name: cyberchef-tailscale
    environment:
      TS_AUTHKEY: ${TS_AUTHKEY:?err}
      TS_HOSTNAME: ${TS_HOSTNAME:-ts-docker}
      TS_STATE_DIR: "/var/lib/tailscale/"
      TS_TAILSCALED_EXTRA_ARGS: ${TS_TAILSCALED_EXTRA_ARGS:-}
      TS_EXTRA_ARGS: ${TS_EXTRA_ARGS:-}
      TS_SERVE_PORT: ${TS_SERVE_PORT:-}
      TS_FUNNEL: ${TS_FUNNEL:-}
    volumes:
      - ./ts_data:/var/lib/tailscale/ 
  cyberchef:
    container_name: cyberchef
    image: mpepping/cyberchef:latest
    restart: unless-stopped
    network_mode: service:tailscale 

I can just bring it online like so:

docker compose up -d 
[+] Running 3/3
 ✔ Network cyberchef_default      Created
 ✔ Container cyberchef-tailscale  Started
 ✔ Container cyberchef            Started

I can review the logs for the `tailscale` service to confirm that the Funnel configuration was applied:

docker compose logs tailscale 
cyberchef-tailscale  | # Health check:
cyberchef-tailscale  | #     - not connected to home DERP region 12
cyberchef-tailscale  | #     - Some peers are advertising routes but --accept-routes is false
cyberchef-tailscale  | 2023/12/30 17:44:48 serve: creating a new proxy handler for http://127.0.0.1:8000
cyberchef-tailscale  | 2023/12/30 17:44:48 Hostinfo.WireIngress changed to true
cyberchef-tailscale  | Available on the internet: 
cyberchef-tailscale  |
cyberchef-tailscale  | https://cyber.tailnet-name.ts.net/
cyberchef-tailscale  | |-- proxy http://127.0.0.1:8000
cyberchef-tailscale  |
cyberchef-tailscale  | Funnel started and running in the background.
cyberchef-tailscale  | To disable the proxy, run: tailscale funnel --https=443 off

And after ~10 minutes or so (it sometimes takes a bit longer for the DNS and SSL to start working outside the tailnet), I'll be able to hit the instance at `https://cyber.tailnet-name.ts.net` from anywhere on the web.

Image: cyberchef

Miniflux

I've lately been playing quite a bit with my omg.lol address [29] and associated services [30], and that's inspired me to revisit the world [31] of curating RSS feeds instead of relying on algorithms to keep me informed. Through that experience, I recently found Miniflux [32], a "Minimalist and opinionated feed reader". It's written in Go, is fast and lightweight, and works really well as a PWA installed on mobile devices, too.

[29] my omg.lol address

[30] associated services

[31] revisit the world

[32] Miniflux

It will be great for keeping track of my feeds, but I need to expose this service publicly. So I'll serve it up inside my tailnet with Tailscale Serve.

Here's the `.env` that I'll use:

DB_USER=db-username
DB_PASS=db-passw0rd
ADMIN_USER=sysadmin
ADMIN_PASS=hunter2
TS_AUTHKEY=tskey-auth-somestring-somelongerstring
TS_HOSTNAME=miniflux
TS_EXTRA_ARGS=--ssh
TS_SERVE_PORT=8080

Funnel will not be configured for this since `TS_FUNNEL` was not defined.

I adapted the example `docker-compose.yml` [33] from Miniflux to add in my Tailscale bits:

services:
  tailscale: 
    image: ghcr.io/jbowdre/tailscale-docker:latest
    restart: unless-stopped
    container_name: miniflux-tailscale
    environment:
      TS_AUTHKEY: ${TS_AUTHKEY:?err}
      TS_HOSTNAME: ${TS_HOSTNAME:-ts-docker}
      TS_STATE_DIR: "/var/lib/tailscale/"
      TS_TAILSCALED_EXTRA_ARGS: ${TS_TAILSCALED_EXTRA_ARGS:-}
      TS_EXTRA_ARGS: ${TS_EXTRA_ARGS:-}
      TS_SERVE_PORT: ${TS_SERVE_PORT:-}
      TS_FUNNEL: ${TS_FUNNEL:-}
    volumes:
      - ./ts_data:/var/lib/tailscale/ 
  miniflux:
    image: miniflux/miniflux:latest
    restart: unless-stopped
    container_name: miniflux
    depends_on:
      db:
        condition: service_healthy
    environment:
      - DATABASE_URL=postgres://${DB_USER}:${DB_PASS}@db/miniflux?sslmode=disable
      - RUN_MIGRATIONS=1
      - CREATE_ADMIN=1
      - ADMIN_USERNAME=${ADMIN_USER}
      - ADMIN_PASSWORD=${ADMIN_PASS}
    network_mode: "service:tailscale" 
  db:
    image: postgres:15
    restart: unless-stopped
    container_name: miniflux-db
    environment:
      - POSTGRES_USER=${DB_USER}
      - POSTGRES_PASSWORD=${DB_PASS}
    volumes:
      - ./mf_data:/var/lib/postgresql/data
    healthcheck:
      test: ["CMD", "pg_isready", "-U", "${DB_USER}"]
      interval: 10s
      start_period: 30s

[33] example `docker-compose.yml`

I can bring it up with:

docker compose up -d 
[+] Running 4/4
 ✔ Network miniflux_default       Created
 ✔ Container miniflux-db          Started
 ✔ Container miniflux-tailscale   Started
 ✔ Container miniflux             Created

And I can hit it at `https://miniflux.tailnet-name.ts.net` from within my tailnet:

Image: miniflux

Nice, right? Now to just convert all of my other containerized apps that don't really need to be public. Fortunately that shouldn't take too long since I've got this nice, portable, repeatable Docker Compose setup I can use.

Maybe I'll write about something *other* than Tailscale soon. Stay tuned!

---

📧 Reply by email

Related articles

SilverBullet: Self-Hosted Knowledge Management Web App

Taking Taildrive for a Testdrive

Automate Packer Builds with GithHub Actions

---

Home

This page on the big web