GitHub Actions are sweet! It's still in beta but it's a great way to automate tasks after things happen like code pushes, comments on Issues, pull requests, etc.

As an author of a GitHub Action, I'm really enamored by the architecture: each "build" is run on a virtual machine in Azure, with each action running inside a Docker container that GitHub executes. As Kyle Daigle mentions on a Changelog podcast::

As a small developer my entire side business could be a Docker container. Not running it, not supporting the payment for it, just a Docker container.

GitHub does this by building and executing your Docker container on the fly. So if you want your Action to finish quickly, you'll want to spend some time optimizing.

These are lessons I've learned building an action to execute a Particle function. I'm not an expert, but they worked for me!

Optimization 1: Use a container deployed to Docker Hub

You can reference an action by pointing to a GitHub repository containing its source:

uses : mattdsteele/particle - action@master

But, the first step for each of your user's workflow will be to checkout the action's source and compile the container. This takes time!

So, if you build the container and publish it to a Docker registry like Docker Hub:

- uses : docker : //mattdsteele/particle - github - action : latest

For my particular action, this improves builds by over 60 seconds!

Of course, you'll need to build and upload your container each time you push a change to your action. Luckily, there's a GitHub Action for that :)

Optimization 2: Trim your Docker images

When referencing a Docker registry, GitHub has to download the image each time. And if you're not careful, Docker containers can be massive, which can slow down how fast your action executes. So the smaller you can make your container, the better.

Don't rely on a big parent image

My first version of the action relied on curl , which even on a large-ish base like Debian wasn't available by default. So here's where I started:

$ docker images REPOSITORY SIZE particle-bash 90 .8MB

Switching from Debian to a smaller Linux base like alpine is a great way to start optimizing:

REPOSITORY SIZE particle-alpine 12 .6MB

Better!

Build a self-contained executable in Go

Going further, we can use Go to build a single, self-contained executable that doesn't rely on any operating system tools! So, I rewrote the action, switched the base to FROM golang:latest , and got this:

REPOSITORY SIZE particle-golang 810MB

Yikes! This is due to Docker images being additive; so the container includes the entire Go build toolchain.

Use a Docker multi-stage build

Let's fix this by setting up a Docker multi-stage build, where we build the Go binary in one stage, and copy it into the empty "scratch" image:

FROM golang : alpine AS build - env RUN apk - - no - cache add build - base git gcc ca - certificates COPY go.mod particle.go /src/ RUN cd /src && go build - o main FROM scratch COPY - - from=build - env /src/main / COPY - - from=build - env /etc/ssl/certs/ca - certificates.crt /etc/ssl/certs/ ENTRYPOINT [ "/main" ]

(I also had to copy over the SSL certificates to the scratch image, since I was making HTTPS calls.)

REPOSITORY SIZE particle-scratch 7 .27MB

Nice!

Optimize the Go binary

We can optimize the image size further by optimizing the Go binary with a couple flags:

RUN cd /src && CGO_ENABLED=0 GOOS=linux GOARCH=amd64 go build - ldflags= "-w -s" - o main

Another improvement:

REPOSITORY SIZE particle-scratch-optim 5 .33MB

In total, these optimizations took my GitHub action from a 6+ minute run down to 30 seconds. Pretty sweet when you're waiting for the action to complete to get candy.