Motivation

In CAD software, we often want to express constraints, e.g.

Point a is 10 cm from point b

is 10 cm from point Line g is perpendicular to line h

is perpendicular to line Line t is tangent to arc c

Drag points in the systems below and see what constraints are obeyed:

In general, constraints can be expressed as systems of equations.

Given a system of equations, e.g.

$$ (2x + 3y) \times (x - y) = 2 $$ $$ 3x + y = 5 $$

How would we go about solving it?

One option is to use a computer algebra system like sage, which performs symbolic manipulation to give exact-form solutions:

sage: solve([(2*x + 3*y) * (x - y) == 2, 3*x + y == 5], x,y) [[x == -1/56*sqrt(401) + 95/56, y == 3/56*sqrt(401) - 5/56], [x == 1/56*sqrt(401) + 95/56, y == -3/56*sqrt(401) - 5/56]]

However, this scales poorly with huge systems of many equations and variables. Instead, we'll approach the problem numerically.

Move your mouse around the graph below. The arrow at your cursor points to a nearby solution: see if you can find it.

There are two possible solutions to this system of equations:

\( x = 1.34 \), \( y = 0.983 \)

\( x = 2.05 \), \( y = -1.16 \)

(which are equal to the closed-form solutions found above)

You probably used the strategy of following the arrow "downhill" towards a solution. We will formalize this technique (known as gradient descent) as a numeric solver for arbitrary systems of equations.

Representing equations

Intuition

To start, we need some way of representing sets of variables and systems of equations.

We'd like a way to evaluate \( f(x,y) \) at a given point \( (x,y) \). Since we're using gradient descent, we'd also like to know all of its partial derivatives \( \partial f/\partial x \), \( \partial f/\partial y \), at that point.

We'll use a technique known as automatic differentiation, which involves keeping track of both the result and its derivatives at every step of a calculation.

Consider the equation \( f(x,y) = (x + 5) \times (x + y) \). If we want to evaluate it at \( x = 2 \), \( y = 3 \), we'll end up with the following computation tree:

In this tree, derivatives are computed with sum and product rules.

Now that we know what we want, let's work out an implementation. As in most of my recent recreational coding, we'll be using Haskell.

Implementation

We'll represent variables (with associated values) as a Data.Map .

type Vars a = Map.Map a Double

The set of variables-and-values \( x=1\), \(y=3\) is represented by

Map.fromList [('x', 1), ('y', 3)]

An Equation has a single function eval . When called, eval returns a value and first derivatives with respect to each variable:

newtype Equation a = Equation {eval :: Vars a -> (Double, Vars a)}

Note the function-fu here: an Equation only returns a value when activated with eval and a map of variable IDs to values.

The simplest equation is a single variable. When called, its value is found in the Map and its derivative is set to 1:

var :: Ord a => a -> Equation a var tag = Equation $ \vars -> (Map.findWithDefault 0 tag vars, Map.singleton tag 1)

We can test this out with eval :

λ> eval (var 'x') $ Map.fromList [('x', 3.0)] (3.0, fromList [('x',1.0)])

As expected, we get back the same value and a partial derivative of 1.

The system is polymorphic in tag types. Above, we used a Char , but we could also use Strings (or anything of typeclass Ord ):

λ> eval (var "distance") $ Map.fromList [("distance", 3.0)] (3.0, fromList [("distance",1.0)])

We'd like to do math on equations, e.g.:

λ> let z = var 'x' + var 'y'

How can we implement this?

The function (+) has type Num a => a -> a -> a , which means we'll need to make our Equation class an instance of Num . We'll also make it an instance of Fractional ; between these two typeclasses, we'll be able to do all kinds of arithmetic on our equations.

As discussed above, we'll also be tracking partial derivatives with respect to each variable. Look closely and you'll spot implementations of the sum, product, and quotient rules for differentiation:

instance Ord a => Num (Equation a) where a + b = Equation $ \vars -> let (ra, das) = eval a vars (rb, dbs) = eval b vars in (ra + rb, Map.unionWith (+) das dbs) a * b = Equation $ \vars -> let (ra, das) = eval a vars (rb, dbs) = eval b vars in (ra * rb, Map.unionWith (+) (Map.map (*rb) das) (Map.map (*ra) dbs)) abs a = Equation $ \vars -> let (r, ds) = eval a vars in (abs r, Map.map (* signum r) ds) negate a = Equation $ \vars -> let (r, ds) = eval a vars in (negate r, Map.map negate ds) signum a = Equation $ \vars -> let (r, _) = eval a vars in (signum r, Map.empty) fromInteger a = Equation $ const (fromInteger a, Map.empty) instance Ord a => Fractional (Equation a) where a / b = Equation $ \vars -> let (ra, das) = eval a vars (rb, dbs) = eval b vars in (ra / rb, Map.map (/rb**2) $ Map.unionWith (+) (Map.map (*rb) das) (Map.map (negate . (*ra)) dbs)) fromRational a = Equation $ const (fromRational a, Map.empty)

With these typesclass instances defined, we get arithmetic!

Let's check our math from the example above:

λ> let x = var 'x' λ> let y = var 'y' λ> eval ((x + 5)*(x + y)) $ Map.fromList [('x', 2), ('y', 3)] (35.0,fromList [('x',12.0),('y',7.0)])

Gradient descent

Now, let's focus on the mechanics of gradient descent.

We start at some point \([x_0,y_0,z_0,...]\) in \(n\)-dimensional space (where \(n\) is the number of variables in the system). Our goal is to find \([x,y,z,...]\) such that a particular function \(f(x,y,z,...)\) is zero.

We'll move through \(n\)-dimensional space until a terminating condition is met. The three terminating conditions are as follows:

The cost function is below some threshold

$$ f(x,y,z,...) < \epsilon$$

All partial derivatives are close to zero (indicating a local minima)

$$ \left|\frac{\partial f}{\partial x}\right| < \epsilon \text{ and } \left|\frac{\partial f}{\partial y}\right| < \epsilon \text{ and } \left|\frac{\partial f}{\partial z}\right| < \epsilon \text{ and ...} $$

The solver fails to converge

$$ \left| f(x_n,y_n,z_n,...) - f(x_{n+1},y_{n+1},z_{n+1},...) \right| < \epsilon$$

The direction of the step is given by the systems' gradient, i.e. $$

abla f(x,y,z,...) = \left[\frac{\partial f}{\partial x}\vec{x}, \frac{\partial f}{\partial y}\vec{y}, \frac{\partial f}{\partial z}\vec{z},...\right]$$ where \(\vec{x}\), \(\vec{y}\), \(\vec{z}\) are unit vectors along that dimension.

The logic of our step function is as follows:

Check for convergence using the conditions above

If we've converged, return Nothing

Otherwise, use a backtracking line search to pick the next point

Return it as Just nextPoint

Here's our implementation:

epsilon :: Double epsilon = 1e-12 -- Solves a single step of gradient descent, -- using a backtracking line search. -- -- Returns Nothing if the descent has converged, -- otherwise Just nextPoint step :: Ord a => Equation a -> Vars a -> Maybe (Vars a) step eqn vars = if r < epsilon || all ((< epsilon) . abs) ds || converged then Nothing else Just next where (r, ds) = eval eqn vars (next, converged) = backtrack 1 threshold = 0.5 * (sum $ Map.map (^2) ds) backtrack stepSize = if r - r' >= stepSize * threshold then (vars', abs (r - r') < epsilon) else backtrack (stepSize * 0.5) where vars' = Map.unionWith (-) vars $ Map.map (*stepSize) ds r' = fst (eval eqn vars')

We can repeat this over and over again until we get Nothing back, indicating that the solver has converged. Because step returns a Maybe , we use the reversed bind operator ( =<< ) to chain function calls:

-- Find a local minima from an Equation and a starting point minimize :: Ord a => Equation a -> Vars a -> Vars a minimize eqn vars = fromJust $ last $ takeWhile isJust $ iterate (step eqn =<<) (return vars)

The grid below contains the same function you tried to solve earlier, but shows the solver's path instead of the local gradient:

Notice that the solver converges to different solutions depending on its starting point. This is actually desirable behavior for a CAD constraint system: given multiple valid solutions, it should pick the one that's closest to the existing state of the drawing.

Solving systems of equations

We've been glossing over how the solver actually solves a system of equations.

We've written a tool that minimizes a single equation then used it to satisfy multiple constraints – how does that work?

The answer: sum-of-squares cost functions.

In our example, we're trying to solve $$ (2x + 3y) \times (x - y) = 2 $$ $$ 3x + y = 5 $$

First, we rephrase them as cost functions: $$ \left((2x + 3y) \times (x - y)\right) - 2 $$ $$ (3x + y) - 5 $$

For ease-of-use, we define an infix operator === :

infixl 5 === (===) :: Ord a => Equation a -> Equation a -> Equation a (===) = (-)

This allows us to construct equations that look like equality expressions but are in fact cost functions, e.g. var "x" === 5 is actually var "x" - 5

Now that we can express cost functions, let's combine a set of them by summing their squares. The result is a function that's always \( \geq 0\); it's only equal to zero when all of the constraints are met: $$ \left(\left((2x + 3y) \times (x - y)\right) - 2\right)^2 + \left((3x + y) - 5\right)^2$$

This is the function that we put into our minimizer.

The simplest solver is thus

\eqns vars -> minimize (sum $ map (^2) eqns) vars

However, this solver is not robust against overconstrainted systems:

If you ask it to solve var 'x' === 3 and var 'x' === 5 , it will give you an answer somewhere in between and fail to satisfy both constraints.

We'd like the solver to satisfy as many constraints as possible; in the example above, it should pick either \(x=3\) or \(x=5\).

Here's one implementation strategy:

-- Returns a list of booleans indicating constraint -- satisfaction and a map of resulting variable values. solveSystem :: Ord a => [Equation a] -> Vars a -> ([Bool], Vars a) solveSystem eqns vars = if and satisfied then (satisfied, vars') else -- If not all constraints are satisfied, drop -- an unsatisfied constraint and recurse let index = fromJust $ elemIndex False satisfied (front, back) = splitAt index eqns (satisfied', out) = solveSystem (front ++ (drop 1 back)) vars' (a, b) = splitAt index satisfied' in (a ++ False:b, out) where vars' = minimize (sum $ map (^2) eqns) vars scores = map (\eqn -> (fst $ eval eqn vars')^2) eqns satisfied = map (< sqrt epsilon) scores

If a constraint is not satisfied, we throw it out and try again. The function returns both a solution and a list of which constraints were satisfied:

λ> let x = var 'x' λ> solveSystem [x === 5, x === 3] $ Map.fromList [('x', 2)] ([False,True],fromList [('x',3.0)])

Constraint solving

With all of this explained, we can now understand the earlier interactive examples. Each one defines a set of constraints-as-equations then uses gradient descent to minimize the total sum-of-squares cost function.

We apply an extra constraint to the dragged point, setting it equal to the cursor's position. This constraint is the first to be discarded if infeasible.

Here's one more for the road – thanks for following along!



$$ a_x^2 + a_y^2 = 1 $$ $$ (a_x - b_x)^2 + (a_y - b_y) ^2 = 2 $$ $$ b_y = c_y = 0 $$ $$ c_x - b_x = 1 $$

About the visualizations

The interactive visualizations are running Haskell code that was cross-compiled into Javascript with Haste. I looked at both Haste and GHCJS; the latter didn't have a good way to make Haskell functions available from Javascript, so it wasn't a suitable choice.

I've been impressed by Haste: it worked out-of-the-box and exporting functions is a single call. However, there's a significant performance penalty: all of these simulations are instantaneous on the desktop, but you'll see a bit of lag when moving points around the diagrams.

The graphics are made with d3.js and lots of amaturish (actual) Javascript.

Fun fact: the solver is 30% less code than the graphics

(146 lines vs. 212, as reported by cloc ).

Thanks to Sam Calisch, Neil Gershenfeld, and Tikhon Jelvis for feedback on a draft of this article; props to Chaoya Li for catching a bug in my implementation of division.