Yet another big problem of the Internet era tackled by Mathematics.

Looping GIFs are a very popular form of art on the Web, with two dedicated forums on Reddit (r/perfectloops and r/cinemagraphs) and countless Tumblr pages.

Finding and extracting well-looping segments from a movie requires much attention and patience, and will likely leave you like this in front of your computer:

To make things easier I wrote a Python script which automates the task. This post explains the math behind the algorithm and provides a few examples of use.

When is a video segment well-looping ?

We will say that a video segment loops well when its first and last video frames are very similar. A video frame can be represented by a sequence of integers whose values indicate the colors of the image’s pixels. For instance, and give the Red, Green, Blue values of the first pixel, , , define the color of the second pixel, etc.

Given two frames , of a same video, we define the difference between these frames as the sum of the differences between their color values:

We will consider that the two frames are similar when is under some arbitrary threshold .

For what follows, it is important to note that defines a distance between the frames, and can be seen as a generalization of the geometrical distance between two points in a plane:

As a consequence has nice mathematical properties which we will use in the next section to speed up computations.

Finding well-looping segments

In this section we want to find the times (start and end) of all the well-looping video segments of duration 3 seconds or less in a given video. A simple way to do this is to compare each frame of the movie with all the frames in the previous three seconds. When we find two similar frames (that is, whose distance in under some pre-defined threshold ), we add their corresponding times to our list.

The problem is that this method requires a huge number of frame comparisons (around ten millions in a standard video) which takes hours. So let us see a few tricks to makes computations faster.

Trick 1: use reduced versions of the frames. HD videos frames can have millions of pixels, so computing the distance between them will require millions of operations. When reduced to small (150-pixel-wide) thumbnails these frames are still detailed enough for our purpose, and their distance can be computed much faster (they also take less place in the RAM).

Trick 2: use triangular inequalities. With this very efficient trick we will be able to deduce whether two frames match, without having to compute their distance. Since defines a mathematical distance between two frames, many results from classical geometry apply, and in particular the following inequalities on the lengths of a triangle:

The first inequality tells us that if A is very close to B which in turn is very close to C, then A is also close to C. In terms of video frames, this becomes:

In practice we will use it as follows: if we already know that a frame is very similar to a frame , and that is very similar to another frame , then we do not need to compute to know that and are also very similar.

The second inequality tells us that if a point A is very near from B, and B is far from C, then A is also far from C. Or in terms of frames:

If is very similar to , and is different from , then we do not need to compute to know that and are also very different.

Now it gets a little more complicated: we will apply these triangular inequalities to get information on the upper and lower bounds of the distances between frames, which will be updated every time we compute a distance between two frames. For instance, after computing the distance , the upper and lower bounds of , denoted and , can be updated as follows:

If after the update we have , we conclude that and are a good match. And if at some point , we know that and don’t match. If we cannot decide whether and match using this technique, we will eventually need to compute , but then knowing will in turn enable us to update the bounds on another distance, , and so on.

As an illustration, suppose that a video has the following frames in this order:

When the algorithm arrives at , it first computes the distance between this frame and and finds that they don’t match. At this point the algorithm has already found thaft is quite similar to and , so it deduces that neither nor match with (and, certainly, neither do the dozen frames before ). In practice, this method avoids computing 80% to 90% of the distances between frames.

Trick 3: use an efficient formula for the distance. When we compute the distance between two frames using the formula from the last section, we need approximately operations: subtractions, products, and additions to obtain the final sum. But the formula for can also be rewritten under this form, known as the law of cosines:

where we used the following notations:

The interesting thing with this expression of is that if we first compute the norm of each frame once, we can obtain the distance between any pair of and simply by computing , which requires only operations and is therefore 50% faster.

Another advantage of computing for each frame is that for two frames and we have

which provides initial values for the upper and lower bounds on the frame distances used in Trick 2:

Final algorithm in pseudo-code. Putting everything together, we obtain the following algorithm:

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 for each frame F1 in the movie: F1 <- downsized( F1 ) previous_frames <- list of frames in the 3 seconds before F1 list_of_matching_couples = [] compute and store |F1| for each frame F2 in previous_frames: compute upper_F1_F2 and lower_F1_F2 using Eq.2 if upper_F1_F2 < T: mark (F1, F2) as matching add (F1, F2) to the list_of_matching_couples if lower_F1_F2 > T: mark (F1, F2) as non-matching for each frame F2 in previous_frames: if couple (F1,F2) isn't already marked matching or non-matching: compute d(F1, F2) for each frame F3 after F2 in previous_frames: update upper_F1_F3 and lower_F1_F3 using Eq.1 if upper_F1_F3 < T: mark (F1, F3) as matching add (F1, F3) to the list_of_matching_couples if lower_F1_F3 > T: mark (F1, F3) as non-matching

Here is the implementation in Python. The computation time may depend on the quality of the video file, but most movies I tried were processed in circa 20 minutes. Impressive, right, Eugene ?

Selecting interesting segments

The algorithm described in the previous section finds all pairs of matching frames, including consecutive frames (which often look very much alike) and frames from still segments (typically, black screens). So we end up with typically a hundred thousand video segments, only a few of which are really interesting, and we must find a way to filter out all the segments we don’t want before extracting GIFs. This filtering operation takes just a few seconds but its success depends greatly on the filtering criteria you use. Here are some examples that work well:

The first and last frames must be separated by at least 0.5 second.

There must be at least one frame in the sequence which doesn’t match at all with the first frame. This criterion enables to eliminate still segments.

The start of the first frame must be at least 0.5 seconds after the start of the last extracted segment. This is to avoid duplicates (segments which start and end almost at the same times).

I try to be not too restrictive (to avoid filtering out good segments by accident) so I generally end up with about 200 GIFs, many of them them only midly interesting (blinking eyes and such). The last step is a manual filtering which looks like this:

Examples of use

I implemented this algorithm as a plugin of my Python video library MoviePy. Here is an example script with much details:

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 from moviepy.editor import VideoFileClip from moviepy.video.tools.cuts import FramesMatches # Open a video file (any format should work) clip = VideoFileClip ( "myvideo.avi" ) # Downsize the clip to a width of 150px to speed up things clip_small = clip . resize ( width = 150 ) # Find all the pairs of matching frames an return their # corresponding start and end times. Takes 15-60 minutes. matches = FramesMatches . from_clip ( clip_small , 5 , 3 ) # (Optional) Save the matches for later use. # matches.save("myvideo_matches.txt") # matches = FramesMatches.load("myvideo_matches.txt") # Filter the scenes: keep only segments with duration >1.5 seconds, # where the first and last frame have a per-pixel distance < 1, # with at least one frame at a distance 2 of the first frame, # and with >0.5 seconds between the starts of the selected segments. selected_scenes = matches . select_scenes ( match_thr = 1 , min_time_span = 1.5 , nomatch_thr = 2 , time_distance = 0.5 ) # The final GIFs will be 450 pixels wide clip_medium = clip . resize ( width = 450 ) # Extract all the selected scenes as GIFs in folder "myfolder" selected_scenes . write_gifs ( clip_medium , "myfolder" )

Here is what be obtain when we try it on Disney’s Snow White:

1 2 3 4 5 6 import moviepy.editor as mp from moviepy.video.tools.cuts import FramesMatches clip = mp . VideoFileClip ( "snowwhite.mp4" ) scenes = FramesMatches . from_clip ( clip . resize ( width = 120 ), 5 , 2 ) selected_scenes = scenes . select_scenes ( 2 , 0.5 , 4 , 0.5 ) selected_scenes . write_gifs ( clip . resize ( width = 270 ), "snow_white" )

Some of these GIFs could be cut better, some are not really interesting (too short), and a few looping segments have been missed. I think the culprits are the parameters in the last filtering step, which could have been tuned better.

As another example, someone recently posted a Youtube video on r/perfectloops and required that it be transformed into a looping GIF. The following script does just that: it downloads the video from Youtube, finds the best times (t1,t2) to cut a looping sequence, and generates a GIF:

1 2 3 4 5 6 7 8 9 10 11 12 13 import moviepy.editor as mpy from moviepy.video.tools.cuts import FramesMatches # Get the video from youtube, save it as "hamac.mp4" mpy . download_webfile ( "NpxD9TZIlv8" , "hamac.mp4" ) clip = mpy . VideoFileClip ( "hamac.mp4" ) . resize ( width = 200 ) matches = FramesMatches . from_clip ( clip , 40 , 3 ) # loose matching # find the best matching pair of frames > 1.5s away best = matches . filter ( lambda x : x . time_span > 1.5 ) . best () # Write the sequence to a GIF (with speed=30% of the original) final = clip . subclip ( best . t1 , best . t2 ) . speedx ( 0.3 ) final . write_gif ( "hamac.gif" , fps = 10 )

With MoviePy you can also post-process your GIFs to add text:

And since you read until there, here is a more advanced trick for you:

Your turn !

The algorithm I presented here is not perfect. It works poorly with low-luminosity clips, and sometimes a slight camera movement or a moving object in the background can prevent a segment from looping. While these segments could be easily corrected by a human, they are more difficult to spot and process with an algorithm.

So my script didn’t completely kill the game, and making looping gifs is still an art. If you have any ideas or remarks on the algorithm, or if you tried it and found some interesting loops in a movie, I’ll be happy to hear about it ! Until then, cheers, and happy GIFing !