Time to go back to the roots! Me and a few colleagues have been travelling across the old continent to attend the renowned Meeting C++ conference in Berlin.

Fully backed by our employer, King, we were able to enjoy a yearly dose of C++ talks at an excellent venue. I was thrilled at the idea of retrying the "German C++ experience" and my expectation were once again met! C++20 being right at the corner, I also had in mind to prioritise talks on this topic as much as possible.

As the tradition goes, I will relate a bit my experience of the event and dress-up a list of the talks I would highly suggest to search for!

A tour of Meeting (2019):

I did a decent introduction of the event two years ago and most of it stand true. Rather than doing another summary, I will compare and contrast my experience at Meeting C++ with the one I had last year at Cppcon.

The venue:

When it comes to the facilities of the venue, I found the rooms in the Andels Hotel to have unequal quality. While the two big rooms on the lower floors are excellent, the upper floor ones are really long and narrow. If the nature made you of a short stature or you don't have the best eyesight (or both like me), you will have a hard time following some of the slides. I cannot remember having any issue with that at CppCon. Having rooms with inclined rows helps a lot!

Note: apparently, the organisers are well-aware of this issue and may come up with a nice solution for the next round of Meeting C++!

Probably to avoid having crammed rooms and irritating noises, Meeting C++ also has this concept of closing doors of the tracks as soon as the talk starts and not letting annoying enter. I am not quite convinced by this idea, it extremely frustrating to not be able to move to another talk if the current one is not to your taste.

The German cuisine:

Being French, a detail often comes to my mind... the food provided during the event! Unlike CppCon, this event offers catering all day long, and tasty one mind you. Obviously you cannot expect conference meals to deserve a Michelin stars's, but there was plenty and with a lot of variety. Is that a must-do for a conference? Probably not, but it is both really convenient and it also brings a LOT of interactions between members of the community. People will tend to have more casual conversations if they are dinning in the same place rather than going out. From what I can recall, CppCon provided snacks here and there, but nothing worth remembering.

The talks:

Last but not least, the speakers! On one hand, Meeting C++ has more diversity: the speakers are coming from all around Europe and you may discover a lot of new names and some hidden gems that didn't/wouldn't make it to CppCon - not everyone is willing or can afford to travel to the USA. On the other hand, the amount of C++ committee members and famous speakers per square kilometre felt higher at CppCon which probably made the talks of a slightly better quality overall.

Meeting C++ offers you the possibility to go to four different tracks for a given time slot. It is enough choices that you will always find a talk to your liking. The talks are also explicitly tagged with a level of experience required to enjoy them: beginner to advanced. Once again, going only to talks with topics that you like is often not the best strategy to fully enjoy this conference. It can be very rewarding to go out of your comfort zone and try a talk not a single word on its title make sense to you. Likewise, do not underestimate the quality of beginner talks. You may learn more than you would expect!

I have had a first hand, albeit short, experience on what it takes to be a speaker at such an event. I gave a lightning talk highly inspired on the performing try emplace post I did at the beginning of this year. While it went rather smoothly (if not for a small technical issue with my slides), I felt intimidated when I went on stage in front of maybe 300 people (best-guess estimation).

You can find a pdf version of the slides right here and maybe a YouTube record of that talk will pop at one point on the Meeting C++ channel. It was interesting to see that the lightning talks were a lot more formal than at CppCon where jokes and bad-puns are a constant. I believe that the lightning talks format was changed recently at Meeting C++: it used to be its own dedicated track on a full afternoon, and it became a more casual evening event. We, the lightning speakers, were also informed a bit late on where, when and how should the talks happen.

Doing a presentation on an innovative topic, interesting and understandable by all requires a lot of upfront work. Having to do it in a foreign language, for one hour, in front of a rather massive crowd is even more demanding. I am still amazed by what all these speakers gave us during these three days!

The chief's suggestions of the year:

Here comes again the menu of the talks I particularly enjoyed. The legend follow the usual rules:

💀 : The difficulty of the talk (💀: Beginner friendly, 💀💀: Intermediate, 💀💀💀: High exposure to C++'s dark corners)

★ : My interest for the talk (★: Good talk, ★★: Tasty talk, ★★★: Legendary talk)

No need to say that I will not spoil everything in the talks, but I will simply try to give an idea of what the talk was about. If the talk seems to be of your taste, I highly encourage you to watch the video record of it or the slides. If your talk or a talk that you liked very much is not part of this menu, do not feel too disenchanted about it. I had to curate this list to a few talks or I would have to write an entire novel. Feel free to post your opinion about the talks in the comment section.

[Keynote] Design Rationale for - Howard Hinnant - 💀💀★★:

The Howard Hinnant, himself, came to present his new-born baby: the support of dates, calendars and time-zones in the chrono header. For those out-of-the-loop, Howard Hinnant has been a massive contributor to what C++ is nowadays. He is known for bringing move semantics to the language, introducing std::unique_ptr , being lead author of libc++... a myriad of things... and the notion of time in the standard through the header <chrono> . It should not come as surprise that Howard is also the man being the dates and calendars coming to the new C++ standard.

Howard's talk was very educating on this addition to C++20. Within two hours, I had a fairly good overview of what we will get and more importantly an answer to why chrono is so verbose. Here is the gist of what we will have:

std::chrono::time_point (already in C++11) get a specialized alias sys_days equal to std::chrono::time_point<std::chrono::system_clock, std::chrono::days> . In other word sys_days represent a time point using the system clock and expressed in days. This sys_days can be thought as an integer representing a value in days.

equal to . In other word represent a time point using the system clock and expressed in days. This can be thought as an integer representing a value in days. We can convert this sys_days time point into a year_month_day representation and vice-versa. This is the same representation of that time point but in a much more human-friendly way: it uses the Gregorian calendar to express that point in time. A time point in the Gregorian calendar is made of three parts: a year, a month and a day of the month. The new standard offers lots of arithmetic operators, conversion operators, helpers structs... to manipulate these parts easily.

time point into a year_month_day representation and vice-versa. This is the same representation of that time point but in a much more human-friendly way: it uses the Gregorian calendar to express that point in time. A time point in the Gregorian calendar is made of three parts: a year, a month and a day of the month. The new standard offers lots of arithmetic operators, conversion operators, helpers structs... to manipulate these parts easily. A new type, with a self-explanatory name, std::chrono::time_zone has been introduced. If you combine a sys_days with a time_zone , you obtain a zoned_time . zoned_time handles conversion from one time_zone to another for you. It is highly convenient!

with a , you obtain a . handles conversion from one to another for you. It is highly convenient! Finally, almost all the types in the <chrono> header have conversion facilities to and from strings. It even works with the new std::format library.

Equipped with this extended <chrono> header, you can easily write such things:

for ( auto d = January / 9 / 2019 ; d . year () < 2020 y ; d = sys_days { d } + weeks { 2 }) { zoned_time london { "Europe/London" , local_days { d } + 18 h }; cout << london << '

' ; cout << zoned_time { "America/New_York" , london } << "



" ; }

In this example, we are displaying the time in "New York" every two weeks, starting from the 9th of January, for the year 2019, when it is 18.00 in "London". As Howard pointed out, this is far from being trivial to do by hand: both the UK and USA have different daylight savings rules.

Some people may complain of the verbosity. Firstly, I find the code very readable even without prior knowledge of this new part of <chrono> Second, safety was a key-point in Howard's design. At multiple time, he gave us example on how his design prevents wrongdoings at compile-time. For instance: what is a time difference? Can this be considered a date? Can 2 hours be translated to a date in a calendar? Probably not!

I am quite eager to get this part of C++20!

[Talk] 10 techniques to understand existing code - Jonathan Boccara - 💀★★★:

Jonathan Boccara, known for his awesome blog fluentcpp.com, is always a safe bet when it comes to C++ talks. You should not necessarily expect learning bleeding-edge techniques with Jonathan, but he is extremely talented at putting words on things!

This time, he was presenting us his favourite approaches to code you have never seen before. Especially, when it comes to intricate code-bases. My favourite analogy was the stronghold one: start your exploration of the code from a place (a function, a routine...) that is crucial in the code-base and that you can easily understand. Like a fog of war in a video-game, the rest of the code is unknown to you at first. But slowly, you can explore the rest of the map/code-base by sending minions/yourself into the callers or the callees related to that function. One strategy to know where to "expand your territory for a maximum of profit" is by looking at call stacks: by jumping in the frames above the current function, this can give you an idea of what is the critical path of your program.

Here is an awesome representation, from Jonathan, of call stacks in a video-game universe:

Not only this applies to decipher a code-base, but it is also a very efficient technique in reverse engineering. When trying to reverse a native application, it is often quite rewarding to put a breakpoint on the system calls: these are your strongholds. Without source-code, you have very little places where you know what the assembly you are reading does. System calls are well documented and exposed, so it is naturally a good place to start.

I believe that I am already applying most of these techniques on a daily-basis. But if I had to explain to a junior programmer what programming on a large code-base is all about, I would probably recommend that person to have a look at this talk.

[Talk] Testing Legacy Code - Fuzzing for Better Input Data - Tina Ulbrich, Niel Waldren - 💀💀★★★:

During such conferences, you may end-up for various reasons (entering the wrong door, following someone, not getting on time to a room...) into a talk that you did not expect on your schedule. Sometimes, you may regret your unfortunate choice as much as std::vector<bool> being in the standard, but it can also turn out into your favour: you may discover one of the best presentation of that day. This is exactly what happened to me with the talk from Tina Ulbrich and Niel Waldren about fuzzing and unit-testing.

Usually, fuzzing consists in feeding an application or part of an application with pseudo-random data to explore as much code-path as possible. The end-goal is to find code-paths that lead to crashes or bugs which can be exploit for malicious purposes.

Once an exploit found, the fuzzer will often try to reduce the "buggy input" to a minimal set to clearly isolate the corner case.

I have heard about fuzzing for a while, but I never had the opportunity to use it myself. This technique is really appealing if you are working in the security business or if your application is critical part of system. If you are developing an application a bit less demanding on the stability, like video-games, fuzzing is not necessarily the number one priority. Tina and Niel twisted the usage of fuzzing to find exploits to improve the test-coverage of their library. They used libFuzzer which is, from what I understand, a LLVM library that combine LLVM's code coverage tool - SanitizerCoverage with a fuzzer engine like AFL. libFuzzer will try to maximize the code-coverage with the minimum input data. This is really appealing for all projects relying on unit-tests.

Tina and Niel did a step by step explanation on how to use libFuzzer and what are its benefits:

How to tweak the random data you receive to valid parameters.

How to reuse the best input data corpus to write new unit-tests for your application.

Why you can use this to detect very subtle changes in your API's behaviour. Changes that you would not necessarily catch with usual unit-tests.

A lot more...

I will have a second watch of their talk as soon as their video is available on the conference's Youtube channel. It was a very innovative topic to me!

[Talk] Modules - The Beginner's Guide - Daniela Engert - 💀💀★★:

Out of the three major language features coming to C++20 (concepts, coroutines and modules), I was the least familiar with modules. The committee is often very active on the standard until the last months before the release of a version. While I had a rough idea of module were about in C++, I did not bother to learn about the syntax and the exact implementation. You never know if the feature you read about in a white paper while look any different in the actual standard! So, I was very grateful that Daniela Engert came up with that talk right before 2020. Being so close to the standard release, the module part was unlikely to change too much!

So what are modules in a few words? Traditionally, a C++ code-base is separated into multiple compilation units which are in most case your .cpp files. These compilation units are compiled separately and combined by the linker. If you want to share functions or objects between two compilation units, you must have a common interface for them: one or more header files (.h/.hpp) that declare what is available. The problem is that header files are:

1) Shared in a very primitive way. The #include directive is doing dumb copy-paste of the header content into your cpp file. 2) Often full of complicated content for the compiler and not just few declarations. In those, you can have macros, templates, inline functions, include of includes...

This results in your compiler doing a lot of unnecessary work parsing these headers for all the compilation units. Wouldn't it be better if compilation units themselves could expose directly what they provide instead of using these hackish headers? This is exactly what modules try to solve!

So writing a very basic module becomes as simple as writing a .cpp file with few annotations to expose what we want:

export module my . module ; // This is the name of our module export int e = 42 ; // Expose a variable. export int bar () { return foo ( e ); // Expose a function. } int foo ( int x ){ return x ; } // No export keyword == no exposition.

When compiling this .cpp file, this will create two others files: an object file that contains binary code (OBJ) and a file for interfacing with that module called Binary Module Interface. Unlike a header file, the BMI file and its associated OBJ file can have a highly optimized representation of what is available in the module.

As a user of a module into another compilation unit, you will pass this BMI file as parameter to your compiler and write an import statement:

import module my . module ; // Ask the compiler to use that module in here. int main () { std :: cout << bar (); // This is using bar from the other module. }

And voilà! These are the new C++ module in all their glory. Except that it gets quite a lot more complicated when you are mixing old header includes, namespaces, special support for standard headers... Daniela was really good at explaining all these quirks that you may encounter in the rough transition that will happen from headers to modules.

As for performance, she observed an improvement from 1546 milliseconds to 62 milliseconds when using a library as a module on a huge-scale project at work. This gives a lot of hope on what modules will offer to us when available on all major compilers!

[Talk] C++20 The small things - Timur Doumler - 💀★★:

You probably have heard of the major features coming to C++20: concepts, coroutines and modules. And you would think that this is enough on your C++ plate for a few years. Well, even without these three features, C++20 still has plenty to offer. This is what Timur Doumler demonstrated to us for one hour!

Amongst a plethora of small improvements, here are my favourite so far:

Designated initialisers

Let's imagine that you have a struct with quite a few members in it. Now let's pretend that we want to initialize this structure with a few variables using the aggregate initilization from C++11. What are the chances that you assign the correct members from the first try?

struct my_struct { int a1 ; int b ; int a2 ; int b2 ; int b3 ; int c ; // ... }; my_struct s { 42 , 43 , 1337 , 54 , /*...*/ };

I tell you, the chances are low. You will have to double-check the struct definition more than once to order things accurately! So what if you could specify which member you designate? This is where C++20 comes to save the day:

my_struct s {. a1 = 42 , . b = 43 , . a2 = 1337 , . b2 = 54 , /*...*/ };

I can hear YOU the C programmer, this has been in C99 for a while... And you will also point with your smug face that you must respect the order (by appearance in the definition) of the members when using C++'s designated initializers unlike C. It happens that C++ has much more complicated rules of evaluation than C. Allowing for random order of assignment could be troublesome. What this is feature is about is 1) safety 2) a nice way to get auto-completion from your IDE on what member you need to fill in.

Improved lambdas

If you want to capture a parameter pack within a lambda before C++20, you will have to be extremely creative. This is even more frustrating that lambdas are great to combine with templates. This very tedious task becomes trivial in C++20:

template < class ... Args > auto foo ( Args ... args ) { return [... args = std :: move ( args )]() { // Do whatever you want with args... }; }

std::mindblowing()! Sprinkling a bit of ... where it should... works as expected!

Lambdas become, in C++20, allowed in a unevaluated contexts. What are unevaluated contexts? Whenever you have an expression within a sizeof(...) or decltype . Is that something you would frequently do? Actually yes. Whenever you want a custom deleter for your std::unique_ptr :

// Before C++20: const auto custom_deleter = []( handle * h ){ /* Do something with h. */ release ( h ); }; std :: unique_ptr < handle , decltype ( custom_deleter ) > p ; // Becomes a one liner: std :: unique_ptr < handle , decltype ([]( handle * h ){ /* Do something with h. */ release ( h ); }) > p ;

Timur came up with other examples like having a custom comparison template parameter for std::set .

Extended Non-Type Template Parameter (NTTP)

If you are a template meta-programming (TMP) enthusiast like me, this will be a game changer! Right now, the standard only allow integer-like (integer, enumeration, pointers...) and types as template parameter. This is highly restrictive and frustrating at times. When pushing template meta-programming to its limit, you will often want to manipulate strings. Right now you will need to decompose your string into char that you pass as template arguments.

For instance, these bits of code would never work in C++17:

struct vec2 { int x ; int y ; }; template < class T , vec2 V > // error: vec2 - illegal type for non-type template parameter struct my_struct { std :: array < std :: array < T , V . y > , V . x > bla ; }; template < std :: string InitValue > // error: std::string - illegal type for non-type template parameter struct my_struct2 { std :: string s = InitValue ; };

Good news! These restrictions have been lifted as long as your non-type template parameter's type (I know the usage of "type" twice here is confusing):

1) Has a comparison operator available ( operator== ) at compile time: it is constexpr . The goal is that the compiler should be able to check if two template instantiations are the same by checking if all template parameters are equal.

) at compile time: it is . The goal is that the compiler should be able to check if two template instantiations are the same by checking if all template parameters are equal. 2) Can be constructed at compile-time: its construction can be done in a constexpr context.

[Talk] Compile Time Regular Expressions with Deterministic Finite Automaton - Hana Dusíková - 💀💀💀★:

Speaking of template meta-programming, NTTP and strings, Hana Dusíková improved her compile-time regular expression library. At last year's CppCon, Hana impressed the crowd with her library: it exploited template meta-programming in C++17 to its maximum to generate a regex parser at compile-time from a string literal. This makes her library ridiculously fast compared to std::regex , which works at runtime. Surprisingly, the library does not affect compilation at all. In fact, it improves a lot the compilation time compared to std::regex ! Does this implies that there is such a thing as "Zero-cost Abstractions"? Maybe...

I have to admit that I was part of the people who missed the chance to see her talk live when I could have... So this year, I took my revenge and went to her follow-up talk. Using some of the C++20 template features, Hana succeeded to make a new regexp engine using a Deterministic Finite Automaton. Do not ask me to summarise properly what I have witnessed during this talk, it was... complicated! But it is also fascinating, it combines a hefty does of meta-programming with language theory: that's a lot of mental stimulation.

If that sounds fun to you, I would probably suggest to watch her initial talk and come back to this one afterwards.

[Other]:

There are a lot of other talks that would be worth using your Google-Fu to find them:

Using C++20's Three-way Comparison <=> - Jonathan Müller

Oh No! More Modern CMake - Deniz Bahadir

C++20 Coroutines - Milosz Warzecha

And many more...

You will surely find the lightning talks on the Meeting C++'s Youtube Channel at one point. These short talks are like snack foods, they vary a lot in their content and quality and don't take long to process.

Conclusion:

This was another very fruitful event for the C++ community. I am glad that our beloved C++ language receive so much attention by its users and the committee. It will be interesting to see how long it will take for the major compiler to be fully C++20 compliant: this release is massive! I was slightly overwhelmed by the amount of new features.

I am also wondering how much of the newly acquired C++20 knowledge will stay true once the standard is released. I guess that I will have to check that by going to another C++ event next year ;)