With the arrival of summer, another TokyoR User Meetup! On May 25th, useRs from all over Tokyo (and some even from further afield - including Kan Nishida of Exploratory, all the way from California!) flocked to Jimbocho, Tokyo for another jam-packed session of R hosted by Mitsui Sumitomo Insurance Group.

Like my previous round up posts (for TokyoR #76 and TokyoR #77) I will be going over around half of all the talks. Hopefully, my efforts will help spread the vast knowledge of Japanese R users to the wider R community. Throughout I will also post helpful blog posts and links from other sources if you are interested in learning more about the topic of a certain talk. You can follow Tokyo.R by searching for the #TokyoR hashtag on Twitter.

Unlike most R Meetups a lot of people present using just their Twitter handles so I’ll mostly be referring to them by those instead. I’ve been going to events here in Japan for a bit over a year but even now sometimes I’m like, “Whoahh that’s what @very_recognizable_twitter_handle_in_the_japan_r_community actually looks like?!”

Anyways…

Let’s get started!

BeginneR Session

As with every TokyoR meetup, we began with a set of beginner user focused talks:

Main Talks

tanakafreelance: Radiant for Data Analysis!

@tanakafreelance talked about Radiant, which is a platform-independent browser-based GUI for business analytics that was developed by Vincent Nijs. It is a tool for business analytics purposes and is based on Shiny . After installation from CRAN you can launch it via using radiant::launcher() . Most of this presentation was a live demo by @tanakafreelance showing a lot of the functionality offered by Radiant such as creating reproducible reports with R Markdown, writing your own R code to use within the GUI, creating and evaluating models (linear/logistic regression, neural networks, naive Bayes, and more), and design of experiments (DOE)!

You can run it from a variety of set ups from online, offline, on shinyapps.io , Shiny server, and even on a cloud service like AWS via a customized Docker container. For a comprehensive introduction to Radiant’s full capabilities you can check out its awesome website here, full of videos and vignettes!

kotaku08: Transitioning a Company to Use R!

@kotaku08 talked about his experiences in data analytics and ways he pushed for the usage of R at his company, VALUES. One of the first things he realized upon entering the company was how the skill set of the team was more of that of data/system engineers rather than data analysts. After some time he found three big problems with his working environment that he wanted to solve:

Mismatch between the tool used and the task needed done easy data manipulation with PHP.

complicated data manipulation with Excel. What was this again?! (Illegible! Non-reproducible! Non-reusable!) Extremely convoluted Excel formulas that look like they could be banned by the Malleus Maleficarum .

. Excel sheets only contain the results .

. If it was a visualization task, it was all done in Tableau… Data extraction being a painful process… Connecting to Redshift is a pain!

While pondering these problems, he came across this article from AirBnB that highlighted their transition to building R tools and teaching R across the company. The key takeaways that @kotaku08 took from the article was:

Most analysts at AirBnB use R.

Intracompany package: Rbnb .

. Efforts put into R education and conducting workshops.

Data analysis is both efficient AND reproducible!

Taking these lessons to heart he decided to implement #rstats learning sessions as well as create a company R package! One of the main functionalities of VALUES’ main R package is being able to access data from Redshift and in tandem with the various packages in the cloudyr project has made getting data much more easier for @kotaku08 and his team.

Another big step was educating fellow employees about #rstats.

For existing employees:

Spread rumors about how accessing data is much easier with R… Those skilled in other scripting languages organically come over to check R out!

For new employees:

Emphasize how R is THE standard at the company,

standard at the company, “Graduate hires”, most of whom have no programming experience, are put into R boot camps

After 3 months of hard work, able to use the tidyverse for analytical tasks!

As a result of these efforts 80% of employees can now use R and the internal company package has two new maintainers (both graduate hires!) to work alongside kotaku08 .

Some other resources:

tomkxy: Making Your Code Faster - Introduction to Vectorisation and Parallel Computing (English with demonstrations)

@tomkxy presented in English (he’s a Kiwi that works for RIKEN!) on vectorizing your code and parallel computing with R. In response to a lot of the accusations that “R is slow”, Tom talked about different techniques to use to make your R code faster along with some some demonstrations (the RMD can be found here).

One of my key take-aways from this talk was, “Code first, optimize later!”. In that it’s important to not get stuck doing premature optimization, especially if you might not actually need to use the code again anyways! Also, sometimes parallel computing may not always be the fastest solution due to overhead costs associated with setting up clusters and communication between clusters.

In addition, the newly developed “Jobs” pane in RStudio 1.2, released last month, means you can keep being productive even while you have your scripts running in the background. A great resource for those interested is the CRAN Task View for high performance and parallel computing available here.

A few other resources:

LTs

ill_identified: Guide to MCMC with the bayesplot package!

@ill_identified presented on using Markov chain Monte Carlo (MCMC) with R, specifically using the bayesplot package. MCMC are a series of methods that contain algorithms for sampling from a probability distribution. These methods involve drawing random samples from a target distributions using algorithms (such as Metropolis-Hastings algorithm, reversible jump, HMC, etc.) then we attempt to construct a Markov chain such that its equilibrium probability distribution is as close to our target distribution as possible by iterating the chain many times.

As I’m not familiar with MCMC very much I won’t go into too much detail here, however for others unfamiliar with MCMC and Bayesian inference, @ill_identified provided a nice list of books to get you started:

Just recently TJ Mahr, one of the authors of bayesplot , presented on the package at Chicago SatRDays. You can check the slides out here. The new version of bayesplot , 1.7.0, will also support tidyselect:

Other resources:

Atsushi776: May I felp you?

@Atsushi776 , known in the Japanese R Community for his “headphones” avatar, created a new package called felp as he was annoyed that he couldn’t look at the source code while looking at the help files of a function. Also there was the added annoyance of having to jump back to the start of the function to type ? back in AND deleting it once you’re done.

source ( "https://install-github.me/atusy/felp" ) library ( felp ) library ( printr ) ## From this: ? help () ## To this: help ? ## Alternatively: felp ( help ) felp ( "help" ) ## Source code is nicely highlighted by `prettycode`: ## Output shortened for brevity... grep () ? .

## function (pattern, x, ignore.case = FALSE, perl = FALSE, value = FALSE, ## fixed = FALSE, useBytes = FALSE, invert = FALSE) ## { ## if (!is.character(x)) ## x <- structure(as.character(x), names = names(x)) ## .Internal(grep(as.character(pattern), x, ignore.case, value, ## perl, fixed, useBytes, invert)) ## } ## <bytecode: 0x000000001750ad50> ## <environment: namespace:base> ## Pattern Matching and Replacement ## ## Description: ## ## 'grep', 'grepl', 'regexpr', 'gregexpr' and 'regexec' search for ## matches to argument 'pattern' within each element of a character ## vector: they differ in the format of and amount of detail in the ## results. ## ## 'sub' and 'gsub' perform replacement of the first and all matches ## respectively. ## ## Usage: ## ## grep(pattern, x, ignore.case = FALSE, perl = FALSE, value = FALSE, ## fixed = FALSE, useBytes = FALSE, invert = FALSE) ## ## grepl(pattern, x, ignore.case = FALSE, perl = FALSE, ## fixed = FALSE, useBytes = FALSE) ## ## sub(pattern, replacement, x, ignore.case = FALSE, perl = FALSE, ## fixed = FALSE, useBytes = FALSE) ## ## gsub(pattern, replacement, x, ignore.case = FALSE, perl = FALSE, ## fixed = FALSE, useBytes = FALSE)

Short for f unctional h elp, he got this to work by modifying the ? operator to show the inner structure of a function along with the help page. This works for both a function as seen above and on packages by package_name?p . You can also use the ? on data set objects to return what you’ll normally get from a str() call in addition the the help page.

iris ? . ## also opens "Help" page for the dataset

## 'data.frame': 150 obs. of 5 variables: ## $ Sepal.Length: num 5.1 4.9 4.7 4.6 5 5.4 4.6 5 4.4 4.9 ... ## $ Sepal.Width : num 3.5 3 3.2 3.1 3.6 3.9 3.4 3.4 2.9 3.1 ... ## $ Petal.Length: num 1.4 1.4 1.3 1.5 1.4 1.7 1.4 1.5 1.4 1.5 ... ## $ Petal.Width : num 0.2 0.2 0.2 0.2 0.2 0.4 0.3 0.2 0.2 0.1 ... ## $ Species : Factor w/ 3 levels "setosa","versicolor",..: 1 1 1 1 1 1 1 1 1 1 ... ## Edgar Anderson's Iris Data ## ## Description: ## ## This famous (Fisher's or Anderson's) iris data set gives the ## measurements in centimeters of the variables sepal length and ## width and petal length and width, respectively, for 50 flowers ## from each of 3 species of iris. The species are _Iris setosa_, ## _versicolor_, and _virginica_. ## ## Usage: ## ## iris ## iris3 ## ## Format: ## ## 'iris' is a data frame with 150 cases (rows) and 5 variables ## (columns) named 'Sepal.Length', 'Sepal.Width', 'Petal.Length', ## 'Petal.Width', and 'Species'. ## ## 'iris3' gives the same data arranged as a 3-dimensional array of ## size 50 by 4 by 3, as represented by S-PLUS. The first dimension ## gives the case number within the species subsample, the second the ## measurements with names 'Sepal L.', 'Sepal W.', 'Petal L.', and ## 'Petal W.', and the third the species. ## ## Source: ## ## Fisher, R. A. (1936) The use of multiple measurements in taxonomic ## problems. _Annals of Eugenics_, *7*, Part II, 179-188. ## ## The data were collected by Anderson, Edgar (1935). The irises of ## the Gaspe Peninsula, _Bulletin of the American Iris Society_, ## *59*, 2-5. ## ## References: ## ## Becker, R. A., Chambers, J. M. and Wilks, A. R. (1988) _The New S ## Language_. Wadsworth & Brooks/Cole. (has 'iris3' as 'iris'.) ## ## See Also: ## ## 'matplot' some examples of which use 'iris'. ## ## Examples: ## ## dni3 <- dimnames(iris3) ## ii <- data.frame(matrix(aperm(iris3, c(1,3,2)), ncol = 4, ## dimnames = list(NULL, sub(" L.",".Length", ## sub(" W.",".Width", dni3[[2]])))), ## Species = gl(3, 50, labels = sub("S", "s", sub("V", "v", dni3[[3]])))) ## all.equal(ii, iris) # TRUE

In the near future @Atsushi776 wants to get rid of not just the . but the ? altogether and wants to work on using a prefix p? in front of the package name to bring up the documentation for an entire package. Go felp yourself by taking a look at the package website!

0_u0: Marketing Science & R!

@0_u0 (better known as きぬいと or Kinuito ) talked about his successful attempt to integrate R into his workflow at the marketing department of a very non-technical traditional Japanese company.

Most of the work being done for customers by his company is descriptive statistics. Nothing fancy or A.I. or even simple linear regression. As such, a lot of the problems that are given to his department can be solved by tables and ggplot s. As a consequence he had been fighting an uphill battle as the company standard is to just use Excel for … well literally everything.

Trying to find some way to incorporate R and Python to make his workflow easier Kinuito started using the tidyverse to simplify the data cleaning processes!

Key takeaways:

Reduce overtime by using the tidyverse to automate a lot of the grunt work involved with cleaning and transforming marketing data.

to automate a lot of the grunt work involved with cleaning and transforming marketing data. Not have to open up extraordinarily large Excel files (as much as before…).

Great success in using ggplot2 and DiagrammeR for creating informative output.

and for creating informative output. Start with descriptive statistics, you can’t do anything more advanced unless you have the infrastructure to do so!

Kinuito also highlighted some things he wanted to do in the near future:

Document R and Python tips for new graduate hires using R Markdown!

Consolidate the company’s R environment: Currently version control is a mess as everybody is still only working in their own local environments. Solution: Docker?



Along with @kotaku08 ’s talk it was great to get more insight into how R is used at various companies. I’ve personally only heard things from an American or English company’s point of view (from the various R conferences/meetups I’ve been to) so it was nice to hear about the differences and similarities in the challenges faced by Japanese corporations at this month’s TokyoR !

Other Talks

Food, Drinks, and Conclusion

Following all of the talks, those who were staying for the after-party were served sushi and drinks! With a loud rendition of “kampai!” (cheers!) R users from all over Tokyo began to talk about their successes and struggles with R. A fun tradition at TokyoR is a Rock-Paper-Scissors tournament with the prize being free data science books!

The prize for this month was: