My name is Jonathan, and I have a coding obsession.
I’ll admit it, the Hadleyverse has ruined me. I can no longer read a blog post or stackoverflow question in base
R and leave it be. There are improvements to make, and I’m somewhat sure that I know what they are. Most of them involve
dplyr. Many involve
data.table. Some involve
This one came up on R-bloggers today (which leads back to MilanoR) and seemed like a good opportunity. The problem raised was; given a list of
data.frames, can you create a list of the variables sorted into those
data.frames? i.e. can you turn this
[code lang=“r”] df_list_in <- list ( df_1 = data.frame(x = 1:5, y = 5:1), df_2 = data.frame(x = 6:10, y = 10:6), df_3 = data.frame(x = 11:15, y = 15:11) ) [/code]
[code lang=“r”] df_list_out <- list ( df_x = data.frame(x_1 = 1:5, x_2 = 6:10, x_3 = 11:15), df_y = data.frame(y_1 = 5:1, y_2 = 10:6, y_3 = 15:11) ) [/code]
That looks like a problem I came across recently. Let’s see…
I managed to replace that function – which, while fast, is a little obtuse and difficult to read – with essentially a one-liner
[code lang=“r”] df_list_in %>% purrr::transpose %>% lapply(as.data.frame) [/code]
You may now proceed to argue over which is easier/simpler/more accessible/requires less knowledge of additional packages/etc… If you ask me, it’s damn-near perfect as long as you can place a cursor on
RStudio and hit
F1 which will bring up the
purrr::transpose help menu and explain exactly what is going on. Anyway, how does it compare? Here’s Michy’s graph (formatting updated and my function added)
and then, just for fun (and because I wanted an excuse to try it out) here’s a
yarrr::pirateplot of the same data
My one-line function (without the
magrittr syntactical sugar) does slightly better than the
arrange_col function (on average), but has a lot less up-front code and is more readable (to me at least). The performance of any of these three doesn’t seem like it would have trouble scaling for any practical use-case.
Scaling up the problem to a list of 100
data.frames each with 1000 observations of 50 variables, the same result pans out as shown in the above
On the giant example (100
data.frames of 1000 observations of 50 variables) the difference is 20ms vs 380ms. Honestly, I don’t know what I’d do with the additional 360ms, but chances are I’d just waste them. I’ll take the efficient code on this one.
Can you do even better than the one-liner? Spot a potential issue? Have I made a mistake? Got comments? You know what to do.