Join us at rstudio::conf(2022) to sharpen your R skills. | July 25-28th in D.C.
Learn More
rstudio::conf(2022) | July 25-28th in D.C. 7/25 - 7/28 in D.C.
The premier IDE for R
RStudio anywhere using a web browser
Put Shiny applications online
Shiny, R Markdown, Tidyverse and more
Next level training for you and your team
Do, share, teach and learn data science
An easy way to access R packages
Let us host your Shiny applications
A single home for R & Python Data Science Teams
Scale, develop, and collaborate across R & Python
Easily share your insights
Control and distribute packages
RStudio
RStudio Server
Shiny Server
R Packages
RStudio Academy
RStudio Cloud
RStudio Public Package Manager
shinyapps.io
RStudio Team
RStudio Workbench
RStudio Connect
RStudio Package Manager
rstudio::global 2021 programming
Easy larger-than-RAM data manipulation with {disk.frame}
January 21, 2021
Learn how to handle 100GBs of data with ease using {disk.frame} - the larger-than-RAM-data manipulation package.
R loads data in its entirety into RAM. However, RAM is a precious resource and often do run out. That's why most R user would have run into the "cannot allocate vector of size xxB." error at some point.
However, the need to handle larger-than-RAM data doesn't go away just because RAM isn't large enough. So many useRs turn to big data tools like Spark for the task. In this talk, I will make the case that {disk.frame} is sufficient and often preferable for manipulating larger-than-RAM data that fit on disk. I will show how you can apply familiar {dplyr}-verbs to manipulate larger-than-RAM data with {disk.frame}.
ZJ, Lucy D'Agostino McGowan, Neal Richardson, and Garrick Aden Buie Q&A
ZJ, Lucy D'Agostino McGowan, and Neal Richardson Q&A
ZJ is a machine learning developer based in Melbourne, Australia. He regularly contributes to open source projects. He has more than 10 years of experience in banking before joining the tech sector. In his free time, he enjoys playing Go/Baduk/Weiqi.