Best practice for working with databases and huge amount of data within RStudio

In an upcoming project I will work with a large amount of data stored on a SQL server. As best practive it can be read everywhere that for performance reasons the arithmetic operations are best done in the database itself, i.e. "push operation - collect results". With the dplyr package this can be easily achieved with tbl(con, "test"). Can I also achieve this approach with the DBI package (which I have not worked with yet)?
I would work with the dyplr package, but when the table in my database exceeds a certain size I unfortunately get an error.


Did you read the documentation provided by RStudio on this topic?

They have tips for any type of commonly used database and how to ensure you run the queries on the database and only retrieve the results so you don't have to load everything into R memory.

Hope this helps,

This topic was automatically closed 21 days after the last reply. New replies are no longer allowed.

If you have a query related to it or one of the replies, start a new topic and refer back with a link.