I am using the read_html command and I get the following error message on a specific website.
library('rvest') library('dplyr') webpage <- read_html("https://www.edmunds.com/ford/escape/2018/cost-to-own/") Error in open.connection(x, "rb") : HTTP error 403.
Are specific websites blocking R from being scraped?
It's not specific to R, but the site might block webscraping. Two places to look are at the site's robots.txt and the Terms of Service https://www.edmunds.com/robots.txt https://www.edmunds.com/about/visitor-agreement.html
Thanks for the response...
This topic was automatically closed 21 days after the last reply. New replies are no longer allowed.