Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Extracting urls from a list of urls #70

Open
e-orlov opened this issue Nov 21, 2019 · 2 comments
Open

Extracting urls from a list of urls #70

e-orlov opened this issue Nov 21, 2019 · 2 comments

Comments

@e-orlov
Copy link

e-orlov commented Nov 21, 2019

Is it in general possible to load a list of urls, from which Rcrawler collects urls?

@smallperks
Copy link

Hey there! Not exactly sure what you are asking here. Could you explain a bit more?

If you would like to crawl a list of URLs that you provide, there is documentation about that here: https://github.com/salimk/Rcrawler/#9-1--scrape-data-from-list-of-urls

The list is passed into the ContentScraper function in a method that is similar to crawling a single URL.

Example provided:

listURLs<-c("http://www.thedermreview.com/la-prairie-reviews/", "http://www.thedermreview.com/arbonne-reviews/", "http://www.thedermreview.com/murad-reviews/")

Reviews<-ContentScraper(Url = listURLs, CssPatterns =c(".entry-title","#comments p"), ManyPerPattern = TRUE)

@e-orlov
Copy link
Author

e-orlov commented Nov 25, 2019 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants