View source: R/GetRealTimeGrib.R
WebCrawler | R Documentation |
Discover all links on a given web page, follow each one, and recursively scan every link found. Return a list of web addresses whose pages contain no links.
WebCrawler(url, depth = NULL, verbose = TRUE)
url |
A URL to scan for links. |
depth |
How many links to return.
This avoids having to recursively scan hundreds of links.
Defaults to |
verbose |
Print out each link as it is discovered.
Defaults to |
CrawlModels
uses this function to get all links present on a model page.
urls.out |
A list of web page addresses, each of which corresponds to a model instance. |
While it might be fun to try WebCrawler
on a large website such as Google, the results will be unpredictable and perhaps disastrous if depth
is not set.
This is because there is no protection against infinite recursion.
Daniel C. Bowman danny.c.bowman@gmail.com
CrawlModels
, ParseModelPage
#Find the first 10 model runs for the
#GFS 0.5x0.5 model
## Not run: urls.out <- WebCrawler(
"http://nomads.ncep.noaa.gov/cgi-bin/filter_gfs_0p50.pl", depth = 10)
## End(Not run)
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.