as.list.robotstxt_text | Convert robotstxt_text to list |
fix_url | Add http protocal if missing from URL |
get_robotstxt | Download a robots.txt file |
get_robotstxt_http_get | Storage for HTTP request response objects |
get_robotstxts | Download multiple robotstxt files |
guess_domain | Guess a domain from path |
http_domain_changed | Check if HTTP domain changed |
http_subdomain_changed | Check if HTTP subdomain changed |
http_was_redirected | Check if HTTP redirect occurred |
is_suspect_robotstxt | Check if file is valid / parsable robots.txt file |
is_valid_robotstxt | Validate if a file is valid / parsable robots.txt file |
list_merge | Merge a number of named lists in sequential order |
named_list | Create a named list |
null_to_default | Return default value if NULL |
parse_robotstxt | Parse a robots.txt file |
parse_url | Parse a URL |
paths_allowed | Check if a bot has permissions to access page(s) |
paths_allowed_worker_spiderbar | Check if a spiderbar bot has permissions to access page(s) |
pipe | re-export magrittr pipe operator |
print.robotstxt | Print robotstxt |
print.robotstxt_text | Print robotstxt's text |
remove_domain | Remove domain from path |
request_handler_handler | Handle robotstxt handlers |
robotstxt | Generate a representation of a robots.txt file |
rt_cache | Get the robotstxt cache |
rt_get_comments | Extract comments from robots.txt |
rt_get_fields | Extract permissions from robots.txt |
rt_get_fields_worker | Extract robotstxt fields |
rt_get_rtxt | Load robots.txt files saved along with the package |
rt_get_useragent | Extract HTTP useragents from robots.txt |
rt_list_rtxt | List robots.txt files saved along with the package |
rt_request_handler | Handle robotstxt object retrieved from HTTP request |
sanitize_path | Make paths uniform |
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.