Nothing
#' @export
#' @import streamR rgdal RSQLite
t2sqlite <- function(fileprefix, import=TRUE, path = ".", pattern = ".json$"){
#list all the files of the folder and get the total number of files
files <- list.files(path, pattern, full.names = TRUE)
files <- files[grep(fileprefix, basename(files))]
#opena database connection
con <- dbConnect(SQLite(), fileprefix)
#--------parse tweets using a loop to append next files to database table--------------#
for(filename in files) {
# ------1) parse the JSON files
#parse tweets and create a data frame
tweets <- parseTweets(filename, simplify = FALSE, verbose = TRUE)
#remove bad character before import it to database
tweets$text<-gsub("[^[:alnum:]///' ]","", tweets$text)
#------2) populate table with tweets
dbWriteTable(con, fileprefix ,tweets,overwrite=FALSE, append=TRUE)
}
#transform to sqlite timestamp format created_at column
dbSendQuery(con,
paste("ALTER TABLE", fileprefix,"ADD t_trans TIMESTAMP WITH TIME ZONE;
UPDATE", fileprefix, "SET t_trans=created_at::TIMESTAMP WITH TIME ZONE;
ALTER TABLE", fileprefix, "drop created_at;
ALTER TABLE ", fileprefix, " rename t_trans TO created_at;"))
#create view only with tweets with only geolocated tweets
dbSendQuery(con,
paste("CREATE TABLE geo",fileprefix," AS
SELECT *
FROM ", fileprefix, "
WHERE lat is not null", sep=""))
if (import==TRUE){
tweets=dbReadTable(con, fileprefix)
}else{
remove(tweets)
}
#close connection with sqlite
dbDisconnect(con)
message(paste("Database created in ", getwd(), "and tweets imported as a data frame.", sep=" "))
return(tweets)
}
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.