.PkgName = "NLP4kec"
.onLoad <- function(libname, pkgname){
library(rJava)
initopt <- c("-Xmx4g", "-Dfile.encoding=UTF-8")
options(java.parameters = initopt)
.jpackage(pkgname, lib.loc = libname)
}
.onAttach <- function(libname, pkgname){
.jinit(system.file(package = .PkgName))
.jaddClassPath(dir( file.path(system.file(package = .PkgName),"java"), full.names=TRUE ))
}
r_parser_r <- function (contentVector, useEn = F, language, korDicPath = "noDic")
{
obj = .jnew("com.namyun.pos.Starter")
props_path = file.path(system.file(package = "NLP4kec"), "java")
if (language == "zh") {
return(.jcall(obj, "[Ljava/lang/String;", "rTextParserFromRtoR",
.jarray(contentVector), as.character(useEn), as.character(language),
paste0(props_path, "/StanfordCoreNLP-chinese.properties"),
"NULL"))
}
else {
return(.jcall(obj, "[Ljava/lang/String;", "rTextParserFromRtoR",
.jarray(contentVector), as.character(useEn), as.character(language),
"NULL", as.character(korDicPath)))
}
}
file_parser_file <- function(path, useEn = F, language, korDicPath = "noDic"){
obj = .jnew("com.namyun.pos.Starter")
props_path = file.path(system.file(package = "NLP4kec"), "java")
if (language == "zh") {
.jcall(obj, "S", "rTextParserToFile", as.character(path),
as.character(useEn), as.character(language), paste0(props_path,
"/StanfordCoreNLP-chinese.properties"), "NULL")
}
else {
.jcall(obj, "S", "rTextParserToFile", as.character(path),
as.character(useEn), as.character(language), "NULL",
as.character(korDicPath))
}
}
file_parser_r <- function(path, useEn = F, language, korDicPath = "noDic"){
obj = .jnew("com.namyun.pos.Starter")
props_path = file.path(system.file(package = "NLP4kec"), "java")
if (language == "zh") {
return(.jcall(obj, "[Ljava/lang/String;", "rTextParserToR",
as.character(path), as.character(useEn), as.character(language),
paste0(props_path, "/StanfordCoreNLP-chinese.properties"),
"NULL"))
}
else {
return(.jcall(obj, "[Ljava/lang/String;", "rTextParserToR",
as.character(path), as.character(useEn), as.character(language),
"NULL", as.character(korDicPath)))
}
}
extract_noun <- function(path, useEn = F, language, korDicPath = "noDic"){
obj = .jnew("com.namyun.pos.Starter")
props_path = file.path(system.file(package = "NLP4kec"),"java")
if(language == "zh"){
return(.jcall(obj, "[Ljava/lang/String;", "rExtractNounToR",as.character(path), as.character(useEn), as.character(language), paste0(props_path,"/StanfordCoreNLP-chinese.properties"), "NULL"))
}else{
return(.jcall(obj, "[Ljava/lang/String;", "rExtractNounToR",as.character(path), as.character(useEn), as.character(language), "NULL", as.character(korDicPath)))
}
}
r_extract_noun <- function(contentVector, useEn = F, language, korDicPath = "noDic"){
obj = .jnew("com.namyun.pos.Starter")
props_path = file.path(system.file(package = "NLP4kec"),"java")
if(language == "zh"){
return(.jcall(obj, "[Ljava/lang/String;", "rExtractNounFromRToR"
, .jarray(contentVector), as.character(useEn), as.character(language), paste0(props_path,"/StanfordCoreNLP-chinese.properties"), "NULL"))
}else{
return(.jcall(obj, "[Ljava/lang/String;", "rExtractNounFromRToR"
,.jarray(contentVector), as.character(useEn), as.character(language), "NULL", as.character(korDicPath)))
}
}
synonym_processing <- function(parsedVector, synonymDic)
{
# 동의어 처리
for (i in 1:nrow(synonymDic)){
targetDocIdx = which(ll <- grepl(synonymDic$originWord[i], parsedVector))
for(j in 1:length(targetDocIdx)){
docNum = targetDocIdx[j]
parsedVector[docNum] = gsub(synonymDic$originWord[i], synonymDic$changeWord[i], parsedVector[docNum])
}
}
return(parsedVector)
}
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.