The hardware and bandwidth for this mirror is donated by dogado GmbH, the Webhosting and Full Service-Cloud Provider. Check out our Wordpress Tutorial.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]dogado.de.
The goal of the wordpredictor package is to provide a flexible and easy to use framework for generating n-gram models for word prediction.
The package allows generating n-gram models from input text files. It also allows exploring n-grams using plots. Additionally it provides methods for measuring n-gram model performance using Perplexity and accuracy.
The n-gram model may be customized using several options such as n-gram size, data cleaning options and options for tokenization.
You can install the released version of wordpredictor from CRAN with:
install.packages("wordpredictor")
And the development version from GitHub with:
# install.packages("devtools")
::install_github("pakjiddat/word-predictor") devtools
The wordpredictor package is based on R6 classes. It is easy to customize and improve. It provides the following classes:
Information about the package can be obtained using the command line or the package website. For example, the command: ?wordpredictor returns information about how the given class works and the parameter details for each class method.
The following code should be run before running the examples.
library(wordpredictor)
# The level of verbosity in the information messages
<- 0
ve
#' @description
#' Used to setup the test environment
#' @param rf The required files.
#' @param ve The verbosity level.
#' @return The list of directories in the test environment
<- function(rf, ve) {
setup_env # An object of class EnvManager is created
<- EnvManager$new(rp = "../", ve = ve)
em # The required files are downloaded
<- em$setup_env(rf)
ed
return(ed)
}
#' @description
#' Used to clean up the test environment
<- function(ve) {
clean_up # An object of class EnvManager is created
<- EnvManager$new(ve = ve)
em # The test environment is removed
$td_env(F)
em }
The following example shows how to generate a n-gram model.
# The required files
<- c("input.txt")
rf # The test environment is setup
<- setup_env(rf, ve)
ed
# The following code generates n-gram model using default options for data
# cleaning and tokenization. See the following section on how to customize these
# options. Note that input.txt is the name of the input data file. It should be
# present in the data directory. ddir is the data directory. mdir is the model
# directory. The output model file, which is def-model.RDS will be placed in
# this directory.
# ModelGenerator class object is created
<- ModelGenerator$new(
mg name = "def-model",
desc = "N-gram model generating using default options",
fn = "def-model.RDS",
df = "input.txt",
n = 4,
ssize = 10,
dir = ed,
dc_opts = list(),
tg_opts = list(),
ve = ve
)
# Generates n-gram model. The output is the file def-model.RDS
$generate_model()
mg
# The test environment is cleaned up
clean_up(ve)
The above code generates the file def-model.RDS. This file represents the n-gram model.
The following example shows how to predict the next word given a set of words:
# The required files
<- c("def-model.RDS")
rf # The test environment is setup
<- setup_env(rf, ve)
ed
# The model file name
<- paste0(ed, "/def-model.RDS")
mfn # An object of class ModelPredictor is created. The mf parameter is the name of
# the model file that was generated in the previous example.
<- ModelPredictor$new(mf = mfn)
mp # Given the words: "how are", the next word is predicted. The top 3 most likely
# next words are returned along with their respective probabilities.
<- mp$predict_word(words = "how are", 3)
res
# The test environment is cleaned up
clean_up(ve)
The wordpredictor package includes a class called DataAnalyzer, that can be used to get an idea of the frequency distribution of n-grams in a model. The model generation process described above, creates an n-gram file in the model directory.
For each n-gram number less than or equal to the n-gram size of the model, a n-gram file is generated. In the example above the n-gram size of the model is 4. So 4 n-gram files are generated in the model folder. These files are: n1.RDS, n2.RDS, n3.RDS and n4.RDS. The n2.RDS file contains n-grams of size 2.
The following example plots the top 10 most occurring bi-grams along with their frequencies:
# The required files
<- c("n2.RDS")
rf # The test environment is setup
<- setup_env(rf, ve)
ed
# The file name
<- paste0(ed, "/n2.RDS")
fn # An object of class DataAnalyzer is created. The fn parameter is the path to
# the n-gram file.
<- DataAnalyzer$new(fn = fn)
da # The top 10 most occurring features are plotted
<- da$plot_n_gram_stats(opts = list(
df "type" = "top_features",
"n" = 10,
"save_to" = NULL,
"dir" = NULL
))
# The test environment is cleaned up
clean_up(ve)
The following example plots the n-gram frequency coverage. It shows the percentage of n-grams with frequency 1, 2 … 10.
# The required files
<- c("n2.RDS")
rf # The test environment is setup
<- setup_env(rf, ve)
ed
# The file name
<- paste0(ed, "/n2.RDS")
fn # An object of class DataAnalyzer is created. The fn parameter is the path to
# the n-gram file.
<- DataAnalyzer$new(fn = fn)
da # The top 10 most occurring features are plotted
<- da$plot_n_gram_stats(opts = list(
df "type" = "coverage",
"n" = 10,
"save_to" = NULL,
"dir" = NULL
))
# The test environment is cleaned up
clean_up(ve)
The following example shows how to get the list of bi-grams starting with “great_” along with their frequencies. It also shows how to get the frequency of the bi-gram “great_deal”.
# The required files
<- c("n2.RDS")
rf # The test environment is setup
<- setup_env(rf, ve)
ed
# The file name
<- paste0(ed, "/n2.RDS")
fn # An object of class DataAnalyzer is created. The fn parameter is the path to
# the n-gram file.
<- DataAnalyzer$new()
da # Bi-grams starting with "great_" are returned
<- da$get_ngrams(fn = fn, c = 10, pre = "^great_*")
df # The data frame is sorted by frequency
<- df[order(df$freq, decreasing = T),]
df # The frequency of the bi-gram "great_deal"
<- as.numeric(df[df$pre == "great_deal", "freq"])
f
# The test environment is cleaned up
clean_up(ve)
The dc_opts parameter to the ModelGenerator class specifies the data cleaning options. The following code shows the data cleaning options and their default values:
# @field dc_opts The options for the data cleaner object.
# min_words -> The minimum number of words per sentence.
# line_count -> The number of lines to read and clean at a time.
# save_data -> If the combined processed lines should be saved.
# output_file -> Name of the output file used to store the data.
# sw_file -> The stop words file path.
# dict_file -> The dictionary file path.
# bad_file -> The bad words file path.
# to_lower -> If the words should be converted to lower case.
# remove_stop -> If stop words should be removed.
# remove_punct -> If punctuation symbols should be removed.
# remove_non_dict -> If non dictionary words should be removed.
# remove_non_alpha -> If non alphabet symbols should be removed.
# remove_extra_space -> If leading, trailing and double spaces
# should be removed.
# remove_bad -> If bad words should be removed
= list(
dc_opts "min_words" = 2,
"line_count" = 1000,
"save_data" = T,
"output_file" = NULL,
"sw_file" = NULL,
"dict_file" = NULL,
"bad_file" = NULL,
"to_lower" = T,
"remove_stop" = F,
"remove_punct" = T,
"remove_non_dict" = T,
"remove_non_alpha" = T,
"remove_extra_space" = T,
"remove_bad" = F
)
The tg_opts parameter to the ModelGenerator class specifies the token generation options. The following code shows the token generation options and their default values:
# @field tg_opts The options for the token generator obj.
# min_freq -> All ngrams with frequency less than min_freq are
# ignored.
# stem_words -> If words should be transformed to their stems.
= list(
tg_opts "stem_words" = F,
"min_freq" = -1
)
The wordpredictor package allows evaluating n-gram model performance. It can measure the performance of a single model as well as compare the performance of multiple models. When evaluating the performance of a model, intrinsic and extrinsic evaluation is performed.
Intrinsic evaluation measures the Perplexity score for each sentence in a validation text file. It returns the minimum, maximum and mean Perplexity score for the sentences.
Extrinsic evaluation measures the accuracy score for the sentences in a validation text file. It tries to predict the last word of each sentence. If the word was correctly predicted, then the accuracy count is increased. The extrinsic evaluation returns the number of valid and invalid predictions.
The following example shows how to evaluate the performance of a model:
# The required files
<- c("def-model.RDS", "validate.txt")
rf # The test environment is setup
<- setup_env(rf, ve)
ed
# The model file name
<- paste0(ed, "/def-model.RDS")
mfn # The validation file name
<- paste0(ed, "/validate.txt")
vfn # ModelEvaluator class object is created
<- ModelEvaluator$new(mf = mfn, ve = 2)
me # The performance evaluation is performed. The performance stats are returned as
# a data frame and also saved within the model file itself.
<- me$evaluate_performance(lc = 20, fn = vfn)
stats
# The test environment is cleaned up
clean_up(ve)
The wordpredictor package includes a demo called “word-predictor”. The demo is a Shiny application that displays the ten most likely words for a given set of words. To access the demo, run the following command from the R shell:
demo("word-predictor", package = "wordpredictor", ask = F)
.
The following is a screenshot of the demo:
The wordpredictor website provides details about how the package works. It includes code samples and details of all the classes and methods.
The wordpredictor package provides an easy to use framework for working with n-gram models. It allows n-gram model generation, performance evaluation and word prediction.
The n-gram language model requires a lot of memory for storing the n-grams. The wordpredictor package has been tested on a machine with dual core processor and 4 GB of RAM. It works well for input data files of size less than 40 Mb and n-gram size 4. For larger data files and n-gram size, more memory and CPU power will be needed.
The wordpredictor package may be extended by adding support for different smoothing techniques such as Good-Turing, Katz-Back-off and handling of Out Of Vocabulary Words.
Support for different types of n-gram models such as Skip-Grams and Syntatic n-grams.
The wordpredictor package is used for predicting words. It may be extended to support other use cases such as spelling correction, biological sequence analysis, data compression and more. This will require further performance optimization.
The source code is organized using R6 classes. It is easy to extend. Contributions are welcome !.
I was motivated to develop the wordpredictor package after taking the courses in the Data Science Specialization offered by John Hopkins university on Coursera. I would like to thank the course instructors for making the courses interesting and motivating for the students.
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.
Health stats visible at Monitor.