Function to do sentiment analysis (on twitter data for example)

# based on

score.sentiment = function(sentences, pos.words, neg.words, .progress=’none’)

# we got a vector of sentences. plyr will handle a list or a vector as an “l” for us
# we want a simple array of scores back, so we use “l” + “a” + “ply” = laply:
scores = laply(sentences, function(sentence, pos.words, neg.words) {

# clean up sentences with R’s regex-driven global substitute, gsub():
sentence = gsub(‘[[:punct:]]’, ”, sentence)
sentence = gsub(‘[[:cntrl:]]’, ”, sentence)
sentence = gsub(‘\\d+’, ”, sentence)
# and convert to lower case:
sentence = tolower(sentence)

# split into words. str_split is in the stringr package
word.list = str_split(sentence, ‘\\s+’)
# sometimes a list() is one level of hierarchy too much
words = unlist(word.list)

# compare our words to the dictionaries of positive & negative terms
pos.matches = match(words, pos.words)
neg.matches = match(words, neg.words)

# match() returns the position of the matched term or NA
# we just want a TRUE/FALSE:
pos.matches = !
neg.matches = !

# and conveniently enough, TRUE/FALSE will be treated as 1/0 by sum():
score = sum(pos.matches) – sum(neg.matches)

}, pos.words, neg.words, .progress=.progress )

scores.df = data.frame(score=scores, text=sentences)

Analysing twitter data with twitteR package

R code:


#Log in to twitter (you can find these details on
#Replace string between “” with the codes from your own twitter account.
consumer_key <- “Consumer Key (API Key)”
consumer_secret <- “Consumer Secret (API Secret)”
access_token <- “Access Token”
access_secret <- “Access Token Secret”

setup_twitter_oauth(consumer_key, consumer_secret, access_token, access_secret)

#Some examples
userTimeline(“utwente”, n=10) # search for tweets of specific user
searchTwitter(“utwente”, n = 10) # search for keywords

#Discover twitteR object
tweetsList <- searchTwitter(“utwente”, n = 10)
tweet <- tweetsList[[1]]

#Harvest tweets based on keyword.
mach_tweets <- searchTwitter(“#prayforparis”, n=1500, lang=”en”)

#Extract the text from the tweets in a vector
#See for an
#approach in Windows.
mach_text <- sapply(mach_tweets, function(x) x$getText())
mach_text <- iconv(mach_text, to = “utf-8-mac”)

###Some initial cleaning
# Remove URLs
mach_text <- gsub(“(f|ht)(tp)(s?)(://)(.*)[.|/](.*)”, “”, mach_text, = TRUE)

# Remove @UserName
#mach_text <- gsub(“@\\w+”, “”, mach_text)

# Create a corpus
mach_corpus <- Corpus(VectorSource(mach_text))

# create document term matrix applying some transformations
tdm <- TermDocumentMatrix(mach_corpus,
control = list(removePunctuation = TRUE,
stopwords = c(“prayforparis”, “paris”, “http”, “https”, stopwords(“english”)),
removeNumbers = TRUE, tolower = TRUE))
## further exploration termd document matrix
#frequent words
findFreqTerms(tdm, lowfreq = 100)

findAssocs(tdm, terms = “syria”, corlimit = 0.3)

## define tdm as matrix
tdMatrix <- as.matrix(tdm)

# get word counts in decreasing order
word_freqs <- sort(rowSums(tdMatrix), decreasing=TRUE)

# create a data frame with words and their frequencies
df <- data.frame(word=names(word_freqs), freq=word_freqs)

# plot wordcloud
wordcloud(df$word, df$freq, random.order=FALSE, colors=brewer.pal(8, “Dark2”), min.freq = 20)
####Sentiment analyses####
# based on
# download Opinion Lexicon (Hu and Liu, KDD-2004)

listPosWords <- scan(“positive-words.txt”, what = “character”, comment.char = “;”)
listNegWords <- scan(“negative-words.txt”, what = “character”, comment.char = “;”)

sentScoreTweets <- score.sentiment(mach_text, listPosWords, listNegWords, .progress = “text”)

Making plots using ggplot

In the February ’14 meeting of the Twente R User group, research master student Karel Kroeze gave us a quick run-through of the many possibilities with the ggplot family of functions and building blocks. Although a bit intimidating at first, the R code is immensely flexible once you get the hang of it. Have a go at his code (click here to get the code) from the presentation and see if you can accommodate his code for your own cool graphics! For the accompanying data files, mail Stephanie. Guus provided some extra helpful links, see Tim Churches, CEB institute and cookbook R.

First TRUG meeting

A sunny sky and a not-too-fresh breeze marked the beginning of the Twente R User Group. While enjoying drinks and tapas in the lush Twente countryside, both students and staff from several departments of the University of Twente met and exchanged experiences and ambitions regarding their use of the statistical programming software R.

The level of experience and type of ambitions varied greatly. Some hardcore programmers want to familiarize themselves with R to broaden their horizon, some use R for Monte Carlo simulation, some merely want to use a platform for data manipulation and analysis that does not require an expensive license, and some use R because it can do types of analyses that can’t be carried out using SPSS. Several members in the group were interested in parallelization and integration with C++ code to speed up analyses.

Martin Schmettow was the first to present. Martin, working at the department of Cognitive Psychology and Ergonomics, presented how to deal with card sorting data, showing some code how to construct graph data from the card sorting data (represented as lists of lists) using a recursive function, and how to best present analysis results: hot-spot plots rather than dendograms. Several libraries and their functions were discussed, like for instance plyr. Powerpoint slides can be found here: Cardsorting Workflow.

The User Group will meet monthly, alternating days and times so that more people are able to join, even if they can only attend bi-monthly. Next presenter, July 11, will be Inga Schwabe, from the Department of Research Methodology, Measurement and Data Analysis. July 11, 1700 hrs, Weleweg 410, Boekelo.