Book Image

R Data Analysis Projects

Book Image

R Data Analysis Projects

Overview of this book

R offers a large variety of packages and libraries for fast and accurate data analysis and visualization. As a result, it’s one of the most popularly used languages by data scientists and analysts, or anyone who wants to perform data analysis. This book will demonstrate how you can put to use your existing knowledge of data analysis in R to build highly efficient, end-to-end data analysis pipelines without any hassle. You’ll start by building a content-based recommendation system, followed by building a project on sentiment analysis with tweets. You’ll implement time-series modeling for anomaly detection, and understand cluster analysis of streaming data. You’ll work through projects on performing efficient market data research, building recommendation systems, and analyzing networks accurately, all provided with easy to follow codes. With the help of these real-world projects, you’ll get a better understanding of the challenges faced when building data analysis pipelines, and see how you can overcome them without compromising on the efficiency or accuracy of your systems. The book covers some popularly used R packages such as dplyr, ggplot2, RShiny, and others, and includes tips on using them effectively. By the end of this book, you’ll have a better understanding of data analysis with R, and be able to put your knowledge to practical use without any hassle.
Table of Contents (15 chapters)
Title Page
Credits
About the Author
About the Reviewer
www.PacktPub.com
Customer Feedback
Preface

Summary


We started this chapter with a discussion about the KDE and its usefulness in understanding the underlying distribution of data. We proceeded by explaining how to extract tweets from Twitter for a given search string in R. Then, we proceeded to explain the sentiment ming, dictionary, and machine learning approaches. Using a dictionary approach, we calculated the sentiment scores for the tweets. We further explained text pre-processing routines required to prepare the text data. We covered weighting schemes for creating document term matrixes. We discussed the classic tfidf and the new Delta TFIDF schemes. We created our training set using the Delta TFIDF scheme. Using this training set, we finally built a Naive Bayes KDE classifier to classify tweets based on the sentiment the text carried.

In the next chapter, we will be working on Record Linakage. A master data management technique to do data dedpulication.