Cookies help us display personalized product recommendations and ensure you have great shopping experience.

By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
SmartData CollectiveSmartData Collective
  • Analytics
    AnalyticsShow More
    data analytics
    How Data Analytics Can Help You Construct A Financial Weather Map
    4 Min Read
    financial analytics
    Financial Analytics Shows The Hidden Cost Of Not Switching Systems
    4 Min Read
    warehouse accidents
    Data Analytics and the Future of Warehouse Safety
    10 Min Read
    stock investing and data analytics
    How Data Analytics Supports Smarter Stock Trading Strategies
    4 Min Read
    predictive analytics risk management
    How Predictive Analytics Is Redefining Risk Management Across Industries
    7 Min Read
  • Big Data
  • BI
  • Exclusive
  • IT
  • Marketing
  • Software
Search
© 2008-25 SmartData Collective. All Rights Reserved.
Reading: Why We Need to Deal with Big Data in R
Share
Notification
Font ResizerAa
SmartData CollectiveSmartData Collective
Font ResizerAa
Search
  • About
  • Help
  • Privacy
Follow US
© 2008-23 SmartData Collective. All Rights Reserved.
SmartData Collective > Analytics > Why We Need to Deal with Big Data in R
AnalyticsR Programming Language

Why We Need to Deal with Big Data in R

DavidMSmith
DavidMSmith
3 Min Read
SHARE

Responding to the birth rates analysis in the post earlier this week on big-data analysis with Revolution R Enterprise, Luis Apiolaza asks at the Quantum Forests blog, do we really need to deal with big data in R?

Responding to the birth rates analysis in the post earlier this week on big-data analysis with Revolution R Enterprise, Luis Apiolaza asks at the Quantum Forests blog, do we really need to deal with big data in R?

My basic question is why would I want to deal with all those 100 million records directly in R? Wouldn’t it make much more sense to reduce the data to a meaningful size using the original database, up there in the cloud, and download the reduced version to continue an in-depth analysis?

As Luis points out (and as most of us know from experience), 90% of statistical data analysis is data preparation. Many “big data” problems are in fact analyses of small data sets, that have been carefully (and often painfully) extracted from a data store we’d refer to today as “Big Data”. And while we could use another tool to do that extraction, personally I’d prefer to do it in R myself. Not just because needing access to another tool probably means delays, authorizations, and probably having to ask a DBA nicely, but also because the extraction process itself (in my opinion) requires a certain level of statistical expertise.

More Read

Image
4 Use Cases for Embedded Analytics Which Boost Business Performance
No Wait in Kuwait – But Some Weight (Part 2)
Investing in smarter infrastructure will create more than 949,000 new jobs in 2009
Choosing the Right Programming Language for A Corporate Database
Merging Big Data and Cloud Computing is the Future of Robotics

For me, at least, it’s often an iterative process of identifying the variables I need, the right way to do the aggregation/smoothing/dimension reduction, how to handle missing values and data quality issues … the list goes on and on. To be able to extract from a large data set using the R language alone is a great boon — especially when the source data set is very large. That’s why we created the rxDataStep function in RevoScaleR. (You can read more about rxDataStep in our new white paper, The RevoScaleR Data Step White Paper.)

Then again, some statistical problems simply do require analysis of very large datasets. wholesale. Some of the commenters to Luis’s post provide their own examples, and Revolution Analytics’ CEO Norman Nie has written a white paper identifying five situations where analysis of large data sets in R is useful:

  1. Use Data Mining to Make Predictions
  2. Make Predictive Models More Powerful
  3. Find and Understand Rare Events
  4. Extract and Analyze ‘Low Incidence Populations’
  5. Avoid Dependence on ‘Statistical Significance’

You can read Norman’s explanations of these uses of Big Data in the white paper, The Rise of Big Data Spurs a Revolution in Big Analytics, available for download at the link below.

Revolution Analytics White Papers: The Rise of Big Data Spurs a Revolution in Big Analytics

TAGGED:big data
Share This Article
Facebook Pinterest LinkedIn
Share

Follow us on Facebook

Latest News

ai kids and their parents
How Cities Use AI to Improve Playground Design
Exclusive News
human resource data
The Integration of Employee Experience with Enterprise Data Tools
Big Data Exclusive
protecting patient data
How to Protect Psychotherapy Data in a Digital Practice
Big Data Exclusive Security
data analytics
How Data Analytics Can Help You Construct A Financial Weather Map
Analytics Exclusive Infographic

Stay Connected

1.2KFollowersLike
33.7KFollowersFollow
222FollowersPin

You Might also Like

big data approaches to marketing
Big DataExclusive

Can Big Data Approaches To Marketing Slash Business Closure Rates?

8 Min Read
reasons to become a data analyst
Data Science

Top 5 Reasons You Should Become a Data Analyst

7 Min Read
big data and Hadoop guide
AnalyticsBig DataExclusiveHadoopSoftware

How Big Data and Hadoop Training Programs Can Make a Big Difference

5 Min Read

How Big Data is Creating the Future of Science Fiction

4 Min Read

SmartData Collective is one of the largest & trusted community covering technical content about Big Data, BI, Cloud, Analytics, Artificial Intelligence, IoT & more.

ai in ecommerce
Artificial Intelligence for eCommerce: A Closer Look
Artificial Intelligence
ai is improving the safety of cars
From Bolts to Bots: How AI Is Fortifying the Automotive Industry
Artificial Intelligence

Quick Link

  • About
  • Contact
  • Privacy
Follow US
© 2008-25 SmartData Collective. All Rights Reserved.
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?