- Links
- Refcardz
- Guides
- About
- Post
- Article
- Link
- Code Snippet
- Announcement
This is SAS's view on big data. The article discusses how big data can be used to take better decisions, cut costs, and gain advantages.
The title of the post is a bit long, but that’s the problem I was facing this morning: importing datasets from files, online. I mean, it was not a “problem”, more a challenge (I should be able to do it in R, directly)
Mazama Science has just finished creating class materials on using R for the AirFire team at the USFS Pacific Wildland Fire Sciences Lab in Seattle, Washington. Autodidacts new to R should take about 20-30 hrs to complete the course.
Make sure you didn't miss anything with this list of the Best of the Week in the Big Data Zone (October 31 - November 07). This week's topics include getting started with Hadoop and MapReduce, data structural integrity, Big Data goals, prediction intervals, JSF versus JSP with CRUD applications.
This article represents some of the basic concepts you need to understand in order to write a Hello world using the R programming language.
Big data has a large role to play in the real estate industry.
I’ve been doing some work with Focused Objective lately, and today the following question came up in our discussion. If you’re sampling from a uniform distribution, how many samples do you need before your sample range has an even chance of covering 90% of the population range?
We make decisions every day; everything we say and do is the result of a decision, whether we make it consciously or not. No matter how big or small is the choice, there's no (easy) formula for making the right decision.
Don’t be afraid of these new applications. They are coming whether you like it or not. Embrace them, understand them as best you can. Then sit back and think about what the network can do for them. You have an ability to significantly impact their ability to perform.
Salesforce the world’s largest enterprise cloud computing company has recently unveiled “Wave”, their new enterprise business intelligence solution.
Hadoop MapReduce framework provides a way to process large data, in parallel, on large clusters of commodity hardware. An edit to an earlier version.
While there is certainly much feel-good hyperbole about the “making the world a better place” nature of big data, that is more than offset with actual real-world details of how data is being used to solve more day-to-day business problems.
Can an image capture an entire system's structural integrity? Can we tell from a graphic whether a system is well-structured? The Blighttown corollary highlights the importance of a good package structure, as this structure will probably constrain the quality of the entire system's structure.
Almost all prediction intervals from time series models are too narrow. This is a well-known phenomenon and arises because they do not account for all sources of uncertainty. When we produce prediction intervals for time series models, we generally only take into account the first of these sources of uncertainty.
Make sure you didn't miss anything with this list of the Best of the Week in the Big Data Zone (October 24 - October 31). This week's topics include Twitter data analysis, running Hadoop on Ubuntu, information retrieval with Apache Lucene, a method for data visualization, and removing references with R.
I’ve been playing around with Python over the last few days while cleaning up a data set and one thing I wanted to do was translate date strings into a timestamp.
The hts package for R allows for forecasting hierarchical and grouped time series data. The idea is to generate forecasts for all series at all levels of aggregation without imposing the aggregation constraints, and then to reconcile the forecasts so they satisfy the aggregation constraints.
In this post my aim is to get Hadoop up and running on a Ubuntu host using Local (Standalone) Mode and on Pseudo-Distributed Mode.