Then we have the question on when data (even big-data) can still be handled in a standard relational database and can still be handled by a "standard" approach. There are some guidelines that can help you. Please do note this is a comparison primarily for handling data in a relational database or in Hadoop. This is not for storing data.
RDBMS | Hadoop / MapReduce | |
Data Size | Gigabytes | Petabytes |
Access | Interactive and batch | Batch |
Structure | Fixed Schema | Unstructured schema |
Language | SQL | Procedural (Java, C++, Ruby, etc.) |
Integrity | High | Low |
Scaling | nonlinear | linear |
Updates | Read and Write | Write ones, read many times |
Latency | Low | High |
By taking this into consideration when you are struggling with the question if you need to use a MapReduce approach or a RDBMS approach it might be a little more easy to make your decision.
No comments:
Post a Comment