Data Craftsman
2012-02-15 18:40:24 UTC
Hello experts,
Based on this blog of Basic Time Series with Cassandra data modeling,
http://rubyscale.com/blog/2011/03/06/basic-time-series-with-cassandra/
"This (wide row column slicing) works well enough for a while, but over
time, this row will get very large. If you are storing sensor data that
updates hundreds of times per second, that row will quickly become gigantic
and unusable. The answer to that is to shard the data up in some way"
There is a limit on how big the row size can be before slowing down the
update and query performance, that is 10MB or less.
Is this still true in Cassandra latest version? or in what release
Cassandra will remove this limit?
Manually sharding the wide row will increase the application complexity, it
would be better if Cassandra can handle it transparently.
Thanks,
Charlie | DBA & Developer
p.s. Quora link,
http://www.quora.com/Cassandra-database/What-are-good-ways-to-design-data-model-in-Cassandra-for-historical-data
Based on this blog of Basic Time Series with Cassandra data modeling,
http://rubyscale.com/blog/2011/03/06/basic-time-series-with-cassandra/
"This (wide row column slicing) works well enough for a while, but over
time, this row will get very large. If you are storing sensor data that
updates hundreds of times per second, that row will quickly become gigantic
and unusable. The answer to that is to shard the data up in some way"
There is a limit on how big the row size can be before slowing down the
update and query performance, that is 10MB or less.
Is this still true in Cassandra latest version? or in what release
Cassandra will remove this limit?
Manually sharding the wide row will increase the application complexity, it
would be better if Cassandra can handle it transparently.
Thanks,
Charlie | DBA & Developer
p.s. Quora link,
http://www.quora.com/Cassandra-database/What-are-good-ways-to-design-data-model-in-Cassandra-for-historical-data