incubator-cassandra-user mailing list archives

> "What CL are you using ?"
>
> I think this can be what causes the issue. I'm writing and reading at CL ONE. I didn't
drain before stopping Cassandra and this may have produce a fail in the current counters (those
which were being written when I stopped a server).
My first thought is to use QUOURM. But with only two nodes it's hard to get strong consistency
using QUOURM.
Can you try it thought, or run a repair ?
> But isn't Cassandra suppose to handle a server crash ? When a server crashes I guess
it don't drain before...
I was asking to understand how you did the upgrade.
Cheers
-----------------
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 1/11/2012, at 11:39 AM, Alain RODRIGUEZ <arodrime@gmail.com> wrote:
> "What version of cassandra are you using ?"
>
> 1.1.2
>
> "Can you explain this further?"
>
> I had an unexplained amount of reads (up to 1800 r/s and 90 Mo/s) on one server the other
was doing about 200 r/s and 5 Mo/s max. I fixed it by rebooting the server. This server is
dedicated to cassandra. I can't tell you more about it 'cause I don't get it... But a simple
Cassandra restart wasn't enough.
>
> "Was something writing to the cluster ?"
>
> Yes we are having some activity and perform about 600 w/s.
>
> "Did you drain for the upgrade ?"
>
> We upgrade a long time ago and to 1.1.2. This warning is about the version 1.1.6.
>
> "What changes did you make ?"
>
> In the cassandra.yaml I just change the "compaction_throughput_mb_per_sec" property to
slow down my compaction a bit. I don't think the problem come from here.
>
> "Are you saying that a particular counter column is giving different values for different
reads ?"
>
> Yes, this is exactly what I was saying. Sorry if something is wrong with my English,
it's not my mother tongue.
>
> "What CL are you using ?"
>
> I think this can be what causes the issue. I'm writing and reading at CL ONE. I didn't
drain before stopping Cassandra and this may have produce a fail in the current counters (those
which were being written when I stopped a server).
>
> But isn't Cassandra suppose to handle a server crash ? When a server crashes I guess
it don't drain before...
>
> Thank you for your time Aaron, once again.
>
> Alain
>
>
>
> 2012/10/31 aaron morton <aaron@thelastpickle.com>
> What version of cassandra are you using ?
>
>> I finally restart Cassandra. It didn't solve the problem so I stopped Cassandra
again on that node and restart my ec2 server. This solved the issue (1800 r/s to 100 r/s).
> Can you explain this further?
> Was something writing to the cluster ?
> Did you drain for the upgrade ? https://github.com/apache/cassandra/blob/cassandra-1.1/NEWS.txt#L17
>
>> Today I changed my cassandra.yml and restart this same server to apply my conf.
>
> What changes did you make ?
>
>> I just noticed that my homepage (which uses a Cassandra counter and refreshes every
sec) shows me 4 different values. 2 of them repeatedly (5000 and 4000) and the 2 other some
rare times (5500 and 3800)
> Are you saying that a particular counter column is giving different values for different
reads ?
> What CL are you using ?
>
> Cheers
>
> -----------------
> Aaron Morton
> Freelance Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 31/10/2012, at 3:39 AM, Jason Wee <peichieh@gmail.com> wrote:
>
>> maybe enable the debug in log4j-server.properties and going through the log to see
what actually happen?
>>
>> On Tue, Oct 30, 2012 at 7:31 PM, Alain RODRIGUEZ <arodrime@gmail.com> wrote:
>> Hi,
>>
>> I have an issue with counters, yesterday I had a lot of ununderstandable reads/sec
on one server. I finally restart Cassandra. It didn't solve the problem so I stopped Cassandra
again on that node and restart my ec2 server. This solved the issue (1800 r/s to 100 r/s).
>>
>> Today I changed my cassandra.yml and restart this same server to apply my conf.
>>
>> I just noticed that my homepage (which uses a Cassandra counter and refreshes every
sec) shows me 4 different values. 2 of them repeatedly (5000 and 4000) and the 2 other some
rare times (5500 and 3800)
>>
>> Only the counters made today and yesterday are concerned.
>>
>> I performed a repair without success. These data are the heart of our business so
if someone had any clue on it, I would be really grateful...
>>
>> The sooner the better, I am in production with these random counters.
>>
>> Alain
>>
>> INFO:
>>
>> My environnement is 2 nodes (EC2 large), RF 2, CL.ONE (R & W), Random Partitioner.
>>
>> xxx.xxx.xxx.241 eu-west 1b Up Normal 151.95 GB 50.00%
0
>> xxx.xxx.xxx.109 eu-west 1b Up Normal 117.71 GB 50.00%
85070591730234615865843651857942052864
>>
>> Here is my conf: http://pastebin.com/5cMuBKDt
>>
>>
>>
>
>