After almost an hour of button clicking i was finally able to report all the tasks on my second machine.
Niether is able to get any new work- every single request resulsts in a Scheduler Timeout. I expect to run out of work completely in the next 40 min on one system, the other later today.

Please, please, please can someone let the satff know that limiting the number of tasks hasn't helped in the slightest. When it does start to help- it will only be becasue everyone is out of work.
Until the Scheduler is fixed they need to stop all AP production & distribution. They need to fix the Scheduler problem.

EDIT- this problem only started 3 (or was it 4?) weeks ago ofter the weekly outage. Whatever changes they did then to cause the problem, please undo them.
____________
Grant
Darwin NT.

Task limits may be helping a little.
BUT, stopping AP production did nothing over the weekend, I was still suffering server side time-outs and very slow delivery with no AP production, not only that the number of ghosts in my possession went up, from less than 5 to about 50.
____________
Bob Smith
Member of Seti PIPPS (Pluto is a Planet Protest Society)
Somewhere in the (un)known Universe?

Task limits may be helping a little.
BUT, stopping AP production did nothing over the weekend, I was still suffering server side time-outs and very slow delivery with no AP production, not only that the number of ghosts in my possession went up, from less than 5 to about 50.

Well, Rob, last AP unit was produced in Sunday, and late Monday/early Tuesday Cricket dropped down and hosts were getting their tasks.

This is free speech forum, so do Your own assumptions.
____________
"Please keep Your signature under four lines so Internet traffic doesn't go up too much"

Task limits may be helping a little.
BUT, stopping AP production did nothing over the weekend, I was still suffering server side time-outs and very slow delivery with no AP production, not only that the number of ghosts in my possession went up, from less than 5 to about 50.

I'm a little concerned to see you report this.

My ability to get to the scheduler went from nothing to everything.

Are you using a proxy and is it still working?

I run a lot of computers and they all seem to have-had a much different experience than you had.

Regarding my earlier post, Message 1306107, I will also add that I did notice the massive increase in tasks sent me in the days leading up to all our current problems. At the time, I thought it might be due to the shorties, but now I realize the schedulers were sending me all those tasks regardless of the estimated run time.

Perhaps I'm making an incorrect assumption here, but it would seem to me that, if the schedulers were sending thousands of hosts nearly three times the correct number of units, it would cause some of the problems we have been getting.

There have always been problems with getting uploads and downloads, and there have also been many times APs were sent out. The real disaster seems to be something newly introduced, along with the newly imposed limits of 100/CPU-100/GPU.

Please, let's get some discussion going on this observation of mine. Can anyone else find a similar history in your logs? It might not be readily apparent to the really big crunchers because you usually have problems keeping your caches full. What about others, though?

I know something went very seriously wrong after an update or maintenance period three or four weeks ago. Things went very well for a few hours and then I stopped getting work. Seti@Home has never been right since. My notebooks without dedicated graphics cards are getting work in and then back out fine. One has a RAC of about 700; another has a RAC of about 1250 and then a single core machine has a RAC of about 175. (That's an older Pentium machine about ten years old). My desktops with Nvidia GTX 550Ti Cards are the ones having a lot of trouble. My Core Duo with one Nvidia GTX 550 Ti is about to run out of work and is getting lots of time outs and some transient https failures. It has a RAC of about 7,400, but normally has been around 8,000 My i7 2600k machine with two Nvidia GTX 550's normally has a RAC of 22,000 to 24,000, but is down to about 18,000 now. Again, lots of time outs and no work available messages, transient https failures, and time outs as well as backoffs. I have had to play with the NNT button a lot with the desktops to keep them in work, but that tends to foul up the getting work routine. I also have a newer notebook with an i7-2670qm processor and an Nvidia 525m graphics processor. It tends to keep its queue full, but sometimes I've had to play with the NNT and update buttons. I think I get upset with it all and kinda forget for a few minutes that it runs pretty much ok if I leave it unattended, but just check in twice a day. It has a RAC of just about 5,000. I don't know for sure about the number of ghosts, but I believe both the desktops have a lot of them because I see a message every so often about recovering lost work or jobs. With the new notebook running at about 5,000 RAC I should have an overall RAC of about 37,000 or close, but I am running several thousand behind that even though we run 24/7. This is another indication to me that something might be pretty wrong in getting work credit or sending work back. I don't know what is wrong. I am not a technical or database expert, but like I said near the start of this message something very wrong happened with a maintenance cycle around 4 or 5 weeks ago and the problem has remained in both sending work out and getting new work. Brother Frank.

I can only get some work from S@H, I've got 40 that won't report, I could for a bit with a proxy get some work, but that went 403 after a day or so, which I assume is an html 403 which means forbidden, so most of the time I've been getting Milkyway work instead. Matt please hurry back, S@H has gone to hell in a handbag...
____________UnderDog, Pluto is a Planet to me.DK-US transportation network has fallen off the rails!

at the time of this posting, with 0 tasks to process and 0 tasks to report in BOINC Manager, i have 134 ghost tasks that have a deadline of 11/17 and 178 ghost tasks that have a deadline of 11/18. these tasks are allocated to mach xxxx033 at a rate of 20 per scheduler request.
i have only successfully contacted project 3 times today due to scheduler time outs, for a total of 60 cpu/gpu tasks downloaded and completed .. at this rate i will have approx. 14 error-ed time outs on 11/17 and 118 error-ed time outs on 11/18 (if my math is correct) :)
the last successful download i received was at 2:41 pm mst today. all uploads are painfully completed utilizing NNT.
____________

Can someone point me to a forum entry detailing how to limit number of files being reported. I've looked thru the boinc & seti XMLs and didn't see a likely entry. I know I've done this before but as they say "the memory is the second thing to go" I forget what's first!

Can someone point me to a forum entry detailing how to limit number of files being reported. I've looked thru the boinc & seti XMLs and didn't see a likely entry. I know I've done this before but as they say "the memory is the second thing to go" I forget what's first!

I got back from work. One system completely out of work, the other running out of CPU work (GPU ran out ages ago).
Every single Scheduler request while i was at work timed out.

I set No New Tasks & managed to report them all, although one machine took almost 5 min before it got a response. I then allowed new work & one machine then requested work & got some, the other requested work and got a Scheduler Timeout.

The graphs show AP work still going out. I'm quite sure if the AP work stopped going out, i'd then be able to contact the Scheduler again.

As someone pointed out some time ago- Synergy has a lot of processes, and i think it's got more than it can hamdle.
____________
Grant
Darwin NT.

Grant, in another thread you suggested using a proxy - have you considered that this problem may be a routing problem between your empty cruncher and the lab and that the cruncher concerned is attempting to use a different one to those whose connection is OK - I have to crunchers sharing the same internet connection, but when I do a tractroute they show they actually, consistently, use different paths through the internet to almost every server I look at....
____________
Bob Smith
Member of Seti PIPPS (Pluto is a Planet Protest Society)
Somewhere in the (un)known Universe?

Grant, in another thread you suggested using a proxy - have you considered that this problem may be a routing problem between your empty cruncher and the lab and that the cruncher concerned is attempting to use a different one to those whose connection is OK

I used the proxy for both systems.
Proxy- both get a response, no proxy- neither get a response when trying to report & request work. No proxy- both sometimes get work if not reporting tasks.

So we'll go with the proxy, till it gets canned & i have to find another one.
____________
Grant
Darwin NT.

My requests in the past 24 hours have been about 50/50 time-out and success. Most of the time-outs end up with a "resent lost task" on the next successful contact. This has been how it has been working all day and evening, until I managed to be issued an AP with a six minute deadline. Well there goes my consecutive valid streak of ~1100. "didn't resend lost task..(expired)" *sigh*
____________
Linux laptop:
Record uptime: 1484d 22h 42m
Ended due to UPS failure, as discovered 14 hours later

Even with all the wierdness going on, i've yet to have that happen- not even once.
You haven't got any software that mucks about with your system clock at all?

No, I don't. Like I said, it happened on the server side! My host didn't even succeed in making a scheduler request. This is the second time it happened. I'm not the only one this happened to; there were other users reporting the same thing.
____________