21) Message boards : Number crunching : Running out of WUs (Message 25065)
Posted 9 Jan 2013 by Christoph
Post:
I found some stuck WUs which mean the number on the server status page may be wrong.
Will report these to Eric Mcintosh the sientist which gave us all the work last year and will have more hopefully end of this month.

With 'stuck' I mean they were validated but there is one more result which was propably late. That one is still waiting for validation. It needs to be re-checked or manualy removed.

Example
22) Message boards : Number crunching : Running out of WUs (Message 25063)
Posted 8 Jan 2013 by Christoph
Post:
They should still be needed. Since I'm a volunteer like you I cannot do anything.
All I did was to sent a PM to Eric whom is one of the project sientist.

Since CERN is closed I guess this will only be adressed after re-opening.
Eric or one of the other scientist will propably be able to shorten the deadline of these units for that they time out and get re-assinged.
If they feel the need for it. If the BOINc server is working correctly it should sort itself out when the deadline is reached.

EDIT: If you know a user how have one of the units you can try to PM this one. Maybe he/she has the email notification turned on and will get the message. If it will help? No idea. Depends to that person.
23) Message boards : Number crunching : Running out of WUs (Message 25053)
Posted 6 Jan 2013 by Christoph
Post:
Have a look at the news. It is a bit short but Eric hit several things which hold him up.
His PC broke down. Really bad without the bakups all would be gone.

Then CERN is closing their doors always one time per year for two weeks.
That is right now so nothing to expect for the next week.

After that we shall see how long it takes the crew to get new data flowing.
But it may be that if Eric has his current studies complete that he will first do the analysis and only then create new WUs with adjusted parameters based on the studies.

If that is the case it may take another.....eh.....hopefully not too long. No idea how long such analysis take.
24) Message boards : Cafe LHC : Merry Christmas and a happy new year! (Message 25043)
Posted 1 Jan 2013 by Christoph
Post:
Happy new year from Germany.
25) Message boards : Cafe LHC : ~~~Last Person To Post Wins~~~ (Message 25042)
Posted 1 Jan 2013 by Christoph
Post:
Happy New Year!!
I think it's my first post in such a thread.
26) Message boards : News : Problems/Status 28th November, 2012 and PAUSE (Message 25004)
Posted 11 Dec 2012 by Christoph
Post:
Due to memory consumtion there is actually no option to run more then one task or more then one virtual machine. When the new wraper goes mainstream that might change. There was a setup instruction how to set up several virtual machines yourself and then run more tasks at the same time. But that discussion somehow disappeared.
27) Message boards : Number crunching : Sorry to have to return those great wu's (Message 24962)
Posted 18 Nov 2012 by Christoph
Post:
Well there was a news item that there would be electric works and therefore a powercut.

They planned to run on their UPSs during that outage but propably it turned out to be longer than expected.

Normal plan was power returning (around) 0800 UTC saturday morning.
28) Message boards : News : Status, Thursday 15th November (Message 24958)
Posted 16 Nov 2012 by Christoph
Post:
Hiccup; mea culpa. On vacation and travelling since Tuesday
and ran out of disk space in BOINC buffer at CERN :-(
I think all is OK again now after corrective actions and more work
is on the way. Sorry about that. Eric.


Thank you for taking time during your vacation. For the future I hope that you can hand over your duties to somebody else during future vacations for that they are REALY vacations.

I hope that fixing and giving us work is the only work related thing you are doing during this vacation.
29) Message boards : Number crunching : WU not being resent to another user (Message 24935)
Posted 5 Nov 2012 by Christoph
Post:
You are welcome. Good to see that my brain still produce good explanations when it is already nearly sleeping.

@Eric: You wrote in your status update that you are the feeder. Ok.
I just deleted my last five minutes work on this post because I recon I took a wrong approach.
Not the feeder is doing the amount of work being available to the users to download on other projects but the splitters.
At least at SETI and Einstein. But then they also do have raw data files which needs pre-processing and need to be split up into workunits to be send out.
That means we don't realy talk bout the feeder process. But the splitter which are not used here.

So then the question for Eric and crew is: Is there a way for you to serve the work to the (maybe) in BOINC Server existing standard splitter / work generator for that it serves the feeder with a limited amount of work where then the resends will be lined up as they appear?

Maybe you ask at the Collatz project for help if necessary. They have a 'work generator' called process and also don't need a splitter.

Hope this gets us a bit closer to reduce the problem of the tail.
30) Message boards : Cafe LHC : Does Sixtrack project makes difference in CERN ?? (Message 24920)
Posted 3 Nov 2012 by Christoph
Post:
Yes. Example: http://lhcathomeclassic.cern.ch/sixtrack/forum_thread.php?id=3536
The author of this message is (as I understand it) the leading scientist when it comes to sixtrack. He is the driving force behind it.

About once a month he is writing a status message in the news section. On newer BOINc clients (7.xx series) these will also be displayed in the message tab.

His last news: http://lhcathomeclassic.cern.ch/sixtrack/forum_thread.php?id=3537
31) Message boards : Number crunching : WU not being resent to another user (Message 24919)
Posted 3 Nov 2012 by Christoph
Post:
In my previous post I gave an explaination of the situatuion the way I remember and understood it reading older posts made by project staff (not necessarily Eric).

It was my intention to help other people to understand the situation and keep patient.

It is nice to see that there is a discussion evolving out of this.

I personally don't care if resends are at the end of the que. If it is helpful for Eric to wait one day or two after one batch completed to give time for the resends to return and complete the study that is also fine with me.

As Jujube said already: we are here to cruch what the project gives us. We don't care if it is one study or three at a time. And I don't realy care about credits. I did also sign up at the BOINC Vbox wrapper test project which will never tell anybody about the earned credits.

About Jujube's point that we need to calculate X cycles to complete one study and that the resends in the middle will delay other work: It is up to Eric. If he can 'sit out' one day or two (Or what ever time frame needed) of crying for new work in the message boards and propably in his message box until also the resends are returned than leave everything as it is.

If he wants to keep the time without work available as short as possible / non existend then the continious feeding could be an option for him because the resends will be trickled in and the studies completed faster then if he keep feeding us batches and some old resends linger around maybe still after the 2nd or 3rd following studie.

That is also the reason why I personally think that the low water / high water feeder system is a good idea.

I hope I didn't forget anything what I planned to comment on. Been a bit long working day and my brain don't like to work any longer.

Looking forward to what will come out of this.
32) Message boards : Number crunching : WU not being resent to another user (Message 24918)
Posted 3 Nov 2012 by Christoph
Post:
If one has 6 studies, and one runs all six in parallel
not one will be complete for 6 weeks say; if one runs them one after
the other then one is finished in 1 week, one in two weeks, etc.
Which is better? Eric.


Better for you? It depends. If your hands are tied until all 6 are complete then it doesn't matter whether you run them in parallel or sequentially. If the data produced by any given study allows you to make progress in some other urgent endeavor then maybe sequential is better. If you're being pulled in a thousand different directions on any given day then it doesn't matter because you never have nothing to do anyway and would still be swamped even if you could work 24/7/365 like a robot.

Better for us? Doesn't matter what's better for us. We're here to donate and do what's best for you and the science. That's all that matters.


+1
33) Message boards : Number crunching : WU not being resent to another user (Message 24907)
Posted 29 Oct 2012 by Christoph
Post:
The 'problem' of this project is that it is not running a feeder which produces the tasks when the que reaches a low water mark and stop it when the high water mark is raeched.
In that case the resends would be somewhere trickled in in the que as they are produced and get crunched when it is their turn to get send out.

Another problem as I take it is the BOINC server code.
It should give the resends a higher priority to get send out earlier.
But it looks like it is not working proper here at LCH@home. For what ever reason.

Since the work gets submitted in bigger batches the resends stack up behind the normal work and are crunched only after all other jobs got done.

Over to Eric to clear up anything which I didn't explain the right way.
34) Message boards : Number crunching : Maximum elapsed time exceeded (Message 24897)
Posted 23 Oct 2012 by Christoph
Post:
Hi Dennis,

I have reported a problem which could be the source of your error WUs. Are you familiar with the cc_config file? I so please activate task_debug.

I did recon on other projects that the GPU apps sometimes fail to report/BOINC fail to recognise when a task complete and it continue to run after reaching 100%.

As mentioned above I send a report to the BOINC Alpha mailing list. I you are willing to help and have this error several times a week you could help the developers to track the bug down a lot!!
35) Message boards : Number crunching : Long WU's (Message 24821)
Posted 13 Sep 2012 by Christoph
Post:
I'm sorry, a 4-day deadline for estimated 100 hours of runtime is too short for my computer. I aborted the task 2813528.

Bye, Grubix.


These WUs are not the standard, mostly you will get shorter ones.
My 100 hours estimate completed after about 39 hours.
I don't know if it hit the wall or completed succesful.
36) Message boards : Number crunching : WU not being resent to another user (Message 24800)
Posted 8 Sep 2012 by Christoph
Post:
I have one of the wlxu7. Nearly 100 hours estimation. Not a re-send.
37) Message boards : News : Very long jobs (Message 24723)
Posted 26 Aug 2012 by Christoph
Post:
One thing other projects do is to bundle up to 8 tasks into one to produce long running WUs.

The long ones are thought especially for GPUs since they cut down the processing time a lot.

Reference for that is Collatz@home. They have 'mini' collatz. That is just one WU.
38) Message boards : News : Status/Plans, 7th August 2012 (Message 24554)
Posted 9 Aug 2012 by Christoph
Post:
In case that you get this problem again in the future, I think somewhere I read that it is also possible to extend the deadline for already sent out tasks.
Propably it was at SETI@home after some of their unplanned-couple-of-days-long outages.
They extended the dealines so that it was possible to report the already overdue tasks.

Sorry that I didn't remembered earlier about that.
39) Message boards : Number crunching : No Tasks ??? (Message 24442)
Posted 25 Jul 2012 by Christoph
Post:
Great news Eric, Thank You!

Now imagine, if you send an EMail to all users (or all inactive last 30 days) and only half of the currently inactive are coming back.......

I guess you should increase the available discspace before that message to not run out of it again ;-)

AND you get it to produce a GPU app. (Dreaming on? Or not?)
40) Message boards : Number crunching : Request for SSE4 and/or OpenCL applications (Message 24441)
Posted 25 Jul 2012 by Christoph
Post:
Oh yea, that part I forgot to mention. I have a small piece of timber under it.


Previous 20 · Next 20


©2024 CERN