1) Message boards : Sixtrack Application : SIXTRACKTEST (Message 39572)
Posted 10 Aug 2019 by mmonnin
Post:
The clue might be in the name. sixtrackTEST, therefore jobs may not work as expected or even at all. If you only want to run reliable jobs then uncheck the "Run test applications" box. You might also consider turning down your work buffer so you don't get so many. Aborting hundreds of tasks isn't helpful.
I have only been able to get 12 of these. Some have been short, some are over 30hrs. The short ones (unstable beam parameters) can be just as useful as the long ones. Any failures can be used to refine the setup and result in better performance of future tasks.


I have a buffer of 0.10 days or 2.4 hours. Barely any buffer at all. But the ETA was only several seconds and hundreds downloaded on a 32t system. I can't help that the project did not set realistic ETA for the longer tasks. Instead of wasting my PCs time and electricity for them to just end up in an error state I sent them back for someone else to hopefully complete. I ran my TEST for sixtractTEST so unless you have suggestions to fix the errors on the one PC (others PCs are still running) ya can step down off your pedestal.
2) Message boards : Sixtrack Application : SIXTRACKTEST (Message 39551)
Posted 9 Aug 2019 by mmonnin
Post:
Have a Computer with only sixtracktest for the Moment:
The wingman had this Error:
exceeded elapsed time limit 13480.39 (1920000000.00G/109981.44G)</message>
https://lhcathome.cern.ch/lhcathome/result.php?resultid=238806725


I have a lot of these today on one PC. Some complete, some error at the same time. Most successful ones complete in a couple of seconds.
<core_client_version>7.9.3</core_client_version>
<![CDATA[
<message>
exceeded elapsed time limit 17457.49 (1920000000.00G/109981.44G)</message>
<stderr_txt>

</stderr_txt>
]]>

They quickly progress to about 50% then go back to 0% and then very slowly progress. Dumped the rest on that PC.
3) Message boards : News : BOINC Pentathlon - Sixtrack sprint (Message 38870)
Posted 16 May 2019 by mmonnin
Post:
There's no assumption. Every project is inquired before hand. If they don't reply (like MilkyWay) they are left out.

Glad to see some new Atlas tasks being sent out still instead of 5 second SixTrack.
4) Message boards : Theory Application : Theory queue currently drying out (Message 38526)
Posted 3 Apr 2019 by mmonnin
Post:
Yet another request to not send tasks w/o jobs...
5) Message boards : ATLAS application : Native Atlas: "Output file missing" (Message 38422)
Posted 24 Mar 2019 by mmonnin
Post:
Hmmm not right now.

When following the guide I mentioned in my first post re: Singularity, it all falls apart after that slightly modified download is unzipped into the /home/username folder. I have been looking around for what and where to cd to in order to "./configure --prefix=/usr/local"


Singularity install:

sudo wget -O- http://neuro.debian.net/lists/xenial.us-ca.full | sudo tee /etc/apt/sources.list.d/neurodebian.sources.list
sudo apt-get install -y singularity-container


And a
singularity --version

will return the version if installed correctly.
6) Message boards : Theory Application : Installation of CVMFS (Message 38421)
Posted 24 Mar 2019 by mmonnin
Post:
The server scheduler is sometimes confused, how to interpret the max # settings.
...

Yeah I already suspected that. As soon as I changed it back enough tasks went in.
Thanks for the app_config but unfortunately when I want to save the file with gedit in the project directory Ubuntu doesn't let me access the 'projects'-folder for some reason...


To open the file manager with elevated access:

sudo nautilus


Or
sudo gedit /var/lib/boinc-client/lhcathome.cern.ch_lhcathome


I run the Theory Native tasks as single threaded with app_config since the only use about 1.1-1.5 threads but are scheduled by default to use 2.[/code]
7) Message boards : News : Warning: possible shortage of CMS jobs - set No New Tasks as a precaution (Message 38389)
Posted 23 Mar 2019 by mmonnin
Post:
[1] How many times do I have to tell people not to touch critical systems on a Friday -- especially Friday afternoon!?


How many times do we have to ask that tasks not be sent to users w/o jobs...
8) Message boards : Theory Application : Installation of CVMFS (Message 38382)
Posted 22 Mar 2019 by mmonnin
Post:
Well, thanks for the tips but I couldn't get this damn CVMFS installed at all no matter what I tried meanwhile. :-\
I'm too inexperienced with Linux to dig into this even more and don't have the time either so I give up here.
Too bad... I hope a Windows app for this will come out sooner or later, if not then so be it...


If you want to try Ubuntu you can use the instructions I gave to my team. It's worked for me on several 18.04 installs. The instructions given at LHC forums are way overly complicated for no reason when CVFMS can be installed from a repository. It's also much quicker than compiling it.

https://www.overclock.net/forum/365-overclock-net-boinc-team/1721752-projects-month-march-2019-a.html#post27876964


Our instructions for the Theory app point to the official CVMFS documentation. Please let me know if you think any improvements are required.


I was referring to the Native ATLAS instructions, which also need CVFMS, but instruct the user to D/L and make the BOINC, CVFMS and singularity apps from source.
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=4840

The official CVFMS instructions are the ones I used at OCN.
9) Message boards : Theory Application : Issues Native Theory application (Message 38369)
Posted 21 Mar 2019 by mmonnin
Post:
I see that the maximum runtime of last 100 tasks is 25.06 hours.

Would be interesting to know what job and result-id it was.


I aborted this one. Running for 2 days while the longest one I've completed was 8 hours.
https://lhcathome.cern.ch/lhcathome/workunit.php?wuid=109410751
10) Message boards : Theory Application : Installation of CVMFS (Message 38368)
Posted 21 Mar 2019 by mmonnin
Post:
Well, thanks for the tips but I couldn't get this damn CVMFS installed at all no matter what I tried meanwhile. :-\
I'm too inexperienced with Linux to dig into this even more and don't have the time either so I give up here.
Too bad... I hope a Windows app for this will come out sooner or later, if not then so be it...


If you want to try Ubuntu you can use the instructions I gave to my team. It's worked for me on several 18.04 installs. The instructions given at LHC forums are way overly complicated for no reason when CVFMS can be installed from a repository. It's also much quicker than compiling it.

https://www.overclock.net/forum/365-overclock-net-boinc-team/1721752-projects-month-march-2019-a.html#post27876964
11) Message boards : Theory Application : Issues Native Theory application (Message 38329)
Posted 20 Mar 2019 by mmonnin
Post:
The task is set to use 2 CPUs by default and barely over 1 is used and the reported time has run time = exactly CPU time. To the second on every task. At most I see 1.5 cores when the task is really short. 6min run time, 8 min CPU time.
Don't trust the values reported in the results, specially when they are equal.
Example your task: https://lhcathome.cern.ch/lhcathome/result.php?resultid=219459914
It reported 51m 1 sec, that is exactly the reported cpu time at the end of the result. 06:18:02 (32596): cranky exited; CPU time 3061.446043,
but when you calculate job finish time minus the job start time (the job should have ran in one flow)
06:18:02 (32596): cranky exited; CPU time 3061.446043
05:38:19 (32596): wrapper (7.15.26016): starting

you'll find the elapsed time is 2383 seconds, so 1 cpu is used far over 100% or 2 cpu's are partial used.


I am looking at BOINCTasks history for the correct run/CPU times. And my statement is still true. At BEST 1.5 threads used but average is like 1.1 threads per task. I set it use 1 per task with app_config so the CPU would be fully used.
12) Message boards : Theory Application : Issues Native Theory application (Message 38284)
Posted 19 Mar 2019 by mmonnin
Post:
The setup went without error, thanks Ivan for the great directions.
The directions are actually by Laurence ;)

2 X rivetvm.exe, 1 at 65% CPU, 1 at ~45% CPU
2 X pythia8.exe, 1 at ~80% CPU, 1 at ~65% CPU
Wahoo!! Very nice to see pythia running native but was hoping to see it using closer to 100% CPU?

To each job there are a lot of processes. Each job needs 1 rivetvm.exe and e.g. pythia8, agile-runmc (=pythia6), sherpa, herwig etc.
So you have to sum 1 jobname-process with a rivetvm and you see they are together >100%, what happens when you have idle cpu's.


The task is set to use 2 CPUs by default and barely over 1 is used and the reported time has run time = exactly CPU time. To the second on every task. At most I see 1.5 cores when the task is really short. 6min run time, 8 min CPU time.
13) Message boards : ATLAS application : Atlas Simulation 1.01 (Vbox64) will not finish (Message 38227)
Posted 12 Mar 2019 by mmonnin
Post:
Something whacky happened..I don't know how another CPU project got activated, but it did.
I shut it down and got ATLAS back and running, everything seemed ok for awhile, but now at 90 some odd percent, while BOINC still counts off the time and the percent and the usual sluggish remaining time vs real time is happening, the log has not update with new target information for 20 mins now.
Not sure what happened. The log refreshes, but no new information comes through.


Note: Due to the interruption by another project switching from GPU to CPU (damn I need to get that second host going, but no time right now) this task I am running now will run another 3hrs before finishing.
Looks like log file caught up finally.

Anyway....dropped ATLAS for now until I can get this host thing set up. All remaining tasks will be finished (2 more).


I run a separate GPU BOINC client on all my PCs.
https://www.overclock.net/forum/18056-boinc-guides-tutorials/1628924-guide-setting-up-multiple-boinc-instances.html
14) Message boards : ATLAS application : No tasks available (Message 38161)
Posted 8 Mar 2019 by mmonnin
Post:
In the past 2 days I've had ~10 native ATLAS canceled by the server on 3 different hosts. None were anywhere near expiry date. Most had run for several hours. This is definitely not a case of caching too many tasks. My watchdog script showed 5 were processing events. I didn't check the other 5.


I had a couple as well.
15) Message boards : ATLAS application : ATLAS issues (Message 38109)
Posted 5 Mar 2019 by mmonnin
Post:
So true.

Tired of having my laptop on for hours for random results...

You may read Don Quixote by Cervantes (a well known author in your locale) followed by Yeti's Checklist. Pay special attention to item #21.


Did you mean another # as the list only goes to 19.
16) Message boards : ATLAS application : Guide for building everything from sources to run native ATLAS on Debian 9 (Stretch) (Message 35420)
Posted 3 Jun 2018 by mmonnin
Post:
From what I recall there was a long setup time. I compared CPU time which doesn't even include setup time where runtime goes up and CPU time does not.

The only benefit I saw was lower memory. I can't run all 32 threads with vbox or native Altas with 32gb of memory so theres basically no benefit at all and only credit penalties with a complicated, incomplete setup.

It is the same app, either setup yourself or bundled in a .vdi.
I've never seen an app vary by 19x due to the credit system. Runtime and credit of native and vbox versions was pretty consistent when comparing within version.
17) Message boards : ATLAS application : Guide for building everything from sources to run native ATLAS on Debian 9 (Stretch) (Message 35417)
Posted 2 Jun 2018 by mmonnin
Post:
How are others seeing performance between Native/VBox Altas apps? Native is 19x slower than VBox on the same PC.

https://lhcathome.cern.ch/lhcathome/results.php?userid=485872&offset=0&show_names=0&state=0&appid=14

Native: 2,892.54 25,856.68 324.96 ----0.75 Point/Min
Vbox: 1,027.45 509.98 121.48 ---- 14.29 Point/Min
18) Message boards : Number crunching : Tips for optimizing BOINC file transfers for LHC@home (Message 35409)
Posted 1 Jun 2018 by mmonnin
Post:
That would only work if you could put the ISP router in bridge mode. Otherwise, the ISP router will still keep track of all connections because it will see all of the different destinations and different source ports and still perform NAT on each connection.


I still don't have dropped connections with the VZ router not in bridge mode and I've added more connections since when I was using it as the main router.

Shouldn't the VZ router only see one device, the aftermarket router since everything else is behind that NAT?
19) Message boards : ATLAS application : Guide for building everything from sources to run native ATLAS on Debian 9 (Stretch) (Message 35408)
Posted 1 Jun 2018 by mmonnin
Post:
Followup: Several tasks completed following the setup.

Skipped:
su testing

Added
sudo apt install git
sudo apt-get install libtool m4 automake

Modified per later post:
CVMFS_HTTP_PROXY=DIRECT

Very nice. thanks for the instructions. Well above my head in Linux.
20) Message boards : ATLAS application : Guide for building everything from sources to run native ATLAS on Debian 9 (Stretch) (Message 35405)
Posted 31 May 2018 by mmonnin
Post:
I just went through the setup.

I needed this to get through the 1st git download. 18.04 min install.

sudo apt install git


Then I also needed this:

sudo apt-get install libtool m4 automake


before this:
./autogen.sh


Now to try and get some work.


Next 20


©2019 CERN