1) Questions and Answers : Preferences : Ready to start... (Message 40140)
Posted 22 hours ago by Gunde
Post:
Rule is to follow keyboard+mouse and any move on these should have effect to boinc. Unplugging and plug in is not counted as in use and as long as system detect hardware it would any move of these.

I would use this, give it a try:
2) Questions and Answers : Preferences : Ready to start... (Message 40139)
Posted 22 hours ago by Gunde
Post:

The setting as it is it pull down gpu also as gpu would require the cpu to be active. That is all it should be doing and design to do. The preferences for gpu would be unnecessary as would need to follow rule of cpu as long as it not have override time to suspend more then cpu.
However on value could give this a value greater then 0. Value as it is would be instant with value 0 of not in 'in use' suggest to increase value to minimum 1 min. This could be bad for any project or any boinc-client to handle request of suspend and resume that frequently. If you run Atlas,Theory or CMS it would take a snapshot of vm machine which include os and application and job into it. Estimated time to shutdown a vm and create a snapshot could take 10-30 sec with good SSD. Suspend and resume task would most likely end up with state 'Postponed'.

Network looks clear and fine. Daily Schedules is also clear so that would not hold up.

We would come up to Status as you show they end up at suspended.

Your issue is probably that it got suspended is probably the setting to 'Suspend when non-BOINC CPU usage is above 25%' in computing. Try increase that to something like 30-50-90% or uncheck that.

Computing
Suspend when computer is in use [0] -> 5 min

Suspend when non-BOINC CPU usage is above [25%] -> 80%




Usage above would include all usage for system except Boinc itself so it could be close in load to system and additional application in background.
Let me know if that changed it.
3) Questions and Answers : Preferences : Ready to start... (Message 40136)
Posted 1 day ago by Gunde
Post:
How would you like it?

If no limit set to [Run always] and [Network activity always]
As i understand you would like to like to keep 'Suspend when computer is in use'. This is based on time you have set in Options->Computing Preferences >Computing ->[When to Suspend]
Sixtrack would change to 'Ready to start' for that that never started and task already started would be 'waiting to run' and task are resume would be in state 'Running'.
Network is limited as based on your rules not as 'in use'.
Go to:Options->Computing Preferences->Network [Usage Limits] & [Others] then check if its limited based on time of day at Options->Computing Preferences->Daily Schedules. Make check that time is set correct if you really need to use this. It would say 'waiting for network access' if time is set for network.

This could also be if you have set on app_config if use that to require network to application If you use this. You point out that no setting is set then i would suggest to set 'Network activity always'.
It looks like you have set `Run based on Preferences` and 'Network based on Preferences' follow another layer of rule.
You could get additional info in Event Log to when and why client change state.

example: If network is limited to run based on time you have set the client will not run. ATLAS will show 'waiting for network access' as it would require to use internet to get jobs inside VM machine. In Event Log also say Suspending network activity - time of day. In that case you would be sure that is set wrong in [Daily Schedules]. If not it would be computer lost internet or app_config is set to hold it.
Application like Sixtrack would not need network while running so it never be suspended by this rule.

Leave suggestion on how you would like to setup boinc and we could help more. If not just try settings.
4) Message boards : ATLAS application : ATLAS native version 2.72 (Message 40129)
Posted 2 days ago by Gunde
Post:
Try remove singularity and use the one build in vm. It worked for me with 18.10.

$ sudo rm -rf \
    /usr/local/libexec/singularity \
    /usr/local/var/singularity \
    /usr/local/etc/singularity \
    /usr/local/bin/singularity \
    /usr/local/bin/run-singularity \
    /usr/local/etc/bash_completion.d/singularity


https://sylabs.io/guides/3.0/user-guide/installation.html#remove-an-old-version
5) Message boards : Number crunching : computation errors (Message 40001)
Posted 21 days ago by Gunde
Post:
I have not manage to build in 19.04 as some libs been deprecated. Cern would need to make it to support 19.04. This could have changed but not sure on that.

If all task failed i would check check setup.
cvmfs_config chksetup


Then also probe.
cvmfs_config probe


Does it say OK?
6) Message boards : Theory Application : New version 263.90 (Message 39783)
Posted 1 Sep 2019 by Gunde
Post:
Did a test and load 2.7 for Theory #Unlimited 32 core task and total for system around 3.6 cores .
Virtualbox would have 16C as supported usage so the do not know it would handle 32C task for Theory.

Other project as Cosmology is effected on this limit and they hand out 32C MT task in docker container but they fail at start if user do not set a limit.
7) Message boards : Theory Application : New version 263.90 (Message 39782)
Posted 31 Aug 2019 by Gunde
Post:
For the project it would help if project admins set count in cores to what the applications could scale up to. For users experience and also to project it could have great balance for MT task. I'm happy to see that Theory that they move out of single core task and also make it possible for native application but still on low processes in jobs to each task (2-4) last tested. I did a test last year and got 40 cores task at that time and host end at 6% in cpu usage. Way out of even possible to reach that count and it could be changed since that time but i would not use more then 4 cores today based on experience of what it could run on my host. Today i use 3 cores each but it would probably be times that only 1 cores are needed when it hit a long runner job.

Looking at Atlas it scale up to even 12 cores and really use up to that count and end faster if when events are done.

Theory could be reduced to max 4 cores as default and when jobs to vm are sorted out it could be increased to core count it could scale up to. This would put away high credits for users that use app_config and project would get more task running and work more effectively.

Remove "unlimited" in users LHC@home preferences to cores the application really could use as max target.
8) Message boards : Theory Application : Could not get X509 credentials (Message 39169)
Posted 22 Jun 2019 by Gunde
Post:
Got this on 3 old host added.
Exit status	206 (0x000000CE) EXIT_INIT_FAILURE

<message>
The filename or extension is too long.
 (0xce) - exit code 206 (0xce)</message>


But log show more why:

[019-06-23 00:33:02 (5080): Guest Log: [INFO] Requesting an X509 credential from LHC@home

2019-06-23 00:33:03 (5080): Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev

2019-06-23 00:33:34 (5080): Guest Log: [INFO] Requesting an X509 credential from LHC@home

2019-06-23 00:33:35 (5080): Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev

2019-06-23 00:34:06 (5080): Guest Log: [INFO] Requesting an X509 credential from LHC@home

2019-06-23 00:34:06 (5080): Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev

2019-06-23 00:34:37 (5080): Guest Log: [INFO] Requesting an X509 credential from LHC@home

2019-06-23 00:34:38 (5080): Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev

2019-06-23 00:35:09 (5080): Guest Log: [INFO] Requesting an X509 credential from LHC@home

2019-06-23 00:35:10 (5080): Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev

2019-06-23 00:35:41 (5080): Guest Log: [INFO] Requesting an X509 credential from LHC@home

2019-06-23 00:35:42 (5080): Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev

2019-06-23 00:36:14 (5080): Guest Log: [DEBUG] 

2019-06-23 00:36:14 (5080): Guest Log: curl: (60) Peer certificate cannot be authenticated with known CA certificates

2019-06-23 00:36:14 (5080): Guest Log: More details here: http://curl.haxx.se/docs/sslcerts.html

2019-06-23 00:36:14 (5080): Guest Log: curl performs SSL certificate verification by default, using a "bundle"

2019-06-23 00:36:14 (5080): Guest Log:  of Certificate Authority (CA) public keys (CA certs). If the default

2019-06-23 00:36:14 (5080): Guest Log:  bundle file isn't adequate, you can specify an alternate file

2019-06-23 00:36:14 (5080): Guest Log:  using the --cacert option.

2019-06-23 00:36:14 (5080): Guest Log: If this HTTPS server uses a certificate signed by a CA represented in

2019-06-23 00:36:14 (5080): Guest Log:  the bundle, the certificate verification probably failed due to a

2019-06-23 00:36:14 (5080): Guest Log:  problem with the certificate (it might be expired, or the name might

2019-06-23 00:36:14 (5080): Guest Log:  not match the domain name in the URL).

2019-06-23 00:36:14 (5080): Guest Log: If you'd like to turn off curl's verification of the certificate, use

2019-06-23 00:36:14 (5080): Guest Log:  the -k (or --insecure) option.

2019-06-23 00:36:14 (5080): Guest Log: [DEBUG] 

2019-06-23 00:36:14 (5080): Guest Log: ERROR: Couldn't find a valid proxy.

2019-06-23 00:36:14 (5080): Guest Log:        globus_sysconfig: File has zero length: File: /tmp/x509up_u0

2019-06-23 00:36:14 (5080): Guest Log: Use -debug for further information.

2019-06-23 00:36:15 (5080): Guest Log: [ERROR] Could not get an x509 credential

2019-06-23 00:36:15 (5080): Guest Log: [ERROR] The x509 proxy creation failed.

2019-06-23 00:36:15 (5080): Guest Log: [INFO] Shutting Down.


Not sure if cert renew in setup files or depended on what server hand out. I let host retry later but if not changed and it would use old certs more host would suffer from it.
These host (win 10 x64) have direct connection to server with no proxy/vpn attached.
9) Message boards : Theory Application : (Native) Theory - Sherpa looooooong runners (Message 38471)
Posted 27 Mar 2019 by Gunde
Post:
I follow a few native task and one of them run sherpa. It was at runtime at 2days 1hour yesterday. Today i saw it got timed out.

We might need to extend time to these sherpa:s.

https://lhcathome.cern.ch/lhcathome/workunit.php?wuid=109636098
10) Message boards : ATLAS application : Native Atlas: "Output file missing" (Message 38412)
Posted 24 Mar 2019 by Gunde
Post:
This is not SLC6, need to run with Singularity....
Checking Singularity...
Singularity is installed
copy /var/lib/boinc-client/slots/5/shared/input.tar.gz
copy /var/lib/boinc-client/slots/5/shared/copy_2LbbWw
copy /var/lib/boinc-client/slots/5/shared/ATLAS.root_0
copy /var/lib/boinc-client/slots/5/shared/RTE.tar.gz
copy /var/lib/boinc-client/slots/5/shared/start_atlas.sh
export ATHENA_PROC_NUMBER=2;start atlas job with 
Testing the function of Singularity...
check singularity with cmd:singularity exec -B /cvmfs /cvmfs/atlas.cern.ch/repo/images/singularity/x86_64-slc6.img hostname

Singularity isnt working...


Looks like Singularity is broken. Did you use version 2.4.2? That guide should work and 2.4.2 been great for my host.
11) Message boards : ATLAS application : Task will not end despite being 100% (Message 38196)
Posted 9 Mar 2019 by Gunde
Post:
Tested 4 task to win host all 4 got suspended for a few hours and resumed late on. All task end after 3.5 hours and got valid (vbox 5.2.26).

Ignore estimated time but if task do not end after 2 days it would be time to abort it.
12) Message boards : Theory Application : Theory queue currently drying out (Message 38176)
Posted 9 Mar 2019 by Gunde
Post:
Thanks Laurence
13) Message boards : Theory Application : Theory queue currently drying out (Message 38169)
Posted 8 Mar 2019 by Gunde
Post:
Everything should be back to normal.
Laurence, I don't think so. There are still no new tasks available for download; and the tasks still running on my computers seem not to get any jobs..

Resume task are still empty on event and as Erich56 say no task are sent out.

Provide us info when if it changed.
14) Message boards : Theory Application : Theory queue currently drying out (Message 38164)
Posted 8 Mar 2019 by Gunde
Post:
Been fine last days except for today for today. Most task idle and some running halftime. Last hours it been a mix on events shared to hosts as i seen a host go from 50% to 10% then up to 100% then drop back to 50%.

It is now stable at 50% in cpu load and keep few task running while i set no new task as the server seems to be busy and i would wait until it recover back. Jobs queues are at 0 and would need a kick to increase the demands to get back to normal stage.

Project admins would be aware of this as they would monitor the status on servers frequently but they may not reach out to boinc page each time. There is a a lot done outside boinc and issues would be take care of when they are available. This could be a mix of issues of in rates of jobs out and network.

It would be up to them if need to share the info into it but most task is now idle with python script as only usage to task is not good to keep the users that contribute to project.

They would would probably deal with as soon as they can and i would wait and focus on others in meantime.
15) Message boards : ATLAS application : Atlas-native App finished in Seconds (Message 37794)
Posted 20 Jan 2019 by Gunde
Post:
When i check log from last task today i found this.
VBoxManage.exe: error: VT-x is disabled in the BIOS for all CPU modes (VERR_VMX_MSR_ALL_VMX_DISABLED)

check your BIOS if VT-x is enabled.
16) Message boards : News : Seasons greetings from LHC@home (Message 37677)
Posted 21 Dec 2018 by Gunde
Post:
Merry Christmas and Happy New Year!
17) Message boards : Cafe LHC : Milestones (Message 37471)
Posted 2 Dec 2018 by Gunde
Post:
Ryzen would use SMT=Simultaneous multithreading much like HT that intel use fo theading.
18) Message boards : Number crunching : Local control of which subprojects run`2 (Message 37274)
Posted 7 Nov 2018 by Gunde
Post:
max_concurrent need to be 1 or higher. To use 0 would not work.
19) Message boards : LHCb Application : exit code 53 /no connection to cern.ch on port 80 (Message 37080)
Posted 23 Oct 2018 by Gunde
Post:
A few hosts lost connections, hope it fixed now.
20) Message boards : LHCb Application : exit code 53 /no connection to cern.ch on port 80 (Message 37079)
Posted 23 Oct 2018 by Gunde
Post:
Getting task but some host with new task from LHCb ends before it reached 5 min

errorcode for task: Exit status 53 (0x00000035) Unknown error code

Stderr output:
2018-10-23 20:08:09 (58259): Guest Log: [INFO] Mounting the shared directory
2018-10-23 20:08:09 (58259): Guest Log: [INFO] Shared directory mounted, enabling vboxmonitor
2018-10-23 20:08:09 (58259): Guest Log: [DEBUG] Testing network connection to cern.ch on port 80
2018-10-23 20:09:10 (58259): Guest Log: [DEBUG] nc: connect to cern.ch port 80 (tcp) timed out: Operation now in progress
2018-10-23 20:09:10 (58259): Guest Log: nc: connect to cern.ch port 80 (tcp) timed out: Operation now in progress
2018-10-23 20:09:10 (58259): Guest Log: [DEBUG] 1
2018-10-23 20:09:10 (58259): Guest Log: [ERROR] Could not connect to cern.ch on port 80
2018-10-23 20:09:10 (58259): Guest Log: [INFO] Shutting Down.
2018-10-23 20:09:10 (58259): VM Completion File Detected.
2018-10-23 20:09:10 (58259): VM Completion Message: Could not connect to cern.ch on port 80

Only some task effected so might be a network change? I donĀ“t any issue on my network and so far only a few host effected. Sorry to that my hosts sent a lot of errors i would stop fetch new thask until it solved/changed.


Next 20


©2019 CERN