Message boards : Theory Application : Some Theory tasks on VirtualBox hang Probing /cvmfs/alice.cern.ch...
Message board moderation

To post messages, you must log in.

AuthorMessage
Glohr

Send message
Joined: 13 Jan 24
Posts: 27
Credit: 4,461,929
RAC: 20,096
Message 51818 - Posted: 6 Apr 2025, 23:22:16 UTC

Some Theory tasks hang with the last thing on the screen "Probing /cvmfs/alice.cern.ch... "

Most tasks continue on and eventually exit normally, but some just sit there never getting the "OK" and so on. The problem tasks don't accumulate any Guest CPU time in VirtualBox after the initial phase. The VM continues to accumulate a few seconds of CPU time per hour, apparently for housekeeping.

Today, all the running.log files in problem tasks visible through the Web application end with
INFO: index summary: size / path
8 /scratch/dat/index/pp_13000_jets_280_-_pythia8_8.306_eetherm.txt

Disk usage: 6792 Kb

CPU usage: 2185 s

Clean tmp ...

Run finished successfully
and contain a line
/cvmfs/sft.cern.ch/lcg/releases/LCG_88b/MCGenerators/pythia8/306/x86_64-centos7-gcc62-opt/pythia8env-genser.sh: line 49: python: command not found

All the successful tasks that I checked were in a different tree and were using CPU time while processing events until they finished and exited.
In the past I've let a couple of the problem tasks go until they timed out after 10 days and exited with "Error while computing" status. That seems rather pointless, so I've been aborting any that I notice going nowhere rather than leaving the dog in the manger blocking other work.

Here are a couple of the problem tasks:
https://lhcathome.cern.ch/lhcathome/result.php?resultid=421021497
https://lhcathome.cern.ch/lhcathome/result.php?resultid=421076672

Is there any way to avoid these, or at least kill them off quickly, automatically?
ID: 51818 · Report as offensive     Reply Quote
Glohr

Send message
Joined: 13 Jan 24
Posts: 27
Credit: 4,461,929
RAC: 20,096
Message 51829 - Posted: 8 Apr 2025, 10:29:51 UTC - in response to Message 51818.  

All of the problem tasks have a running.log that begins with the same line with the same old timestamp:
===> [runRivet] Wed Apr 2 05:28:25 PM UTC 2025 [boinc pp jets 13000 280 - pythia8 8.306 eetherm 100000 42]
I've killed off a couple more today.
https://lhcathome.cern.ch/lhcathome/result.php?resultid=421143601
https://lhcathome.cern.ch/lhcathome/result.php?resultid=421141211
ID: 51829 · Report as offensive     Reply Quote
Crystal Pellet
Volunteer moderator
Volunteer tester

Send message
Joined: 14 Jan 10
Posts: 1450
Credit: 9,747,300
RAC: 593
Message 51831 - Posted: 8 Apr 2025, 15:55:21 UTC - in response to Message 51829.  
Last modified: 8 Apr 2025, 15:56:35 UTC

All of the problem tasks have a running.log that begins with the same line with the same old timestamp:
===> [runRivet] Wed Apr 2 05:28:25 PM UTC 2025 [boinc pp jets 13000 280 - pythia8 8.306 eetherm 100000 42]
I've killed off a couple more today.
https://lhcathome.cern.ch/lhcathome/result.php?resultid=421143601
https://lhcathome.cern.ch/lhcathome/result.php?resultid=421141211

That is very weird. It looks like you're not using the default vdi, when different tasks come with the same job desciption (remnant of an old log?).
You could consider to reset the LHC project in your BOINC Manager.
ID: 51831 · Report as offensive     Reply Quote
CloverField

Send message
Joined: 17 Oct 06
Posts: 92
Credit: 59,893,637
RAC: 10,184
Message 51833 - Posted: 13 Apr 2025, 17:57:02 UTC

This happened to me as well but a project reset appears to have fixed it.
ID: 51833 · Report as offensive     Reply Quote

Message boards : Theory Application : Some Theory tasks on VirtualBox hang Probing /cvmfs/alice.cern.ch...


©2025 CERN