1) Message boards : Theory Application : MadGraph5 (Message 50269)
Posted 28 May 2024 by Henry Nebrensky
Post:
Thanks... I wasn't sure if there was some clean-up check that would catch it, apart from the 10-day mop-up.

I should have said earlier: upshot was that the task was taking up a slot but using zero CPU - hence my dredging up an old thread about madgraph doing exactly that.
2) Message boards : Theory Application : MadGraph5 (Message 50267)
Posted 28 May 2024 by Henry Nebrensky
Post:
It also has significant stretches of not actually using CPU at all.
e.g I recently killed task 281349801 precisely because it was holding two cores but idle - it's reported as using just 50 mins in 20 hours :(
Just aborted task 411417290 as runRivet.log shows
===> [runRivet] Tue May 28 13:26:22 UTC 2024 [boinc pp zinclusive 13000 - - madgraph5amc 2.6.0.atlas nlo 100000 160]
and then
INFO: Result for check_poles:
INFO: Poles successfully cancel for 20 points over 20 (tolerance=1.0e-05)
INFO: Starting run
INFO: Using 2 cores
INFO: Cleaning previous results
INFO: Generating events without running the shower.
INFO: Setting up grids
WARNING: program /shared/tmp/tmp.CDpOJU3qAk/MG5RUN/SubProcesses/P0_uux_epem/ajob1 1 F 0 0 launch ends with non zero status: 1. Stop all computation
/shared/tmp/tmp.CDpOJU3qAk/MG5RUN/SubProcesses/P0_uux_epem/ajob1: line 34: 10644 Terminated ../madevent_mintMC > log.txt < input_app.txt 2>&1
INFO: Idle: 6, Running: 0, Completed: 2 [ current time: 13h38 ]

I'd have left it running to see what happens, but the machine will get powered off soon. Looks like internal fault detection isn't being passed up the chain and stopping the task, or did I just get too impatient?

Edit: upshot was that the task was taking up a slot but using zero CPU - hence my revisiting an old thread about madgraph doing exactly that.
3) Message boards : LHC@home Science : Collision energies (Message 49951)
Posted 16 Apr 2024 by Henry Nebrensky
Post:
Thanks for that - I'll have to read it through properly.
It's nice to see LHC experiments other than the big four get a mention.

For reference., p21 answers a similar question I think: Lead-Lead has a total of 1150 TeV available but there is actually only 5.02 TeV available to the colliding pair of nucleons within; possibly I'm misunderstanding the physics but shouldn't there be a similar downshift between the overall proton(nucleon) energy and that of the actual colliding constituents?
4) Message boards : LHC@home Science : Collision energies (Message 49898)
Posted 6 Apr 2024 by Henry Nebrensky
Post:
Hi all,

Does anybody here know roughly what the maximum "physics energy" is for an LHC proton-proton collision?

Since protons are composite particles the collisions are actually between individual quarks/gluons/whatever, and I'd like to get an idea of the "inefficiency" compared to the beam proton energy.

I couldn't see this in a quick look at Wikipedia, probably because I'm not sure what the technical terms would be!

Thanks
5) Message boards : Cafe LHC : Eviction of Russian scientists (Message 48897)
Posted 4 Nov 2023 by Henry Nebrensky
Post:
(Who's "we"?)
That's odd considering we welcome doctors from India/Pakistan.
Only because they accept NHS pay rates...

So scientists aren't considered important?
No - (foreign) scientists are just as much seen as cheap exploitable disposable labour. That's why the UK Government's R&D people and culture strategy avoids any commitment to retaining experienced staff or supporting returns after a career break, but when it comes to getting in naive junior staff there's bullet points a-plenty!
6) Message boards : Cafe LHC : CMS: I'm (almost) back (Message 48896)
Posted 4 Nov 2023 by Henry Nebrensky
Post:
Hi Ivan, glad to hear you've made it home at last...

I've been trying to arrange another trip into London to visit en route, but not been able to stitch (geddit!) all the various stop-offs needed together :(

Focus on your recovery!
7) Message boards : Theory Application : How long may Native-Theory-Tasks run (Message 48154)
Posted 30 May 2023 by Henry Nebrensky
Post:
Theory native logs can be checked, e.g for a task running in slot 0
.../slots/0/cernvm/shared/runRivet.log

If you "head" the runRivet.log, it will tell you the code in use and how many events that specific task is to generate:
[boinc pp jets 8000 170,-,2960 - pythia8 8.301 dire-default 57000 482]
57k in this case. If you then "tail" the log you can see how far it's got and if it's making progress...
8) Message boards : LHC@home Science : Ready to restart! (Message 47245)
Posted 12 Sep 2022 by Henry Nebrensky
Post:
It looks like they are indeed training the magnets these days, there are magnet quenches with no beam.
Possibly, but given the potential side-effects of a quench - remember those magnets scattered around the tunnel after the first startup - I know I would be wary of trying to train to 14TeV current settings before getting 13TeV data in the can. I'm intrigued by the cryptic references to UFOs in that blog post - sounds suspiciously like someone's dropping their spanners on the machine. (Unless that's a dated baguette reference?)
9) Message boards : LHC@home Science : Ready to restart! (Message 47195)
Posted 27 Aug 2022 by Henry Nebrensky
Post:
Meanwhile the LHC magnets remain at 2K, allowing more time to train the magnets.
What does that involve?
10) Message boards : LHC@home Science : Proton Physics : Stable Beams (Message 47054)
Posted 30 Jul 2022 by Henry Nebrensky
Post:
Why does the luminosity at an experiment's IP vary relative to the circulating beam? (e.g. what happened at Alice before/after 20:00-ish?)

Does this mean that the LHC dynamically alters the lattice even within a fill?
11) Message boards : Theory Application : Sherpa tasks run okay for long time, then they fail (Message 47024)
Posted 15 Jul 2022 by Henry Nebrensky
Post:
7.100 events (8d 17h) atm.
But isn't that going to take more like 14 days to complete, rather than 10?
12) Message boards : Number crunching : CentOS9 squid Win11pro (Message 46857)
Posted 8 Jun 2022 by Henry Nebrensky
Post:
CentOS8 have no Mirror, out of support since 2022/01.
As announced a couple of years ago, CentOS became "CentOS Stream". Conversion instructions for CentOS8 to CentOS Stream8 are at https://www.centos.org/centos-stream/.
(I've not tried them; I've been reverting to C7 instead).
13) Questions and Answers : Getting started : How To Setup VirtualBox With LHC@HOME On Command Line Only Server? (Message 46645)
Posted 15 Apr 2022 by Henry Nebrensky
Post:
I got some cloud VMs and want to contribute to the project, but the native applications always fail, and I couldn't fix it. I am wondering how can I setup VirtualBox with LHC.
As these are already VMs, I doubt Virtualbox will run correctly inside them. Never tried it though.

I did simply `sudo apt install virtualbox -y`, is that all?

Answering the subject line, my old notes for CentOS7/8 (Minimal install) say:

    VBox needs gcc, make, elfutils-libelf-devel, kernel-devel and kernel-headers; not a dependency in the RPM. May also need: setools-libs libsemanage-python
    yum install gcc make elfutils-libelf-devel kernel-devel  kernel-headers setools-libs libsemanage-python libseccomp

    Install the downloaded VirtualBox RPM:
    yum localinstall VBox-version.RPM

    user running BOINCmust be in VBox group:
    usermod --append --groups vboxusers boinc

    Rebuild the kernel modules:
    vboxconfig

    Start VBox services:
    service vboxautostart-service start
    service vboxdrv restart

    Optional:
    VBoxManage extpack install /tmp/Oracle_VM_VirtualBox_Extension_Pack-version.vbox-extpack


All as root, via SSH. That should should give a working VirtualBox testable by restarting the BOINC client.
If working, remember to chkconfig the services on before rebooting.
When updating the kernel, remember to re-run vboxconfig to build the new modules before rebooting.

Other useful bits:

VBoxManage list extpacks

VBoxManage list vms
VBoxManage showvminfo 1e6cf289-3176-476f-b80c-6fa3fbe42922
VBoxManage unregistervm 1e6cf289-3176-476f-b80c-6fa3fbe42922 --delete


Distro-specific and SysV-to-systemd translation left as an exercise for the reader.

Note that I've not tested this in a couple of years as I don't have any suitable machine at the moment. Any problems see Yeti's Checklist
14) Message boards : Number crunching : CentOS9 (Message 46488)
Posted 20 Mar 2022 by Henry Nebrensky
Post:
Mirror for CentOS8 is out of Service since 22/01/31.
Isn't that part of the change from CentOS to CentOS Stream? This has caused some "discussion" in the community, including proposals for LHC to move away from CentOS to some other distro. I don't know if there's any clear route on this yet.

CVMFS is not supported for CentOS9 (only CentOS7 and CentOS8).
May well be low-priority due to above.
15) Message boards : ATLAS application : Atlas task slowing right down near the end but still using all cores - continue? (Message 46414)
Posted 3 Mar 2022 by Henry Nebrensky
Post:
I would vote for 240.
Actually, I'd vote for 360 - them Babylonians knew what they were doing ...
Are they the ones responsible for clocks?
... and angles, (which is where clock faces came from?).
But, actually computezrmle was right: 240 also gets you a division by 16, for future expansion.
16) Message boards : ATLAS application : Atlas task slowing right down near the end but still using all cores - continue? (Message 46403)
Posted 2 Mar 2022 by Henry Nebrensky
Post:
I would vote for 240.
Actually, I'd vote for 360 - them Babylonians knew what they were doing - but if people are already struggling with compute times then it would be better to stick to low hanging fruit. :(
17) Message boards : ATLAS application : Atlas task slowing right down near the end but still using all cores - continue? (Message 46402)
Posted 2 Mar 2022 by Henry Nebrensky
Post:
Even within a task, as the number of threads is reduced then each thread must run more events, and it's more likely - but not guaranteed - that they will average out across the threads.
Even with the full 8 cores, that's 25 events per thread, which is more than enough to average things out.
Doesn't that depend on the variance, which I've never studied? In any case, the aim is that the averaging overcomes the variance, which is why divisibility is important for avoiding a small number of events left over.

Also, do you have a figure for how much time is wasted? Since it's 200 in the pool, the wasted cores at the end are probably a fraction of a percent of inefficiency.
IIRC, back when I was running 8-core native Atlas I would generally see the active threads reduce over usually 1-2 minutes, 5 if slow, for tasks of about 4 hrs total wall-clock. (There might be numbers in some ancient post here, but the laptop's tired tonight :( )
A few minutes in 4 hours is nothing.
Thank you - I did put some effort into setting those machines up...
18) Message boards : ATLAS application : Atlas task slowing right down near the end but still using all cores - continue? (Message 46399)
Posted 2 Mar 2022 by Henry Nebrensky
Post:
... they are random sizes.
You are looking on just 1 task but you would have to look at the long term averages.
Really huge numbers!
... I'm surprised there's a difference though, considering the wide variance in event times.
Even within a task, as the number of threads is reduced then each thread must run more events, and it's more likely - but not guaranteed - that they will average out across the threads.

Was the wide variance I saw unusual?
I've no idea.

Also, do you have a figure for how much time is wasted? Since it's 200 in the pool, the wasted cores at the end are probably a fraction of a percent of inefficiency.
IIRC, back when I was running 8-core native Atlas I would generally see the active threads reduce over usually 1-2 minutes, 5 if slow, for tasks of about 4 hrs total wall-clock. (There might be numbers in some ancient post here, but the laptop's tired tonight :( )
19) Message boards : ATLAS application : Atlas task slowing right down near the end but still using all cores - continue? (Message 46390)
Posted 1 Mar 2022 by Henry Nebrensky
Post:
As CP mentioned each ATLAS task processes 200 events from a pool.
It has struck me before, that changing the task's pool size to 180 or 240 events would give better divisibility.
20) Message boards : Sixtrack Application : Sixtrack BOINC application on track to be ported to GPUs? (Message 46205)
Posted 9 Feb 2022 by Henry Nebrensky
Post:
... and now SQUID. That is a lot to ask of volunteer crunchers.
Set up once, runs for months...
More to the point, you don't even need to start worrying about Squid until after you've been through the VirtualBox/CVMFS palaver at least once (to create useful traffic for a Squid to cache), and then only for the larger scale crunchers.
If Sixtrack becomes as painful to get running as, say, Atlas, then surely this will be a big turn-off when trying to encourage new, casual volunteers to join the project.


Next 20


©2024 CERN