Message boards :
CMS Application :
CMS tasks failing
Message board moderation
Previous · 1 · 2 · 3
| Author | Message |
|---|---|
|
Send message Joined: 27 Apr 24 Posts: 21 Credit: 1,253,898 RAC: 1,544 |
you should be aware though that they are of no use for the science. I know that. It's their problem, not mine. |
|
Send message Joined: 15 Jun 08 Posts: 2710 Credit: 292,062,887 RAC: 145,435 |
CMS seem to work on some hosts but mine get this error: 2025-11-02 08:12:45 (1029094): Guest Log: [INFO] Requesting an X509 credential from LHC@home 2025-11-02 08:12:46 (1029094): Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev 2025-11-02 08:13:17 (1029094): Guest Log: [DEBUG] % Total % Received % Xferd Average Speed Time Time Time Current 2025-11-02 08:13:17 (1029094): Guest Log: Dload Upload Total Spent Left Speed 2025-11-02 08:13:17 (1029094): Guest Log: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2025-11-02 08:13:17 (1029094): Guest Log: 100 54 0 54 0 0 70 0 --:--:-- --:--:-- --:--:-- 70 2025-11-02 08:13:17 (1029094): Guest Log: 100 54 0 54 0 0 66 0 --:--:-- --:--:-- --:--:-- 66 2025-11-02 08:13:17 (1029094): Guest Log: [DEBUG] 2025-11-02 08:13:17 (1029094): Guest Log: ERROR: Couldn't read proxy from: /tmp/x509up_u0 2025-11-02 08:13:17 (1029094): Guest Log: globus_credential: Error reading proxy credential 2025-11-02 08:13:17 (1029094): Guest Log: globus_credential: Error reading proxy credential: Couldn't read PEM from bio 2025-11-02 08:13:17 (1029094): Guest Log: OpenSSL Error: pem_lib.c:707: in library: PEM routines, function PEM_read_bio: no start line 2025-11-02 08:13:17 (1029094): Guest Log: Use -debug for further information. 2025-11-02 08:13:17 (1029094): Guest Log: [ERROR] Could not get an x509 credential 2025-11-02 08:13:17 (1029094): Guest Log: [ERROR] The x509 proxy creation failed. |
|
Send message Joined: 18 Dec 15 Posts: 1923 Credit: 149,489,868 RAC: 143,780 |
CMS seem to work on some hosts but mine get this error:same problem here :-( |
|
Send message Joined: 14 Jan 10 Posts: 1469 Credit: 9,927,016 RAC: 1,807 |
On the development system CMS is running OK:CMS seem to work on some hosts but mine get this error:same problem here :-( 00:01:21.379155 VMMDev: Guest Log: [INFO] Reading volunteer information 00:01:26.181010 VMMDev: Guest Log: [INFO] Requesting an X509 credential from LHC@home 00:01:26.975471 VMMDev: Guest Log: [INFO] Requesting an X509 credential from vLHC@home-dev 00:01:28.156063 VMMDev: Guest Log: [INFO] Requesting an idtoken from LHC@home 00:01:28.755910 VMMDev: Guest Log: [INFO] Requesting an idtoken from vLHC@home-dev 00:01:29.491935 VMMDev: Guest Log: [INFO] CMS application starting. Check log files. |
|
Send message Joined: 13 May 20 Posts: 52 Credit: 2,795,997 RAC: 3,164 |
bonsoir, je ne sais pas si cela pourra aider mais il y a quelques jours ,toutes mes taches Cms partaient en erreur.J'ai réduit la limite d'unités téléchargées a 8 au lieu de"pas de limite". J'ai demandé a Claude ai qui m'as fait faire une manipulation dans le terminal-sudo usermod -aG vboxusers $USER- et depuis tout fonctionne correctement. Je suis sous linux mint 22.2 et virtualbox 7.24 et j'ai retiré kvm intel du noyau dans -sudo nano /etc/modprobe.d/blacklist-kvm.conf blacklist kvm_intel blacklist kvm_amd blacklist kvm" good evening, I don’t know if it will help but a few days ago, all my tasks Cms was leaving in error. I have reduced the limit of downloaded units to 8 instead of 'no limit'. I asked Claude AI who made me do a manipulation in the terminal-sudo usermod -aG vboxusers $USER- and since then everything works correctly. I am on linux mint 22.2 and virtualbox 7.24 and I removed kvm intel from the kernel in -sudo nano /etc/modprobe. d/blacklist-kvm.conf kvm_intel blacklist blacklist kvm_amd blacklist kvm" |
|
Send message Joined: 29 Aug 05 Posts: 1119 Credit: 10,397,050 RAC: 19,398 |
Hello everyone. Thanks for your patience these last weeks. We were finally able to get our "supply chain" sorted out last Friday and get jobs flowing again. I refrained from posting a celebration, because Hallowe'en... Sure enough we had a little hiccup that I was able to get fixed on Saturday, but things went bad again this morning with failures in getting certificate proxies -- funnily enough it didn't affect my running machine, I must have got my last task just before the problem arose. The failure was in a CA server run by CERN IT, who were able to fix it soon after we raised a ticket. So, fingers crossed, we are now back in action again and you can resume getting new tasks if you have been holding off. |
|
Send message Joined: 15 Jun 08 Posts: 2710 Credit: 292,062,887 RAC: 145,435 |
Hint for volunteers using a local firewall: CMS now requires TCP port 9620 to be open for outgoing connections to HTCondor CCB. |
|
Send message Joined: 15 Jun 08 Posts: 2710 Credit: 292,062,887 RAC: 145,435 |
Hint for volunteers using a local firewall: This afternoon a test for port 9620 has been added to the CMS bootstrap script. Tasks passing the test report something like this to stderr.txt: 2025-11-04 10:29:28 (113263): Guest Log: [INFO] Testing connection to HTCondor-Collector 2025-11-04 10:29:29 (113263): Guest Log: [INFO] Testing connection to HTCondor-CCB |
|
Send message Joined: 24 Jan 06 Posts: 7 Credit: 9,452,791 RAC: 10,393 |
Looks like the x509 errors are back? This is on a Windows VM. 2025-11-11 04:43:39 (6964): Guest Log: [INFO] Testing connection to http://cms-frontier.openhtc.io:8080/FrontierProd/Frontier/ Getting quite a few of them again over the last few days. FWIW: The connection check to HTCondor CCB is passing. Theory tasks seem to be working fine and even had one CMS task get to completion, amongst the failures. |
|
Send message Joined: 29 Aug 05 Posts: 1119 Credit: 10,397,050 RAC: 19,398 |
|
|
Send message Joined: 29 Aug 05 Posts: 1119 Credit: 10,397,050 RAC: 19,398 |
We've had a lot of jobs fail this morning. From the logs it seems to be a network problem. I suspect it's the Cloudflare outage. |
|
Send message Joined: 29 Aug 05 Posts: 1119 Credit: 10,397,050 RAC: 19,398 |
We've had a lot of jobs fail this morning. From the logs it seems to be a network problem. We seem to be recovering now, according to the Running Jobs graph. |
©2025 CERN