Message boards :
ATLAS application :
ATLAS native_mt fail
Message board moderation
Author | Message |
---|---|
Send message Joined: 6 Dec 14 Posts: 3 Credit: 130,968 RAC: 0 |
I cannot for the life of me find where to Show Computers I do not know if you can see them. :-/ Why are the tasks failing on this setup? Thanks! <core_client_version>7.6.33</core_client_version> <![CDATA[ <stderr_txt> 14:18:54 (20335): wrapper (7.7.26015): starting 14:18:54 (20335): wrapper: running run_atlas (--nthreads 2) singularity image is /cvmfs/atlas.cern.ch/repo/images/singularity/x86_64-slc6.img sys.argv = ['run_atlas', '--nthreads', '2'] THREADS=2 Checking for CVMFS CVMFS is installed OS:cat: /etc/redhat-release: No such file or directory This is not SLC6, need to run with Singularity.... Checking Singularity... Singularity is installed copy /var/lib/boinc-client/slots/2/shared/start_atlas.sh copy /var/lib/boinc-client/slots/2/shared/RTE.tar.gz copy /var/lib/boinc-client/slots/2/shared/input.tar.gz copy /var/lib/boinc-client/slots/2/shared/ATLAS.root_0 export ATHENA_PROC_NUMBER=2;start atlas job with PandaID=3957346728 Testing the function of Singularity... check singularity with cmd:singularity exec -B /cvmfs /cvmfs/atlas.cern.ch/repo/images/singularity/x86_64-slc6.img hostname Singularity Works... cmd = singularity exec --pwd /var/lib/boinc-client/slots/2 -B /cvmfs,/var /cvmfs/atlas.cern.ch/repo/images/singularity/x86_64-slc6.img sh start_atlas.sh > runtime_log 2> runtime_log.err running cmd return value is 0 ***********************log_extracts.txt************************* - Last 10 lines from /var/lib/boinc-client/slots/2/Panda_Pilot_20784_1528571937/PandaJob/athena_stdout.txt - PyJobTransforms.trfExe.preExecute 2018-06-09 14:19:36,673 INFO Batch/grid running - command outputs will not be echoed. Logs for EVNTtoHITS are in log.EVNTtoHITS PyJobTransforms.trfExe.preExecute 2018-06-09 14:19:36,675 INFO Now writing wrapper for substep executor EVNTtoHITS PyJobTransforms.trfExe._writeAthenaWrapper 2018-06-09 14:19:36,676 INFO Valgrind not engaged PyJobTransforms.trfExe.preExecute 2018-06-09 14:19:36,676 INFO Athena will be executed in a subshell via ['./runwrapper.EVNTtoHITS.sh'] PyJobTransforms.trfExe.execute 2018-06-09 14:19:36,676 INFO Starting execution of EVNTtoHITS (['./runwrapper.EVNTtoHITS.sh']) PyJobTransforms.trfExe.execute 2018-06-09 14:23:34,791 INFO EVNTtoHITS executor returns 64 PyJobTransforms.trfExe.validate 2018-06-09 14:23:35,700 ERROR Validation of return code failed: Non-zero return code from EVNTtoHITS (64) (Error code 65) PyJobTransforms.trfExe.validate 2018-06-09 14:23:35,732 INFO Scanning logfile log.EVNTtoHITS for errors PyJobTransforms.transform.execute 2018-06-09 14:23:36,121 CRITICAL Transform executor raised TransformValidationException: Non-zero return code from EVNTtoHITS (64) PyJobTransforms.transform.execute 2018-06-09 14:23:39,295 WARNING Transform now exiting early with exit code 65 (Non-zero return code from EVNTtoHITS (64)) - Walltime - JobRetrival=3, StageIn=10, Execution=273, StageOut=0, CleanUp=14 ***********************pilot_error_report.json********************* { "3957346728": { "2": [ { "pilotErrorCode": 0, "pilotErrorDiag": "Job failed: Non-zero failed job return code: 65" } ] } } *****************The last 100 lines of the pilot log****************** "seopt": "token:ATLASDATADISK:srm://srm.ndgf.org:8443/srm/managerv2?SFN=", "sepath": "/atlas/disk/atlasdatadisk/rucio", "seprodpath": "/atlas/disk/atlasdatadisk/rucio", "setokens": "ATLASDATADISK", "site": "BOINC", "siteid": "BOINC_MCORE", "sitershare": null, "space": 0, "special_par": null, "stageinretry": 2, "stageoutretry": 2, "status": "brokeroff", "statusoverride": "offline", "sysconfig": "manual", "system": "arc", "tags": "arc", "tier": "T3", "timefloor": 0, "tmpdir": null, "transferringlimit": 20000, "tspace": "2070-01-01T00:00:00", "use_newmover": "True", "validatedreleases": "True", "version": null, "wansinklimit": null, "wansourcelimit": null, "wnconnectivity": "full", "wntmpdir": null } 2018-06-09 19:18:57|20784|SiteInformat| Queuedata was successfully downloaded by pilot wrapper script 2018-06-09 19:18:57|20784|ATLASSiteInf| curl command returned valid queuedata 2018-06-09 19:18:57|20784|ATLASSiteInf| Site BOINC_MCORE is currently in brokeroff mode 2018-06-09 19:18:57|20784|ATLASSiteInf| Job recovery turned off 2018-06-09 19:18:57|20784|ATLASSiteInf| Confirmed correctly formatted rucio sepath 2018-06-09 19:18:57|20784|ATLASSiteInf| Confirmed correctly formatted rucio seprodpath 2018-06-09 19:18:57|20784|SiteInformat| Evaluating queuedata 2018-06-09 19:18:57|20784|SiteInformat| Setting unset pilot variables using queuedata 2018-06-09 19:18:57|20784|SiteInformat| appdir: 2018-06-09 19:18:57|20784|pUtil.py | File registration will be done by server 2018-06-09 19:18:57|20784|pUtil.py | Updated stage-in retry number to 2 2018-06-09 19:18:57|20784|pUtil.py | Updated stage-out retry number to 2 2018-06-09 19:18:57|20784|pUtil.py | Detected unset (NULL) release/homepackage string 2018-06-09 19:18:57|20784|ATLASExperim| Application dir confirmed: /var/lib/boinc-client/slots/2/ 2018-06-09 19:18:57|20784|pilot.py | Pilot will serve experiment: Nordugrid-ATLAS 2018-06-09 19:18:57|20784|ATLASExperim| Architecture information: 2018-06-09 19:18:57|20784|ATLASExperim| Excuting command: lsb_release -a 2018-06-09 19:18:57|20784|ATLASExperim| sh: lsb_release: command not found 2018-06-09 19:18:57|20784|pUtil.py | getSiteInformation: got experiment=ATLAS 2018-06-09 19:18:57|20784|ATLASExperim| appdirs = ['/cvmfs/atlas.cern.ch/repo/sw'] 2018-06-09 19:18:57|20784|ATLASExperim| head of /cvmfs/atlas.cern.ch/repo/sw/ChangeLog: -------------------------------------------------------------------------------- 2018-06-09 21:00:23 Alessandro De Salvo * + AGISData 20180609210023 2018-06-09 20:01:16 Alessandro De Salvo * + GroupData 201806092001 2018-06-09 20:00:27 Alessandro De Salvo * + AGISData 20180609200027 2018-06-09 19:00:17 Alessandro De Salvo -------------------------------------------------------------------------------- 2018-06-09 19:18:57|20784|ATLASExperim| ATLAS_PYTHON_PILOT set to /usr/bin/python 2018-06-09 19:18:57|20784|pUtil.py | getSiteInformation: got experiment=ATLAS 2018-06-09 19:18:57|20784|ATLASExperim| Executing command: export ATLAS_LOCAL_ROOT_BASE=/cvmfs/atlas.cern.ch/repo/ATLASLocalRootBase;$ATLAS_LOCAL_ROOT_BASE/utilities/checkValidity.sh (time-out: 300) 2018-06-09 19:18:57|20784|pUtil.py | Executing command: export ATLAS_LOCAL_ROOT_BASE=/cvmfs/atlas.cern.ch/repo/ATLASLocalRootBase;$ATLAS_LOCAL_ROOT_BASE/utilities/checkValidity.sh (protected by timed_command, timeout: 300 s) 2018-06-09 19:18:58|20784|pUtil.py | Elapsed time: 0 2018-06-09 19:18:58|20784|ATLASExperim| Diagnostics tool has verified CVMFS 2018-06-09 19:18:58|20784|Node.py | Collecting machine features 2018-06-09 19:18:58|20784|Node.py | $MACHINEFEATURES not defined locally 2018-06-09 19:18:58|20784|Node.py | $JOBFEATURES not defined locally 2018-06-09 19:18:58|20784|Node.py | Executing command: hostname -i 2018-06-09 19:18:58|20784|Node.py | IP number of worker node: 127.0.1.1 2018-06-09 19:18:58|20784|pUtil.py | getSiteInformation: got experiment=Nordugrid-ATLAS 2018-06-09 19:18:58|20784|pilot.py | Using site information for experiment: Nordugrid-ATLAS 2018-06-09 19:18:58|20784|pilot.py | Will attempt to create workdir: /var/lib/boinc-client/slots/2/Panda_Pilot_20784_1528571937 2018-06-09 19:18:58|20784|pilot.py | Creating file: /var/lib/boinc-client/slots/2/CURRENT_SITEWORKDIR 2018-06-09 19:18:58|20784|pUtil.py | Wrote string "/var/lib/boinc-client/slots/2/Panda_Pilot_20784_1528571937" to file: /var/lib/boinc-client/slots/2/CURRENT_SITEWORKDIR 2018-06-09 19:18:58|20784|ATLASExperim| ATLAS_POOLCOND_PATH not set by wrapper 2018-06-09 19:18:58|20784|pilot.py | Preparing to execute Cleaner 2018-06-09 19:18:58|20784|pilot.py | Cleaning /var/lib/boinc-client/slots/2 2018-06-09 19:18:58|20784|Cleaner.py | Cleaner initialized with clean-up limit: 2 hours 2018-06-09 19:18:58|20784|Cleaner.py | Cleaner will scan for lost directories in verified path: /var/lib/boinc-client/slots/2 2018-06-09 19:18:58|20784|Cleaner.py | Executing empty dirs clean-up, stage 1/5 2018-06-09 19:18:58|20784|Cleaner.py | Purged 0 empty directories 2018-06-09 19:18:58|20784|Cleaner.py | Executing work dir clean-up, stage 2/5 2018-06-09 19:18:58|20784|Cleaner.py | Purged 0 single workDirs directories 2018-06-09 19:18:58|20784|Cleaner.py | Executing maxed-out dirs clean-up, stage 3/5 2018-06-09 19:18:58|20784|Cleaner.py | Purged 0 empty directories 2018-06-09 19:18:58|20784|Cleaner.py | Executing AthenaMP clean-up, stage 4/5 <SKIPPED> 2018-06-09 19:18:58|20784|Cleaner.py | Executing PanDA Pilot dir clean-up, stage 5/5 2018-06-09 19:18:58|20784|Cleaner.py | Number of found job state files: 0 2018-06-09 19:18:58|20784|Cleaner.py | No job state files were found, aborting clean-up 2018-06-09 19:18:58|20784|pilot.py | Update frequencies: 2018-06-09 19:18:58|20784|pilot.py | ...Processes: 300 s 2018-06-09 19:18:58|20784|pilot.py | .......Space: 600 s 2018-06-09 19:18:58|20784|pilot.py | ......Server: 1800 s 2018-06-09 19:18:58|20784|pUtil.py | Timefloor set to zero in queuedata (multi-jobs disabled) ***************diag file************ runtimeenvironments=APPS/HEP/ATLAS-SITE; Processors=1 WallTime=411.32s KernelTime=18.39s UserTime=252.80s CPUUsage=65% MaxResidentMemory=1807372kB AverageResidentMemory=0kB AverageTotalMemory=0kB AverageUnsharedMemory=0kB AverageUnsharedStack=0kB AverageSharedMemory=0kB PageSize=4096B MajorPageFaults=6937 MinorPageFaults=2270894 Swaps=0 ForcedSwitches=24219 WaitSwitches=487507 Inputs=2706816 Outputs=65056 SocketReceived=0 SocketSent=0 Signals=0 nodename=PoppaGeek@Dev9400 exitcode=0 ******************************WorkDir*********************** total 263632 drwxrwx--x 6 boinc boinc 4096 Jun 9 14:25 . drwxrwx--x 5 boinc boinc 4096 Jun 9 13:30 .. -rw------- 1 boinc boinc 6739364 Jun 9 14:19 agis_ddmendpoints.cvmfs.json -rw------- 1 boinc boinc 5359206 Jun 9 14:19 agis_schedconf.cvmfs.json drwx------ 2 boinc boinc 4096 Jun 9 14:19 .alrb drwxr-xr-x 3 boinc boinc 4096 Jun 9 14:18 APPS -rwx------ 1 boinc boinc 2435 Jun 9 10:31 ARCpilot -rw------- 1 boinc boinc 549 Jun 9 14:19 .asetup -rw------- 1 boinc boinc 10994 Jun 9 14:19 .asetup.save -rw-r--r-- 1 boinc boinc 0 Jun 9 14:18 boinc_lockfile -rw-r--r-- 1 boinc boinc 8192 Jun 9 14:25 boinc_mmap_file -rw-r--r-- 1 boinc boinc 526 Jun 9 14:23 boinc_task_state.xml -rw------- 1 boinc boinc 58 Jun 9 14:18 CURRENT_SITEWORKDIR -rw-r--r-- 1 boinc boinc 256192482 Jun 9 14:18 EVNT.13837267._001172.pool.root.1 -rw-r--r-- 1 boinc boinc 5744 Jun 9 14:18 init_data.xml -rw-r--r-- 1 boinc boinc 1091389 Jun 9 14:18 input.tar.gz -rw------- 1 boinc boinc 488 Jun 9 14:25 IUWLDmW1ulsnlyackoJh5iwnABFKDmABFKDmqz7XDmABFKDmOvp3Fm.diag -rw------- 1 boinc boinc 3467 Jun 9 14:25 jobSmallFiles.tgz -rw-r--r-- 1 boinc boinc 105 Jun 9 14:18 job.xml -rw------- 1 boinc boinc 170277 Jun 9 14:25 log.14322886._074314.job.log.1 -rw------- 1 boinc boinc 152071 Jun 9 14:24 log.14322886._074314.job.log.tgz.1 -rw------- 1 boinc boinc 1490 Jun 9 14:24 log_extracts.txt -rw------- 1 boinc boinc 306 Jun 9 14:23 memory_monitor_summary.json -rw------- 1 boinc boinc 599 Jun 9 14:25 metadata-surl.xml -rw------- 1 boinc boinc 241 Jun 9 14:24 output.list -rw------- 1 boinc boinc 11 Jun 9 14:19 pandaIDs.out -rw------- 1 boinc boinc 2951 Jun 9 14:19 pandaJobData_1.out -rw------- 1 boinc boinc 2951 Jun 9 14:18 pandaJobData.out -rw------- 1 boinc boinc 8158 Jun 9 14:24 panda_node_struct.pickle -rw------- 1 boinc boinc 203 Jun 9 14:24 pilot_error_report.json -rw------- 1 boinc boinc 29 Jun 9 14:18 PILOT_INITDIR -rw------- 1 boinc boinc 139 Jun 9 14:25 pilotlog-last.txt -rw------- 1 boinc boinc 11387 Jun 9 14:18 pilotlog.txt drwx------ 3 boinc boinc 4096 Jun 9 14:19 .pki -rw------- 1 boinc boinc 3751 Jun 9 14:19 queuedata.json -rw-r--r-- 1 boinc boinc 4376 Jun 9 10:32 queuedata.pilot.json -rw-r--r-- 1 boinc boinc 606 Jun 9 14:18 RTE.tar.gz -rwxr-xr-x 1 boinc boinc 8356 Jun 9 14:18 run_atlas -rw-r--r-- 1 boinc boinc 604 Jun 9 14:25 runtime_log -rw-r--r-- 1 boinc boinc 10385 Jun 9 14:25 runtime_log.err drwxrwx--x 2 boinc boinc 4096 Jun 9 14:25 shared -rw-r--r-- 1 boinc boinc 14425 Jun 9 14:18 start_atlas.sh -rw------- 1 boinc boinc 19 Jun 9 14:19 START_TIME_3957346728 -rw------- 1 boinc boinc 1 Jun 9 14:18 STATUSCODE -rw-r--r-- 1 boinc boinc 9737 Jun 9 14:25 stderr.txt -rw------- 1 boinc boinc 47 Jun 9 14:24 workdir_size-3957346728.json -rw-r--r-- 1 boinc boinc 100 Jun 9 14:18 wrapper_26015_x86_64-pc-linux-gnu -rw-r--r-- 1 boinc boinc 24 Jun 9 14:25 wrapper_checkpoint.txt running start_atlas return value is 0 Parent exit 0 child process exit 0 14:25:47 (20335): run_atlas exited; CPU time 253.180000 14:25:47 (20335): called boinc_finish(0) </stderr_txt> ]]> |
Send message Joined: 6 Dec 14 Posts: 3 Credit: 130,968 RAC: 0 |
Found this: This is the error that we have seen before: "No events to process: 4050 (skipEvents) >= 2000 (inputEvents of EVNT)" https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=4179&postid=33433 So case closed? |
Send message Joined: 15 Nov 14 Posts: 602 Credit: 24,371,321 RAC: 0 |
I cannot for the life of me find where to Show Computers I do not know if you can see them. :-/ Yes, I can see them. In your "Primary (default) preferences", there is a checkbox for it (Should LHC@home show your computers on its web site?). It is probably enabled by default. I don't see the problem, so I can't comment on it, but it should not last long unless there is something wrong. |
Send message Joined: 6 Dec 14 Posts: 3 Credit: 130,968 RAC: 0 |
PyJobTransforms.trfExe.validate 2018-06-09 14:23:35,700 ERROR Validation of return code failed: Non-zero return code from EVNTtoHITS (64) (Error code 65) ***********************pilot_error_report.json********************* { "3957346728": { "2": [ { "pilotErrorCode": 0, "pilotErrorDiag": "Job failed: Non-zero failed job return code: 65" } ] } } *****************The last 100 lines of the pilot log****************** 6 work units all completed and validated with runtime less than 500 seconds. |
Send message Joined: 24 Jul 16 Posts: 88 Credit: 239,917 RAC: 0 |
I see other hosts where native application fails. Host 1 ID: 10511353 Details | Tasks Cross-project stats: BOINCstats.com Free-DC 9 9,352.80 1,304,278 7.8.4 GenuineIntel Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz [Family 6 Model 63 Stepping 2] (32 processors) --- Linux CentOS CentOS Linux release 7.4.1708 (Core) [3.10.0-514.26.2.el7.x86_64] 17 Jun 2018, 14:00:28 UTC Host 2 ID: 10511351 Details | Tasks Cross-project stats: BOINCstats.com Free-DC 254 10,101.40 1,276,884 7.8.4 GenuineIntel Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz [Family 6 Model 63 Stepping 2] (32 processors) --- Linux CentOS CentOS Linux release 7.4.1708 (Core) [3.10.0-514.26.2.el7.x86_64] 17 Jun 2018, 13:05:03 UTC Host 3 ID: 10511349 Details | Tasks Cross-project stats: BOINCstats.com Free-DC 346 10,209.53 1,243,281 7.8.4 GenuineIntel Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz [Family 6 Model 63 Stepping 2] (32 processors) --- Linux CentOS CentOS Linux release 7.4.1708 (Core) [3.10.0-514.26.2.el7.x86_64] 17 Jun 2018, 12:42:55 UTC Host 4 ID: 10511348 Details | Tasks Cross-project stats: BOINCstats.com Free-DC 539 9,979.68 1,237,588 7.8.4 GenuineIntel Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz [Family 6 Model 63 Stepping 2] (32 processors) --- Linux CentOS CentOS Linux release 7.4.1708 (Core) [3.10.0-514.26.2.el7.x86_64] 17 Jun 2018, 11:39:32 UTC There are the same lines in the log which displays : Host 5 ID: 10511352 Details | Tasks Cross-project stats: BOINCstats.com Free-DC 601 10,301.15 1,275,983 7.8.4 GenuineIntel Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz [Family 6 Model 63 Stepping 2] (32 processors) --- Linux CentOS CentOS Linux release 7.4.1708 (Core) [3.10.0-514.26.2.el7.x86_64] 17 Jun 2018, 11:07:29 UTC Host 6 ID: 10511347 Details | Tasks Cross-project stats: BOINCstats.com Free-DC 682 10,034.29 1,281,656 7.8.4 GenuineIntel Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz [Family 6 Model 63 Stepping 2] (32 processors) --- Linux CentOS CentOS Linux release 7.4.1708 (Core) [3.10.0-514.26.2.el7.x86_64] 17 Jun 2018, 10:23:36 UTC In their logs appears the same lines : <core_client_version>7.8.4</core_client_version> Other similar hosts works fine but have this line instead of the red above : <core_client_version>7.8.4</core_client_version> Is it the reason for the constant failure ? |
Send message Joined: 15 Jun 08 Posts: 2411 Credit: 226,067,265 RAC: 126,878 |
@PHILIPPE Regarding the hosts mentioned in your post: Independent from the error I highly recommend to reduce the #threads (currently 12) as this would be extremely inefficient. #threads may not exceed 4. @PHILIPPE @PoppaGeek The error will most likely disappear if the owner (Agile Boincers or PoppaGeek) runs "sudo cvmfs_config wipecache" and "cvmfs_config probe" immediately before the next WU starts. |
Send message Joined: 4 Dec 15 Posts: 7 Credit: 1,076,791 RAC: 307 |
Got some problems with properties, I think: https://lhcathome.cern.ch/lhcathome/result.php?resultid=199031637 Tasks crash after ten minutes. Can I do anything about that? - - - - - - - - - - Greetings, Jens |
Send message Joined: 9 Dec 14 Posts: 202 Credit: 2,533,875 RAC: 0 |
Got some problems with properties, I think:looks like your cvmfs installation is not correct: Checking for CVMFS ls: Zugriff auf '/cvmfs/atlas.cern.ch/repo/sw' nicht möglich: Datei oder Verzeichnis nicht gefunden cvmfs_config doesn't exist, check cvmfs with cmd ls /cvmfs/atlas.cern.ch/repo/sw ls /cvmfs/atlas.cern.ch/repo/sw failed,aborting the jobs Did you install cvmfs, and if yes, how? Btw, you also need to install singularity (if ot already done) for running native ATLAS on most Linux OS's. |
Send message Joined: 4 Dec 15 Posts: 7 Credit: 1,076,791 RAC: 307 |
Hi. Thanks for your answer. I have neither cvmfs nor singularity (or is it singular?) installed. Just put some plain Debian onto my machine. - - - - - - - - - - Greetings, Jens |
Send message Joined: 15 Nov 14 Posts: 602 Credit: 24,371,321 RAC: 0 |
I have neither cvmfs nor singularity (or is it singular?) installed. Here is my latest version for Ubuntu 16.04. I suppose it is the same, or similar, for Debian. CVMFS: In order to add the apt repository, run wget https://ecsft.cern.ch/dist/cvmfs/cvmfs-release/cvmfs-release-latest_all.deb sudo dpkg -i cvmfs-release-latest_all.deb rm -f cvmfs-release-latest_all.deb sudo apt update ================================================= Installation: Step 1 Install the CernVM-FS packages: sudo apt install cvmfs cvmfs-config-default NOTE: If error about unmet dependencies related to "curl", run: sudo apt remove libcurl3 and then run Step 1 again. ------------------------------------------------- Step 2 Base setup: sudo cvmfs_config setup ------------------------------------------------- Step 3 Create /etc/cvmfs/default.local and open the file for editing. (e.g., using sudo gedit) sudo gedit /etc/cvmfs/default.local ------------------------------------------------- Step 4 Place the desired repositories in "default.local". For ATLAS, for instance, set CVMFS_REPOSITORIES=atlas.cern.ch,atlas-condb.cern.ch,grid.cern.ch CVMFS_HTTP_PROXY=DIRECT ------------------------------------------------- Step 5 Check if CernVM-FS mounts the specified repositories by: sudo cvmfs_config probe sudo cvmfs_config chksetup If the probe fails, try to restart autofs with sudo service autofs restart. Singularity First, check the latest version on GitHub: https://github.com/singularityware/singularity/releases And substitute that for "$VERSION" $VERSION=2.5.1 (as of 6 May 2018) wget https://github.com/singularityware/singularity/releases/download/$VERSION/singularity-$VERSION.tar.gz => wget https://github.com/singularityware/singularity/releases/download/2.5.1/singularity-2.5.1.tar.gz tar xvf singularity-$VERSION.tar.gz => tar xvf singularity-2.5.1.tar.gz cd singularity-$VERSION => cd singularity-2.5.1 Install package libarchive sudo apt install libarchive-dev ./configure --prefix=/usr/local make sudo make install To check the version installed: singularity --version To check the usage: singularity --help It should work, though you may have to enable "Run test applications?" in your settings to get the native ATLAS. EDIT: Also, I should point out that if you have VirtualBox installed, you will get both the VBox and native versions of ATLAS, depending on the whims of the LHC server. If you want only native ATLAS, then you must remove or deactivate VBox. |
Send message Joined: 4 Dec 15 Posts: 7 Credit: 1,076,791 RAC: 307 |
Thanks a lot for this input. The most important thing to me might be this: EDIT: Also, I should point out that if you have VirtualBox installed, you will get both the VBox and native versions of ATLAS, depending on the whims of the LHC server. If you want only native ATLAS, then you must remove or deactivate VBox. I've been getting mainly or only native ATLAS although I'm running Virtual Box. I thought it was using vbox, so I wondered what was wrong. On my Macs I just have to add the project and vbox, and everything is fine. If I should get into problems installing those two (Singularity obviously not being the game I found in the Debian packages), I can still configure Boinc to not run native ATLAS, so it will use the vbox version. Will give all this a try later-on. So tired I think I might mess something up. Thanks again. - - - - - - - - - - Greetings, Jens |
©2024 CERN