Message boards :
ATLAS application :
Problems with BOINC:Virtualbox (Bad page map in process ...)
Message board moderation
Author | Message |
---|---|
Send message Joined: 30 Oct 18 Posts: 16 Credit: 192,743,156 RAC: 0 |
Hi everybody, i have an unusual problem to run ATLAS applications in BOINC. I got a Scientific Linux 7, but i try with Ubuntu 18 and i got the same problem. I follow the Yeti checklist to install and configure BOINC with Virtualbox, but i can't solve the problem. The problem is within the Virtualbox. My tasks apparently run, but when i look at the Virtualbox machine, i see that: It's says that i have a BUG: Bad page map in process ... The result of this is an computing error after 6-7 days, and i never complete one task. Could someone tell me how to fix it? Thank you! (Excuse me if my english is not very well) |
Send message Joined: 15 Jun 08 Posts: 2541 Credit: 254,608,838 RAC: 56,545 |
You may try the following steps: 1. Set all BOINC projects to "no new tasks". 2. Set all currently NOT running ATLAS tasks to "pause". 3. Cancel all currently running ATLAS tasks and wait a minute until the VMs are shut down. 4. Gracefully shutdown your BOINC client. 5. Check your VirtualBox manager and remove all suspect VMs. 6. Check the folder "/var/lib/boinc-client/slots/". If you only run ATLAS, it should be empty. 7- Check if your BOINC client account has access rights to the slots folder 8. Set your project preferences to "max #CPUs = 2" 9. Create an app_config.xml in "your_boinc-client_basic_folder/projects/lhcathome.cern.ch_lhcathome/" <app_config> <app_version> <app_name>ATLAS</app_name> <plan_class>vbox64_mt_mcore_atlas</plan_class> <avg_ncpus>2.0</avg_ncpus> <cmdline>--memory_size_mb 4800</cmdline> </app_version> <project_max_concurrent>2</project_max_concurrent> </app_config> 10. Restart your host 11. Resume all paused tasks 12. Allow your client to request new tasks If this doesn't work you may check your firewall for closed ports (see: http://lhcathome.web.cern.ch/test4theory/my-firewall-complaining-which-ports-does-project-use) and/or do a project reset to get a fresh *.vdi file. |
Send message Joined: 30 Oct 18 Posts: 16 Credit: 192,743,156 RAC: 0 |
I do all steps that you recommend me, but i got the same error. The firewall is disable, and i got VirtualBox Extension pack installed and VT-x enable. The message is the same: So i don't know what is the matter. Any more ideas? Thank you for your help! |
Send message Joined: 15 Jun 08 Posts: 2541 Credit: 254,608,838 RAC: 56,545 |
Some rare comments you can find with google point out that the error may be a result of bad physical RAM. Could you check this with memtest? |
Send message Joined: 2 Sep 04 Posts: 455 Credit: 201,266,333 RAC: 9,771 |
|
Send message Joined: 30 Oct 18 Posts: 16 Credit: 192,743,156 RAC: 0 |
Some rare comments you can find with google point out that the error may be a result of bad physical RAM. Yes, i ran memtest for a one day and everything was right. When i looked the problem in google, i only found a bad RAM or bad kernel, but i think that it's is impossible, because i try with a three different PC's with the same hardware and all have the same problem. |
Send message Joined: 30 Oct 18 Posts: 16 Credit: 192,743,156 RAC: 0 |
I had some similar happenings. The only but working solution was to run Atlas-WUs as Single-Core and this worked fine And.. How can i change the boinc configuration to run Atlas-WUs as Single-Core? I think it's done through Project Preferences --> Max # CPUs, but i'm not sure, is it correct? |
Send message Joined: 14 Jan 10 Posts: 1422 Credit: 9,484,585 RAC: 1,882 |
And.. How can i change the boinc configuration to run Atlas-WUs as Single-Core? I think it's done through Project Preferences --> Max # CPUs, but i'm not sure, is it correct? That's correct. Max # CPUs is used to create a VM with # cores and also used by BOINC to reserve the needed memory. |
Send message Joined: 30 Oct 18 Posts: 16 Credit: 192,743,156 RAC: 0 |
Maybe is not the only way. I found other possibility with app_config.xml file. If you put a <cmdline> --nthreats 1</cmdline> you'll have the same result. But if you have enough RAM memory to run 2 tasks, you can run both tasks at the same time. But i don't understand why my single-core tasks consume more than 2,1 GB of memory RAM, as indicated in Yeti's Checklist (Each SingeCore-Atlas-Task needs 2,1 GB free RAM, MultiCore-WUs need 3,0 GB + 0,9 GB * number of cores (Last Update from 01.08.2018) So 7,5 GB for a 5-Core WU.). Any ideas? Thank you all! |
Send message Joined: 15 Jun 08 Posts: 2541 Credit: 254,608,838 RAC: 56,545 |
To change the web preferences is the easiest way to get a good setup as it also adjusts the RAM calculation. The server sends the correct values within the scheduler_reply. This method is recommended if you run only 1 subproject, e.g. ATLAS. #cpus and RAM settings from the server can be overruled by a local app_config.xml. If you use "<avg_ncpus>x</avg_ncpus>", "<cmdline>--nthreads ..." should be obsolete. BTW: It's "nthreads" instead of "nthreats" But i don't understand why my single-core tasks consume more than 2,1 GB of memory RAM, as indicated in Yeti's Checklist ... It was necessary to change the RAM fromula a couple of times during the last years. Yeti also posted the recent one: Yeti wrote: ... MultiCore-WUs need 3,0 GB + 0,9 GB * number of cores |
Send message Joined: 2 Sep 04 Posts: 455 Credit: 201,266,333 RAC: 9,771 |
But i don't understand why my single-core tasks consume more than 2,1 GB of memory RAM, as indicated in Yeti's Checklist (Each SingeCore-Atlas-Task needs 2,1 GB free RAM, MultiCore-WUs need 3,0 GB + 0,9 GB * number of cores (Last Update from 01.08.2018) So 7,5 GB for a 5-Core WU.). Aahh, time has overtaken this. In former times there has been a separat app for SIngleCore-WUs, these needed 2,1 GB. But this app has been cancelled. When you run Atlas with 1-Core, than it is still a MulticoreWU, only reduced to 1-Core. This will leed, as the Fomula calculates, to 3,9 GB for 1-Cor WU I will try to find a better wording in the Checklist Supporting BOINC, a great concept ! |
Send message Joined: 2 Sep 04 Posts: 455 Credit: 201,266,333 RAC: 9,771 |
|
Send message Joined: 30 Oct 18 Posts: 16 Credit: 192,743,156 RAC: 0 |
To change the web preferences is the easiest way to get a good setup as it also adjusts the RAM calculation. Ahhh, now i understand. Sorry for my bad english, i wanted to say threads not threats. So now i know that my computer don't have Hyper-Threading Technology and i need to put --nthreads flag to work fine. But i think that <avg_ncpus>x<avg_ncpus> is different that <cmdline>--ntheads</cmdline>, because the cores of CPU can run multiple threads, if you have a Hyper-Threading Technology in your computer. But it's my opinion. Thank you for all! |
Send message Joined: 30 Oct 18 Posts: 16 Credit: 192,743,156 RAC: 0 |
HM, I had already made a note about this: Okey, so now all it's fine. i thought that this formula was currently valid. Thank you for time! |
©2024 CERN