Message boards :
Theory Application :
Move TheoryN Back Into Theory.
Message board moderation
Previous · 1 · 2 · 3 · 4
Author | Message |
---|---|
Send message Joined: 2 May 07 Posts: 2192 Credit: 173,352,756 RAC: 52,278 |
Is this a possibility for get only one task in -native? <fetch_minimal_work>0|1</fetch_minimal_work> Fetch one job per device (see --fetch_minimal_work). https://boinc.berkeley.edu/wiki/Client_configuration#Logging_flags |
Send message Joined: 20 Jun 14 Posts: 380 Credit: 238,712 RAC: 0 |
This will affect all VBox apps. I will investigate how to disable Max # of CPUs for single threaded apps. I believe the relevant line is here. I don't think we can just disable this as it is used for ATLAS and CMS to select the number of CPUs to use for a VM so there needs to be an AND statement with something where that something is essentially !Theory. |
Send message Joined: 20 Jun 14 Posts: 380 Credit: 238,712 RAC: 0 |
This will affect all VBox apps. I will investigate how to disable Max # of CPUs for single threaded apps. So thinking about this a bit more, I think the use of Max # CPUs is a mistake. For a start from the BOINC scheduling perspective this is threads. In the vboxwrapper, this parameter is interpreted as CPUs. Also in the current implementation this affects the whole project where you may want to defined the VM size by host and project. The best way to do this is in the app_config.xml on the client, for example: <app_config> <app_version> <app_name>ATLAS</app_name> <plan_class>vbox64_mt_mcore_atlas</plan_class> <avg_ncpus>x</avg_ncpus> </app_version> </app_config> or <app_config> <app_version> <app_name>ATLAS</app_name> <plan_class>vbox64_mt_mcore_atlas</plan_class> <cmdline>--nthreads 7</cmdline> </app_version>] </app_config> My suggestion would be therefore to disable the Max # CPU functionality and control it with the app_config.xml. Comments? |
Send message Joined: 14 Jan 10 Posts: 1376 Credit: 9,162,540 RAC: 5,071 |
My suggestion would be therefore to disable the Max # CPU functionality and control it with the app_config.xml. Comments?There is one big BUT. When using only single core from server perspective, BOINC credit will calculate and grant credit based on the elapsed time times reported GFLOPS. When a user setup VM's as dual, quadcore whatever by using app_config.xml, the elapsed time will reduce and his credit will be significant lower. A lot of crunchers will not appreciate that. We know that only ATLAS real benefits from multi-core. As long as ATLAS tasks are rather equal one could change to fixed credit / task for ATLAS. |
Send message Joined: 15 Jun 08 Posts: 2506 Credit: 248,881,528 RAC: 129,054 |
One of the reasons why Max#CPUs has been introduced was to simplify the ATLAS multicore configuration for users who didn't want to deal with an app_config.xml When a workunit is sent to the client via scheduler reply it includes an <app_version> section and within this section <avg_ncpus> is set. The client copies this <avg_ncpus> value to it's client_state.xml but overwrites it with the value from app_config.xml. From now on the client as well as the vboxwrapper use the value that has been set last. Unfortunately <avg_ncpus> is never reported back to the server via scheduler request. It's in the user's responsibility to keep the values in sync. If server and client are not in sync this affects credit calculation as well as the amount of work the server will send in future requests (at least until the FLOPS value is adjusted). If Max#CPUs will be deactivated, every workunit will be treated as singlecore on the server, even ATLAS. Not a problem for other apps since all of them are now singlecore but ATLAS may require a reworked default policy, e.g. a singlecore default combined with a new method to identify if the client runs tasks with n threads. |
Send message Joined: 2 May 07 Posts: 2192 Credit: 173,352,756 RAC: 52,278 |
native Theory(300.02) and native-Atlas(2.73) get only ONE Task for me. native Theory(1.01) got two tasks with allways ONE Cpu in use. Had nothing changed in prefs or app_config. |
Send message Joined: 20 Jun 14 Posts: 380 Credit: 238,712 RAC: 0 |
native Theory(300.02) and native-Atlas(2.73) get only ONE Task for me. This agrees with the configuration. Both ATLAS and Theory set a limit of one task per cpu and TheoryN had two. |
Send message Joined: 20 Jun 14 Posts: 380 Credit: 238,712 RAC: 0 |
My suggestion would be therefore to disable the Max # CPU functionality and control it with the app_config.xml. Comments?There is one big BUT. Thanks for pointing this out. I didn't appreciate the affect on the credit. |
Send message Joined: 20 Jun 14 Posts: 380 Credit: 238,712 RAC: 0 |
One of the reasons why Max#CPUs has been introduced was to simplify the ATLAS multicore configuration for users who didn't want to deal with an app_config.xml This leads to a wider discussion but essentially there are policies and the implementation of those polices. We should first understand the policy that we need, how it should be implemented, then how it can be implemented within the limitations of the existing code base.
Thanks, I didn't appreciate this subtlety.
I think the current implementation is wrong. This sets max_cpus to be effective_ncpus but you are talking about avg_ncpus. Will dig a little more into the code. |
Send message Joined: 15 Jun 08 Posts: 2506 Credit: 248,881,528 RAC: 129,054 |
... you are talking about avg_ncpus See: https://boinc.berkeley.edu/wiki/Client_configuration https://github.com/BOINC/boinc/blob/3a132e1ee2964fc8a03adcff7b369bd32231377b/client/app_config.cpp Only <avg_ncpus> can be modified by app_config.xml. Other (CPU-)options not mentioned at the referenced pages will be ignored by the client. |
Send message Joined: 12 Jun 18 Posts: 126 Credit: 53,906,164 RAC: 0 |
Unfortunately <avg_ncpus> is never reported back to the server via scheduler request.Should clients be adding a line to our ATLAS app_config files??? |
Send message Joined: 15 Jun 08 Posts: 2506 Credit: 248,881,528 RAC: 129,054 |
Should clients be adding a line to our ATLAS app_config files??? Would better be asked in the ATLAS threads. The lower the #threads the more efficient an ATLAS task will be. 2 reasons lead to higher #threads: - ATLAS (vbox) requires lots of RAM per task - less #threads ==> higher overall runtimes If your web preferences are set to "unlimited" then control the #threads using <avg_ncpus> in your app_config.xml The server will set up the task as an n-core task with n = the number of reported CPUs but with an upper limit of 12. This affects only ATLAS as meanwhile all other LHC@home apps are singlecore. |
Send message Joined: 20 Jun 14 Posts: 380 Credit: 238,712 RAC: 0 |
I am going to move this discussion to the number crunching topic as it affects all apps. |
©2024 CERN