Changes between Version 18 and Version 19 of AppCoprocessor
- Timestamp:
- Aug 21, 2009, 11:38:19 AM (16 years ago)
Legend:
- Unmodified
- Added
- Removed
- Modified
-
AppCoprocessor
v18 v19 1 1 = Applications that use coprocessors = 2 2 3 This document describes BOINC's support for applications that use coprocessors such as 4 5 * GPUs 6 * Cell SPEs 7 8 We'll assume that these resources are allocated rather than scheduled: 9 i.e., an application using a coprocessor has it locked while the app is in memory, 10 even if the app is suspended by BOINC or descheduled by the OS. 3 BOINC supports applications that use coprocessors. 4 The supported coprocessor types (as of [18892])are NVIDIA and API GPUs. 11 5 12 6 The BOINC client probes for coprocessors and reports them in scheduler requests. … … 14 8 It only runs an app if enough instances are available. 15 9 10 You can develop your application using any programming system, e.g. 11 CUDA (for NVIDIA), Brook+ (for ATI) or OpenCL. 12 13 == Command-line arguments == 14 15 Some hosts have multiple GPUs. 16 When your application is run by BOINC, it will be passed a command-line argument 17 {{{ 18 --device N 19 }}} 20 where N is the device number of the GPU that is to be used. 21 If your application uses multiple GPUs, 22 it will be passed multiple --device arguments, e.g. 23 {{{ 24 --device 0 --device 3 25 }}} 26 16 27 == Deploying a coprocessor app == 17 28 18 BOINC uses the [AppPlan application planning] mechanism to 19 coordinate the scheduling of multi-threaded applications. 29 When you deploy a coprocessor app you must specify: 20 30 21 Suppose you've developed a coprocessor program, 22 that it uses a CUDA GPU and 1 GFLOPS of the CPU, 23 and produces a total of 100 GFLOPS. 24 To deploy it: 31 * its hardware and software requirements 32 * an estimate of what fraction of a CPU it will use 33 * an estimate of its performance on individual hosts 25 34 26 * Choose a "planning class" name for the program, say "cuda" (see below). 35 This information is specified in an 36 [AppPlan application planning function] that you link into your scheduler. 37 Specifically, you must: 38 39 * Choose a "plan class" name for your program, say "cuda" (see below). 27 40 * Create an [UpdateVersions app version], specifying its plan class as "cuda". 28 * Link the following function into your scheduler (customize as needed): 41 * Edit the function '''app_plan()''' in '''sched/sched_customize.cpp''' so that it contains a clause for your plan class. 42 43 The default '''app_plan()''' contains a clause for plan class '''cuda'''. 44 We will explain its logic; you may need to modify it for your CUDA app. 45 46 First, we check if the host has an NVIDIA GPU. 29 47 {{{ 30 48 int app_plan(SCHEDULER_REQUEST& sreq, char* plan_class, HOST_USAGE& hu) { 49 ... 31 50 if (!strcmp(plan_class, "cuda")) { 32 // the following is for an app that uses a CUDA GPU33 //34 51 COPROC_CUDA* cp = (COPROC_CUDA*)sreq.coprocs.lookup("CUDA"); 35 52 if (!cp) { … … 41 58 return PLAN_REJECT_CUDA_NO_DEVICE; 42 59 } 60 }}} 61 62 Check the compute capability (1.0 or better): 63 {{{ 43 64 int v = (cp->prop.major)*100 + cp->prop.minor; 44 65 if (v < 100) { … … 50 71 return PLAN_REJECT_CUDA_VERSION; 51 72 } 73 }}} 52 74 53 if (cp->drvVersion && cp->drvVersion < PLAN_CUDA_MIN_DRIVER_VERSION) { 54 if (config.debug_version_select) { 55 log_messages.printf(MSG_NORMAL, 56 "[version] NVIDIA driver version %d < PLAN_CUDA_MIN_DRIVER_VERSION\n", 57 cp->drvVersion 58 ); 75 Check the CUDA runtime version. 76 As of client version 6.10, all clients report the CUDA runtime version 77 (cp->cuda_version); use that if it's present. 78 In 6.8 and earlier, the CUDA runtime version isn't reported. 79 Windows clients report the driver version, 80 from which the CUDA version can be inferred; 81 Linux clients don't return the driver version, 82 so we don't know what the CUDA version is. 83 {{{ 84 // for CUDA 2.3, we need to check the CUDA RT version. 85 // Old BOINC clients report display driver version; 86 // newer ones report CUDA RT version 87 // 88 if (!strcmp(plan_class, "cuda23")) { 89 if (cp->cuda_version) { 90 if (cp->cuda_version < 2030) { 91 return PLAN_REJECT_CUDA_VERSION; 92 } 93 } else if (cp->display_driver_version) { 94 if (cp->display_driver_version < PLAN_CUDA23_MIN_DRIVER_VERSION) { 95 return PLAN_REJECT_CUDA_VERSION; 96 } 97 } else { 98 return PLAN_REJECT_CUDA_VERSION; 59 99 } 60 return PLAN_REJECT_NVIDIA_DRIVER_VERSION; 61 } 100 }}} 62 101 102 Check for the amount of video RAM: 103 {{{ 63 104 if (cp->prop.dtotalGlobalMem < PLAN_CUDA_MIN_RAM) { 64 105 if (config.debug_version_select) { … … 70 111 return PLAN_REJECT_CUDA_MEM; 71 112 } 113 }}} 114 115 Estimate the FLOPS: 116 {{{ 72 117 hu.flops = cp->flops_estimate(); 118 }}} 73 119 120 Estimate its CPU usage: 121 {{{ 74 122 // assume we'll need 0.5% as many CPU FLOPS as GPU FLOPS 75 123 // to keep the GPU fed. … … 78 126 hu.avg_ncpus = x; 79 127 hu.max_ncpus = x; 80 81 hu.ncudas = 1;82 83 if (config.debug_version_select) {84 log_messages.printf(MSG_NORMAL,85 "[version] CUDA app estimated %.2f GFLOPS (clock %d count %d)\n",86 hu.flops/1e9, cp->prop.clockRate,87 cp->prop.multiProcessorCount88 );89 }90 return 0;91 }92 log_messages.printf(MSG_CRITICAL,93 "Unknown plan class: %s\n", plan_class94 );95 return PLAN_REJECT_UNKNOWN;96 }97 128 }}} 98 129 99 == Questions == 130 Indicate the number of GPUs used. 131 Typically this will be 1. 132 If your application uses only a fraction X<1 of the CPU processors, 133 and a fraction Y<1 of video RAM, 134 reports the number of GPUs as min(X, Y). 135 In this case BOINC will attempt to run multiple jobs per GPU is possible. 136 {{{ 137 hu.ncudas = 1; 138 }}} 100 139 101 * How does BOINC know if non-BOINC applications are using resources? 102 103 140 Return 0 to indicate that the application can be run on the host: 141 {{{ 142 return 0; 143 }}}