boinc/sched/sched_send.cpp

1806 lines
58 KiB
C++

// This file is part of BOINC.
// http://boinc.berkeley.edu
// Copyright (C) 2008 University of California
//
// BOINC is free software; you can redistribute it and/or modify it
// under the terms of the GNU Lesser General Public License
// as published by the Free Software Foundation,
// either version 3 of the License, or (at your option) any later version.
//
// BOINC is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
// See the GNU Lesser General Public License for more details.
//
// You should have received a copy of the GNU Lesser General Public License
// along with BOINC. If not, see <http://www.gnu.org/licenses/>.
// scheduler code related to sending jobs.
// NOTE: there should be nothing here specific to particular
// scheduling policies (array scan, matchmaking, locality)
#include "config.h"
#include <vector>
#include <list>
#include <string>
#include <ctime>
#include <cstdio>
#include <cstring>
#include <stdlib.h>
#include <sys/time.h>
#include <unistd.h>
#include "error_numbers.h"
#include "parse.h"
#include "util.h"
#include "str_util.h"
#include "synch.h"
#include "credit.h"
#include "sched_types.h"
#include "sched_shmem.h"
#include "sched_config.h"
#include "sched_util.h"
#include "sched_main.h"
#include "sched_array.h"
#include "sched_msgs.h"
#include "sched_hr.h"
#include "hr.h"
#include "sched_locality.h"
#include "sched_timezone.h"
#include "sched_assign.h"
#include "sched_customize.h"
#include "sched_version.h"
#include "sched_send.h"
#ifdef _USING_FCGI_
#include "boinc_fcgi.h"
#endif
// if host sends us an impossible RAM size, use this instead
//
const double DEFAULT_RAM_SIZE = 64000000;
void send_work_matchmaker();
int preferred_app_message_index=0;
const char* infeasible_string(int code) {
switch (code) {
case INFEASIBLE_MEM: return "Not enough memory";
case INFEASIBLE_DISK: return "Not enough disk";
case INFEASIBLE_CPU: return "CPU too slow";
case INFEASIBLE_APP_SETTING: return "App not selected";
case INFEASIBLE_WORKLOAD: return "Existing workload";
case INFEASIBLE_DUP: return "Already in reply";
case INFEASIBLE_HR: return "Homogeneous redundancy";
case INFEASIBLE_BANDWIDTH: return "Download bandwidth too low";
}
return "Unknown";
}
const double MIN_REQ_SECS = 0;
const double MAX_REQ_SECS = (28*SECONDS_IN_DAY);
const int MAX_GPUS = 8;
// don't believe clients who claim they have more GPUs than this
// get limits on:
// # jobs per day
// # jobs per RPC
// # jobs in progress
//
void WORK_REQ::get_job_limits() {
int n;
n = g_reply->host.p_ncpus;
if (g_request->global_prefs.max_ncpus_pct && g_request->global_prefs.max_ncpus_pct < 100) {
n = (int)((n*g_request->global_prefs.max_ncpus_pct)/100.);
}
if (n > config.max_ncpus) n = config.max_ncpus;
if (n < 1) n = 1;
effective_ncpus = n;
n = g_request->coprocs.nvidia.count + g_request->coprocs.ati.count;
if (n > MAX_GPUS) n = MAX_GPUS;
effective_ngpus = n;
int mult = effective_ncpus + config.gpu_multiplier * effective_ngpus;
if (config.non_cpu_intensive) {
mult = 1;
effective_ncpus = 1;
if (effective_ngpus) effective_ngpus = 1;
}
if (config.max_wus_to_send) {
g_wreq->max_jobs_per_rpc = mult * config.max_wus_to_send;
} else {
g_wreq->max_jobs_per_rpc = 999999;
}
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] effective ncpus %d ngpus %d\n",
effective_ncpus, effective_ngpus
);
}
config.max_jobs_in_progress.reset(effective_ncpus, effective_ngpus);
}
static const char* find_user_friendly_name(int appid) {
APP* app = ssp->lookup_app(appid);
if (app) return app->user_friendly_name;
return "deprecated application";
}
// Compute the max additional disk usage we can impose on the host.
// Depending on the client version, it can either send us
// - d_total and d_free (pre 4 oct 2005)
// - the above plus d_boinc_used_total and d_boinc_used_project
//
double max_allowable_disk() {
HOST host = g_request->host;
GLOBAL_PREFS prefs = g_request->global_prefs;
double x1, x2, x3, x;
// defaults are from config.xml
// if not there these are used:
// -default_max_used_gb= 100
// -default_max_used_pct = 50
// -default_min_free_gb = .001
//
if (prefs.disk_max_used_gb == 0) {
prefs.disk_max_used_gb = config.default_disk_max_used_gb;
}
if (prefs.disk_max_used_pct == 0) {
prefs.disk_max_used_pct = config.default_disk_max_used_pct;
}
if (prefs.disk_min_free_gb < config.default_disk_min_free_gb) {
prefs.disk_min_free_gb = config.default_disk_min_free_gb;
}
// no defaults for total/free disk space (host.d_total, d_free)
// if they're zero, client will get no work.
//
if (host.d_boinc_used_total) {
// The post 4 oct 2005 case.
// Compute the max allowable additional disk usage based on prefs
//
x1 = prefs.disk_max_used_gb*GIGA - host.d_boinc_used_total;
x2 = host.d_total*prefs.disk_max_used_pct/100.
- host.d_boinc_used_total;
x3 = host.d_free - prefs.disk_min_free_gb*GIGA; // may be negative
x = std::min(x1, std::min(x2, x3));
// see which bound is the most stringent
//
if (x==x1) {
g_reply->disk_limits.max_used = x;
} else if (x==x2) {
g_reply->disk_limits.max_frac = x;
} else {
g_reply->disk_limits.min_free = x;
}
} else {
// here we don't know how much space BOINC is using.
// so we're kinda screwed.
// All we can do is assume that BOINC is using zero space.
// We can't honor the max_used for max_used_pct preferences.
// We can only honor the min_free pref.
//
x = host.d_free - prefs.disk_min_free_gb*GIGA; // may be negative
g_reply->disk_limits.min_free = x;
x1 = x2 = x3 = 0;
}
if (x < 0) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] No disk space available: disk_max_used_gb %.2fGB disk_max_used_pct %.2f disk_min_free_gb %.2fGB\n",
prefs.disk_max_used_gb/GIGA,
prefs.disk_max_used_pct,
prefs.disk_min_free_gb/GIGA
);
log_messages.printf(MSG_NORMAL,
"[send] No disk space available: host.d_total %.2fGB host.d_free %.2fGB host.d_boinc_used_total %.2fGB\n",
host.d_total/GIGA,
host.d_free/GIGA,
host.d_boinc_used_total/GIGA
);
log_messages.printf(MSG_NORMAL,
"[send] No disk space available: x1 %.2fGB x2 %.2fGB x3 %.2fGB x %.2fGB\n",
x1/GIGA, x2/GIGA, x3/GIGA, x/GIGA
);
}
g_wreq->disk.set_insufficient(-x);
x = 0;
}
return x;
}
static double estimate_duration_unscaled(WORKUNIT& wu, BEST_APP_VERSION& bav) {
double rsc_fpops_est = wu.rsc_fpops_est;
if (rsc_fpops_est <= 0) rsc_fpops_est = 1e12;
return rsc_fpops_est/bav.host_usage.projected_flops;
}
static inline void get_running_frac() {
double rf;
if (g_request->core_client_version<=41900) {
rf = g_reply->host.on_frac;
} else {
rf = g_reply->host.active_frac * g_reply->host.on_frac;
}
// clamp running_frac to a reasonable range
//
if (rf > 1) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL, "[send] running_frac=%f; setting to 1\n", rf);
}
rf = 1;
} else if (rf < .1) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL, "[send] running_frac=%f; setting to 0.1\n", rf);
}
rf = .1;
}
g_wreq->running_frac = rf;
}
// estimate the amount of real time to complete this WU,
// taking into account active_frac etc.
// Note: don't factor in resource_share_fraction.
// The core client doesn't necessarily round-robin across all projects.
//
double estimate_duration(WORKUNIT& wu, BEST_APP_VERSION& bav) {
double edu = estimate_duration_unscaled(wu, bav);
double ed = edu/g_wreq->running_frac;
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] est. duration for WU %d: unscaled %.2f scaled %.2f\n",
wu.id, edu, ed
);
}
return ed;
}
static void get_prefs_info() {
char buf[8096];
std::string str;
unsigned int pos = 0;
int temp_int=0;
bool flag;
extract_venue(g_reply->user.project_prefs, g_reply->host.venue, buf);
str = buf;
// scan user's project prefs for elements of the form <app_id>N</app_id>,
// indicating the apps they want to run.
//
g_wreq->preferred_apps.clear();
while (parse_int(str.substr(pos,str.length()-pos).c_str(), "<app_id>", temp_int)) {
APP_INFO ai;
ai.appid = temp_int;
ai.work_available = false;
g_wreq->preferred_apps.push_back(ai);
pos = str.find("<app_id>", pos) + 1;
}
if (parse_bool(buf,"allow_non_preferred_apps", flag)) {
g_wreq->allow_non_preferred_apps = flag;
}
if (parse_bool(buf,"allow_beta_work", flag)) {
g_wreq->allow_beta_work = flag;
}
if (parse_bool(buf,"no_gpus", flag)) {
// deprecated, but need to handle
if (flag) {
g_wreq->no_cuda = true;
g_wreq->no_ati = true;
}
}
if (parse_bool(buf,"no_cpu", flag)) {
g_wreq->no_cpu = flag;
}
if (parse_bool(buf,"no_cuda", flag)) {
g_wreq->no_cuda = flag;
}
if (parse_bool(buf,"no_ati", flag)) {
g_wreq->no_ati = flag;
}
}
// Decide whether or not this app version is 'reliable'
// An app version is reliable if the following conditions are true
// (for those that are set in the config file)
// 1) The host average turnaround is less than a threshold
// 2) consecutive_valid is above a threshold
// 3) The host results per day is equal to the max value
//
void get_reliability_version(HOST_APP_VERSION& hav, double multiplier) {
if (hav.turnaround.n > MIN_HOST_SAMPLES && config.reliable_max_avg_turnaround) {
if (hav.turnaround.get_avg() > config.reliable_max_avg_turnaround*multiplier) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [AV#%d] not reliable; avg turnaround: %.3f > %.3f hrs\n",
hav.app_version_id,
hav.turnaround.get_avg()/3600,
config.reliable_max_avg_turnaround*multiplier/3600
);
}
hav.reliable = false;
return;
}
}
if (hav.consecutive_valid < CONS_VALID_RELIABLE) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [AV#%d] not reliable; cons valid %d < %d\n",
hav.app_version_id,
hav.consecutive_valid, CONS_VALID_RELIABLE
);
}
hav.reliable = false;
return;
}
if (config.daily_result_quota) {
if (hav.max_jobs_per_day < config.daily_result_quota) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [AV#%d] not reliable; max_jobs_per_day %d>%d\n",
hav.app_version_id,
hav.max_jobs_per_day,
config.daily_result_quota
);
}
hav.reliable = false;
return;
}
}
hav.reliable = true;
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [HOST#%d] app version %d is reliable\n",
g_reply->host.id, hav.app_version_id
);
}
g_wreq->has_reliable_version = true;
}
// decide whether do unreplicated jobs with this app version
//
static void set_trust(DB_HOST_APP_VERSION& hav) {
hav.trusted = false;
if (hav.consecutive_valid < CONS_VALID_UNREPLICATED) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] set_trust: cons valid %d < %d, don't use single replication\n",
hav.consecutive_valid, CONS_VALID_UNREPLICATED
);
}
return;
}
double x = 1./hav.consecutive_valid;
if (drand() > x) hav.trusted = true;
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] set_trust: random choice for cons valid %d: %s\n",
hav.consecutive_valid, hav.trusted?"yes":"no"
);
}
}
static void update_quota(DB_HOST_APP_VERSION& hav) {
if (config.daily_result_quota) {
if (hav.max_jobs_per_day == 0) {
hav.max_jobs_per_day = config.daily_result_quota;
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] [HAV#%d] Initializing max_results_day to %d\n",
hav.app_version_id,
config.daily_result_quota
);
}
}
}
if (g_request->last_rpc_dayofyear != g_request->current_rpc_dayofyear) {
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] [HOST#%d] [HAV#%d] Resetting n_jobs_today\n",
g_reply->host.id, hav.app_version_id
);
}
hav.n_jobs_today = 0;
}
}
void update_n_jobs_today() {
for (unsigned int i=0; i<g_wreq->host_app_versions.size(); i++) {
DB_HOST_APP_VERSION& hav = g_wreq->host_app_versions[i];
update_quota(hav);
}
}
static void get_reliability_and_trust() {
// Platforms other than Windows, Linux and Intel Macs need a
// larger set of computers to be marked reliable
//
double multiplier = 1.0;
if (strstr(g_reply->host.os_name,"Windows")
|| strstr(g_reply->host.os_name,"Linux")
|| (strstr(g_reply->host.os_name,"Darwin")
&& !(strstr(g_reply->host.p_vendor,"Power Macintosh"))
)) {
multiplier = 1.0;
} else {
multiplier = 1.8;
}
for (unsigned int i=0; i<g_wreq->host_app_versions.size(); i++) {
DB_HOST_APP_VERSION& hav = g_wreq->host_app_versions[i];
get_reliability_version(hav, multiplier);
set_trust(hav);
}
}
// Return true if the user has set application preferences,
// and this job is not for a selected app
//
bool app_not_selected(WORKUNIT& wu) {
unsigned int i;
if (g_wreq->preferred_apps.size() == 0) return false;
for (i=0; i<g_wreq->preferred_apps.size(); i++) {
if (wu.appid == g_wreq->preferred_apps[i].appid) {
g_wreq->preferred_apps[i].work_available = true;
return false;
}
}
return true;
}
// see how much RAM we can use on this machine
//
static inline void get_mem_sizes() {
g_wreq->ram = g_reply->host.m_nbytes;
if (g_wreq->ram <= 0) g_wreq->ram = DEFAULT_RAM_SIZE;
g_wreq->usable_ram = g_wreq->ram;
double busy_frac = g_request->global_prefs.ram_max_used_busy_frac;
double idle_frac = g_request->global_prefs.ram_max_used_idle_frac;
double frac = 1;
if (busy_frac>0 && idle_frac>0) {
frac = std::max(busy_frac, idle_frac);
if (frac > 1) frac = 1;
g_wreq->usable_ram *= frac;
}
}
static inline int check_memory(WORKUNIT& wu) {
double diff = wu.rsc_memory_bound - g_wreq->usable_ram;
if (diff > 0) {
char message[256];
sprintf(message,
"%s needs %0.2f MB RAM but only %0.2f MB is available for use.",
find_user_friendly_name(wu.appid),
wu.rsc_memory_bound/MEGA, g_wreq->usable_ram/MEGA
);
add_no_work_message(message);
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [WU#%d %s] needs %0.2fMB RAM; [HOST#%d] has %0.2fMB, %0.2fMB usable\n",
wu.id, wu.name, wu.rsc_memory_bound/MEGA,
g_reply->host.id, g_wreq->ram/MEGA, g_wreq->usable_ram/MEGA
);
}
g_wreq->mem.set_insufficient(wu.rsc_memory_bound);
g_reply->set_delay(DELAY_NO_WORK_TEMP);
return INFEASIBLE_MEM;
}
return 0;
}
static inline int check_disk(WORKUNIT& wu) {
double diff = wu.rsc_disk_bound - g_wreq->disk_available;
if (diff > 0) {
char message[256];
sprintf(message,
"%s needs %0.2fMB more disk space. You currently have %0.2f MB available and it needs %0.2f MB.",
find_user_friendly_name(wu.appid),
diff/MEGA, g_wreq->disk_available/MEGA, wu.rsc_disk_bound/MEGA
);
add_no_work_message(message);
g_wreq->disk.set_insufficient(diff);
return INFEASIBLE_DISK;
}
return 0;
}
static inline int check_bandwidth(WORKUNIT& wu) {
if (wu.rsc_bandwidth_bound == 0) return 0;
// if n_bwdown is zero, the host has never downloaded anything,
// so skip this check
//
if (g_reply->host.n_bwdown == 0) return 0;
double diff = wu.rsc_bandwidth_bound - g_reply->host.n_bwdown;
if (diff > 0) {
char message[256];
sprintf(message,
"%s requires %0.2f KB/sec download bandwidth. Your computer has been measured at %0.2f KB/sec.",
find_user_friendly_name(wu.appid),
wu.rsc_bandwidth_bound/KILO, g_reply->host.n_bwdown/KILO
);
add_no_work_message(message);
g_wreq->bandwidth.set_insufficient(diff);
return INFEASIBLE_BANDWIDTH;
}
return 0;
}
// Determine if the app is "hard",
// and we should send it only to high-end hosts.
// Currently this is specified by setting weight=-1;
// this is a kludge for SETI@home/Astropulse.
//
static inline bool hard_app(APP& app) {
return (app.weight == -1);
}
static inline double get_estimated_delay(BEST_APP_VERSION& bav) {
if (bav.host_usage.ncudas) {
return g_request->coprocs.nvidia.estimated_delay;
} else if (bav.host_usage.natis) {
return g_request->coprocs.ati.estimated_delay;
} else {
return g_request->cpu_estimated_delay;
}
}
static inline void update_estimated_delay(BEST_APP_VERSION& bav, double dt) {
g_request->coprocs.nvidia.estimated_delay += dt*bav.host_usage.ncudas/g_request->coprocs.nvidia.count;
g_request->coprocs.ati.estimated_delay += dt*bav.host_usage.natis/g_request->coprocs.ati.count;
g_request->cpu_estimated_delay += dt*bav.host_usage.avg_ncpus/g_request->host.p_ncpus;
}
// return the delay bound to use for this job/host.
// Actually, return two: optimistic (lower) and pessimistic (higher).
// If the deadline check with the optimistic bound fails,
// try the pessimistic bound.
// TODO: clean up this mess
//
static void get_delay_bound_range(
WORKUNIT& wu,
int res_server_state, int res_priority, double res_report_deadline,
BEST_APP_VERSION& bav,
double& opt, double& pess
) {
if (res_server_state == RESULT_SERVER_STATE_IN_PROGRESS) {
double now = dtime();
if (res_report_deadline < now) {
// if original deadline has passed, return zeros
// This will skip deadline check.
opt = pess = 0;
}
opt = res_report_deadline - now;
pess = wu.delay_bound;
} else {
opt = pess = wu.delay_bound;
// If the workunit needs reliable and is being sent to a reliable host,
// then shorten the delay bound by the percent specified
//
if (config.reliable_on_priority && res_priority >= config.reliable_on_priority && config.reliable_reduced_delay_bound > 0.01
) {
opt = wu.delay_bound*config.reliable_reduced_delay_bound;
double est_wallclock_duration = estimate_duration(wu, bav);
// Check to see how reasonable this reduced time is.
// Increase it to twice the estimated delay bound
// if all the following apply:
//
// 1) Twice the estimate is longer then the reduced delay bound
// 2) Twice the estimate is less then the original delay bound
// 3) Twice the estimate is less then the twice the reduced delay bound
if (est_wallclock_duration*2 > opt
&& est_wallclock_duration*2 < wu.delay_bound
&& est_wallclock_duration*2 < wu.delay_bound*config.reliable_reduced_delay_bound*2
) {
opt = est_wallclock_duration*2;
}
}
}
}
// return 0 if the job, with the given delay bound,
// will complete by its deadline, and won't cause other jobs to miss deadlines.
//
static inline int check_deadline(
WORKUNIT& wu, APP& app, BEST_APP_VERSION& bav
) {
if (config.ignore_delay_bound) return 0;
// skip delay check if host currently doesn't have any work
// and it's not a hard app.
// (i.e. everyone gets one result, no matter how slow they are)
//
if (get_estimated_delay(bav) == 0 && !hard_app(app)) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] est delay 0, skipping deadline check\n"
);
}
return 0;
}
// if it's a hard app, don't send it to a host with no credit
//
if (hard_app(app) && g_reply->host.total_credit == 0) {
return INFEASIBLE_CPU;
}
// do EDF simulation if possible; else use cruder approximation
//
if (config.workload_sim && g_request->have_other_results_list) {
double est_dur = estimate_duration(wu, bav);
if (g_reply->wreq.edf_reject_test(est_dur, wu.delay_bound)) {
return INFEASIBLE_WORKLOAD;
}
IP_RESULT candidate("", wu.delay_bound, est_dur);
strcpy(candidate.name, wu.name);
if (check_candidate(candidate, g_wreq->effective_ncpus, g_request->ip_results)) {
// it passed the feasibility test,
// but don't add it to the workload yet;
// wait until we commit to sending it
} else {
g_reply->wreq.edf_reject(est_dur, wu.delay_bound);
g_reply->wreq.speed.set_insufficient(0);
return INFEASIBLE_WORKLOAD;
}
} else {
double ewd = estimate_duration(wu, bav);
if (hard_app(app)) ewd *= 1.3;
double est_completion_delay = get_estimated_delay(bav) + ewd;
double est_report_delay = std::max(
est_completion_delay,
g_request->global_prefs.work_buf_min()
);
double diff = est_report_delay - wu.delay_bound;
if (diff > 0) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [WU#%d] deadline miss %d > %d\n",
wu.id, (int)est_report_delay, wu.delay_bound
);
}
g_reply->wreq.speed.set_insufficient(diff);
return INFEASIBLE_CPU;
} else {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [WU#%d] meets deadline: %.2f + %.2f < %d\n",
wu.id, get_estimated_delay(bav), ewd, wu.delay_bound
);
}
}
}
return 0;
}
// Fast checks (no DB access) to see if the job can be sent to the host.
// Reasons why not include:
// 1) the host doesn't have enough memory;
// 2) the host doesn't have enough disk space;
// 3) based on CPU speed, resource share and estimated delay,
// the host probably won't get the result done within the delay bound
// 4) app isn't in user's "approved apps" list
//
// If the job is feasible, return 0 and fill in wu.delay_bound
// with the delay bound we've decided to use.
//
int wu_is_infeasible_fast(
WORKUNIT& wu,
int res_server_state, int res_priority, double res_report_deadline,
APP& app, BEST_APP_VERSION& bav
) {
int retval;
// project-specific check
//
if (wu_is_infeasible_custom(wu, app, bav)) {
return INFEASIBLE_CUSTOM;
}
if (config.user_filter) {
if (wu.batch && wu.batch != g_reply->user.id) {
return INFEASIBLE_USER_FILTER;
}
}
// homogeneous redundancy: can't send if app uses HR and
// 1) host is of unknown HR class
//
if (app_hr_type(app)) {
if (hr_unknown_class(g_reply->host, app_hr_type(app))) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [HOST#%d] [WU#%d %s] host is of unknown class in HR type %d\n",
g_reply->host.id, wu.id, wu.name, app_hr_type(app)
);
}
return INFEASIBLE_HR;
}
if (already_sent_to_different_platform_quick(wu, app)) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [HOST#%d] [WU#%d %s] failed quick HR check: WU is class %d, host is class %d\n",
g_reply->host.id, wu.id, wu.name, wu.hr_class, hr_class(g_request->host, app_hr_type(app))
);
}
return INFEASIBLE_HR;
}
}
if (config.one_result_per_user_per_wu || config.one_result_per_host_per_wu) {
if (wu_already_in_reply(wu)) {
return INFEASIBLE_DUP;
}
}
retval = check_memory(wu);
if (retval) return retval;
retval = check_disk(wu);
if (retval) return retval;
retval = check_bandwidth(wu);
if (retval) return retval;
if (config.non_cpu_intensive) {
return 0;
}
// do deadline check last because EDF sim uses some CPU
//
double opt, pess;
get_delay_bound_range(
wu, res_server_state, res_priority, res_report_deadline, bav, opt, pess
);
wu.delay_bound = (int)opt;
if (opt == 0) {
// this is a resend; skip deadline check
return 0;
}
retval = check_deadline(wu, app, bav);
if (retval && (opt != pess)) {
wu.delay_bound = (int)pess;
retval = check_deadline(wu, app, bav);
}
return retval;
}
// insert "text" right after "after" in the given buffer
//
static int insert_after(char* buffer, const char* after, const char* text) {
char* p;
char temp[BLOB_SIZE];
if (strlen(buffer) + strlen(text) >= BLOB_SIZE-1) {
log_messages.printf(MSG_CRITICAL,
"insert_after: overflow: %d %d\n",
(int)strlen(buffer),
(int)strlen(text)
);
return ERR_BUFFER_OVERFLOW;
}
p = strstr(buffer, after);
if (!p) {
log_messages.printf(MSG_CRITICAL,
"insert_after: %s not found in %s\n", after, buffer
);
return ERR_XML_PARSE;
}
p += strlen(after);
strcpy(temp, p);
strcpy(p, text);
strcat(p, temp);
return 0;
}
// add elements to WU's xml_doc,
// in preparation for sending it to a client
//
static int insert_wu_tags(WORKUNIT& wu, APP& app) {
char buf[BLOB_SIZE];
sprintf(buf,
" <rsc_fpops_est>%f</rsc_fpops_est>\n"
" <rsc_fpops_bound>%f</rsc_fpops_bound>\n"
" <rsc_memory_bound>%f</rsc_memory_bound>\n"
" <rsc_disk_bound>%f</rsc_disk_bound>\n"
" <name>%s</name>\n"
" <app_name>%s</app_name>\n",
wu.rsc_fpops_est,
wu.rsc_fpops_bound,
wu.rsc_memory_bound,
wu.rsc_disk_bound,
wu.name,
app.name
);
return insert_after(wu.xml_doc, "<workunit>\n", buf);
}
// Add the given workunit, app, and app version to a reply.
//
static int add_wu_to_reply(
WORKUNIT& wu, SCHEDULER_REPLY&, APP* app, BEST_APP_VERSION* bavp
) {
int retval;
WORKUNIT wu2, wu3;
APP_VERSION* avp = bavp->avp;
// add the app, app_version, and workunit to the reply,
// but only if they aren't already there
//
if (avp) {
APP_VERSION av2=*avp, *avp2=&av2;
if (strlen(config.replace_download_url_by_timezone)) {
process_av_timezone(avp, av2);
}
g_reply->insert_app_unique(*app);
av2.bavp = bavp;
g_reply->insert_app_version_unique(*avp2);
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] Sending app_version %s %d %d %s; projected %.2f GFLOPS\n",
app->name,
avp2->platformid, avp2->version_num, avp2->plan_class,
bavp->host_usage.projected_flops/1e9
);
}
}
// modify the WU's xml_doc; add <name>, <rsc_*> etc.
//
wu2 = wu; // make copy since we're going to modify its XML field
// adjust FPOPS figures for anonymous platform
//
if (bavp->cavp) {
wu2.rsc_fpops_est *= bavp->cavp->rsc_fpops_scale;
wu2.rsc_fpops_bound *= bavp->cavp->rsc_fpops_scale;
}
retval = insert_wu_tags(wu2, *app);
if (retval) {
log_messages.printf(MSG_CRITICAL,
"insert_wu_tags failed: %s\n", boincerror(retval)
);
return retval;
}
wu3 = wu2;
if (strlen(config.replace_download_url_by_timezone)) {
process_wu_timezone(wu2, wu3);
}
g_reply->insert_workunit_unique(wu3);
// switch to tighter policy for estimating delay
//
return 0;
}
// add <name> tags to result's xml_doc_in
//
static int insert_name_tags(RESULT& result, WORKUNIT const& wu) {
char buf[256];
int retval;
sprintf(buf, "<name>%s</name>\n", result.name);
retval = insert_after(result.xml_doc_in, "<result>\n", buf);
if (retval) return retval;
sprintf(buf, "<wu_name>%s</wu_name>\n", wu.name);
retval = insert_after(result.xml_doc_in, "<result>\n", buf);
if (retval) return retval;
return 0;
}
static int insert_deadline_tag(RESULT& result) {
char buf[256];
sprintf(buf, "<report_deadline>%d</report_deadline>\n", result.report_deadline);
int retval = insert_after(result.xml_doc_in, "<result>\n", buf);
if (retval) return retval;
return 0;
}
// update workunit when send an instance of it:
// - transition time
// - app_version_id, if app uses homogeneous app version
//
int update_wu_on_send(WORKUNIT wu, time_t x, APP& app, BEST_APP_VERSION& bav) {
DB_WORKUNIT dbwu;
char buf[256];
dbwu.id = wu.id;
// SQL note: can't use min() here
//
sprintf(buf,
"transition_time=if(transition_time<%d, transition_time, %d)",
(int)x, (int)x
);
if (app.homogeneous_app_version && (bav.avp->id != wu.app_version_id)) {
char buf2[256];
sprintf(buf2, ", app_version_id=%d", bav.avp->id);
strcat(buf, buf2);
}
return dbwu.update_field(buf);
}
// return true iff a result for same WU is already being sent
//
bool wu_already_in_reply(WORKUNIT& wu) {
unsigned int i;
for (i=0; i<g_reply->results.size(); i++) {
if (wu.id == g_reply->results[i].workunitid) {
return true;
}
}
return false;
}
void lock_sema() {
lock_semaphore(sema_key);
}
void unlock_sema() {
unlock_semaphore(sema_key);
}
// return true if additional work is needed,
// and there's disk space left,
// and we haven't exceeded result per RPC limit,
// and we haven't exceeded results per day limit
//
bool work_needed(bool locality_sched) {
if (locality_sched) {
// if we've failed to send a result because of a transient condition,
// return false to preserve invariant
//
if (g_wreq->disk.insufficient || g_wreq->speed.insufficient || g_wreq->mem.insufficient || g_wreq->no_allowed_apps_available) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] stopping work search - locality condition\n"
);
}
return false;
}
}
// see if we've reached limits on in-progress jobs
//
bool some_type_allowed = false;
// check GPU limit
//
if (config.max_jobs_in_progress.exceeded(NULL, true)) {
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] reached limit on GPU jobs in progress\n"
);
}
g_wreq->clear_gpu_req();
if (g_wreq->effective_ngpus) {
g_wreq->max_jobs_on_host_gpu_exceeded = true;
}
} else {
some_type_allowed = true;
}
// check CPU limit
//
if (config.max_jobs_in_progress.exceeded(NULL, false)) {
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] reached limit on CPU jobs in progress\n"
);
}
g_wreq->clear_cpu_req();
g_wreq->max_jobs_on_host_cpu_exceeded = true;
} else {
some_type_allowed = true;
}
if (!some_type_allowed) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] in-progress job limit exceeded\n"
);
}
g_wreq->max_jobs_on_host_exceeded = true;
return false;
}
// see if we've reached max jobs per RPC
//
if (g_wreq->njobs_sent >= g_wreq->max_jobs_per_rpc) {
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] stopping work search - njobs %d >= max_jobs_per_rpc %d\n",
g_wreq->njobs_sent, g_wreq->max_jobs_per_rpc
);
}
return false;
}
#if 0
log_messages.printf(MSG_NORMAL,
"work_needed: spec req %d sec to fill %.2f; CPU (%.2f, %.2f) CUDA (%.2f, %.2f) ATI(%.2f, %.2f)\n",
g_wreq->rsc_spec_request,
g_wreq->seconds_to_fill,
g_wreq->cpu_req_secs, g_wreq->cpu_req_instances,
g_wreq->cuda_req_secs, g_wreq->cuda_req_instances,
g_wreq->ati_req_secs, g_wreq->ati_req_instances
);
#endif
if (g_wreq->rsc_spec_request) {
if (g_wreq->need_cpu()) {
return true;
}
if (g_wreq->need_cuda()) {
return true;
}
if (g_wreq->need_ati()) {
return true;
}
} else {
if (g_wreq->seconds_to_fill > 0) {
return true;
}
}
if (config.debug_send) {
log_messages.printf(MSG_NORMAL, "[send] don't need more work\n");
}
return false;
}
// return the app version ID, or -2/-3/-4 if anonymous platform
//
inline static int get_app_version_id(BEST_APP_VERSION* bavp) {
if (bavp->avp) {
return bavp->avp->id;
} else {
return bavp->cavp->host_usage.resource_type();
}
}
int add_result_to_reply(
SCHED_DB_RESULT& result, WORKUNIT& wu, BEST_APP_VERSION* bavp,
bool locality_scheduling
) {
int retval;
bool resent_result = false;
APP* app = ssp->lookup_app(wu.appid);
retval = add_wu_to_reply(wu, *g_reply, app, bavp);
if (retval) return retval;
// Adjust available disk space.
// In the scheduling locality case,
// reduce the available space by less than the workunit rsc_disk_bound,
// if the host already has the file or the file was not already sent.
//
if (!locality_scheduling || decrement_disk_space_locality(wu)) {
g_wreq->disk_available -= wu.rsc_disk_bound;
}
// update the result in DB
//
result.hostid = g_reply->host.id;
result.userid = g_reply->user.id;
result.sent_time = time(0);
result.report_deadline = result.sent_time + wu.delay_bound;
result.flops_estimate = bavp->host_usage.peak_flops;
result.app_version_id = get_app_version_id(bavp);
int old_server_state = result.server_state;
if (result.server_state != RESULT_SERVER_STATE_IN_PROGRESS) {
// We're sending this result for the first time
//
result.server_state = RESULT_SERVER_STATE_IN_PROGRESS;
} else {
// Result was already sent to this host but was lost,
// so we're resending it.
//
resent_result = true;
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [RESULT#%d] [HOST#%d] (resend lost work)\n",
result.id, g_reply->host.id
);
}
}
retval = result.mark_as_sent(old_server_state, config.report_grace_period);
if (retval == ERR_DB_NOT_FOUND) {
log_messages.printf(MSG_CRITICAL,
"[RESULT#%d] [HOST#%d]: CAN'T SEND, already sent to another host\n",
result.id, g_reply->host.id
);
} else if (retval) {
log_messages.printf(MSG_CRITICAL,
"add_result_to_reply: can't update result: %s\n", boincerror(retval)
);
}
if (retval) return retval;
double est_dur = estimate_duration(wu, *bavp);
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[HOST#%d] Sending [RESULT#%d %s] (est. dur. %.2f seconds)\n",
g_reply->host.id, result.id, result.name, est_dur
);
}
retval = update_wu_on_send(
wu, result.report_deadline + config.report_grace_period, *app, *bavp
);
if (retval) {
log_messages.printf(MSG_CRITICAL,
"add_result_to_reply: can't update WU transition time: %d\n",
retval
);
return retval;
}
// The following overwrites the result's xml_doc field.
// But that's OK cuz we're done with DB updates
//
retval = insert_name_tags(result, wu);
if (retval) {
log_messages.printf(MSG_CRITICAL,
"add_result_to_reply: can't insert name tags: %d\n",
retval
);
return retval;
}
retval = insert_deadline_tag(result);
if (retval) {
log_messages.printf(MSG_CRITICAL,
"add_result_to_reply: can't insert deadline tag: %s\n", boincerror(retval)
);
return retval;
}
result.bav = *bavp;
g_reply->insert_result(result);
if (g_wreq->rsc_spec_request) {
if (bavp->host_usage.ncudas) {
g_wreq->cuda_req_secs -= est_dur;
g_wreq->cuda_req_instances -= bavp->host_usage.ncudas;
} else if (bavp->host_usage.natis) {
g_wreq->ati_req_secs -= est_dur;
g_wreq->ati_req_instances -= bavp->host_usage.natis;
} else {
g_wreq->cpu_req_secs -= est_dur;
g_wreq->cpu_req_instances -= bavp->host_usage.avg_ncpus;
}
} else {
g_wreq->seconds_to_fill -= est_dur;
}
update_estimated_delay(*bavp, est_dur);
g_wreq->njobs_sent++;
config.max_jobs_in_progress.register_job(app, bavp->host_usage.uses_gpu());
if (!resent_result) {
DB_HOST_APP_VERSION* havp = bavp->host_app_version();
if (havp) {
havp->n_jobs_today++;
}
}
// add this result to workload for simulation
//
if (config.workload_sim && g_request->have_other_results_list) {
IP_RESULT ipr ("", time(0)+wu.delay_bound, est_dur);
g_request->ip_results.push_back(ipr);
}
// mark job as done if debugging flag is set;
// this is used by sched_driver.C (performance testing)
//
if (mark_jobs_done) {
DB_WORKUNIT dbwu;
char buf[256];
sprintf(buf,
"server_state=%d outcome=%d",
RESULT_SERVER_STATE_OVER, RESULT_OUTCOME_SUCCESS
);
result.update_field(buf);
dbwu.id = wu.id;
sprintf(buf, "transition_time=%ld", time(0));
dbwu.update_field(buf);
}
// If we're sending an unreplicated job to an untrusted host,
// mark it as replicated
//
if (wu.target_nresults == 1 && app->target_nresults > 1) {
if (bavp->trusted) {
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [WU#%d] using trusted app version, not replicating\n", wu.id
);
}
} else {
DB_WORKUNIT dbwu;
char buf[256];
sprintf(buf,
"target_nresults=%d, min_quorum=%d, transition_time=%ld",
app->target_nresults, app->target_nresults, time(0)
);
dbwu.id = wu.id;
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] [WU#%d] sending to untrusted host, replicating\n", wu.id
);
}
retval = dbwu.update_field(buf);
if (retval) {
log_messages.printf(MSG_CRITICAL,
"WU update failed: %s", boincerror(retval)
);
}
}
}
return 0;
}
// Send high-priority messages about things the user can change easily
// (namely the driver version)
// and low-priority messages about things that can't easily be changed,
// but which may be interfering with getting tasks or latest apps
//
static void send_gpu_messages(
GPU_REQUIREMENTS& req, double ram, int version, const char* rsc_name
) {
char buf[256];
if (ram < req.min_ram) {
sprintf(buf,
"A minimum of %d MB (preferably %d MB) of video RAM is needed to process tasks using your computer's %s",
(int) (req.min_ram/MEGA),
(int) (req.opt_ram/MEGA),
rsc_name
);
g_reply->insert_message(buf, "low");
} else {
if (version) {
if (version < req.min_driver_version) {
sprintf(buf,
"%s: %s",
rsc_name,
_("Upgrade to the latest driver to process tasks using your computer's GPU")
);
g_reply->insert_message(buf, "notice");
} else if (version < req.opt_driver_version) {
sprintf(buf,
"%s: %s",
rsc_name,
_("Upgrade to the latest driver to use all of this project's GPU applications")
);
g_reply->insert_message(buf, "low");
}
}
}
}
// send messages to user about why jobs were or weren't sent,
// recommendations for GPU driver upgrades, etc.
//
static void send_user_messages() {
char buf[512];
unsigned int i;
int j;
// Mac client with GPU but too-old client
//
if (g_request->coprocs.nvidia.count
&& ssp->have_cuda_apps
&& strstr(g_request->host.os_name, "Darwin")
&& g_request->core_client_version < 61028
) {
g_reply->insert_message(
_("A newer version of BOINC is needed to use your NVIDIA GPU; please upgrade to the current version"),
"notice"
);
}
// GPU-only project, client lacks GPU
//
bool usable_gpu = (ssp->have_cuda_apps && g_request->coprocs.nvidia.count)
|| (ssp->have_ati_apps && g_request->coprocs.ati.count);
if (!ssp->have_cpu_apps && !usable_gpu) {
if (ssp->have_cuda_apps) {
if (ssp->have_ati_apps) {
g_reply->insert_message(
_("An NVIDIA or ATI GPU is required to run tasks for this project"),
"notice"
);
} else {
g_reply->insert_message(
_("An NVIDIA GPU is required to run tasks for this project"),
"notice"
);
}
} else if (ssp->have_ati_apps) {
g_reply->insert_message(
_("An ATI GPU is required to run tasks for this project"),
"notice"
);
}
}
if (g_request->coprocs.nvidia.count && ssp->have_cuda_apps) {
send_gpu_messages(cuda_requirements,
g_request->coprocs.nvidia.prop.dtotalGlobalMem,
g_request->coprocs.nvidia.display_driver_version,
"NVIDIA GPU"
);
}
if (g_request->coprocs.ati.count && ssp->have_ati_apps) {
send_gpu_messages(ati_requirements,
g_request->coprocs.ati.attribs.localRAM*MEGA,
g_request->coprocs.ati.version_num,
"ATI GPU"
);
}
// If work was sent from apps the user did not select, explain.
// NOTE: this will have to be done differently with matchmaker scheduling
//
if (!config.locality_scheduling && !config.locality_scheduler_fraction && !config.matchmaker) {
if (g_wreq->njobs_sent && !g_wreq->user_apps_only) {
g_reply->insert_message(
"No tasks are available for the applications you have selected",
"low"
);
// Inform the user about applications with no work
//
for (i=0; i<g_wreq->preferred_apps.size(); i++) {
if (!g_wreq->preferred_apps[i].work_available) {
APP* app = ssp->lookup_app(g_wreq->preferred_apps[i].appid);
// don't write message if the app is deprecated
//
if (app) {
char explanation[256];
sprintf(explanation,
"No tasks are available for %s",
find_user_friendly_name(g_wreq->preferred_apps[i].appid)
);
g_reply->insert_message( explanation, "low");
}
}
}
// Tell the user about applications they didn't qualify for
//
for (j=0; j<preferred_app_message_index; j++){
g_reply->insert_message(g_wreq->no_work_messages.at(j));
}
g_reply->insert_message(
"Your preferences allow tasks from applications other than those selected",
"low"
);
g_reply->insert_message(
"Sending tasks from other applications", "low"
);
}
}
// if client asked for work and we're not sending any, explain why
//
if (g_wreq->njobs_sent == 0) {
g_reply->set_delay(DELAY_NO_WORK_TEMP);
g_reply->insert_message("No tasks sent", "low");
// Tell the user about applications with no work
//
for (i=0; i<g_wreq->preferred_apps.size(); i++) {
if (!g_wreq->preferred_apps[i].work_available) {
APP* app = ssp->lookup_app(g_wreq->preferred_apps[i].appid);
// don't write message if the app is deprecated
if (app != NULL) {
sprintf(buf, "No tasks are available for %s",
find_user_friendly_name(
g_wreq->preferred_apps[i].appid
)
);
g_reply->insert_message(buf, "low");
}
}
}
// Tell the user about applications they didn't qualify for
//
for (i=0; i<g_wreq->no_work_messages.size(); i++){
g_reply->insert_message(g_wreq->no_work_messages.at(i));
}
if (g_wreq->no_allowed_apps_available) {
g_reply->insert_message(
_("No tasks are available for the applications you have selected."),
"low"
);
}
if (g_wreq->speed.insufficient) {
if (g_request->core_client_version>41900) {
sprintf(buf,
"Tasks won't finish in time: BOINC runs %.1f%% of the time; computation is enabled %.1f%% of that",
100*g_reply->host.on_frac, 100*g_reply->host.active_frac
);
} else {
sprintf(buf,
"Tasks won't finish in time: Computer available %.1f%% of the time",
100*g_reply->host.on_frac
);
}
g_reply->insert_message(buf, "low");
}
if (g_wreq->hr_reject_temp) {
g_reply->insert_message(
"Tasks are committed to other platforms",
"low"
);
}
if (g_wreq->hr_reject_perm) {
g_reply->insert_message(
_("Your computer type is not supported by this project"),
"notice"
);
}
if (g_wreq->outdated_client) {
g_reply->insert_message(
_("Newer BOINC version required; please install current version"),
"notice"
);
g_reply->set_delay(DELAY_NO_WORK_PERM);
log_messages.printf(MSG_NORMAL,
"Not sending tasks because newer client version required\n"
);
}
if (g_wreq->no_cuda_prefs) {
g_reply->insert_message(
_("Tasks for NVIDIA GPU are available, but your preferences are set to not accept them"),
"low"
);
}
if (g_wreq->no_ati_prefs) {
g_reply->insert_message(
_("Tasks for ATI GPU are available, but your preferences are set to not accept them"),
"low"
);
}
if (g_wreq->no_cpu_prefs) {
g_reply->insert_message(
_("Tasks for CPU are available, but your preferences are set to not accept them"),
"low"
);
}
DB_HOST_APP_VERSION* havp = quota_exceeded_version();
if (havp) {
sprintf(buf, "This computer has finished a daily quota of %d tasks)",
havp->max_jobs_per_day
);
g_reply->insert_message(buf, "low");
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] Daily quota %d exceeded for app version %d\n",
havp->max_jobs_per_day, havp->app_version_id
);
}
g_reply->set_delay(DELAY_NO_WORK_CACHE);
}
if (g_wreq->max_jobs_on_host_exceeded
|| g_wreq->max_jobs_on_host_cpu_exceeded
|| g_wreq->max_jobs_on_host_gpu_exceeded
) {
sprintf(buf, "This computer has reached a limit on tasks in progress");
g_reply->insert_message(buf, "low");
g_reply->set_delay(DELAY_NO_WORK_CACHE);
}
}
}
static double clamp_req_sec(double x) {
if (x < MIN_REQ_SECS) return MIN_REQ_SECS;
if (x > MAX_REQ_SECS) return MAX_REQ_SECS;
return x;
}
// prepare to send jobs, both resent and new;
// decipher request type, fill in WORK_REQ
//
void send_work_setup() {
unsigned int i;
g_wreq->seconds_to_fill = clamp_req_sec(g_request->work_req_seconds);
g_wreq->cpu_req_secs = clamp_req_sec(g_request->cpu_req_secs);
g_wreq->cpu_req_instances = g_request->cpu_req_instances;
g_wreq->anonymous_platform = is_anonymous(g_request->platforms.list[0]);
if (g_wreq->anonymous_platform) {
estimate_flops_anon_platform();
}
cuda_requirements.clear();
ati_requirements.clear();
g_wreq->disk_available = max_allowable_disk();
get_mem_sizes();
get_running_frac();
g_wreq->get_job_limits();
if (g_request->coprocs.nvidia.count) {
g_wreq->cuda_req_secs = clamp_req_sec(g_request->coprocs.nvidia.req_secs);
g_wreq->cuda_req_instances = g_request->coprocs.nvidia.req_instances;
if (g_request->coprocs.nvidia.estimated_delay < 0) {
g_request->coprocs.nvidia.estimated_delay = g_request->cpu_estimated_delay;
}
}
if (g_request->coprocs.ati.count) {
g_wreq->ati_req_secs = clamp_req_sec(g_request->coprocs.ati.req_secs);
g_wreq->ati_req_instances = g_request->coprocs.ati.req_instances;
if (g_request->coprocs.ati.estimated_delay < 0) {
g_request->coprocs.ati.estimated_delay = g_request->cpu_estimated_delay;
}
}
if (g_wreq->cpu_req_secs || g_wreq->cuda_req_secs || g_wreq->ati_req_secs) {
g_wreq->rsc_spec_request = true;
} else {
g_wreq->rsc_spec_request = false;
}
for (i=0; i<g_request->other_results.size(); i++) {
OTHER_RESULT& r = g_request->other_results[i];
APP* app = NULL;
bool uses_gpu = false;
bool have_cav = false;
if (r.app_version >= 0
&& r.app_version < (int)g_request->client_app_versions.size()
) {
CLIENT_APP_VERSION& cav = g_request->client_app_versions[r.app_version];
app = cav.app;
if (app) {
have_cav = true;
uses_gpu = cav.host_usage.uses_gpu();
}
}
if (!have_cav) {
if (r.have_plan_class && app_plan_uses_gpu(r.plan_class)) {
uses_gpu = true;
}
}
config.max_jobs_in_progress.register_job(app, uses_gpu);
}
// print details of request to log
//
if (config.debug_quota) {
log_messages.printf(MSG_NORMAL,
"[quota] max jobs per RPC: %d\n", g_wreq->max_jobs_per_rpc
);
config.max_jobs_in_progress.print_log();
}
if (config.debug_send) {
log_messages.printf(MSG_NORMAL,
"[send] %s matchmaker scheduling; %s EDF sim\n",
config.matchmaker?"Using":"Not using",
config.workload_sim?"Using":"Not using"
);
log_messages.printf(MSG_NORMAL,
"[send] CPU: req %.2f sec, %.2f instances; est delay %.2f\n",
g_wreq->cpu_req_secs, g_wreq->cpu_req_instances,
g_request->cpu_estimated_delay
);
if (g_request->coprocs.nvidia.count) {
log_messages.printf(MSG_NORMAL,
"[send] CUDA: req %.2f sec, %.2f instances; est delay %.2f\n",
g_wreq->cuda_req_secs, g_wreq->cuda_req_instances,
g_request->coprocs.nvidia.estimated_delay
);
}
if (g_request->coprocs.ati.count) {
log_messages.printf(MSG_NORMAL,
"[send] ATI: req %.2f sec, %.2f instances; est delay %.2f\n",
g_wreq->ati_req_secs, g_wreq->ati_req_instances,
g_request->coprocs.ati.estimated_delay
);
}
log_messages.printf(MSG_NORMAL,
"[send] work_req_seconds: %.2f secs\n",
g_wreq->seconds_to_fill
);
log_messages.printf(MSG_NORMAL,
"[send] available disk %.2f GB, work_buf_min %d\n",
g_wreq->disk_available/GIGA,
(int)g_request->global_prefs.work_buf_min()
);
log_messages.printf(MSG_NORMAL,
"[send] active_frac %f on_frac %f\n",
g_reply->host.active_frac,
g_reply->host.on_frac
);
if (g_wreq->anonymous_platform) {
log_messages.printf(MSG_NORMAL,
"Anonymous platform app versions:\n"
);
for (i=0; i<g_request->client_app_versions.size(); i++) {
CLIENT_APP_VERSION& cav = g_request->client_app_versions[i];
log_messages.printf(MSG_NORMAL,
" app: %s version %d cpus %.2f cudas %.2f atis %.2f flops %fG\n",
cav.app_name,
cav.version_num,
cav.host_usage.avg_ncpus,
cav.host_usage.ncudas,
cav.host_usage.natis,
cav.host_usage.projected_flops/1e9
);
}
}
}
}
// If a record is not in DB, create it.
//
int update_host_app_versions(vector<SCHED_DB_RESULT>& results, int hostid) {
vector<DB_HOST_APP_VERSION> new_havs;
unsigned int i, j;
int retval;
for (i=0; i<results.size(); i++) {
RESULT& r = results[i];
int gavid = generalized_app_version_id(r.app_version_id, r.appid);
DB_HOST_APP_VERSION* havp = gavid_to_havp(gavid);
if (!havp) {
bool found = false;
for (j=0; j<new_havs.size(); j++) {
DB_HOST_APP_VERSION& hav = new_havs[j];
if (hav.app_version_id == gavid) {
found = true;
hav.n_jobs_today++;
}
}
if (!found) {
DB_HOST_APP_VERSION hav;
hav.clear();
hav.host_id = hostid;
hav.app_version_id = gavid;
hav.n_jobs_today = 1;
new_havs.push_back(hav);
}
}
}
// create new records
//
for (i=0; i<new_havs.size(); i++) {
DB_HOST_APP_VERSION& hav = new_havs[i];
retval = hav.insert();
if (retval) {
log_messages.printf(MSG_CRITICAL,
"hav.insert(): %s\n", boincerror(retval)
);
} else {
if (config.debug_credit) {
log_messages.printf(MSG_NORMAL,
"[credit] created host_app_version record (%d, %d)\n",
hav.host_id, hav.app_version_id
);
}
}
}
return 0;
}
void send_work() {
int retval;
if (!work_needed(false)) {
send_user_messages();
return;
}
g_wreq->no_jobs_available = true;
if (!g_wreq->rsc_spec_request && g_wreq->seconds_to_fill == 0) {
return;
}
if (all_apps_use_hr && hr_unknown_platform(g_request->host)) {
log_messages.printf(MSG_NORMAL,
"Not sending work because unknown HR class\n"
);
g_wreq->hr_reject_perm = true;
return;
}
// decide on attributes of HOST_APP_VERSIONS
//
get_reliability_and_trust();
get_prefs_info();
if (config.enable_assignment) {
if (send_assigned_jobs()) {
if (config.debug_assignment) {
log_messages.printf(MSG_NORMAL,
"[assign] [HOST#%d] sent assigned jobs\n", g_reply->host.id
);
}
goto done;
}
}
if (config.workload_sim && g_request->have_other_results_list) {
init_ip_results(
g_request->global_prefs.work_buf_min(),
g_wreq->effective_ncpus, g_request->ip_results
);
}
if (config.locality_scheduler_fraction > 0) {
if (drand() < config.locality_scheduler_fraction) {
if (config.debug_locality) {
log_messages.printf(MSG_NORMAL,
"[mixed] sending locality work first\n"
);
}
send_work_locality();
if (config.debug_locality) {
log_messages.printf(MSG_NORMAL,
"[mixed] sending non-locality work second\n"
);
}
send_work_old();
} else {
if (config.debug_locality) {
log_messages.printf(MSG_NORMAL,
"[mixed] sending non-locality work first\n"
);
}
send_work_old();
if (config.debug_locality) {
log_messages.printf(MSG_NORMAL,
"[mixed] sending locality work second\n"
);
}
send_work_locality();
}
} else if (config.locality_scheduling) {
send_work_locality();
} else if (config.matchmaker) {
send_work_matchmaker();
} else {
send_work_old();
}
done:
retval = update_host_app_versions(g_reply->results, g_reply->host.id);
if (retval) {
log_messages.printf(MSG_CRITICAL,
"update_host_app_versions() failed: %s\n", boincerror(retval)
);
}
send_user_messages();
}
const char *BOINC_RCSID_32dcd335e7 = "$Id$";