// The contents of this file are subject to the BOINC Public License // Version 1.0 (the "License"); you may not use this file except in // compliance with the License. You may obtain a copy of the License at // http://boinc.berkeley.edu/license_1.0.txt // // Software distributed under the License is distributed on an "AS IS" // basis, WITHOUT WARRANTY OF ANY KIND, either express or implied. See the // License for the specific language governing rights and limitations // under the License. // // The Original Code is the Berkeley Open Infrastructure for Network Computing. // // The Initial Developer of the Original Code is the SETI@home project. // Portions created by the SETI@home project are Copyright (C) 2002 // University of California at Berkeley. All Rights Reserved. // // Contributor(s): // // scheduler code related to sending work #include #include #include #include using namespace std; #include "error_numbers.h" #include "server_types.h" #include "sched_shmem.h" #include "sched_config.h" #include "sched_util.h" #include "main.h" #include "sched_msgs.h" #include "sched_send.h" #include "sched_locality.h" #ifdef _USING_FCGI_ #include "fcgi_stdio.h" #endif const int MIN_SECONDS_TO_SEND = 0; const int MAX_SECONDS_TO_SEND = (28*SECONDS_IN_DAY); const double MIN_POSSIBLE_RAM = 64000000; bool anonymous(PLATFORM& platform) { return (!strcmp(platform.name, "anonymous")); } bool SCHEDULER_REQUEST::has_version(APP& app) { unsigned int i; for (i=0; i= app.min_version) { return true; } } return false; } // compute the max additional disk usage we can impose on the host // double max_allowable_disk(SCHEDULER_REQUEST& req) { // ROMW: Reverting back to older implementation until all clients are 4.x // or higher. #if 1 HOST host = req.host; GLOBAL_PREFS prefs = req.global_prefs; double x1, x2, x3, x; // fill in default values for missing prefs // if (prefs.disk_max_used_gb == 0) prefs.disk_max_used_gb = 0.1; // 100 MB if (prefs.disk_max_used_pct == 0) prefs.disk_max_used_pct = 10; // min_free_gb can be zero // default values for BOINC disk usage (project and total) is zero // // no defaults for total/free disk space (host.d_total, d_free) // if they're zero, project will get no work. // x1 = prefs.disk_max_used_gb*1e9 - req.total_disk_usage; x2 = host.d_total*prefs.disk_max_used_pct/100.; x3 = host.d_free - prefs.disk_min_free_gb*1e9; // may be negative x = min(x1, min(x2, x3)); if (x < 0) { log_messages.printf( SCHED_MSG_LOG::NORMAL, "disk_max_used_gb %f disk_max_used_pct %f disk_min_free_gb %f\n", prefs.disk_max_used_gb, prefs.disk_max_used_pct, prefs.disk_min_free_gb ); log_messages.printf( SCHED_MSG_LOG::NORMAL, "req.total_disk_usage %f host.d_total %f host.d_free %f\n", req.total_disk_usage, host.d_total, host.d_free ); log_messages.printf( SCHED_MSG_LOG::NORMAL, "x1 %f x2 %f x3 %f x %f\n", x1, x2, x3, x ); } return x; #else double x1, x2, x3; HOST host = req.host; x1 = req.project_disk_free; x2 = req.potentially_free_offender; x3 = req.potentially_free_self; if (x1 < 0) { log_messages.printf( SCHED_MSG_LOG::NORMAL, "req.project_disk_free_gb %f\n", x1 ); } return max(max(x1,x2), x3); #endif } // if a host has active_frac < 0.1, assume 0.1 so we don't deprive it of work. // const double HOST_ACTIVE_FRAC_MIN = 0.1; // estimate the number of seconds that a workunit requires running 100% on a // single CPU of this host. // // TODO: improve this. take memory bandwidth into account // static double estimate_cpu_duration(WORKUNIT& wu, HOST& host) { if (host.p_fpops <= 0) host.p_fpops = 1e9; if (wu.rsc_fpops_est <= 0) wu.rsc_fpops_est = 1e12; return wu.rsc_fpops_est/host.p_fpops; } // estimate the amount of real time to complete this WU, // taking into account active_frac and resource_share_fraction // static double estimate_wallclock_duration( WORKUNIT& wu, HOST& host, double resource_share_fraction ) { return estimate_cpu_duration(wu, host) / (max(HOST_ACTIVE_FRAC_MIN, host.active_frac)*resource_share_fraction) ; } // return false if the WU can't be executed on the host // because of insufficient memory, CPU speed, or resource share // static bool wu_is_feasible( WORKUNIT& wu, HOST& host, WORK_REQ& wreq, double resource_share_fraction, double estimated_delay ) { double m_nbytes = host.m_nbytes; if (m_nbytes < MIN_POSSIBLE_RAM) m_nbytes = MIN_POSSIBLE_RAM; if (wu.rsc_memory_bound > m_nbytes) { log_messages.printf( SCHED_MSG_LOG::DEBUG, "[WU#%d %s] needs %f mem; [HOST#%d] has %f\n", wu.id, wu.name, wu.rsc_memory_bound, host.id, m_nbytes ); wreq.insufficient_mem = true; return false; } if (config.enforce_delay_bound) { double wu_wallclock_time = estimate_wallclock_duration(wu, host, resource_share_fraction); double host_remaining_time = estimated_delay; if (host_remaining_time + wu_wallclock_time > wu.delay_bound) { log_messages.printf( SCHED_MSG_LOG::DEBUG, "[WU#%d %s] needs requires %d seconds on [HOST#%d]; delay_bound is %d\n", wu.id, wu.name, (int)wu_wallclock_time, host.id, wu.delay_bound ); wreq.insufficient_speed = true; return false; } } return true; } // insert "text" right after "after" in the given buffer // int insert_after(char* buffer, char* after, char* text) { char* p; char temp[LARGE_BLOB_SIZE]; if (strlen(buffer) + strlen(text) > LARGE_BLOB_SIZE-1) { log_messages.printf(SCHED_MSG_LOG::NORMAL, "insert_after: overflow\n"); return ERR_BUFFER_OVERFLOW; } p = strstr(buffer, after); if (!p) { log_messages.printf(SCHED_MSG_LOG::CRITICAL, "insert_after: %s not found in %s\n", after, buffer); return ERR_NULL; } p += strlen(after); strcpy(temp, p); strcpy(p, text); strcat(p, temp); return 0; } // add elements to WU's xml_doc, in preparation for sending // it to a client // int insert_wu_tags(WORKUNIT& wu, APP& app) { char buf[LARGE_BLOB_SIZE]; sprintf(buf, " %f\n" " %f\n" " %f\n" " %f\n" " %s\n" " %s\n", wu.rsc_fpops_est, wu.rsc_fpops_bound, wu.rsc_memory_bound, wu.rsc_disk_bound, wu.name, app.name ); return insert_after(wu.xml_doc, "\n", buf); } // return the APP and APP_VERSION for the given WU, for the given platform. // return false if none // bool find_app_version( WORK_REQ& wreq, WORKUNIT& wu, PLATFORM& platform, SCHED_SHMEM& ss, APP*& app, APP_VERSION*& avp ) { app = ss.lookup_app(wu.appid); if (!app) { log_messages.printf( SCHED_MSG_LOG::CRITICAL, "Can't find APP#%d\n", wu.appid ); return false; } avp = ss.lookup_app_version(app->id, platform.id, app->min_version); if (!avp) { log_messages.printf( SCHED_MSG_LOG::DEBUG, "no app version available: APP#%d PLATFORM#%d min_version %d\n", app->id, platform.id, app->min_version ); wreq.no_app_version = true; return false; } return true; } // verify that the given APP_VERSION will work with the core client // bool app_core_compatible(WORK_REQ& wreq, APP_VERSION& av) { if (wreq.core_client_version < av.min_core_version) { #if 0 log_messages.printf( SCHED_MSG_LOG::DEBUG, "Outdated core version: wanted %d, got %d\n", av.min_core_version, wreq.core_client_version ); #endif wreq.outdated_core = true; return false; } return true; } // add the given workunit to a reply. // look up its app, and make sure there's a version for this platform. // Add the app and app_version to the reply also. // int add_wu_to_reply( WORKUNIT& wu, SCHEDULER_REPLY& reply, PLATFORM& platform, WORK_REQ& wreq, APP* app, APP_VERSION* avp ) { int retval; WORKUNIT wu2; // add the app, app_version, and workunit to the reply, // but only if they aren't already there // if (avp) { reply.insert_app_unique(*app); reply.insert_app_version_unique(*avp); log_messages.printf( SCHED_MSG_LOG::DEBUG, "[HOST#%d] Sending app_version %s %s %d\n", reply.host.id, app->name, platform.name, avp->version_num ); } // add time estimate to reply // wu2 = wu; // make copy since we're going to modify its XML field retval = insert_wu_tags(wu2, *app); if (retval) { log_messages.printf(SCHED_MSG_LOG::NORMAL, "insert_wu_tags failed\n"); return retval; } reply.insert_workunit_unique(wu2); return 0; } int insert_name_tags(RESULT& result, WORKUNIT const& wu) { char buf[256]; int retval; sprintf(buf, "%s\n", result.name); retval = insert_after(result.xml_doc_in, "\n", buf); if (retval) return retval; sprintf(buf, "%s\n", wu.name); retval = insert_after(result.xml_doc_in, "\n", buf); if (retval) return retval; return 0; } int insert_deadline_tag(RESULT& result) { char buf[256]; sprintf(buf, "%d\n", result.report_deadline); int retval = insert_after(result.xml_doc_in, "\n", buf); if (retval) return retval; return 0; } static int update_wu_transition_time(WORKUNIT wu, time_t x) { DB_WORKUNIT dbwu; char buf[256]; dbwu.id = wu.id; sprintf(buf, "transition_time=%d", (int)x); return dbwu.update_field(buf); } // return true iff a result for same WU is already being sent // static bool wu_already_in_reply(WORKUNIT& wu, SCHEDULER_REPLY& reply) { unsigned int i; for (i=0; i= config.max_wus_to_send) return false; if (config.daily_result_quota) { if (reply.host.nresults_today >= config.daily_result_quota) { daily_result_quota_exceeded = true; return false; } } return true; } int add_result_to_reply( DB_RESULT& result, WORKUNIT& wu, SCHEDULER_REPLY& reply, PLATFORM& platform, WORK_REQ& wreq, APP* app, APP_VERSION* avp ) { int retval; double wu_seconds_filled; retval = add_wu_to_reply(wu, reply, platform, wreq, app, avp); if (retval) return retval; wreq.disk_available -= wu.rsc_disk_bound; // update the result in DB // result.server_state = RESULT_SERVER_STATE_IN_PROGRESS; result.hostid = reply.host.id; result.userid = reply.user.id; result.sent_time = time(0); result.report_deadline = result.sent_time + wu.delay_bound; result.update_subset(); wu_seconds_filled = estimate_cpu_duration(wu, reply.host); log_messages.printf( SCHED_MSG_LOG::NORMAL, "[HOST#%d] Sending [RESULT#%d %s] (fills %d seconds)\n", reply.host.id, result.id, result.name, int(wu_seconds_filled) ); retval = update_wu_transition_time(wu, result.report_deadline); if (retval) { log_messages.printf( SCHED_MSG_LOG::CRITICAL, "send_work: can't update WU transition time\n" ); } // The following overwrites the result's xml_doc field. // But that's OK cuz we're done with DB updates // retval = insert_name_tags(result, wu); if (retval) { log_messages.printf( SCHED_MSG_LOG::CRITICAL, "send_work: can't insert name tags\n" ); } retval = insert_deadline_tag(result); if (retval) { log_messages.printf( SCHED_MSG_LOG::CRITICAL, "send_work: can't insert deadline tag\n" ); } reply.insert_result(result); wreq.seconds_to_fill -= wu_seconds_filled; wreq.nresults++; reply.host.nresults_today++; return 0; } // Make a pass through the wu/results array, sending work. // If "infeasible_only" is true, send only results that were // previously infeasible for some host // static void scan_work_array( WORK_REQ& wreq, SCHEDULER_REQUEST& sreq, SCHEDULER_REPLY& reply, PLATFORM& platform, SCHED_SHMEM& ss ) { int i, retval, n; WORKUNIT wu; DB_RESULT result; char buf[256]; APP* app; APP_VERSION* avp; bool found; if (wreq.disk_available < 0) wreq.insufficient_disk = true; lock_sema(); for (i=0; i wreq.disk_available) { wreq.insufficient_disk = true; wu_result.infeasible_count++; continue; } // don't send if we're already sending a result for same WU // if (config.one_result_per_user_per_wu) { if (wu_already_in_reply(wu_result.workunit, reply)) { continue; } } // don't send if host can't handle it // wu = wu_result.workunit; if (!wu_is_feasible( wu, reply.host, wreq, sreq.resource_share_fraction, sreq.estimated_delay )) { log_messages.printf( SCHED_MSG_LOG::DEBUG, "[HOST#%d] [WU#%d %s] WU is infeasible\n", reply.host.id, wu.id, wu.name ); wu_result.infeasible_count++; continue; } // Find the app and app_version for the client's platform. // If none, treat the WU as infeasible // if (anonymous(platform)) { app = ss.lookup_app(wu.appid); found = sreq.has_version(*app); if (!found) { continue; } avp = NULL; } else { found = find_app_version(wreq, wu, platform, ss, app, avp); if (!found) { wu_result.infeasible_count++; continue; } // see if the core client is too old. // don't bump the infeasible count because this // isn't the result's fault // if (!app_core_compatible(wreq, *avp)) { continue; } } // end of fast checks - mark wu_result as checked out and release sema. // from here on in this loop, don't continue on failure; // instead, goto dont_send (so that we reacquire semaphore) wu_result.state = WR_STATE_CHECKED_OUT; unlock_sema(); // Don't send if we've already sent a result of this WU to this user. // if (config.one_result_per_user_per_wu) { sprintf(buf, "where workunitid=%d and userid=%d", wu_result.workunit.id, reply.user.id ); retval = result.count(n, buf); if (retval) { log_messages.printf( SCHED_MSG_LOG::CRITICAL, "send_work: can't get result count (%d)\n", retval ); goto dont_send; } else { if (n>0) { #if 0 log_messages.printf( SCHED_MSG_LOG::NORMAL, "send_work: user %d already has %d result(s) for WU %d\n", reply.user.id, n, wu_result.workunit.id ); #endif goto dont_send; } } } // if desired, make sure redundancy is homogeneous // if (config.homogeneous_redundancy || app->homogeneous_redundancy) { if (already_sent_to_different_platform( sreq, wu_result.workunit, wreq )) { goto dont_send; } } result.id = wu_result.resultid; // mark slot as empty AFTER we've copied out of it // (since otherwise feeder might overwrite it) // wu_result.state = WR_STATE_EMPTY; // reread result from DB, make sure it's still unsent // TODO: from here to update() should be a transaction // retval = result.lookup_id(result.id); if (retval) { log_messages.printf(SCHED_MSG_LOG::CRITICAL, "[RESULT#%d] result.lookup_id() failed %d\n", result.id, retval ); goto done; } if (result.server_state != RESULT_SERVER_STATE_UNSENT) { log_messages.printf(SCHED_MSG_LOG::DEBUG, "[RESULT#%d] expected to be unsent; instead, state is %d\n", result.id, result.server_state ); goto done; } if (result.workunitid != wu.id) { log_messages.printf(SCHED_MSG_LOG::CRITICAL, "[RESULT#%d] wrong WU ID: wanted %d, got %d\n", result.id, wu.id, result.workunitid ); goto done; } // ****** HERE WE'VE COMMITTED TO SENDING THIS RESULT TO HOST ****** // retval = add_result_to_reply( result, wu, reply, platform, wreq, app, avp ); if (!retval) goto done; dont_send: // here we couldn't send the result for some reason -- // set its state back to PRESENT // wu_result.state = WR_STATE_PRESENT; done: lock_sema(); } unlock_sema(); } int send_work( SCHEDULER_REQUEST& sreq, SCHEDULER_REPLY& reply, PLATFORM& platform, SCHED_SHMEM& ss ) { WORK_REQ wreq; memset(&wreq, 0, sizeof(wreq)); // ROMW: Reverting back to older implementation until all clients are 4.x // or higher. #if 1 wreq.disk_available = max_allowable_disk(sreq); #else wreq.disk_available = sreq.project_disk_free; #endif wreq.insufficient_disk = false; wreq.insufficient_mem = false; wreq.insufficient_speed = false; wreq.no_app_version = false; wreq.homogeneous_redundancy_reject = false; wreq.daily_result_quota_exceeded = false; wreq.core_client_version = sreq.core_client_major_version*100 + sreq.core_client_minor_version; wreq.nresults = 0; log_messages.printf( SCHED_MSG_LOG::NORMAL, "[HOST#%d] got request for %d seconds of work; available disk %f GB\n", reply.host.id, sreq.work_req_seconds, wreq.disk_available/1e9 ); if (sreq.work_req_seconds <= 0) return 0; wreq.seconds_to_fill = sreq.work_req_seconds; if (wreq.seconds_to_fill > MAX_SECONDS_TO_SEND) { wreq.seconds_to_fill = MAX_SECONDS_TO_SEND; } if (wreq.seconds_to_fill < MIN_SECONDS_TO_SEND) { wreq.seconds_to_fill = MIN_SECONDS_TO_SEND; } if (config.locality_scheduling) { wreq.infeasible_only = false; send_work_locality(sreq, reply, platform, wreq, ss); } else { // give priority to results that were infeasible for some other host // wreq.infeasible_only = true; scan_work_array(wreq, sreq, reply, platform, ss); wreq.infeasible_only = false; scan_work_array(wreq, sreq, reply, platform, ss); } #if 0 // huh??? if (wreq.nresults == 0) { wreq.disk_available = sreq.potentially_free_offender; scan_work_array(wreq, sreq, reply, platform, ss); } if (wreq.nresults == 0 && config.delete_from_self) { wreq.disk_available = sreq.potentially_free_self; scan_work_array(wreq, sreq, reply, platform, ss); } #endif log_messages.printf( SCHED_MSG_LOG::NORMAL, "[HOST#%d] Sent %d results\n", reply.host.id, wreq.nresults ); if (wreq.nresults == 0) { reply.request_delay = 3600; strcpy(reply.message, "No work available"); if (wreq.no_app_version) { strcat(reply.message, " (there was work for other platforms)" ); reply.request_delay = 3600*24; } if (wreq.insufficient_disk) { strcat(reply.message, " (there was work but you don't have enough disk space allocated)" ); } if (wreq.insufficient_mem) { strcat(reply.message, " (there was work but your computer doesn't have enough memory)" ); } if (wreq.insufficient_speed) { strcat(reply.message, " (there was work but your computer would not finish it before it is due" ); } if (wreq.homogeneous_redundancy_reject) { strcat(reply.message, " (there was work but it was committed to other platforms" ); } if (wreq.outdated_core) { strcat(reply.message, " (your core client is out of date - please upgrade)" ); reply.request_delay = 3600*24; log_messages.printf( SCHED_MSG_LOG::NORMAL, "Not sending work because core client is outdated\n" ); } if (wreq.daily_result_quota_exceeded) { strcat(reply.message, " (daily quota exceeded)"); log_messages.printf( SCHED_MSG_LOG::NORMAL, "Daily result quota exceeded for host %d\n", reply.host.id ); } strcpy(reply.message_priority, "high"); log_messages.printf( SCHED_MSG_LOG::NORMAL, "[HOST#%d] %s\n", reply.host.id, reply.message ); } return 0; }