// Berkeley Open Infrastructure for Network Computing // http://boinc.berkeley.edu // Copyright (C) 2005 University of California // // This is free software; you can redistribute it and/or // modify it under the terms of the GNU Lesser General Public // License as published by the Free Software Foundation; // either version 2.1 of the License, or (at your option) any later version. // // This software is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. // See the GNU Lesser General Public License for more details. // // To view the GNU Lesser General Public License visit // http://www.gnu.org/copyleft/lesser.html // or write to the Free Software Foundation, Inc., // 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA // scheduler code related to sending work #include "config.h" #include "main.h" #include "server_types.h" #include "sched_shmem.h" #include "sched_hr.h" #include "sched_config.h" #include "sched_util.h" #include "sched_msgs.h" #include "sched_send.h" #include "sched_array.h" #ifdef _USING_FCGI_ #include "fcgi_stdio.h" #else #define FCGI_ToFILE(x) (x) #endif // Make a pass through the wu/results array, sending work. // If reply.wreq.infeasible_only is true, // send only results that were previously infeasible for some host // void scan_work_array( SCHEDULER_REQUEST& sreq, SCHEDULER_REPLY& reply, PLATFORM_LIST& platforms, SCHED_SHMEM& ss ) { int i, j, retval, n, rnd_off; WORKUNIT wu; DB_RESULT result; char buf[256]; APP* app; APP_VERSION* avp; bool found; lock_sema(); rnd_off = rand() % ss.max_wu_results; for (j=0; jbeta) { continue; } log_messages.printf(SCHED_MSG_LOG::MSG_DEBUG, "[HOST#%d] beta work found. [RESULT#%d]\n", reply.host.id, wu_result.resultid ); } else { if (app->beta) { continue; } } // If this is a reliable host and we are checking for results that // need a reliable host, then continue if the result is a normal result // skip if the app is beta (beta apps don't use the reliable mechanism) // if (!app->beta) { if (reply.wreq.reliable_only && (!wu_result.need_reliable)) { continue; } else if (!reply.wreq.reliable_only && wu_result.need_reliable) { continue; } } // don't send if we are looking for infeasible results // and the result is not infeasible // if (reply.wreq.infeasible_only && (wu_result.infeasible_count==0)) { continue; } // don't send if host can't handle it // wu = wu_result.workunit; retval = wu_is_infeasible(wu, sreq, reply, *app); if (retval) { log_messages.printf( SCHED_MSG_LOG::MSG_DEBUG, "[HOST#%d] [WU#%d %s] WU is infeasible: %s\n", reply.host.id, wu.id, wu.name, infeasible_string(retval) ); continue; } // Find the app and app_version for the client's platform. // If none, treat the WU as infeasible // if (anonymous(platforms.list[0])) { app = ss.lookup_app(wu.appid); found = sreq.has_version(*app); if (!found) { continue; } avp = NULL; } else { found = find_app_version(reply.wreq, wu, platforms, ss, app, avp); if (!found) { continue; } // see if the core client is too old. // don't bump the infeasible count because this // isn't the result's fault // if (!app_core_compatible(reply.wreq, *avp)) { continue; } } // End of fast checks; // mark wu_result as checked out and release semaphore. // from here on in this loop, don't continue on failure; // instead, goto dont_send (so that we reacquire semaphore) // // Note: without the semaphore we don't have mutual exclusion; // ideally we should use a transaction from now until when // we commit to sending the results. wu_result.state = g_pid; unlock_sema(); // Don't send if we've already sent a result of this WU to this user. // if (config.one_result_per_user_per_wu) { sprintf(buf, "where workunitid=%d and userid=%d", wu_result.workunit.id, reply.user.id ); retval = result.count(n, buf); if (retval) { log_messages.printf( SCHED_MSG_LOG::MSG_CRITICAL, "send_work: can't get result count (%d)\n", retval ); goto dont_send; } else { if (n>0) { log_messages.printf( SCHED_MSG_LOG::MSG_DEBUG, "send_work: user %d already has %d result(s) for WU %d\n", reply.user.id, n, wu_result.workunit.id ); goto dont_send; } } } else if (config.one_result_per_host_per_wu) { // Don't send if we've already sent a result // of this WU to this host. // We only have to check this // if we don't send one result per user. // sprintf(buf, "where workunitid=%d and hostid=%d", wu_result.workunit.id, reply.host.id ); retval = result.count(n, buf); if (retval) { log_messages.printf( SCHED_MSG_LOG::MSG_CRITICAL, "send_work: can't get result count (%d)\n", retval ); goto dont_send; } else { if (n>0) { log_messages.printf( SCHED_MSG_LOG::MSG_DEBUG, "send_work: host %d already has %d result(s) for WU %d\n", reply.host.id, n, wu_result.workunit.id ); goto dont_send; } } } if (app_hr_type(*app)) { if (already_sent_to_different_platform_careful( sreq, reply.wreq, wu_result.workunit, *app )) { log_messages.printf(SCHED_MSG_LOG::MSG_DEBUG, "[HOST#%d] [WU#%d %s] WU is infeasible (assigned to different platform)\n", reply.host.id, wu.id, wu.name ); // Mark the workunit as infeasible. // This ensures that jobs already assigned to a platform // are processed first. // wu_result.infeasible_count++; goto dont_send; } } result.id = wu_result.resultid; // mark slot as empty AFTER we've copied out of it // (since otherwise feeder might overwrite it) // wu_result.state = WR_STATE_EMPTY; // reread result from DB, make sure it's still unsent // TODO: from here to add_result_to_reply() // (which updates the DB record) should be a transaction // retval = result.lookup_id(result.id); if (retval) { log_messages.printf(SCHED_MSG_LOG::MSG_CRITICAL, "[RESULT#%d] result.lookup_id() failed %d\n", result.id, retval ); goto done; } if (result.server_state != RESULT_SERVER_STATE_UNSENT) { log_messages.printf(SCHED_MSG_LOG::MSG_DEBUG, "[RESULT#%d] expected to be unsent; instead, state is %d\n", result.id, result.server_state ); goto done; } if (result.workunitid != wu.id) { log_messages.printf(SCHED_MSG_LOG::MSG_CRITICAL, "[RESULT#%d] wrong WU ID: wanted %d, got %d\n", result.id, wu.id, result.workunitid ); goto done; } retval = add_result_to_reply( result, wu, sreq, reply, platforms, app, avp ); // add_result_to_reply() fails only in fairly pathological cases - // e.g. we couldn't update the DB record or modify XML fields. // If this happens, don't replace the record in the array // (we can't anyway, since we marked the entry as "empty"). // The feeder will eventually pick it up again, // and hopefully the problem won't happen twice. // goto done; dont_send: // here we couldn't send the result for some reason -- // set its state back to PRESENT // wu_result.state = WR_STATE_PRESENT; done: lock_sema(); } unlock_sema(); } const char *BOINC_RCSID_d9f764fd14="$Id$";