X-Git-Url: https://main.carlh.net/gitweb/?p=dcpomatic.git;a=blobdiff_plain;f=src%2Flib%2Fjob_manager.cc;h=d95f95a2481ba860ae4ee0c304d3190cf09f53f9;hp=2db91a177cc08c194644e7b35efc7095e577cddd;hb=ea6b2dae46caa1da829fbf499e83cd6ae3b3773a;hpb=b5001080a3e5b414f6cad1c52926ed757f2d8574 diff --git a/src/lib/job_manager.cc b/src/lib/job_manager.cc index 2db91a177..d95f95a24 100644 --- a/src/lib/job_manager.cc +++ b/src/lib/job_manager.cc @@ -1,19 +1,20 @@ /* - Copyright (C) 2012 Carl Hetherington + Copyright (C) 2012-2018 Carl Hetherington - This program is free software; you can redistribute it and/or modify + This file is part of DCP-o-matic. + + DCP-o-matic is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. - This program is distributed in the hope that it will be useful, + DCP-o-matic is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License - along with this program; if not, write to the Free Software - Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. + along with DCP-o-matic. If not, see . */ @@ -21,38 +22,92 @@ * @brief A simple scheduler for jobs. */ -#include -#include #include "job_manager.h" #include "job.h" #include "cross.h" +#include "analyse_audio_job.h" +#include "film.h" +#include +#include +#include -using namespace std; -using namespace boost; +using std::string; +using std::list; +using std::cout; +using boost::shared_ptr; +using boost::weak_ptr; +using boost::function; +using boost::dynamic_pointer_cast; +using boost::optional; +using boost::bind; JobManager* JobManager::_instance = 0; JobManager::JobManager () + : _terminate (false) + , _paused (false) { - boost::thread (boost::bind (&JobManager::scheduler, this)); + +} + +void +JobManager::start () +{ + _scheduler = boost::thread (boost::bind(&JobManager::scheduler, this)); +#ifdef DCPOMATIC_LINUX + pthread_setname_np (_scheduler.native_handle(), "job-scheduler"); +#endif +} + +JobManager::~JobManager () +{ + BOOST_FOREACH (boost::signals2::connection& i, _connections) { + i.disconnect (); + } + + { + boost::mutex::scoped_lock lm (_mutex); + _terminate = true; + _empty_condition.notify_all (); + } + + if (_scheduler.joinable()) { + try { + _scheduler.join(); + } catch (...) { + + } + } } shared_ptr JobManager::add (shared_ptr j) { - boost::mutex::scoped_lock lm (_mutex); - _jobs.push_back (j); + { + boost::mutex::scoped_lock lm (_mutex); + _jobs.push_back (j); + _empty_condition.notify_all (); + } + + emit (boost::bind (boost::ref (JobAdded), weak_ptr (j))); + return j; } -void +shared_ptr JobManager::add_after (shared_ptr after, shared_ptr j) { - boost::mutex::scoped_lock lm (_mutex); - list >::iterator i = find (_jobs.begin(), _jobs.end(), after); - assert (i != _jobs.end ()); - ++i; - _jobs.insert (i, j); + { + boost::mutex::scoped_lock lm (_mutex); + list >::iterator i = find (_jobs.begin(), _jobs.end(), after); + DCPOMATIC_ASSERT (i != _jobs.end()); + _jobs.insert (i, j); + _empty_condition.notify_all (); + } + + emit (boost::bind (boost::ref (JobAdded), weak_ptr (j))); + + return j; } list > @@ -78,50 +133,225 @@ bool JobManager::errors () const { boost::mutex::scoped_lock lm (_mutex); - for (list >::const_iterator i = _jobs.begin(); i != _jobs.end(); ++i) { - if ((*i)->finished_in_error ()) { + BOOST_FOREACH (shared_ptr i, _jobs) { + if (i->finished_in_error ()) { return true; } } return false; -} - +} void JobManager::scheduler () { - while (1) { - { - boost::mutex::scoped_lock lm (_mutex); - for (list >::iterator i = _jobs.begin(); i != _jobs.end(); ++i) { - if ((*i)->running ()) { - /* Something is already happening */ - break; + while (true) { + + boost::mutex::scoped_lock lm (_mutex); + + while (true) { + bool have_new = false; + bool have_running = false; + BOOST_FOREACH (shared_ptr i, _jobs) { + if (i->running()) { + have_running = true; } - - if ((*i)->is_new()) { - shared_ptr r = (*i)->required (); - if (!r || r->finished_ok ()) { - (*i)->start (); - - /* Only start one job at once */ - break; - } + if (i->is_new()) { + have_new = true; } } + + if ((!have_running && have_new) || _terminate) { + break; + } + + _empty_condition.wait (lm); + } + + if (_terminate) { + break; } - dvdomatic_sleep (1); + BOOST_FOREACH (shared_ptr i, _jobs) { + if (i->is_new()) { + _connections.push_back (i->FinishedImmediate.connect(bind(&JobManager::job_finished, this))); + i->start (); + emit (boost::bind (boost::ref (ActiveJobsChanged), _last_active_job, i->json_name())); + _last_active_job = i->json_name (); + /* Only start one job at once */ + break; + } + } } } +void +JobManager::job_finished () +{ + { + boost::mutex::scoped_lock lm (_mutex); + emit (boost::bind (boost::ref (ActiveJobsChanged), _last_active_job, optional())); + _last_active_job = optional(); + } + + _empty_condition.notify_all (); +} + JobManager * JobManager::instance () { if (_instance == 0) { _instance = new JobManager (); + _instance->start (); } return _instance; } + +void +JobManager::drop () +{ + delete _instance; + _instance = 0; +} + +void +JobManager::analyse_audio ( + shared_ptr film, + shared_ptr playlist, + bool from_zero, + boost::signals2::connection& connection, + function ready + ) +{ + { + boost::mutex::scoped_lock lm (_mutex); + + BOOST_FOREACH (shared_ptr i, _jobs) { + shared_ptr a = dynamic_pointer_cast (i); + if (a && a->path() == film->audio_analysis_path(playlist)) { + i->when_finished (connection, ready); + return; + } + } + } + + shared_ptr job; + + { + boost::mutex::scoped_lock lm (_mutex); + + job.reset (new AnalyseAudioJob (film, playlist, from_zero)); + connection = job->Finished.connect (ready); + _jobs.push_back (job); + _empty_condition.notify_all (); + } + + emit (boost::bind (boost::ref (JobAdded), weak_ptr (job))); +} + +void +JobManager::increase_priority (shared_ptr job) +{ + bool changed = false; + + { + boost::mutex::scoped_lock lm (_mutex); + list >::iterator last = _jobs.end (); + for (list >::iterator i = _jobs.begin(); i != _jobs.end(); ++i) { + if (*i == job && last != _jobs.end()) { + swap (*i, *last); + changed = true; + break; + } + last = i; + } + } + + if (changed) { + priority_changed (); + } +} + +void +JobManager::priority_changed () +{ + { + boost::mutex::scoped_lock lm (_mutex); + + bool first = true; + BOOST_FOREACH (shared_ptr i, _jobs) { + if (first) { + if (i->is_new ()) { + i->start (); + } else if (i->paused_by_priority ()) { + i->resume (); + } + first = false; + } else { + if (i->running ()) { + i->pause_by_priority (); + } + } + } + } + + emit (boost::bind (boost::ref (JobsReordered))); +} + +void +JobManager::decrease_priority (shared_ptr job) +{ + bool changed = false; + + { + boost::mutex::scoped_lock lm (_mutex); + for (list >::iterator i = _jobs.begin(); i != _jobs.end(); ++i) { + list >::iterator next = i; + ++next; + if (*i == job && next != _jobs.end()) { + swap (*i, *next); + changed = true; + break; + } + } + } + + if (changed) { + priority_changed (); + } +} + +void +JobManager::pause () +{ + boost::mutex::scoped_lock lm (_mutex); + + if (_paused) { + return; + } + + BOOST_FOREACH (shared_ptr i, _jobs) { + if (i->pause_by_user()) { + _paused_job = i; + } + } + + _paused = true; +} + +void +JobManager::resume () +{ + boost::mutex::scoped_lock lm (_mutex); + if (!_paused) { + return; + } + + if (_paused_job) { + _paused_job->resume (); + } + + _paused_job.reset (); + _paused = false; +}