This project is mirrored from https://:*****@github.com/Ensembl/ensembl-hive.git.
Pull mirroring updated .
- Jun 13, 2005
-
-
Jessica Severin authored
have not been run before (< retry_count)
-
Jessica Severin authored
don't have any ensembl-hive extensions. To get access to extended hive functionality one must inherit from Hive::Process. A multiple inheritance is allowed like: our @ISA = qw( Bio::EnsEMBL::Hive::Process Bio::EnsEMBL::Pipeline::RunnableDB );
-
Jessica Severin authored
into the Worker object (and out of the Process) - added Process::worker method so that running processes can talk to the worker that is currently running itself. - modified system so that if a process subclass uses Process::dataflow_output_id on branch_code 1, it will turn off the automatic of flowing of the input_job out on branch_code 1. This will make coding much cleaner so that processes no longer need to modifiy the input_id of the input_job - added method Process::autoflow_inputjob which toggles this autoflow behaviour if a subclass would like to modify this directly - auto_dataflow now happens right after the Process::write_output stage
-
- May 26, 2005
-
-
Jessica Severin authored
-
- Apr 18, 2005
-
-
Jessica Severin authored
thus allowing one to do hybrid URL/Registry code like... if($url) { $dbc = Bio::EnsEMBL::Hive::URLFactory->fetch($url, 'compara')->dbc; $dbname = $dbc->dbname(); } else { $dbc = Bio::EnsEMBL::Registry->get_DBAdaptor($dbname,'compara')->dbc; } $fa = Bio::EnsEMBL::Registry->get_adaptor($dbname,'compara','Family');
-
- Mar 22, 2005
-
-
Jessica Severin authored
which is a simple paired file like. If file is missing it does nothing ia64e:3306 127.0.0.1:3371
-
- Mar 11, 2005
-
-
Jessica Severin authored
-
Jessica Severin authored
-lsf_options <string> which passes whatever is specified in the string to the bsub command thus allowing the user complete flexibility
-
- Mar 08, 2005
-
-
Jessica Severin authored
program has evolved into being the primary portal for user interaction with the Hive, so 95% of functions are compute resource agnostic so it makes more sense to have internal switches for which compute resource to submit/check. Expanded local logic to allow multiple local cpus, submit into background(fork), check with 'ps', kill with 'kill -9'.
-
Jessica Severin authored
-
- Mar 04, 2005
-
-
Jessica Severin authored
-
Jessica Severin authored
added columns hive_id and retry. Allows user to join to failed workers in the hive table, and to see which retry level the job was at when the STDOUT/STDERR files were generated. Sets at beginning of job run, and deletes those for 'empty' files at job end.
-
Jessica Severin authored
in a more filesystem friendly manner (creates at 256 layer hash which distributes the directories evenly and reduces concurrent directory modification. Also reordered how the job output files are saved (done at the beginning right after redirection starts, and at the end right before it's closed).
-
Jessica Severin authored
any problems related to setting undef or '0' values.
-
Jessica Severin authored
made the worse (Tim Cutts). This will do until we figure this out.... I like the '>/dev/null + rerun failed jobs manually with debug' option personally :)
-
Jessica Severin authored
-
- Mar 03, 2005
-
-
Jessica Severin authored
each digit becomes a directory with a final directory created with the full hive_id hive_id=1234 => <base_dir>/1/2/3/4/hive_id_1234/ hive_id=12 => <base_dir>/1/2/hive_id_12/ this should distribute the output directories
-
Jessica Severin authored
is calculated. If batch_size>0 use batch_size, else use avg_msec_per_job equation.
-
Jessica Severin authored
to RunnableDB to allow full benefit of dataflow graph capabilities. - Removed from Extension.pm branch_code, analysis_job_id, reset_job extensions to RunnableDB (no longer trying to shoe-horn hive 'extra' functions into them) - Bio::EnsEMBL::Hive::Process mirrors some of the RunnableDB interface (new, analysis, fetch_input, run, write_output) but uses a new job interface (input_job, dataflow_output_id) instead of input_id (but provides convenience method $self->input_id which redirects to $self->input_job->input_id to simplify porting) - Changed Worker to only use hive 'extended' function if the processing module isa(Bio::EnsEMBL::Hive::Process). Also allows all RunnableDB modules to still be used (or any object which implements a minimal 'RunnableDB interface') (new, input_id, db, fetch_input, run, write_output)
-
Jessica Severin authored
reordered where the blocking checks are done (added, deleted, moved).
-
Jessica Severin authored
-
Jessica Severin authored
-
Jessica Severin authored
needed workers after this worker is done. Useful in debugging one's dataflow and blocking_ctrl graphs by running one worker at a time (like stepping in a debugger)
-
Jessica Severin authored
-
- Mar 02, 2005
-
-
Jessica Severin authored
a job that has been flowed into an analysis/process
-
Jessica Severin authored
-
Jessica Severin authored
-
- Feb 23, 2005
-
-
Jessica Severin authored
-
Jessica Severin authored
added option -no_pend which ignores the pending_count when figuring out how many workers to submit removed some superfluous calls to Queen::get_num_running_workers
-
Jessica Severin authored
-
Jessica Severin authored
when debugging an analysis which fails and would increment the retry_count.
-
Jessica Severin authored
-
Jessica Severin authored
-
Jessica Severin authored
-
Jessica Severin authored
to be promoted to 'DONE'
-
- Feb 22, 2005
-
-
Jessica Severin authored
-
- Feb 21, 2005
-
-
Jessica Severin authored
needed to better manage the hive system's load on the database housing all the hive related tables (in case the database is overloaded by multiple users). Added analysis_stats.sync_lock column (and correspondly in Object and Adaptor) Added Queen::safe_synchronize_AnalysisStats method which wraps over the synchronize_AnalysisStats method and does various checks and locks to ensure that only one worker is trying to do a 'synchronize' on a given analysis at any given moment. Cleaned up API between Queen/Worker so that worker only talks directly to the Queen, rather than getting the underlying database adaptor. Added analysis_job columns runtime_msec, query_count to provide more data on how the jobs hammer a database (queries/sec).
-
- Feb 17, 2005
-
-
Jessica Severin authored
called when worker dies to replace itself in the needed_workers count since it's decremented when it's born, and it's counted as living (and subtracted) as long as it's running. This gunarantees that another worker will quickly be created after this one dies (and not need to wait for a synch to happen)
-
Jessica Severin authored
-
- Feb 16, 2005
-
-
Jessica Severin authored
is when there are lots of workers 'WORKING' so as to avoid them falling over each other. The 'WORKING' state only exists in the middle of a large run. When the last worker dies, the state is 'ALL_CLAIMED' so the sync on death will happen properly. As the last pile of workers die they will all do a synch, but that's OK since the system needs to be properly synched when the last one dies since there won't be anybody left to do it. Also added 10 minute check for if already 'SYNCHING' to deal with case if worker dies in the middle of 'SYNCHING'.
-