diff --git a/modules/Bio/EnsEMBL/Hive/PipeConfig/ApplyToDatabases_conf.pm b/modules/Bio/EnsEMBL/Hive/PipeConfig/ApplyToDatabases_conf.pm index beee91facf8f8b0b3a888d09a4dbe61c91617206..cf58f706407d35ab584dfe57375529f24a0b1cb5 100644 --- a/modules/Bio/EnsEMBL/Hive/PipeConfig/ApplyToDatabases_conf.pm +++ b/modules/Bio/EnsEMBL/Hive/PipeConfig/ApplyToDatabases_conf.pm @@ -67,7 +67,6 @@ sub pipeline_analyses { 'inputquery' => q{SHOW DATABASES LIKE "}.$self->o('only_databases').q{"}, 'column_names' => [ 'dbname' ], }, - -hive_capacity => 5, # allow several workers to perform identical tasks in parallel -input_ids => [ { 'db_conn' => $self->o('source_server1'), 'input_id' => { 'db_conn' => {'-host' => $self->o('source_server1', '-host'), '-port' => $self->o('source_server1', '-port'), '-user' => $self->o('source_server1', '-user'), '-pass' => $self->o('source_server1', '-pass'), '-dbname' => '#dbname#'}, }, }, { 'db_conn' => $self->o('source_server2'), 'input_id' => { 'db_conn' => {'-host' => $self->o('source_server2', '-host'), '-port' => $self->o('source_server2', '-port'), '-user' => $self->o('source_server2', '-user'), '-pass' => $self->o('source_server2', '-pass'), '-dbname' => '#dbname#'}, }, }, @@ -81,7 +80,7 @@ sub pipeline_analyses { -module => 'Bio::EnsEMBL::Hive::RunnableDB::Dummy', # use SqlCmd.pm to run your query or another JobFactory.pm to make another fan on table names -parameters => { }, - -hive_capacity => 10, # allow several workers to perform identical tasks in parallel + -analysis_capacity => 10, # allow several workers to perform identical tasks in parallel -input_ids => [ # (jobs for this analysis will be flown_into via branch-2 from 'get_databases' jobs above) ], diff --git a/modules/Bio/EnsEMBL/Hive/PipeConfig/FileZipperUnzipper_conf.pm b/modules/Bio/EnsEMBL/Hive/PipeConfig/FileZipperUnzipper_conf.pm index a8c41f47cec8b6ce97a7fe6a2bb19ee7d4f8b757..d7ea1adc3827c9c79cc1d7e0636f3ee9bbb9b427 100644 --- a/modules/Bio/EnsEMBL/Hive/PipeConfig/FileZipperUnzipper_conf.pm +++ b/modules/Bio/EnsEMBL/Hive/PipeConfig/FileZipperUnzipper_conf.pm @@ -109,7 +109,7 @@ sub pipeline_analyses { -parameters => { 'cmd' => 'gzip '.($self->o('unzip')?'-d ':'').'#filename#', }, - -hive_capacity => $self->o('zipping_capacity'), # allow several workers to perform identical tasks in parallel + -analysis_capacity => $self->o('zipping_capacity'), # allow several workers to perform identical tasks in parallel -input_ids => [ # (jobs for this analysis will be flown_into via branch-2 from 'get_tables' jobs above) ], diff --git a/modules/Bio/EnsEMBL/Hive/PipeConfig/RunListOfCommandsOnFarm_conf.pm b/modules/Bio/EnsEMBL/Hive/PipeConfig/RunListOfCommandsOnFarm_conf.pm index d48db110dec6c402024be44428ab9342c61fef99..d4f40fab44b3f60aabb27ac730fb421945d7f3f5 100644 --- a/modules/Bio/EnsEMBL/Hive/PipeConfig/RunListOfCommandsOnFarm_conf.pm +++ b/modules/Bio/EnsEMBL/Hive/PipeConfig/RunListOfCommandsOnFarm_conf.pm @@ -87,7 +87,7 @@ sub pipeline_analyses { { -logic_name => 'run_cmd', -module => 'Bio::EnsEMBL::Hive::RunnableDB::SystemCmd', -parameters => { }, - -hive_capacity => $self->o('capacity'), # allow several workers to perform identical tasks in parallel + -analysis_capacity => $self->o('capacity'), # allow several workers to perform identical tasks in parallel }, ]; } diff --git a/modules/Bio/EnsEMBL/Hive/PipeConfig/TableDumperZipper_conf.pm b/modules/Bio/EnsEMBL/Hive/PipeConfig/TableDumperZipper_conf.pm index 6079c67a082482234272e2216913d7a15be29a01..e5eeb79eabb9a8a4e5740c9c8c8e923710a62860 100644 --- a/modules/Bio/EnsEMBL/Hive/PipeConfig/TableDumperZipper_conf.pm +++ b/modules/Bio/EnsEMBL/Hive/PipeConfig/TableDumperZipper_conf.pm @@ -125,7 +125,7 @@ sub pipeline_analyses { 'target_dir' => $self->o('target_dir'), 'cmd' => 'mysqldump '.$self->dbconn_2_mysql('source_db', 0).' '.$self->o('source_db','-dbname').($self->o('with_schema')?'':' -t').' #table_name# | gzip >#target_dir#/#table_name#.sql.gz', }, - -hive_capacity => $self->o('dumping_capacity'), # allow several workers to perform identical tasks in parallel + -analysis_capacity => $self->o('dumping_capacity'), # allow several workers to perform identical tasks in parallel -input_ids => [ # (jobs for this analysis will be flown_into via branch-2 from 'get_tables' jobs above) ],