X-Git-Url: https://git.sur5r.net/?a=blobdiff_plain;f=bacula%2Fsrc%2Fdird%2Fmigrate.c;h=357250b3348bd826814674f494d8bae965cdf0a8;hb=04d2e746550ef971dcbd3253026b1c97979f3bbe;hp=65ebb0b7f8cfd375cd35fb2d32897f73b3050a05;hpb=c6a1b2d710b28a7a5801698d1b908fc1d3fd561e;p=bacula%2Fbacula diff --git a/bacula/src/dird/migrate.c b/bacula/src/dird/migrate.c index 65ebb0b7f8..357250b334 100644 --- a/bacula/src/dird/migrate.c +++ b/bacula/src/dird/migrate.c @@ -1,7 +1,36 @@ +/* + Bacula® - The Network Backup Solution + + Copyright (C) 2004-2010 Free Software Foundation Europe e.V. + + The main author of Bacula is Kern Sibbald, with contributions from + many others, a complete list can be found in the file AUTHORS. + This program is Free Software; you can redistribute it and/or + modify it under the terms of version three of the GNU Affero General Public + License as published by the Free Software Foundation and included + in the file LICENSE. + + This program is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + General Public License for more details. + + You should have received a copy of the GNU Affero General Public License + along with this program; if not, write to the Free Software + Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA + 02110-1301, USA. + + Bacula® is a registered trademark of Kern Sibbald. + The licensor of Bacula is the Free Software Foundation Europe + (FSFE), Fiduciary Program, Sumatrastrasse 25, 8006 Zürich, + Switzerland, email:ftf@fsfeurope.org. +*/ /* * * Bacula Director -- migrate.c -- responsible for doing - * migration jobs. + * migration and copy jobs. + * + * Also handles Copy jobs (March MMVIII) * * Kern Sibbald, September MMIV * @@ -12,21 +41,6 @@ * to do the backup. * When the Storage daemon finishes the job, update the DB. * - * Version $Id$ - */ -/* - Copyright (C) 2004-2006 Kern Sibbald - - This program is free software; you can redistribute it and/or - modify it under the terms of the GNU General Public License - version 2 as amended with additional clauses defined in the - file LICENSE in the main source directory. - - This program is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - the file LICENSE for additional details. - */ #include "bacula.h" @@ -38,139 +52,202 @@ #include #endif -static char OKbootstrap[] = "3000 OK bootstrap\n"; -static bool get_job_to_migrate(JCR *jcr); +static const int dbglevel = 10; + +static int getJob_to_migrate(JCR *jcr); +struct idpkt; +static bool regex_find_jobids(JCR *jcr, idpkt *ids, const char *query1, + const char *query2, const char *type); +static bool find_mediaid_then_jobids(JCR *jcr, idpkt *ids, const char *query1, + const char *type); +static bool find_jobids_from_mediaid_list(JCR *jcr, idpkt *ids, const char *type); +static bool find_jobids_of_pool_uncopied_jobs(JCR *jcr, idpkt *ids); +static void start_migration_job(JCR *jcr); +static int get_next_dbid_from_list(char **p, DBId_t *DBId); +static bool set_migration_next_pool(JCR *jcr, POOL **pool); /* * Called here before the job is run to do the job - * specific setup. + * specific setup. Note, one of the important things to + * complete in this init code is to make the definitive + * choice of input and output storage devices. This is + * because immediately after the init, the job is queued + * in the jobq.c code, and it checks that all the resources + * (storage resources in particular) are available, so these + * must all be properly defined. + * + * previous_jr refers to the job DB record of the Job that is + * going to be migrated. + * prev_job refers to the job resource of the Job that is + * going to be migrated. + * jcr is the jcr for the current "migration" job. It is a + * control job that is put in the DB as a migration job, which + * means that this job migrated a previous job to a new job. + * No Volume or File data is associated with this control + * job. + * mig_jcr refers to the newly migrated job that is run by + * the current jcr. It is a backup job that moves (migrates) the + * data written for the previous_jr into the new pool. This + * job (mig_jcr) becomes the new backup job that replaces + * the original backup job. Note, this jcr is not really run. It + * is simply attached to the current jcr. It will show up in + * the Director's status output, but not in the SD or FD, both of + * which deal only with the current migration job (i.e. jcr). */ bool do_migration_init(JCR *jcr) { - POOL_DBR pr; + POOL *pool = NULL; + JOB *job, *prev_job; + JCR *mig_jcr; /* newly migrated job */ + int count; - if (!get_job_to_migrate(jcr)) { - return false; - } - if (jcr->previous_jr.JobId == 0) { - return true; /* no work */ - } + apply_pool_overrides(jcr); - if (!get_or_create_fileset_record(jcr)) { + if (!allow_duplicate_job(jcr)) { return false; } + jcr->jr.PoolId = get_or_create_pool_record(jcr, jcr->pool->name()); + if (jcr->jr.PoolId == 0) { + Dmsg1(dbglevel, "JobId=%d no PoolId\n", (int)jcr->JobId); + Jmsg(jcr, M_FATAL, 0, _("Could not get or create a Pool record.\n")); + return false; + } /* - * Get the Pool record -- first apply any level defined pools + * Note, at this point, pool is the pool for this job. We + * transfer it to rpool (read pool), and a bit later, + * pool will be changed to point to the write pool, + * which comes from pool->NextPool. */ - switch (jcr->previous_jr.JobLevel) { - case L_FULL: - if (jcr->full_pool) { - jcr->pool = jcr->full_pool; - } - break; - case L_INCREMENTAL: - if (jcr->inc_pool) { - jcr->pool = jcr->inc_pool; - } - break; - case L_DIFFERENTIAL: - if (jcr->dif_pool) { - jcr->pool = jcr->dif_pool; - } - break; - } - memset(&pr, 0, sizeof(pr)); - bstrncpy(pr.Name, jcr->pool->hdr.name, sizeof(pr.Name)); + jcr->rpool = jcr->pool; /* save read pool */ + pm_strcpy(jcr->rpool_source, jcr->pool_source); - while (!db_get_pool_record(jcr, jcr->db, &pr)) { /* get by Name */ - /* Try to create the pool */ - if (create_pool(jcr, jcr->db, jcr->pool, POOL_OP_CREATE) < 0) { - Jmsg(jcr, M_FATAL, 0, _("Pool %s not in database. %s"), pr.Name, - db_strerror(jcr->db)); - return false; - } else { - Jmsg(jcr, M_INFO, 0, _("Pool %s created in database.\n"), pr.Name); - } - } - jcr->jr.PoolId = pr.PoolId; - /* If pool storage specified, use it instead of job storage */ - copy_storage(jcr, jcr->pool->storage); + Dmsg2(dbglevel, "Read pool=%s (From %s)\n", jcr->rpool->name(), jcr->rpool_source); - if (!jcr->storage) { - Jmsg(jcr, M_FATAL, 0, _("No Storage specification found in Job or Pool.\n")); + if (!get_or_create_fileset_record(jcr)) { + Dmsg1(dbglevel, "JobId=%d no FileSet\n", (int)jcr->JobId); + Jmsg(jcr, M_FATAL, 0, _("Could not get or create the FileSet record.\n")); return false; } - if (!create_restore_bootstrap_file(jcr)) { + /* If we find a job or jobs to migrate it is previous_jr.JobId */ + count = getJob_to_migrate(jcr); + if (count < 0) { return false; } - return true; -} + if (count == 0) { + set_migration_next_pool(jcr, &pool); + return true; /* no work */ + } -/* - * Do a Migration of a previous job - * - * Returns: false on failure - * true on success - */ -bool do_migration(JCR *jcr) -{ - POOL_DBR pr; - POOL *pool; - char ed1[100]; - BSOCK *sd; - JOB *job, *tjob; - JCR *tjcr; + Dmsg1(dbglevel, "Back from getJob_to_migrate JobId=%d\n", (int)jcr->JobId); if (jcr->previous_jr.JobId == 0) { - jcr->JobStatus = JS_Terminated; - migration_cleanup(jcr, jcr->JobStatus); + Dmsg1(dbglevel, "JobId=%d no previous JobId\n", (int)jcr->JobId); + Jmsg(jcr, M_INFO, 0, _("No previous Job found to %s.\n"), jcr->get_ActionName(0)); + set_migration_next_pool(jcr, &pool); + return true; /* no work */ + } + + if (create_restore_bootstrap_file(jcr) < 0) { + Jmsg(jcr, M_FATAL, 0, _("Create bootstrap file failed.\n")); + return false; + } + + if (jcr->previous_jr.JobId == 0 || jcr->ExpectedFiles == 0) { + set_jcr_job_status(jcr, JS_Terminated); + Dmsg1(dbglevel, "JobId=%d expected files == 0\n", (int)jcr->JobId); + if (jcr->previous_jr.JobId == 0) { + Jmsg(jcr, M_INFO, 0, _("No previous Job found to %s.\n"), jcr->get_ActionName(0)); + } else { + Jmsg(jcr, M_INFO, 0, _("Previous Job has no data to %s.\n"), jcr->get_ActionName(0)); + } + set_migration_next_pool(jcr, &pool); return true; /* no work */ } - Dmsg4(000, "Target: Name=%s JobId=%d Type=%c Level=%c\n", - jcr->previous_jr.Name, jcr->previous_jr.JobId, - jcr->previous_jr.JobType, jcr->previous_jr.JobLevel); - Dmsg4(000, "Current: Name=%s JobId=%d Type=%c Level=%c\n", - jcr->jr.Name, jcr->jr.JobId, + + Dmsg5(dbglevel, "JobId=%d: Current: Name=%s JobId=%d Type=%c Level=%c\n", + (int)jcr->JobId, + jcr->jr.Name, (int)jcr->jr.JobId, jcr->jr.JobType, jcr->jr.JobLevel); LockRes(); job = (JOB *)GetResWithName(R_JOB, jcr->jr.Name); - tjob = (JOB *)GetResWithName(R_JOB, jcr->previous_jr.Name); + prev_job = (JOB *)GetResWithName(R_JOB, jcr->previous_jr.Name); UnlockRes(); - if (!job || !tjob) { + if (!job) { + Jmsg(jcr, M_FATAL, 0, _("Job resource not found for \"%s\".\n"), jcr->jr.Name); + return false; + } + if (!prev_job) { + Jmsg(jcr, M_FATAL, 0, _("Previous Job resource not found for \"%s\".\n"), + jcr->previous_jr.Name); return false; } - /* - * Target jcr is the new Job that corresponds to the original - * target job. It "runs" at the same time as the current - * migration job and becomes a new backup job that replaces - * the original backup job. Most operations on the current - * migration jcr are also done on the target jcr. - */ - tjcr = jcr->previous_jcr = new_jcr(sizeof(JCR), dird_free_jcr); - memcpy(&tjcr->previous_jr, &jcr->previous_jr, sizeof(tjcr->previous_jr)); + jcr->spool_data = job->spool_data; /* turn on spooling if requested in job */ + + /* Create a migration jcr */ + mig_jcr = jcr->mig_jcr = new_jcr(sizeof(JCR), dird_free_jcr); + memcpy(&mig_jcr->previous_jr, &jcr->previous_jr, sizeof(mig_jcr->previous_jr)); - /* Turn the tjcr into a "real" job */ - set_jcr_defaults(tjcr, tjob); - if (!setup_job(tjcr)) { + /* + * Turn the mig_jcr into a "real" job that takes on the aspects of + * the previous backup job "prev_job". + */ + set_jcr_defaults(mig_jcr, prev_job); + if (!setup_job(mig_jcr)) { + Jmsg(jcr, M_FATAL, 0, _("setup job failed.\n")); return false; } - /* Set output PoolId and FileSetId. */ - tjcr->jr.PoolId = jcr->jr.PoolId; - tjcr->jr.FileSetId = jcr->jr.FileSetId; + + /* Now reset the job record from the previous job */ + memcpy(&mig_jcr->jr, &jcr->previous_jr, sizeof(mig_jcr->jr)); + /* Update the jr to reflect the new values of PoolId and JobId. */ + mig_jcr->jr.PoolId = jcr->jr.PoolId; + mig_jcr->jr.JobId = mig_jcr->JobId; + + /* Don't let WatchDog checks Max*Time value on this Job */ + mig_jcr->no_maxtime = true; + + Dmsg4(dbglevel, "mig_jcr: Name=%s JobId=%d Type=%c Level=%c\n", + mig_jcr->jr.Name, (int)mig_jcr->jr.JobId, + mig_jcr->jr.JobType, mig_jcr->jr.JobLevel); + + if (set_migration_next_pool(jcr, &pool)) { + /* If pool storage specified, use it for restore */ + copy_rstorage(mig_jcr, pool->storage, _("Pool resource")); + copy_rstorage(jcr, pool->storage, _("Pool resource")); + + mig_jcr->pool = jcr->pool; + mig_jcr->jr.PoolId = jcr->jr.PoolId; + } + + return true; +} + + +/* + * set_migration_next_pool() called by do_migration_init() + * at differents stages. + * The idea here is tofactorize the NextPool's search code and + * to permit do_migration_init() to return with NextPool set in jcr struct. + */ +static bool set_migration_next_pool(JCR *jcr, POOL **retpool) +{ + POOL_DBR pr; + POOL *pool; + char ed1[100]; /* * Get the PoolId used with the original job. Then * find the pool name from the database record. */ memset(&pr, 0, sizeof(pr)); - pr.PoolId = tjcr->previous_jr.PoolId; + pr.PoolId = jcr->jr.PoolId; if (!db_get_pool_record(jcr, jcr->db, &pr)) { Jmsg(jcr, M_FATAL, 0, _("Pool for JobId %s not in database. ERR=%s\n"), edit_int64(pr.PoolId, ed1), db_strerror(jcr->db)); @@ -178,62 +255,81 @@ bool do_migration(JCR *jcr) } /* Get the pool resource corresponding to the original job */ pool = (POOL *)GetResWithName(R_POOL, pr.Name); + *retpool = pool; if (!pool) { Jmsg(jcr, M_FATAL, 0, _("Pool resource \"%s\" not found.\n"), pr.Name); return false; } - /* Check Migration time and High/Low water marks */ - /* ***FIXME*** */ - - /* If pool storage specified, use it for restore */ - copy_storage(tjcr, pool->storage); - - /* If the original backup pool has a NextPool, make sure a - * record exists in the database. + /* + * If the original backup pool has a NextPool, make sure a + * record exists in the database. Note, in this case, we + * will be migrating from pool to pool->NextPool. */ if (pool->NextPool) { - memset(&pr, 0, sizeof(pr)); - bstrncpy(pr.Name, pool->NextPool->hdr.name, sizeof(pr.Name)); - - while (!db_get_pool_record(jcr, jcr->db, &pr)) { /* get by Name */ - /* Try to create the pool */ - if (create_pool(jcr, jcr->db, pool->NextPool, POOL_OP_CREATE) < 0) { - Jmsg(jcr, M_FATAL, 0, _("Pool \"%s\" not in database. %s"), pr.Name, - db_strerror(jcr->db)); - return false; - } else { - Jmsg(jcr, M_INFO, 0, _("Pool \"%s\" created in database.\n"), pr.Name); - } + jcr->jr.PoolId = get_or_create_pool_record(jcr, pool->NextPool->name()); + if (jcr->jr.PoolId == 0) { + return false; } - /* - * put the "NextPool" resource pointer in our jcr so that we - * can pull the Storage reference from it. - */ - tjcr->pool = jcr->pool = pool->NextPool; - tjcr->jr.PoolId = jcr->jr.PoolId = pr.PoolId; } + if (!set_migration_wstorage(jcr, pool)) { + return false; + } + jcr->pool = pool->NextPool; + pm_strcpy(jcr->pool_source, _("Job Pool's NextPool resource")); - /* If pool storage specified, use it instead of job storage for backup */ - copy_storage(jcr, jcr->pool->storage); + Dmsg2(dbglevel, "Write pool=%s read rpool=%s\n", jcr->pool->name(), jcr->rpool->name()); - /* Print Job Start message */ - Jmsg(jcr, M_INFO, 0, _("Start Migration JobId %s, Job=%s\n"), - edit_uint64(jcr->JobId, ed1), jcr->Job); + return true; +} - set_jcr_job_status(jcr, JS_Running); - set_jcr_job_status(tjcr, JS_Running); - Dmsg2(000, "JobId=%d JobLevel=%c\n", jcr->jr.JobId, jcr->jr.JobLevel); - if (!db_update_job_start_record(jcr, jcr->db, &jcr->jr)) { - Jmsg(jcr, M_FATAL, 0, "%s", db_strerror(jcr->db)); - return false; + +/* + * Do a Migration of a previous job + * + * Returns: false on failure + * true on success + */ +bool do_migration(JCR *jcr) +{ + char ed1[100]; + BSOCK *sd; + JCR *mig_jcr = jcr->mig_jcr; /* newly migrated job */ + + /* + * If mig_jcr is NULL, there is nothing to do for this job, + * so set a normal status, cleanup and return OK. + */ + if (!mig_jcr) { + set_jcr_job_status(jcr, JS_Terminated); + migration_cleanup(jcr, jcr->JobStatus); + return true; } - if (!db_update_job_start_record(tjcr, tjcr->db, &tjcr->jr)) { - Jmsg(jcr, M_FATAL, 0, "%s", db_strerror(tjcr->db)); - return false; + if (!db_get_job_record(jcr, jcr->db, &jcr->previous_jr)) { + Jmsg(jcr, M_FATAL, 0, _("Could not get job record for JobId %s to %s. ERR=%s"), + edit_int64(jcr->previous_jr.JobId, ed1), + jcr->get_ActionName(0), + db_strerror(jcr->db)); + set_jcr_job_status(jcr, JS_Terminated); + migration_cleanup(jcr, jcr->JobStatus); + return true; + } + /* Make sure this job was not already migrated */ + if (jcr->previous_jr.JobType != JT_BACKUP && + jcr->previous_jr.JobType != JT_JOB_COPY) { + Jmsg(jcr, M_INFO, 0, _("JobId %s already %s probably by another Job. %s stopped.\n"), + edit_int64(jcr->previous_jr.JobId, ed1), + jcr->get_ActionName(1), + jcr->get_OperationName()); + set_jcr_job_status(jcr, JS_Terminated); + migration_cleanup(jcr, jcr->JobStatus); + return true; } + /* Print Job Start message */ + Jmsg(jcr, M_INFO, 0, _("Start %s JobId %s, Job=%s\n"), + jcr->get_OperationName(), edit_uint64(jcr->JobId, ed1), jcr->Job); /* * Open a message channel connection with the Storage @@ -243,7 +339,7 @@ bool do_migration(JCR *jcr) */ Dmsg0(110, "Open connection with storage daemon\n"); set_jcr_job_status(jcr, JS_WaitSD); - set_jcr_job_status(tjcr, JS_WaitSD); + set_jcr_job_status(mig_jcr, JS_WaitSD); /* * Start conversation with Storage daemon */ @@ -254,56 +350,144 @@ bool do_migration(JCR *jcr) /* * Now start a job with the Storage daemon */ - Dmsg2(000, "Read store=%s, write store=%s\n", - ((STORE *)tjcr->storage->first())->hdr.name, - ((STORE *)jcr->storage->first())->hdr.name); - if (!start_storage_daemon_job(jcr, tjcr->storage, jcr->storage)) { + Dmsg2(dbglevel, "Read store=%s, write store=%s\n", + ((STORE *)jcr->rstorage->first())->name(), + ((STORE *)jcr->wstorage->first())->name()); + + if (!start_storage_daemon_job(jcr, jcr->rstorage, jcr->wstorage, /*send_bsr*/true)) { return false; } Dmsg0(150, "Storage daemon connection OK\n"); - if (!send_bootstrap_file(jcr, sd) || - !response(jcr, sd, OKbootstrap, "Bootstrap", DISPLAY_ERROR)) { + + /* + * We re-update the job start record so that the start + * time is set after the run before job. This avoids + * that any files created by the run before job will + * be saved twice. They will be backed up in the current + * job, but not in the next one unless they are changed. + * Without this, they will be backed up in this job and + * in the next job run because in that case, their date + * is after the start of this run. + */ + jcr->start_time = time(NULL); + jcr->jr.StartTime = jcr->start_time; + jcr->jr.JobTDate = jcr->start_time; + set_jcr_job_status(jcr, JS_Running); + + /* Update job start record for this migration control job */ + if (!db_update_job_start_record(jcr, jcr->db, &jcr->jr)) { + Jmsg(jcr, M_FATAL, 0, "%s", db_strerror(jcr->db)); + return false; + } + + + mig_jcr->start_time = time(NULL); + mig_jcr->jr.StartTime = mig_jcr->start_time; + mig_jcr->jr.JobTDate = mig_jcr->start_time; + set_jcr_job_status(mig_jcr, JS_Running); + + /* Update job start record for the real migration backup job */ + if (!db_update_job_start_record(mig_jcr, mig_jcr->db, &mig_jcr->jr)) { + Jmsg(jcr, M_FATAL, 0, "%s", db_strerror(mig_jcr->db)); return false; } + Dmsg4(dbglevel, "mig_jcr: Name=%s JobId=%d Type=%c Level=%c\n", + mig_jcr->jr.Name, (int)mig_jcr->jr.JobId, + mig_jcr->jr.JobType, mig_jcr->jr.JobLevel); + + /* - * Now start a Storage daemon message thread + * Start the job prior to starting the message thread below + * to avoid two threads from using the BSOCK structure at + * the same time. */ - if (!start_storage_daemon_message_thread(jcr)) { + if (!sd->fsend("run")) { return false; } - if (!bnet_fsend(sd, "run")) { + /* + * Now start a Storage daemon message thread + */ + if (!start_storage_daemon_message_thread(jcr)) { return false; } + set_jcr_job_status(jcr, JS_Running); - set_jcr_job_status(tjcr, JS_Running); + set_jcr_job_status(mig_jcr, JS_Running); /* Pickup Job termination data */ - /* Note, the SD stores in jcr->JobFiles/ReadBytes/JobBytes/Errors */ + /* Note, the SD stores in jcr->JobFiles/ReadBytes/JobBytes/JobErrors */ wait_for_storage_daemon_termination(jcr); + set_jcr_job_status(jcr, jcr->SDJobStatus); + db_write_batch_file_records(jcr); /* used by bulk batch file insert */ + if (jcr->JobStatus != JS_Terminated) { + return false; + } - jcr->JobStatus = jcr->SDJobStatus; - if (jcr->JobStatus == JS_Terminated) { - migration_cleanup(jcr, jcr->JobStatus); - return true; + migration_cleanup(jcr, jcr->JobStatus); + + return true; +} + +struct idpkt { + POOLMEM *list; + uint32_t count; +}; + +/* Add an item to the list if it is unique */ +static void add_unique_id(idpkt *ids, char *item) +{ + const int maxlen = 30; + char id[maxlen+1]; + char *q = ids->list; + + /* Walk through current list to see if each item is the same as item */ + for ( ; *q; ) { + id[0] = 0; + for (int i=0; icount == 0) { + ids->list[0] = 0; + } else { + pm_strcat(ids->list, ","); } - return false; + pm_strcat(ids->list, item); + ids->count++; +// Dmsg3(0, "add_uniq count=%d Ids=%p %s\n", ids->count, ids->list, ids->list); + return; } /* - * Callback handler make list of JobIds + * Callback handler make list of DB Ids */ -static int jobid_handler(void *ctx, int num_fields, char **row) +static int unique_dbid_handler(void *ctx, int num_fields, char **row) { - POOLMEM *JobIds = (POOLMEM *)ctx; + idpkt *ids = (idpkt *)ctx; - if (JobIds[0] != 0) { - pm_strcat(JobIds, ","); + /* Sanity check */ + if (!row || !row[0]) { + Dmsg0(dbglevel, "dbid_hdlr error empty row\n"); + return 1; /* stop calling us */ } - pm_strcat(JobIds, row[0]); + + add_unique_id(ids, row[0]); + Dmsg3(dbglevel, "dbid_hdlr count=%d Ids=%p %s\n", ids->count, ids->list, ids->list); return 0; } @@ -329,7 +513,7 @@ static int unique_name_handler(void *ctx, int num_fields, char **row) memset(new_item, 0, sizeof(uitem)); new_item->item = bstrdup(row[0]); - Dmsg1(000, "Item=%s\n", row[0]); + Dmsg1(dbglevel, "Unique_name_hdlr Item=%s\n", row[0]); item = (uitem *)list->binary_insert((void *)new_item, item_compare); if (item != new_item) { /* already in list */ free(new_item->item); @@ -339,345 +523,598 @@ static int unique_name_handler(void *ctx, int num_fields, char **row) return 0; } +/* Get Job names in Pool */ +const char *sql_job = + "SELECT DISTINCT Job.Name from Job,Pool" + " WHERE Pool.Name='%s' AND Job.PoolId=Pool.PoolId"; + +/* Get JobIds from regex'ed Job names */ +const char *sql_jobids_from_job = + "SELECT DISTINCT Job.JobId,Job.StartTime FROM Job,Pool" + " WHERE Job.Name='%s' AND Pool.Name='%s' AND Job.PoolId=Pool.PoolId" + " ORDER by Job.StartTime"; + +/* Get Client names in Pool */ +const char *sql_client = + "SELECT DISTINCT Client.Name from Client,Pool,Job" + " WHERE Pool.Name='%s' AND Job.ClientId=Client.ClientId AND" + " Job.PoolId=Pool.PoolId"; + +/* Get JobIds from regex'ed Client names */ +const char *sql_jobids_from_client = + "SELECT DISTINCT Job.JobId,Job.StartTime FROM Job,Pool,Client" + " WHERE Client.Name='%s' AND Pool.Name='%s' AND Job.PoolId=Pool.PoolId" + " AND Job.ClientId=Client.ClientId AND Job.Type IN ('B','C')" + " AND Job.JobStatus IN ('T','W')" + " ORDER by Job.StartTime"; + +/* Get Volume names in Pool */ +const char *sql_vol = + "SELECT DISTINCT VolumeName FROM Media,Pool WHERE" + " VolStatus in ('Full','Used','Error') AND Media.Enabled=1 AND" + " Media.PoolId=Pool.PoolId AND Pool.Name='%s'"; + +/* Get JobIds from regex'ed Volume names */ +const char *sql_jobids_from_vol = + "SELECT DISTINCT Job.JobId,Job.StartTime FROM Media,JobMedia,Job" + " WHERE Media.VolumeName='%s' AND Media.MediaId=JobMedia.MediaId" + " AND JobMedia.JobId=Job.JobId AND Job.Type IN ('B','C')" + " AND Job.JobStatus IN ('T','W') AND Media.Enabled=1" + " ORDER by Job.StartTime"; + const char *sql_smallest_vol = - "SELECT MediaId FROM Media,Pool WHERE" - " VolStatus in ('Full','Used','Error') AND" + "SELECT Media.MediaId FROM Media,Pool,JobMedia WHERE" + " Media.MediaId in (SELECT DISTINCT MediaId from JobMedia) AND" + " Media.VolStatus in ('Full','Used','Error') AND Media.Enabled=1 AND" " Media.PoolId=Pool.PoolId AND Pool.Name='%s'" " ORDER BY VolBytes ASC LIMIT 1"; const char *sql_oldest_vol = - "SELECT MediaId FROM Media,Pool WHERE" - " VolStatus in ('Full','Used','Error') AND" + "SELECT Media.MediaId FROM Media,Pool,JobMedia WHERE" + " Media.MediaId in (SELECT DISTINCT MediaId from JobMedia) AND" + " Media.VolStatus in ('Full','Used','Error') AND Media.Enabled=1 AND" " Media.PoolId=Pool.PoolId AND Pool.Name='%s'" " ORDER BY LastWritten ASC LIMIT 1"; +/* Get JobIds when we have selected MediaId */ const char *sql_jobids_from_mediaid = - "SELECT DISTINCT Job.JobId FROM JobMedia,Job" - " WHERE JobMedia.JobId=Job.JobId AND JobMedia.MediaId=%s" + "SELECT DISTINCT Job.JobId,Job.StartTime FROM JobMedia,Job" + " WHERE JobMedia.JobId=Job.JobId AND JobMedia.MediaId IN (%s)" + " AND Job.Type IN ('B','C') AND Job.JobStatus IN ('T','W')" " ORDER by Job.StartTime"; +/* Get the number of bytes in the pool */ const char *sql_pool_bytes = - "SELECT SUM(VolBytes) FROM Media,Pool WHERE" - " VolStatus in ('Full','Used','Error','Append') AND" - " Media.PoolId=Pool.PoolId AND Pool.Name='%s'"; - -const char *sql_vol_bytes = + "SELECT SUM(JobBytes) FROM Job WHERE JobId IN" + " (SELECT DISTINCT Job.JobId from Pool,Job,Media,JobMedia WHERE" + " Pool.Name='%s' AND Media.PoolId=Pool.PoolId AND" + " VolStatus in ('Full','Used','Error','Append') AND Media.Enabled=1 AND" + " Job.Type IN ('B','C') AND Job.JobStatus IN ('T','W') AND" + " JobMedia.JobId=Job.JobId AND Job.PoolId=Media.PoolId)"; + +/* Get the number of bytes in the Jobs */ +const char *sql_job_bytes = + "SELECT SUM(JobBytes) FROM Job WHERE JobId IN (%s)"; + +/* Get Media Ids in Pool */ +const char *sql_mediaids = "SELECT MediaId FROM Media,Pool WHERE" - " VolStatus in ('Full','Used','Error') AND" - " Media.PoolId=Pool.PoolId AND Pool.Name='%s' AND" - " VolBytes<%s ORDER BY LastWritten ASC LIMIT 1"; - -const char *sql_client = - "SELECT DISTINCT Client.Name from Client,Pool,Media,Job,JobMedia " - " WHERE Media.PoolId=Pool.PoolId AND Pool.Name='%s' AND" - " JobMedia.JobId=Job.JobId AND Job.ClientId=Client.ClientId AND" - " Job.PoolId=Media.PoolId"; - -const char *sql_job = - "SELECT DISTINCT Job.Name from Job,Pool" - " WHERE Pool.Name='%s' AND Job.PoolId=Pool.PoolId"; - -const char *sql_jobids_from_job = - "SELECT DISTINCT Job.JobId FROM Job,Pool" - " WHERE Job.Name=%s AND Pool.Name='%s' AND Job.PoolId=Pool.PoolId" + " VolStatus in ('Full','Used','Error') AND Media.Enabled=1 AND" + " Media.PoolId=Pool.PoolId AND Pool.Name='%s' ORDER BY LastWritten ASC"; + +/* Get JobIds in Pool longer than specified time */ +const char *sql_pool_time = + "SELECT DISTINCT Job.JobId FROM Pool,Job,Media,JobMedia WHERE" + " Pool.Name='%s' AND Media.PoolId=Pool.PoolId AND" + " VolStatus IN ('Full','Used','Error') AND Media.Enabled=1 AND" + " Job.Type IN ('B','C') AND Job.JobStatus IN ('T','W') AND" + " JobMedia.JobId=Job.JobId AND Job.PoolId=Media.PoolId" + " AND Job.RealEndTime<='%s'"; + +/* Get JobIds from successfully completed backup jobs which have not been copied before */ +const char *sql_jobids_of_pool_uncopied_jobs = + "SELECT DISTINCT Job.JobId,Job.StartTime FROM Job,Pool" + " WHERE Pool.Name = '%s' AND Pool.PoolId = Job.PoolId" + " AND Job.Type = 'B' AND Job.JobStatus IN ('T','W')" + " AND Job.jobBytes > 0" + " AND Job.JobId NOT IN" + " (SELECT PriorJobId FROM Job WHERE" + " Type IN ('B','C') AND Job.JobStatus IN ('T','W')" + " AND PriorJobId != 0)" " ORDER by Job.StartTime"; - -const char *sql_ujobid = - "SELECT DISTINCT Job.Job from Client,Pool,Media,Job,JobMedia " - " WHERE Media.PoolId=Pool.PoolId AND Pool.Name='%s' AND" - " JobMedia.JobId=Job.JobId AND Job.PoolId=Media.PoolId"; - -const char *sql_vol = - "SELECT DISTINCT VolumeName FROM Media,Pool WHERE" - " VolStatus in ('Full','Used','Error') AND" - " Media.PoolId=Pool.PoolId AND Pool.Name='%s'"; - +/* +* const char *sql_ujobid = +* "SELECT DISTINCT Job.Job from Client,Pool,Media,Job,JobMedia " +* " WHERE Media.PoolId=Pool.PoolId AND Pool.Name='%s' AND" +* " JobMedia.JobId=Job.JobId AND Job.PoolId=Media.PoolId"; +*/ /* - * Returns: false on error - * true if OK and jcr->previous_jr filled in + * + * This is the central piece of code that finds a job or jobs + * actually JobIds to migrate. It first looks to see if one + * has been "manually" specified in jcr->MigrateJobId, and if + * so, it returns that JobId to be run. Otherwise, it + * examines the Selection Type to see what kind of migration + * we are doing (Volume, Job, Client, ...) and applies any + * Selection Pattern if appropriate to obtain a list of JobIds. + * Finally, it will loop over all the JobIds found, except the last + * one starting a new job with MigrationJobId set to that JobId, and + * finally, it returns the last JobId to the caller. + * + * Returns: -1 on error + * 0 if no jobs to migrate + * 1 if OK and jcr->previous_jr filled in */ -static bool get_job_to_migrate(JCR *jcr) +static int getJob_to_migrate(JCR *jcr) { - char ed1[30]; + char ed1[30], ed2[30]; POOL_MEM query(PM_MESSAGE); - POOLMEM *JobIds = get_pool_memory(PM_MESSAGE); JobId_t JobId; - int stat, rc; + DBId_t DBId = 0; + int stat; char *p; - dlist *item_chain; - uitem *item = NULL; - uitem *last_item = NULL; - char prbuf[500]; - regex_t preg; + idpkt ids, mid, jids; + db_int64_ctx ctx; + int64_t pool_bytes; + time_t ttime; + struct tm tm; + char dt[MAX_TIME_LENGTH]; + int count = 0; + + ids.list = get_pool_memory(PM_MESSAGE); + ids.list[0] = 0; + ids.count = 0; + mid.list = get_pool_memory(PM_MESSAGE); + mid.list[0] = 0; + mid.count = 0; + jids.list = get_pool_memory(PM_MESSAGE); + jids.list[0] = 0; + jids.count = 0; + - JobIds[0] = 0; + /* + * If MigrateJobId is set, then we migrate only that Job, + * otherwise, we go through the full selection of jobs to + * migrate. + */ if (jcr->MigrateJobId != 0) { - jcr->previous_jr.JobId = jcr->MigrateJobId; - Dmsg1(000, "previous jobid=%u\n", jcr->MigrateJobId); + Dmsg1(dbglevel, "At Job start previous jobid=%u\n", jcr->MigrateJobId); + edit_uint64(jcr->MigrateJobId, ids.list); + ids.count = 1; } else { switch (jcr->job->selection_type) { case MT_JOB: - if (!jcr->job->selection_pattern) { - Jmsg(jcr, M_FATAL, 0, _("No Migration Job selection pattern specified.\n")); + if (!regex_find_jobids(jcr, &ids, sql_job, sql_jobids_from_job, "Job")) { goto bail_out; - } - Dmsg1(000, "Job regex=%s\n", jcr->job->selection_pattern); - /* Complie regex expression */ - rc = regcomp(&preg, jcr->job->selection_pattern, REG_EXTENDED); - if (rc != 0) { - regerror(rc, &preg, prbuf, sizeof(prbuf)); - Jmsg(jcr, M_FATAL, 0, _("Could not compile regex pattern \"%s\" ERR=%s\n"), - jcr->job->selection_pattern, prbuf); + } + break; + case MT_CLIENT: + if (!regex_find_jobids(jcr, &ids, sql_client, sql_jobids_from_client, "Client")) { goto bail_out; - } - item_chain = New(dlist(item, &item->link)); - /* Basic query for Job names */ - Mmsg(query, sql_job, jcr->pool->hdr.name); - Dmsg1(000, "query=%s\n", query.c_str()); - if (!db_sql_query(jcr->db, query.c_str(), unique_name_handler, - (void *)item_chain)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Job failed. ERR=%s\n"), db_strerror(jcr->db)); + } + break; + case MT_VOLUME: + if (!regex_find_jobids(jcr, &ids, sql_vol, sql_jobids_from_vol, "Volume")) { goto bail_out; - } - /* Now apply the regex to the job names and remove any item not matched */ - foreach_dlist(item, item_chain) { - const int nmatch = 30; - regmatch_t pmatch[nmatch]; - if (last_item) { - free(last_item->item); - Dmsg1(000, "Remove item %s\n", last_item->item); - item_chain->remove(last_item); - } - Dmsg1(000, "Jobitem=%s\n", item->item); - rc = regexec(&preg, item->item, nmatch, pmatch, 0); - if (rc == 0) { - last_item = NULL; /* keep this one */ - } else { - last_item = item; - } - } - if (last_item) { - free(last_item->item); - Dmsg1(000, "Remove item %s\n", last_item->item); - item_chain->remove(last_item); - } - regfree(&preg); - /* - * At this point, we have a list of items in item_chain - * that have been matched by the regex, so now we need - * to look up their jobids. - */ - JobIds = get_pool_memory(PM_MESSAGE); - JobIds[0] = 0; - foreach_dlist(item, item_chain) { - Dmsg1(000, "Got Job: %s\n", item->item); - Mmsg(query, sql_jobids_from_job, item->item, jcr->pool->hdr.name); - if (!db_sql_query(jcr->db, query.c_str(), jobid_handler, (void *)JobIds)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); - goto bail_out; - } - } - if (JobIds[0] == 0) { - Jmsg(jcr, M_INFO, 0, _("No jobs found to migrate.\n")); - goto ok_out; - } - Dmsg1(000, "Job Jobids=%s\n", JobIds); - free_pool_memory(JobIds); - delete item_chain; + } break; - case MT_SMALLEST_VOL: - Mmsg(query, sql_smallest_vol, jcr->pool->hdr.name); - JobIds = get_pool_memory(PM_MESSAGE); - JobIds[0] = 0; - if (!db_sql_query(jcr->db, query.c_str(), jobid_handler, (void *)JobIds)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Volume failed. ERR=%s\n"), db_strerror(jcr->db)); + case MT_SQLQUERY: + if (!jcr->job->selection_pattern) { + Jmsg(jcr, M_FATAL, 0, _("No %s SQL selection pattern specified.\n"), jcr->get_OperationName()); goto bail_out; } - if (JobIds[0] == 0) { - Jmsg(jcr, M_INFO, 0, _("No Volumes found to migrate.\n")); - goto ok_out; - } - /* ***FIXME*** must loop over JobIds */ - Mmsg(query, sql_jobids_from_mediaid, JobIds); - JobIds[0] = 0; - if (!db_sql_query(jcr->db, query.c_str(), jobid_handler, (void *)JobIds)) { + Dmsg1(dbglevel, "SQL=%s\n", jcr->job->selection_pattern); + if (!db_sql_query(jcr->db, jcr->job->selection_pattern, + unique_dbid_handler, (void *)&ids)) { Jmsg(jcr, M_FATAL, 0, - _("SQL to get Volume failed. ERR=%s\n"), db_strerror(jcr->db)); + _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); goto bail_out; } - Dmsg1(000, "Smallest Vol Jobids=%s\n", JobIds); break; - case MT_OLDEST_VOL: - Mmsg(query, sql_oldest_vol, jcr->pool->hdr.name); - JobIds = get_pool_memory(PM_MESSAGE); - JobIds[0] = 0; - if (!db_sql_query(jcr->db, query.c_str(), jobid_handler, (void *)JobIds)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Volume failed. ERR=%s\n"), db_strerror(jcr->db)); + case MT_SMALLEST_VOL: + if (!find_mediaid_then_jobids(jcr, &ids, sql_smallest_vol, "Smallest Volume")) { goto bail_out; } - if (JobIds[0] == 0) { - Jmsg(jcr, M_INFO, 0, _("No Volume found to migrate.\n")); - goto ok_out; - } - Mmsg(query, sql_jobids_from_mediaid, JobIds); - JobIds[0] = 0; - if (!db_sql_query(jcr->db, query.c_str(), jobid_handler, (void *)JobIds)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Volume failed. ERR=%s\n"), db_strerror(jcr->db)); + break; + case MT_OLDEST_VOL: + if (!find_mediaid_then_jobids(jcr, &ids, sql_oldest_vol, "Oldest Volume")) { goto bail_out; } - Dmsg1(000, "Oldest Vol Jobids=%s\n", JobIds); break; case MT_POOL_OCCUPANCY: - Mmsg(query, sql_pool_bytes, jcr->pool->hdr.name); - JobIds = get_pool_memory(PM_MESSAGE); - JobIds[0] = 0; - if (!db_sql_query(jcr->db, query.c_str(), jobid_handler, (void *)JobIds)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Volume failed. ERR=%s\n"), db_strerror(jcr->db)); + ctx.count = 0; + /* Find count of bytes in pool */ + Mmsg(query, sql_pool_bytes, jcr->rpool->name()); + if (!db_sql_query(jcr->db, query.c_str(), db_int64_handler, (void *)&ctx)) { + Jmsg(jcr, M_FATAL, 0, _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); goto bail_out; } - if (JobIds[0] == 0) { - Jmsg(jcr, M_INFO, 0, _("No jobs found to migrate.\n")); + if (ctx.count == 0) { + Jmsg(jcr, M_INFO, 0, _("No Volumes found to %s.\n"), jcr->get_ActionName(0)); goto ok_out; } - Dmsg1(000, "Pool Occupancy Jobids=%s\n", JobIds); - break; - case MT_POOL_TIME: - Dmsg0(000, "Pool time not implemented\n"); - break; - case MT_CLIENT: - if (!jcr->job->selection_pattern) { - Jmsg(jcr, M_FATAL, 0, _("No Migration Client selection pattern specified.\n")); - goto bail_out; - } - Dmsg1(000, "Client regex=%s\n", jcr->job->selection_pattern); - rc = regcomp(&preg, jcr->job->selection_pattern, REG_EXTENDED); - if (rc != 0) { - regerror(rc, &preg, prbuf, sizeof(prbuf)); - Jmsg(jcr, M_FATAL, 0, _("Could not compile regex pattern \"%s\" ERR=%s\n"), - jcr->job->selection_pattern, prbuf); + pool_bytes = ctx.value; + Dmsg2(dbglevel, "highbytes=%lld pool=%lld\n", jcr->rpool->MigrationHighBytes, + pool_bytes); + if (pool_bytes < (int64_t)jcr->rpool->MigrationHighBytes) { + Jmsg(jcr, M_INFO, 0, _("No Volumes found to %s.\n"), jcr->get_ActionName(0)); + goto ok_out; } - item_chain = New(dlist(item, &item->link)); - Mmsg(query, sql_client, jcr->pool->hdr.name); - Dmsg1(100, "query=%s\n", query.c_str()); - if (!db_sql_query(jcr->db, query.c_str(), unique_name_handler, - (void *)item_chain)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Client failed. ERR=%s\n"), db_strerror(jcr->db)); + Dmsg0(dbglevel, "We should do Occupation migration.\n"); + + ids.count = 0; + /* Find a list of MediaIds that could be migrated */ + Mmsg(query, sql_mediaids, jcr->rpool->name()); + Dmsg1(dbglevel, "query=%s\n", query.c_str()); + if (!db_sql_query(jcr->db, query.c_str(), unique_dbid_handler, (void *)&ids)) { + Jmsg(jcr, M_FATAL, 0, _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); goto bail_out; } - /* Now apply the regex and create the jobs */ - foreach_dlist(item, item_chain) { - const int nmatch = 30; - regmatch_t pmatch[nmatch]; - rc = regexec(&preg, item->item, nmatch, pmatch, 0); - if (rc == 0) { - Dmsg1(000, "Do Client=%s\n", item->item); - } - free(item->item); - } - regfree(&preg); - delete item_chain; - break; - case MT_VOLUME: - if (!jcr->job->selection_pattern) { - Jmsg(jcr, M_FATAL, 0, _("No Migration Volume selection pattern specified.\n")); - goto bail_out; - } - Dmsg1(000, "Volume regex=%s\n", jcr->job->selection_pattern); - rc = regcomp(&preg, jcr->job->selection_pattern, REG_EXTENDED); - if (rc != 0) { - regerror(rc, &preg, prbuf, sizeof(prbuf)); - Jmsg(jcr, M_FATAL, 0, _("Could not compile regex pattern \"%s\" ERR=%s\n"), - jcr->job->selection_pattern, prbuf); + if (ids.count == 0) { + Jmsg(jcr, M_INFO, 0, _("No Volumes found to %s.\n"), jcr->get_ActionName(0)); + goto ok_out; } - item_chain = New(dlist(item, &item->link)); - Mmsg(query, sql_vol, jcr->pool->hdr.name); - Dmsg1(100, "query=%s\n", query.c_str()); - if (!db_sql_query(jcr->db, query.c_str(), unique_name_handler, - (void *)item_chain)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Job failed. ERR=%s\n"), db_strerror(jcr->db)); + Dmsg2(dbglevel, "Pool Occupancy ids=%d MediaIds=%s\n", ids.count, ids.list); + + if (!find_jobids_from_mediaid_list(jcr, &ids, "Volume")) { goto bail_out; } - /* Now apply the regex and create the jobs */ - foreach_dlist(item, item_chain) { - const int nmatch = 30; - regmatch_t pmatch[nmatch]; - rc = regexec(&preg, item->item, nmatch, pmatch, 0); - if (rc == 0) { - Dmsg1(000, "Do Vol=%s\n", item->item); + /* ids == list of jobs */ + p = ids.list; + for (int i=0; i < (int)ids.count; i++) { + stat = get_next_dbid_from_list(&p, &DBId); + Dmsg2(dbglevel, "get_next_dbid stat=%d JobId=%u\n", stat, (uint32_t)DBId); + if (stat < 0) { + Jmsg(jcr, M_FATAL, 0, _("Invalid JobId found.\n")); + goto bail_out; + } else if (stat == 0) { + break; + } + + mid.count = 1; + Mmsg(mid.list, "%s", edit_int64(DBId, ed1)); + if (jids.count > 0) { + pm_strcat(jids.list, ","); + } + pm_strcat(jids.list, mid.list); + jids.count += mid.count; + + /* Find count of bytes from Jobs */ + Mmsg(query, sql_job_bytes, mid.list); + Dmsg1(dbglevel, "Jobbytes query: %s\n", query.c_str()); + if (!db_sql_query(jcr->db, query.c_str(), db_int64_handler, (void *)&ctx)) { + Jmsg(jcr, M_FATAL, 0, _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); + goto bail_out; + } + pool_bytes -= ctx.value; + Dmsg2(dbglevel, "Total %s Job bytes=%s\n", jcr->get_ActionName(0), edit_int64_with_commas(ctx.value, ed1)); + Dmsg2(dbglevel, "lowbytes=%s poolafter=%s\n", + edit_int64_with_commas(jcr->rpool->MigrationLowBytes, ed1), + edit_int64_with_commas(pool_bytes, ed2)); + if (pool_bytes <= (int64_t)jcr->rpool->MigrationLowBytes) { + Dmsg0(dbglevel, "We should be done.\n"); + break; } - free(item->item); } - regfree(&preg); - delete item_chain; + /* Transfer jids to ids, where the jobs list is expected */ + ids.count = jids.count; + pm_strcpy(ids.list, jids.list); + Dmsg2(dbglevel, "Pool Occupancy ids=%d JobIds=%s\n", ids.count, ids.list); break; - case MT_SQLQUERY: - JobIds[0] = 0; - if (!jcr->job->selection_pattern) { - Jmsg(jcr, M_FATAL, 0, _("No Migration SQL selection pattern specified.\n")); - goto bail_out; - } - Dmsg1(000, "SQL=%s\n", jcr->job->selection_pattern); - if (!db_sql_query(jcr->db, query.c_str(), jobid_handler, (void *)JobIds)) { - Jmsg(jcr, M_FATAL, 0, - _("SQL to get Volume failed. ERR=%s\n"), db_strerror(jcr->db)); + case MT_POOL_TIME: + ttime = time(NULL) - (time_t)jcr->rpool->MigrationTime; + (void)localtime_r(&ttime, &tm); + strftime(dt, sizeof(dt), "%Y-%m-%d %H:%M:%S", &tm); + + ids.count = 0; + Mmsg(query, sql_pool_time, jcr->rpool->name(), dt); + Dmsg1(dbglevel, "query=%s\n", query.c_str()); + if (!db_sql_query(jcr->db, query.c_str(), unique_dbid_handler, (void *)&ids)) { + Jmsg(jcr, M_FATAL, 0, _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); goto bail_out; } - if (JobIds[0] == 0) { - Jmsg(jcr, M_INFO, 0, _("No jobs found to migrate.\n")); + if (ids.count == 0) { + Jmsg(jcr, M_INFO, 0, _("No Volumes found to %s.\n"), jcr->get_ActionName(0)); goto ok_out; } - Dmsg1(000, "Jobids=%s\n", JobIds); - goto bail_out; + Dmsg2(dbglevel, "PoolTime ids=%d JobIds=%s\n", ids.count, ids.list); + break; + case MT_POOL_UNCOPIED_JOBS: + if (!find_jobids_of_pool_uncopied_jobs(jcr, &ids)) { + goto bail_out; + } break; default: - Jmsg(jcr, M_FATAL, 0, _("Unknown Migration Selection Type.\n")); + Jmsg(jcr, M_FATAL, 0, _("Unknown %s Selection Type.\n"), jcr->get_OperationName()); goto bail_out; } } - p = JobIds; + /* + * Loop over all jobids except the last one, sending + * them to start_migration_job(), which will start a job + * for each of them. For the last JobId, we handle it below. + */ + p = ids.list; + if (ids.count == 0) { + Jmsg(jcr, M_INFO, 0, _("No JobIds found to %s.\n"), jcr->get_ActionName(0)); + goto ok_out; + } + + Jmsg(jcr, M_INFO, 0, _("The following %u JobId%s chosen to be %s: %s\n"), + ids.count, (ids.count < 2) ? _(" was") : _("s were"), + jcr->get_ActionName(1), ids.list); + + Dmsg2(dbglevel, "Before loop count=%d ids=%s\n", ids.count, ids.list); + for (int i=1; i < (int)ids.count; i++) { + JobId = 0; + stat = get_next_jobid_from_list(&p, &JobId); + Dmsg3(dbglevel, "getJobid_no=%d stat=%d JobId=%u\n", i, stat, JobId); + if (stat < 0) { + Jmsg(jcr, M_FATAL, 0, _("Invalid JobId found.\n")); + goto bail_out; + } else if (stat == 0) { + Jmsg(jcr, M_INFO, 0, _("No JobIds found to %s.\n"), jcr->get_ActionName(0)); + goto ok_out; + } + jcr->MigrateJobId = JobId; + start_migration_job(jcr); + Dmsg0(dbglevel, "Back from start_migration_job\n"); + } + + /* Now get the last JobId and handle it in the current job */ JobId = 0; stat = get_next_jobid_from_list(&p, &JobId); - Dmsg2(000, "get_next_jobid stat=%d JobId=%u\n", stat, JobId); + Dmsg2(dbglevel, "Last get_next_jobid stat=%d JobId=%u\n", stat, (int)JobId); if (stat < 0) { Jmsg(jcr, M_FATAL, 0, _("Invalid JobId found.\n")); goto bail_out; } else if (stat == 0) { - Jmsg(jcr, M_INFO, 0, _("No JobIds found to migrate.\n")); + Jmsg(jcr, M_INFO, 0, _("No JobIds found to %s.\n"), jcr->get_ActionName(0)); goto ok_out; } - + jcr->previous_jr.JobId = JobId; - Dmsg1(000, "Last jobid=%d\n", jcr->previous_jr.JobId); + Dmsg1(dbglevel, "Previous jobid=%d\n", (int)jcr->previous_jr.JobId); if (!db_get_job_record(jcr, jcr->db, &jcr->previous_jr)) { - Jmsg(jcr, M_FATAL, 0, _("Could not get job record for JobId %s to migrate. ERR=%s"), + Jmsg(jcr, M_FATAL, 0, _("Could not get job record for JobId %s to %s. ERR=%s"), edit_int64(jcr->previous_jr.JobId, ed1), + jcr->get_ActionName(0), db_strerror(jcr->db)); goto bail_out; } - Jmsg(jcr, M_INFO, 0, _("Migration using JobId=%d Job=%s\n"), - jcr->previous_jr.JobId, jcr->previous_jr.Job); + Jmsg(jcr, M_INFO, 0, _("%s using JobId=%s Job=%s\n"), + jcr->get_OperationName(), + edit_int64(jcr->previous_jr.JobId, ed1), jcr->previous_jr.Job); + Dmsg4(dbglevel, "%s JobId=%d using JobId=%s Job=%s\n", + jcr->get_OperationName(), + jcr->JobId, + edit_int64(jcr->previous_jr.JobId, ed1), jcr->previous_jr.Job); + count = 1; ok_out: - free_pool_memory(JobIds); - return true; + goto out; + +bail_out: + count = -1; + +out: + free_pool_memory(ids.list); + free_pool_memory(mid.list); + free_pool_memory(jids.list); + return count; +} + +static void start_migration_job(JCR *jcr) +{ + UAContext *ua = new_ua_context(jcr); + char ed1[50]; + ua->batch = true; + Mmsg(ua->cmd, "run job=\"%s\" jobid=%s", jcr->job->name(), + edit_uint64(jcr->MigrateJobId, ed1)); + Dmsg2(dbglevel, "=============== %s cmd=%s\n", jcr->get_OperationName(), ua->cmd); + parse_ua_args(ua); /* parse command */ + JobId_t jobid = run_cmd(ua, ua->cmd); + if (jobid == 0) { + Jmsg(jcr, M_ERROR, 0, _("Could not start migration job.\n")); + } else { + Jmsg(jcr, M_INFO, 0, _("%s JobId %d started.\n"), jcr->get_OperationName(), (int)jobid); + } + free_ua_context(ua); +} + +static bool find_mediaid_then_jobids(JCR *jcr, idpkt *ids, const char *query1, + const char *type) +{ + bool ok = false; + POOL_MEM query(PM_MESSAGE); + + ids->count = 0; + /* Basic query for MediaId */ + Mmsg(query, query1, jcr->rpool->name()); + if (!db_sql_query(jcr->db, query.c_str(), unique_dbid_handler, (void *)ids)) { + Jmsg(jcr, M_FATAL, 0, _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); + goto bail_out; + } + if (ids->count == 0) { + Jmsg(jcr, M_INFO, 0, _("No %s found to %s.\n"), type, jcr->get_ActionName(0)); + ok = true; /* Not an error */ + goto bail_out; + } else if (ids->count != 1) { + Jmsg(jcr, M_FATAL, 0, _("SQL error. Expected 1 MediaId got %d\n"), ids->count); + goto bail_out; + } + Dmsg2(dbglevel, "%s MediaIds=%s\n", type, ids->list); + + ok = find_jobids_from_mediaid_list(jcr, ids, type); + +bail_out: + return ok; +} + +/* + * This routine returns: + * false if an error occurred + * true otherwise + * ids.count number of jobids found (may be zero) + */ +static bool find_jobids_from_mediaid_list(JCR *jcr, idpkt *ids, const char *type) +{ + bool ok = false; + POOL_MEM query(PM_MESSAGE); + + Mmsg(query, sql_jobids_from_mediaid, ids->list); + ids->count = 0; + if (!db_sql_query(jcr->db, query.c_str(), unique_dbid_handler, (void *)ids)) { + Jmsg(jcr, M_FATAL, 0, _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); + goto bail_out; + } + if (ids->count == 0) { + Jmsg(jcr, M_INFO, 0, _("No %ss found to %s.\n"), type, jcr->get_ActionName(0)); + } + ok = true; + +bail_out: + return ok; +} + +/* + * This routine returns: + * false if an error occurred + * true otherwise + * ids.count number of jobids found (may be zero) + */ +static bool find_jobids_of_pool_uncopied_jobs(JCR *jcr, idpkt *ids) +{ + bool ok = false; + POOL_MEM query(PM_MESSAGE); + + /* Only a copy job is allowed */ + if (jcr->getJobType() != JT_COPY) { + Jmsg(jcr, M_FATAL, 0, + _("Selection Type 'pooluncopiedjobs' only applies to Copy Jobs")); + goto bail_out; + } + + Dmsg1(dbglevel, "copy selection pattern=%s\n", jcr->rpool->name()); + Mmsg(query, sql_jobids_of_pool_uncopied_jobs, jcr->rpool->name()); + Dmsg1(dbglevel, "get uncopied jobs query=%s\n", query.c_str()); + if (!db_sql_query(jcr->db, query.c_str(), unique_dbid_handler, (void *)ids)) { + Jmsg(jcr, M_FATAL, 0, + _("SQL to get uncopied jobs failed. ERR=%s\n"), db_strerror(jcr->db)); + goto bail_out; + } + ok = true; bail_out: - free_pool_memory(JobIds); - return false; + return ok; } +static bool regex_find_jobids(JCR *jcr, idpkt *ids, const char *query1, + const char *query2, const char *type) +{ + dlist *item_chain; + uitem *item = NULL; + uitem *last_item = NULL; + regex_t preg; + char prbuf[500]; + int rc; + bool ok = false; + POOL_MEM query(PM_MESSAGE); + + item_chain = New(dlist(item, &item->link)); + if (!jcr->job->selection_pattern) { + Jmsg(jcr, M_FATAL, 0, _("No %s %s selection pattern specified.\n"), + jcr->get_OperationName(), type); + goto bail_out; + } + Dmsg1(dbglevel, "regex-sel-pattern=%s\n", jcr->job->selection_pattern); + /* Basic query for names */ + Mmsg(query, query1, jcr->rpool->name()); + Dmsg1(dbglevel, "get name query1=%s\n", query.c_str()); + if (!db_sql_query(jcr->db, query.c_str(), unique_name_handler, + (void *)item_chain)) { + Jmsg(jcr, M_FATAL, 0, + _("SQL to get %s failed. ERR=%s\n"), type, db_strerror(jcr->db)); + goto bail_out; + } + Dmsg1(dbglevel, "query1 returned %d names\n", item_chain->size()); + if (item_chain->size() == 0) { + Jmsg(jcr, M_INFO, 0, _("Query of Pool \"%s\" returned no Jobs to %s.\n"), + jcr->rpool->name(), jcr->get_ActionName(0)); + ok = true; + goto bail_out; /* skip regex match */ + } else { + /* Compile regex expression */ + rc = regcomp(&preg, jcr->job->selection_pattern, REG_EXTENDED); + if (rc != 0) { + regerror(rc, &preg, prbuf, sizeof(prbuf)); + Jmsg(jcr, M_FATAL, 0, _("Could not compile regex pattern \"%s\" ERR=%s\n"), + jcr->job->selection_pattern, prbuf); + goto bail_out; + } + /* Now apply the regex to the names and remove any item not matched */ + foreach_dlist(item, item_chain) { + const int nmatch = 30; + regmatch_t pmatch[nmatch]; + if (last_item) { + Dmsg1(dbglevel, "Remove item %s\n", last_item->item); + free(last_item->item); + item_chain->remove(last_item); + } + Dmsg1(dbglevel, "get name Item=%s\n", item->item); + rc = regexec(&preg, item->item, nmatch, pmatch, 0); + if (rc == 0) { + last_item = NULL; /* keep this one */ + } else { + last_item = item; + } + } + if (last_item) { + free(last_item->item); + Dmsg1(dbglevel, "Remove item %s\n", last_item->item); + item_chain->remove(last_item); + } + regfree(&preg); + } + if (item_chain->size() == 0) { + Jmsg(jcr, M_INFO, 0, _("Regex pattern matched no Jobs to %s.\n"), jcr->get_ActionName(0)); + ok = true; + goto bail_out; /* skip regex match */ + } + + /* + * At this point, we have a list of items in item_chain + * that have been matched by the regex, so now we need + * to look up their jobids. + */ + ids->count = 0; + foreach_dlist(item, item_chain) { + Dmsg2(dbglevel, "Got %s: %s\n", type, item->item); + Mmsg(query, query2, item->item, jcr->rpool->name()); + Dmsg1(dbglevel, "get id from name query2=%s\n", query.c_str()); + if (!db_sql_query(jcr->db, query.c_str(), unique_dbid_handler, (void *)ids)) { + Jmsg(jcr, M_FATAL, 0, + _("SQL failed. ERR=%s\n"), db_strerror(jcr->db)); + goto bail_out; + } + } + if (ids->count == 0) { + Jmsg(jcr, M_INFO, 0, _("No %ss found to %s.\n"), type, jcr->get_ActionName(0)); + } + ok = true; + +bail_out: + Dmsg2(dbglevel, "Count=%d Jobids=%s\n", ids->count, ids->list); + foreach_dlist(item, item_chain) { + free(item->item); + } + delete item_chain; + return ok; +} /* * Release resources allocated during backup. @@ -686,73 +1123,122 @@ void migration_cleanup(JCR *jcr, int TermCode) { char sdt[MAX_TIME_LENGTH], edt[MAX_TIME_LENGTH]; char ec1[30], ec2[30], ec3[30], ec4[30], ec5[30], elapsed[50]; + char ec6[50], ec7[50], ec8[50]; char term_code[100], sd_term_msg[100]; const char *term_msg; - int msg_type; + int msg_type = M_INFO; MEDIA_DBR mr; double kbps; utime_t RunTime; - JCR *prev_jcr = jcr->previous_jcr; + JCR *mig_jcr = jcr->mig_jcr; POOL_MEM query(PM_MESSAGE); Dmsg2(100, "Enter migrate_cleanup %d %c\n", TermCode, TermCode); - dequeue_messages(jcr); /* display any queued messages */ + update_job_end(jcr, TermCode); memset(&mr, 0, sizeof(mr)); - set_jcr_job_status(jcr, TermCode); - update_job_end_record(jcr); /* update database */ - - /* Check if we actually did something */ - if (!prev_jcr) { - prev_jcr->JobFiles = jcr->JobFiles = jcr->SDJobFiles; - prev_jcr->JobBytes = jcr->JobBytes = jcr->SDJobBytes; - prev_jcr->VolSessionId = jcr->VolSessionId; - prev_jcr->VolSessionTime = jcr->VolSessionTime; - - set_jcr_job_status(prev_jcr, TermCode); - - - update_job_end_record(prev_jcr); + /* + * Check if we actually did something. + * mig_jcr is jcr of the newly migrated job. + */ + if (mig_jcr) { + char old_jobid[50], new_jobid[50]; + + edit_uint64(jcr->previous_jr.JobId, old_jobid); + edit_uint64(mig_jcr->jr.JobId, new_jobid); + + mig_jcr->JobFiles = jcr->JobFiles = jcr->SDJobFiles; + mig_jcr->JobBytes = jcr->JobBytes = jcr->SDJobBytes; + mig_jcr->VolSessionId = jcr->VolSessionId; + mig_jcr->VolSessionTime = jcr->VolSessionTime; + mig_jcr->jr.RealEndTime = 0; + mig_jcr->jr.PriorJobId = jcr->previous_jr.JobId; + + update_job_end(mig_jcr, TermCode); + + /* Update final items to set them to the previous job's values */ Mmsg(query, "UPDATE Job SET StartTime='%s',EndTime='%s'," "JobTDate=%s WHERE JobId=%s", jcr->previous_jr.cStartTime, jcr->previous_jr.cEndTime, edit_uint64(jcr->previous_jr.JobTDate, ec1), - edit_uint64(prev_jcr->jr.JobId, ec2)); - db_sql_query(prev_jcr->db, query.c_str(), NULL, NULL); + new_jobid); + db_sql_query(mig_jcr->db, query.c_str(), NULL, NULL); + + /* + * If we terminated a migration normally: + * - mark the previous job as migrated + * - move any Log records to the new JobId + * - Purge the File records from the previous job + */ + if (jcr->getJobType() == JT_MIGRATE && jcr->JobStatus == JS_Terminated) { + Mmsg(query, "UPDATE Job SET Type='%c' WHERE JobId=%s", + (char)JT_MIGRATED_JOB, old_jobid); + db_sql_query(mig_jcr->db, query.c_str(), NULL, NULL); + UAContext *ua = new_ua_context(jcr); + /* Move JobLog to new JobId */ + Mmsg(query, "UPDATE Log SET JobId=%s WHERE JobId=%s", + new_jobid, old_jobid); + db_sql_query(mig_jcr->db, query.c_str(), NULL, NULL); + /* Purge all old file records, but leave Job record */ + purge_files_from_jobs(ua, old_jobid); + free_ua_context(ua); + } + + /* + * If we terminated a Copy (rather than a Migration) normally: + * - copy any Log records to the new JobId + * - set type="Job Copy" for the new job + */ + if (jcr->getJobType() == JT_COPY && jcr->JobStatus == JS_Terminated) { + /* Copy JobLog to new JobId */ + Mmsg(query, "INSERT INTO Log (JobId, Time, LogText ) " + "SELECT %s, Time, LogText FROM Log WHERE JobId=%s", + new_jobid, old_jobid); + db_sql_query(mig_jcr->db, query.c_str(), NULL, NULL); + Mmsg(query, "UPDATE Job SET Type='%c' WHERE JobId=%s", + (char)JT_JOB_COPY, new_jobid); + db_sql_query(mig_jcr->db, query.c_str(), NULL, NULL); + } if (!db_get_job_record(jcr, jcr->db, &jcr->jr)) { - Jmsg(jcr, M_WARNING, 0, _("Error getting job record for stats: %s"), + Jmsg(jcr, M_WARNING, 0, _("Error getting Job record for Job report: ERR=%s"), db_strerror(jcr->db)); set_jcr_job_status(jcr, JS_ErrorTerminated); } - bstrncpy(mr.VolumeName, jcr->VolumeName, sizeof(mr.VolumeName)); - if (!db_get_media_record(jcr, jcr->db, &mr)) { - Jmsg(jcr, M_WARNING, 0, _("Error getting Media record for Volume \"%s\": ERR=%s"), - mr.VolumeName, db_strerror(jcr->db)); - set_jcr_job_status(jcr, JS_ErrorTerminated); - } - - update_bootstrap_file(prev_jcr); + update_bootstrap_file(mig_jcr); - if (!db_get_job_volume_names(prev_jcr, prev_jcr->db, prev_jcr->jr.JobId, &prev_jcr->VolumeName)) { + if (!db_get_job_volume_names(mig_jcr, mig_jcr->db, mig_jcr->jr.JobId, &mig_jcr->VolumeName)) { /* - * Note, if the job has erred, most likely it did not write any + * Note, if the job has failed, most likely it did not write any * tape, so suppress this "error" message since in that case * it is normal. Or look at it the other way, only for a * normal exit should we complain about this error. */ if (jcr->JobStatus == JS_Terminated && jcr->jr.JobBytes) { - Jmsg(jcr, M_ERROR, 0, "%s", db_strerror(prev_jcr->db)); + Jmsg(jcr, M_ERROR, 0, "%s", db_strerror(mig_jcr->db)); + } + mig_jcr->VolumeName[0] = 0; /* none */ + } + + if (mig_jcr->VolumeName[0]) { + /* Find last volume name. Multiple vols are separated by | */ + char *p = strrchr(mig_jcr->VolumeName, '|'); + if (p) { + p++; /* skip | */ + } else { + p = mig_jcr->VolumeName; /* no |, take full name */ + } + bstrncpy(mr.VolumeName, p, sizeof(mr.VolumeName)); + if (!db_get_media_record(jcr, jcr->db, &mr)) { + Jmsg(jcr, M_WARNING, 0, _("Error getting Media record for Volume \"%s\": ERR=%s"), + mr.VolumeName, db_strerror(jcr->db)); } - prev_jcr->VolumeName[0] = 0; /* none */ } - } - msg_type = M_INFO; /* by default INFO message */ - switch (jcr->JobStatus) { + switch (jcr->JobStatus) { case JS_Terminated: - if (jcr->Errors || jcr->SDErrors) { + if (jcr->JobErrors || jcr->SDErrors) { term_msg = _("%s OK -- with warnings"); } else { term_msg = _("%s OK"); @@ -781,31 +1267,44 @@ void migration_cleanup(JCR *jcr, int TermCode) default: term_msg = _("Inappropriate %s term code"); break; + } + } else { + if (jcr->getJobType() == JT_MIGRATE && jcr->previous_jr.JobId != 0) { + /* Mark previous job as migrated */ + Mmsg(query, "UPDATE Job SET Type='%c' WHERE JobId=%s", + (char)JT_MIGRATED_JOB, edit_uint64(jcr->previous_jr.JobId, ec1)); + db_sql_query(jcr->db, query.c_str(), NULL, NULL); + } + term_msg = _("%s -- no files to %s"); } - bsnprintf(term_code, sizeof(term_code), term_msg, "Migration"); + + bsnprintf(term_code, sizeof(term_code), term_msg, jcr->get_OperationName(), jcr->get_ActionName(0)); bstrftimes(sdt, sizeof(sdt), jcr->jr.StartTime); bstrftimes(edt, sizeof(edt), jcr->jr.EndTime); RunTime = jcr->jr.EndTime - jcr->jr.StartTime; if (RunTime <= 0) { kbps = 0; } else { - kbps = (double)jcr->jr.JobBytes / (1000 * RunTime); + kbps = (double)jcr->SDJobBytes / (1000 * RunTime); } - jobstatus_to_ascii(jcr->SDJobStatus, sd_term_msg, sizeof(sd_term_msg)); -// bmicrosleep(15, 0); /* for debugging SIGHUP */ - - Jmsg(jcr, msg_type, 0, _("Bacula %s (%s): %s\n" -" Old Backup JobId: %u\n" -" New Backup JobId: %u\n" -" JobId: %u\n" -" Job: %s\n" + Jmsg(jcr, msg_type, 0, _("%s %s %s (%s):\n" +" Build OS: %s %s %s\n" +" Prev Backup JobId: %s\n" +" Prev Backup Job: %s\n" +" New Backup JobId: %s\n" +" Current JobId: %s\n" +" Current Job: %s\n" " Backup Level: %s%s\n" " Client: %s\n" " FileSet: \"%s\" %s\n" -" Pool: \"%s\"\n" +" Read Pool: \"%s\" (From %s)\n" +" Read Storage: \"%s\" (From %s)\n" +" Write Pool: \"%s\" (From %s)\n" +" Write Storage: \"%s\" (From %s)\n" +" Catalog: \"%s\" (From %s)\n" " Start time: %s\n" " End time: %s\n" " Elapsed time: %s\n" @@ -820,25 +1319,32 @@ void migration_cleanup(JCR *jcr, int TermCode) " SD Errors: %d\n" " SD termination status: %s\n" " Termination: %s\n\n"), - VERSION, - LSMDATE, - edt, - prev_jcr ? jcr->previous_jr.JobId : 0, - jcr->jr.JobId, + BACULA, my_name, VERSION, LSMDATE, + HOST_OS, DISTNAME, DISTVER, + edit_uint64(jcr->previous_jr.JobId, ec6), + jcr->previous_jr.Job, + mig_jcr ? edit_uint64(mig_jcr->jr.JobId, ec7) : "0", + edit_uint64(jcr->jr.JobId, ec8), jcr->jr.Job, - level_to_str(jcr->JobLevel), jcr->since, - jcr->client->hdr.name, - jcr->fileset->hdr.name, jcr->FSCreateTime, - jcr->pool->hdr.name, + level_to_str(jcr->getJobLevel()), jcr->since, + jcr->client->name(), + jcr->fileset->name(), jcr->FSCreateTime, + jcr->rpool->name(), jcr->rpool_source, + jcr->rstore?jcr->rstore->name():"*None*", + NPRT(jcr->rstore_source), + jcr->pool->name(), jcr->pool_source, + jcr->wstore?jcr->wstore->name():"*None*", + NPRT(jcr->wstore_source), + jcr->catalog->name(), jcr->catalog_source, sdt, edt, edit_utime(RunTime, elapsed, sizeof(elapsed)), jcr->JobPriority, edit_uint64_with_commas(jcr->SDJobFiles, ec1), edit_uint64_with_commas(jcr->SDJobBytes, ec2), - edit_uint64_with_suffix(jcr->jr.JobBytes, ec3), + edit_uint64_with_suffix(jcr->SDJobBytes, ec3), (float)kbps, - prev_jcr ? prev_jcr->VolumeName : "", + mig_jcr ? mig_jcr->VolumeName : "", jcr->VolSessionId, jcr->VolSessionTime, edit_uint64_with_commas(mr.VolBytes, ec4), @@ -847,8 +1353,66 @@ void migration_cleanup(JCR *jcr, int TermCode) sd_term_msg, term_code); - Dmsg1(100, "Leave migrate_cleanup() previous_jcr=0x%x\n", jcr->previous_jcr); - if (jcr->previous_jcr) { - free_jcr(jcr->previous_jcr); + Dmsg1(100, "migrate_cleanup() mig_jcr=0x%x\n", jcr->mig_jcr); + if (jcr->mig_jcr) { + free_jcr(jcr->mig_jcr); + jcr->mig_jcr = NULL; } + Dmsg0(100, "Leave migrate_cleanup()\n"); +} + +/* + * Return next DBId from comma separated list + * + * Returns: + * 1 if next DBId returned + * 0 if no more DBIds are in list + * -1 there is an error + */ +static int get_next_dbid_from_list(char **p, DBId_t *DBId) +{ + const int maxlen = 30; + char id[maxlen+1]; + char *q = *p; + + id[0] = 0; + for (int i=0; iNextPool; + + if (!wpool) { + Jmsg(jcr, M_FATAL, 0, _("No Next Pool specification found in Pool \"%s\".\n"), + pool->hdr.name); + return false; + } + + if (!wpool->storage || wpool->storage->size() == 0) { + Jmsg(jcr, M_FATAL, 0, _("No Storage specification found in Next Pool \"%s\".\n"), + wpool->name()); + return false; + } + + /* If pool storage specified, use it instead of job storage for backup */ + copy_wstorage(jcr, wpool->storage, _("Storage from Pool's NextPool resource")); + return true; }