fis-gtm/sr_unix/jnlpool_init.c

921 lines
44 KiB
C

/****************************************************************
* *
* Copyright 2001, 2012 Fidelity Information Services, Inc *
* *
* This source code contains the intellectual property *
* of its copyright holder(s), and is made available *
* under a license. If you do not know the terms of *
* the license, please stop and do not read further. *
* *
****************************************************************/
#include "mdef.h"
#include "gtm_ipc.h"
#include <errno.h>
#include "gtm_fcntl.h"
#include "gtm_unistd.h"
#include "gtm_inet.h"
#include "gtm_stdlib.h"
#include "gtm_string.h"
#include <sys/sem.h>
#include <sys/shm.h>
#include <stddef.h>
#include "gdsroot.h"
#include "gdsblk.h"
#include "gtm_facility.h"
#include "fileinfo.h"
#include "gdsbt.h"
#include "gdsfhead.h"
#include "filestruct.h"
#include "iosp.h"
#include "repl_msg.h"
#include "gtmsource.h"
#include "gtmrecv.h"
#include "gtm_logicals.h"
#include "mutex.h"
#include "jnl.h"
#include "repl_sem.h"
#include "gtmimagename.h"
#include "io.h"
#include "do_shmat.h"
#include "repl_instance.h"
#include "mu_gv_cur_reg_init.h"
#include "gtmmsg.h"
#include "gtm_sem.h"
#include "ipcrmid.h"
#include "ftok_sems.h"
#include "is_proc_alive.h"
#include "repl_shutdcode.h"
#include "send_msg.h"
#include "heartbeat_timer.h"
#include "lockconst.h"
GBLREF jnlpool_addrs jnlpool;
GBLREF recvpool_addrs recvpool;
GBLREF jnlpool_ctl_ptr_t jnlpool_ctl;
GBLREF sm_uc_ptr_t jnldata_base;
GBLREF uint4 process_id;
GBLREF gd_region *gv_cur_region;
GBLREF jnlpool_ctl_ptr_t temp_jnlpool_ctl;
GBLREF gtmsource_options_t gtmsource_options;
GBLREF boolean_t pool_init;
GBLREF uint4 process_id;
GBLREF seq_num seq_num_zero;
GBLREF enum gtmImageTypes image_type;
GBLREF node_local_ptr_t locknl;
GBLREF uint4 log_interval;
GBLREF boolean_t is_updproc;
GBLREF uint4 mutex_per_process_init_pid;
GBLREF repl_conn_info_t *this_side, *remote_side;
GBLREF int4 strm_index;
LITREF char gtm_release_name[];
LITREF int4 gtm_release_name_len;
error_def(ERR_ACTIVATEFAIL);
error_def(ERR_JNLPOOLBADSLOT);
error_def(ERR_JNLPOOLSETUP);
error_def(ERR_NOJNLPOOL);
error_def(ERR_PRIMARYISROOT);
error_def(ERR_PRIMARYNOTROOT);
error_def(ERR_REPLINSTNMSAME);
error_def(ERR_REPLINSTNOHIST);
error_def(ERR_REPLINSTSECNONE);
error_def(ERR_REPLINSTSEQORD);
error_def(ERR_REPLREQROLLBACK);
error_def(ERR_REPLREQRUNDOWN);
error_def(ERR_REPLWARN);
error_def(ERR_SRCSRVEXISTS);
error_def(ERR_SRCSRVNOTEXIST);
error_def(ERR_SRCSRVTOOMANY);
error_def(ERR_TEXT);
#define REMOVE_OR_RELEASE_SEM(NEW_IPC, UDI) \
{ \
if (NEW_IPC) \
remove_sem_set(SOURCE); \
else \
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM); \
}
#define DETACH_AND_REMOVE_SHM_AND_SEM \
{ \
if (new_ipc) \
{ \
assert(!IS_GTM_IMAGE); /* Since "gtm_putmsg" is done below ensure it is never GT.M */ \
if (NULL != jnlpool.jnlpool_ctl) \
{ \
if (-1 == shmdt((caddr_t)jnlpool.jnlpool_ctl)) \
gtm_putmsg(VARLSTCNT(5) ERR_REPLWARN, 2, \
RTS_ERROR_LITERAL("Could not detach from journal pool"), errno); \
jnlpool_ctl = NULL; \
jnlpool.jnlpool_ctl = NULL; \
jnlpool.repl_inst_filehdr = NULL; \
jnlpool.gtmsrc_lcl_array = NULL; \
jnlpool.gtmsource_local_array = NULL; \
jnlpool.jnldata_base = NULL; \
pool_init = FALSE; \
} \
assert(INVALID_SHMID != udi->shmid); \
if (0 != shm_rmid(udi->shmid)) \
gtm_putmsg(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2, \
RTS_ERROR_LITERAL("Error removing jnlpool "), errno); \
remove_sem_set(SOURCE); \
} \
}
#define CHECK_SLOT(gtmsourcelocal_ptr) \
{ \
if ((GTMSOURCE_DUMMY_STATE != gtmsourcelocal_ptr->gtmsource_state) || (0 != gtmsourcelocal_ptr->gtmsource_pid)) \
{ /* Slot is in an out-of-design situation. Send an operator log message with enough debugging detail */ \
send_msg(VARLSTCNT(7) ERR_JNLPOOLBADSLOT, 5, LEN_AND_STR((char *)gtmsourcelocal_ptr->secondary_instname), \
gtmsourcelocal_ptr->gtmsource_pid, gtmsourcelocal_ptr->gtmsource_state, \
gtmsourcelocal_ptr->gtmsrc_lcl_array_index); \
} \
}
void jnlpool_init(jnlpool_user pool_user, boolean_t gtmsource_startup, boolean_t *jnlpool_creator)
{
boolean_t new_ipc, is_src_srvr, slot_needs_init, reset_gtmsrclcl_info, hold_onto_ftok_sem, srv_alive;
char machine_name[MAX_MCNAMELEN], instfilename[MAX_FN_LEN + 1];
gd_region *r_save, *reg;
int status, save_errno;
int4 index;
union semun semarg;
struct semid_ds semstat;
struct shmid_ds shmstat;
repl_inst_hdr repl_instance;
sm_long_t status_l;
unsigned int full_len;
mutex_spin_parms_ptr_t jnlpool_mutex_spin_parms;
unix_db_info *udi;
sgmnt_addrs *csa;
gd_segment *seg;
gtmsrc_lcl_ptr_t gtmsrclcl_ptr;
gtmsource_local_ptr_t gtmsourcelocal_ptr, reuse_slot_ptr;
uint4 gtmsource_pid, gtmrecv_pid;
gtmsource_state_t gtmsource_state;
seq_num reuse_slot_seqnum, instfilehdr_seqno;
repl_histinfo last_histinfo;
jnlpool_ctl_ptr_t tmp_jnlpool_ctl;
DEBUG_ONLY(int4 semval;)
assert(gtmsource_startup == gtmsource_options.start);
memset(machine_name, 0, SIZEOF(machine_name));
if (GETHOSTNAME(machine_name, MAX_MCNAMELEN, status))
rts_error(VARLSTCNT(5) ERR_TEXT, 2, RTS_ERROR_TEXT("Unable to get the hostname"), errno);
if (NULL == recvpool.recvpool_dummy_reg)
{
r_save = gv_cur_region;
mu_gv_cur_reg_init();
jnlpool.jnlpool_dummy_reg = reg = gv_cur_region;
gv_cur_region = r_save;
ASSERT_IN_RANGE(MIN_RN_LEN, SIZEOF(JNLPOOL_DUMMY_REG_NAME) - 1, MAX_RN_LEN);
MEMCPY_LIT(reg->rname, JNLPOOL_DUMMY_REG_NAME);
reg->rname_len = STR_LIT_LEN(JNLPOOL_DUMMY_REG_NAME);
reg->rname[reg->rname_len] = 0;
} else
{ /* Have already attached to the receive pool. Use the receive pool region for the journal pool as well as they
* both correspond to one and the same file (replication instance file). We need to do this to ensure that an
* "ftok_sem_get" done with either "jnlpool.jnlpool_dummy_reg" region or "recvpool.recvpool_dummy_reg" region
* locks the same entity.
*/
assert(is_updproc); /* Should have already attached to receive pool only in case of update process */
reg = jnlpool.jnlpool_dummy_reg = recvpool.recvpool_dummy_reg;
}
udi = FILE_INFO(reg);
csa = &udi->s_addrs;
seg = reg->dyn.addr;
assert(!udi->s_addrs.hold_onto_crit); /* so that we can do unconditional grab_locks and rel_locks */
assertpro(repl_inst_get_name(instfilename, &full_len, MAX_FN_LEN + 1, issue_rts_error)); /* rts_error should have been
* issued by repl_inst_get_name */
assert((recvpool.recvpool_dummy_reg != jnlpool.jnlpool_dummy_reg)
|| (seg->fname_len == full_len) && !STRCMP(seg->fname, instfilename));
if (recvpool.recvpool_dummy_reg != jnlpool.jnlpool_dummy_reg)
{ /* Fill in fields only if this is the first time this process is opening the replication instance file */
memcpy((char *)seg->fname, instfilename, full_len);
udi->fn = (char *)seg->fname;
seg->fname_len = full_len;
seg->fname[full_len] = '\0';
}
/* First grab ftok semaphore for replication instance file. Once we have it locked, no one else can start up
* or shut down replication for this instance. We will release ftok semaphore when initialization is done.
*/
if (!ftok_sem_get(jnlpool.jnlpool_dummy_reg, TRUE, REPLPOOL_ID, FALSE))
rts_error(VARLSTCNT(1) ERR_JNLPOOLSETUP);
repl_inst_read(udi->fn, (off_t)0, (sm_uc_ptr_t)&repl_instance, SIZEOF(repl_inst_hdr));
is_src_srvr = (GTMSOURCE == pool_user);
/* If caller is source server and secondary instance name has been specified check if it is different from THIS instance */
if (is_src_srvr && gtmsource_options.instsecondary)
{
if (0 == STRCMP(repl_instance.inst_info.this_instname, gtmsource_options.secondary_instname))
{
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(4) ERR_REPLINSTNMSAME, 2, LEN_AND_STR((char *)repl_instance.inst_info.this_instname));
}
}
new_ipc = FALSE;
if (INVALID_SEMID == repl_instance.jnlpool_semid)
{ /* First process to do "jnlpool_init". Create the journal pool. */
assertpro(INVALID_SHMID == repl_instance.jnlpool_shmid);
/* Source server startup is the only command that can create the journal pool. Check that. */
if (!is_src_srvr || !gtmsource_options.start)
{
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(4) ERR_NOJNLPOOL, 2, full_len, udi->fn);
}
if (repl_instance.crash)
{
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(4) ERR_REPLREQROLLBACK, 2, full_len, udi->fn);
}
new_ipc = TRUE;
assert(NUM_SRC_SEMS == NUM_RECV_SEMS);
if (INVALID_SEMID == (udi->semid = init_sem_set_source(IPC_PRIVATE, NUM_SRC_SEMS, RWDALL | IPC_CREAT)))
{
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error creating journal pool semaphore"), REPL_SEM_ERRNO);
}
/* Following will set semaphore SOURCE_ID_SEM value as GTM_ID. In case we have orphaned semaphore
* for some reason, mupip rundown will be able to identify GTM semaphores checking the value and can remove.
*/
semarg.val = GTM_ID;
if (-1 == semctl(udi->semid, SOURCE_ID_SEM, SETVAL, semarg))
{
save_errno = errno;
remove_sem_set(SOURCE); /* Remove what we created */
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error with jnlpool semctl SETVAL"), save_errno);
}
/* Warning: We must read the sem_ctime using IPC_STAT after SETVAL, which changes it. We must NOT do any
* more SETVAL after this. Our design is to use sem_ctime as creation time of semaphore.
*/
semarg.buf = &semstat;
if (-1 == semctl(udi->semid, SOURCE_ID_SEM, IPC_STAT, semarg))
{
save_errno = errno;
remove_sem_set(SOURCE); /* Remove what we created */
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error with jnlpool semctl IPC_STAT"), save_errno);
}
udi->gt_sem_ctime = semarg.buf->sem_ctime;
} else
{ /* find create time of semaphore from the file header and check if the id is reused by others */
assert(repl_instance.crash);
semarg.buf = &semstat;
if (-1 == semctl(repl_instance.jnlpool_semid, 0, IPC_STAT, semarg))
{
save_errno = errno;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(5) ERR_REPLREQROLLBACK, 2, full_len, udi->fn, save_errno);
} else if (semarg.buf->sem_ctime != repl_instance.jnlpool_semid_ctime)
{
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(8) ERR_REPLREQROLLBACK, 2, full_len, udi->fn,
ERR_TEXT, 2, RTS_ERROR_TEXT("jnlpool sem_ctime does not match"));
}
udi->semid = repl_instance.jnlpool_semid;
udi->gt_sem_ctime = repl_instance.jnlpool_semid_ctime;
set_sem_set_src(udi->semid); /* repl_sem.c has some functions which needs some static variable to have the id */
}
assert((INVALID_SEMID != udi->semid) && (0 != udi->gt_sem_ctime));
assert(!udi->grabbed_access_sem);
status = grab_sem(SOURCE, JNL_POOL_ACCESS_SEM);
if (SS_NORMAL != status)
{
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error with journal pool access semaphore"), REPL_SEM_ERRNO);
}
udi->grabbed_access_sem = TRUE;
if (INVALID_SHMID == repl_instance.jnlpool_shmid)
{ /* We have an INVALID shmid in the file header. There are three ways this can happen
*
* 1. A rollback started off with either no journal pool or rundown'd an existing journal pool and created new
* semaphores, but got killed in the middle. At this point, if a new source server starts up, it will notice
* a valid usable semid, but will find an invalid shmid.
*
* 2. A rollback started off with either no journal pool or rundown'd an existing journal pool and created new
* semaphores. Before it goes to mur_close_files, lets say a source server started. It will acquire the ftok
* semaphore, but will be waiting for the access control semaphore which rollback is holding. Rollback, on the
* other hand, will see if the ftok semaphore is available BEFORE removing the semaphores from the system. But,
* since source server is holding the ftok, Rollback, will not remove the access control semaphore. But, would
* just let go of them and exit (repl_instance.file_corrupt can be either TRUE or FALSE depending on whether
* Rollback completed successfully or not).
*
* 3. A fresh startup.
*/
/* Ensure that NO one has yet incremented the SRC_SERV_COUNT_SEM (as implied by all the 3 cases above) */
assert(0 == (semval = semctl(udi->semid, SRC_SERV_COUNT_SEM, GETVAL))); /* semval = number of processes attached */
new_ipc = TRUE; /* need to create new IPC */
} else if (-1 == shmctl(repl_instance.jnlpool_shmid, IPC_STAT, &shmstat))
{ /* shared memory ID was removed form the system by an IPCRM command or we have a permission issue (or such) */
save_errno = errno;
REMOVE_OR_RELEASE_SEM(new_ipc, udi);
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(5) ERR_REPLREQROLLBACK, 2, full_len, udi->fn, save_errno);
} else if (shmstat.shm_ctime != repl_instance.jnlpool_shmid_ctime)
{ /* shared memory was possibly reused (causing shm_ctime and jnlpool_shmid_ctime to be different. We can't rely
* on the shmid as it could be connected to a valid instance file in a different environment. Create new IPCs
*/
new_ipc = TRUE; /* need to create new IPC */
} else
{
udi->shmid = repl_instance.jnlpool_shmid;
udi->gt_shm_ctime = repl_instance.jnlpool_shmid_ctime;
}
/* Source server startup is the only command that can create the journal pool. Check that. */
if (new_ipc && (!is_src_srvr || !gtmsource_options.start))
{
REMOVE_OR_RELEASE_SEM(new_ipc, udi);
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(4) ERR_NOJNLPOOL, 2, full_len, udi->fn);
}
if (repl_instance.file_corrupt)
{ /* A prior rollback on this instance have been killed in the middle. Can't proceed until rollback is-rerun */
REMOVE_OR_RELEASE_SEM(new_ipc, udi);
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(8) ERR_REPLREQROLLBACK, 2, full_len, udi->fn,
ERR_TEXT, 2, LEN_AND_LIT("file_corrupt field in instance file header is set to TRUE"));
}
if (new_ipc)
{ /* create new shared memory */
if (-1 == (udi->shmid = shmget(IPC_PRIVATE, gtmsource_options.buffsize, RWDALL | IPC_CREAT)))
{
udi->shmid = INVALID_SHMID;
save_errno = errno;
remove_sem_set(SOURCE);
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error with journal pool creation"), save_errno);
}
if (-1 == shmctl(udi->shmid, IPC_STAT, &shmstat))
{
save_errno = errno;
DETACH_AND_REMOVE_SHM_AND_SEM; /* remove any sem/shm we had created */
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error with jnlpool shmctl IPC_STAT"), save_errno);
}
udi->gt_shm_ctime = shmstat.shm_ctime;
}
assert((INVALID_SHMID != udi->shmid) && (0 != udi->gt_shm_ctime));
status_l = (sm_long_t)(tmp_jnlpool_ctl = (jnlpool_ctl_ptr_t)do_shmat(udi->shmid, 0, 0));
if (-1 == status_l)
{
save_errno = errno;
DETACH_AND_REMOVE_SHM_AND_SEM; /* remove any sem/shm we had created */
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
/* Assert below ensures we dont try to clean up the journal pool even though we errored out while attaching to it */
assert(NULL == jnlpool.jnlpool_ctl);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error with journal pool shmat"), save_errno);
}
jnlpool.jnlpool_ctl = tmp_jnlpool_ctl;
/* Set a flag to indicate the journal pool is uninitialized. Do this as soon as attaching to shared memory.
* This flag will be reset by "gtmsource_seqno_init" when it is done with setting the jnl_seqno fields.
*/
if (new_ipc)
jnlpool.jnlpool_ctl->pool_initialized = FALSE;
assert(SIZEOF(jnlpool_ctl_struct) % 16 == 0); /* enforce 16-byte alignment for this structure */
/* Since seqno is an 8-byte quantity and is used in most of the sections below, we require all sections to
* be at least 8-byte aligned. In addition we expect that the beginning of the journal data (JNLDATA_BASE_OFF) is
* aligned at a boundary that is suitable for journal records (defined by JNL_WRT_END_MASK).
*/
assert(JNLPOOL_CTL_SIZE % 8 == 0);
assert(JNLPOOL_CRIT_SIZE % 8 == 0);
assert(SIZEOF(repl_inst_hdr) % 8 == 0);
assert(SIZEOF(gtmsrc_lcl) % 8 == 0);
assert(SIZEOF(gtmsource_local_struct) % 8 == 0);
assert(REPL_INST_HDR_SIZE % 8 == 0);
assert(GTMSRC_LCL_SIZE % 8 == 0);
assert(GTMSOURCE_LOCAL_SIZE % 8 == 0);
assert(JNLDATA_BASE_OFF % JNL_WRT_END_MODULUS == 0);
csa->critical = (mutex_struct_ptr_t)((sm_uc_ptr_t)jnlpool.jnlpool_ctl + JNLPOOL_CTL_SIZE); /* secshr_db_clnup uses this
* relationship */
jnlpool_mutex_spin_parms = (mutex_spin_parms_ptr_t)((sm_uc_ptr_t)csa->critical + CRIT_SPACE);
csa->nl = (node_local_ptr_t)((sm_uc_ptr_t)jnlpool_mutex_spin_parms + SIZEOF(mutex_spin_parms_struct));
if (new_ipc)
memset(csa->nl, 0, SIZEOF(node_local)); /* Make csa->nl->glob_sec_init FALSE */
csa->now_crit = FALSE;
csa->onln_rlbk_cycle = jnlpool.jnlpool_ctl->onln_rlbk_cycle; /* Take a copy of the latest onln_rlbk_cycle */
jnlpool.repl_inst_filehdr = (repl_inst_hdr_ptr_t)((sm_uc_ptr_t)csa->critical + JNLPOOL_CRIT_SIZE);
jnlpool.gtmsrc_lcl_array = (gtmsrc_lcl_ptr_t)((sm_uc_ptr_t)jnlpool.repl_inst_filehdr + REPL_INST_HDR_SIZE);
jnlpool.gtmsource_local_array = (gtmsource_local_ptr_t)((sm_uc_ptr_t)jnlpool.gtmsrc_lcl_array + GTMSRC_LCL_SIZE);
jnldata_base = jnlpool.jnldata_base = (sm_uc_ptr_t)jnlpool.jnlpool_ctl + JNLDATA_BASE_OFF;
jnlpool_ctl = jnlpool.jnlpool_ctl;
jnlpool_ctl->critical_off = (sm_uc_ptr_t)csa->critical - (sm_uc_ptr_t)jnlpool_ctl;
jnlpool_ctl->filehdr_off = (sm_uc_ptr_t)jnlpool.repl_inst_filehdr - (sm_uc_ptr_t)jnlpool_ctl;
jnlpool_ctl->srclcl_array_off = (sm_uc_ptr_t)jnlpool.gtmsrc_lcl_array - (sm_uc_ptr_t)jnlpool_ctl;
jnlpool_ctl->sourcelocal_array_off = (sm_uc_ptr_t)jnlpool.gtmsource_local_array - (sm_uc_ptr_t)jnlpool_ctl;
assert(!mutex_per_process_init_pid || mutex_per_process_init_pid == process_id);
if (!mutex_per_process_init_pid)
mutex_per_process_init();
START_HEARTBEAT_IF_NEEDED;
if (new_ipc)
{ /* Need to initialize the different sections of journal pool. Start with the FILE HEADER section */
repl_instance.jnlpool_semid = udi->semid;
repl_instance.jnlpool_shmid = udi->shmid;
repl_instance.jnlpool_semid_ctime = udi->gt_sem_ctime;
repl_instance.jnlpool_shmid_ctime = udi->gt_shm_ctime;
memcpy(jnlpool.repl_inst_filehdr, &repl_instance, REPL_INST_HDR_SIZE); /* Initialize FILE HEADER */
jnlpool.repl_inst_filehdr->crash = TRUE;
/* Since we are creating the journal pool, initialize the mutex structures in the shared memory for later
* grab_locks to work correctly
*/
DEBUG_ONLY(locknl = csa->nl;) /* for DEBUG_ONLY LOCK_HIST macro */
gtm_mutex_init(reg, NUM_CRIT_ENTRY, FALSE);
DEBUG_ONLY(locknl = NULL;) /* restore "locknl" to default value */
jnlpool_mutex_spin_parms->mutex_hard_spin_count = MUTEX_HARD_SPIN_COUNT;
jnlpool_mutex_spin_parms->mutex_sleep_spin_count = MUTEX_SLEEP_SPIN_COUNT;
jnlpool_mutex_spin_parms->mutex_spin_sleep_mask = MUTEX_SPIN_SLEEP_MASK;
GRAB_LOCK(jnlpool.jnlpool_dummy_reg, ASSERT_NO_ONLINE_ROLLBACK);
/* Flush the file header to disk so future callers of "jnlpool_init" see the jnlpool_semid and jnlpool_shmid */
repl_inst_flush_filehdr();
/* Initialize GTMSRC_LCL section in journal pool */
repl_inst_read(udi->fn, (off_t)REPL_INST_HDR_SIZE, (sm_uc_ptr_t)jnlpool.gtmsrc_lcl_array, GTMSRC_LCL_SIZE);
rel_lock(jnlpool.jnlpool_dummy_reg);
/* Initialize GTMSOURCE_LOCAL section in journal pool */
memset(jnlpool.gtmsource_local_array, 0, GTMSOURCE_LOCAL_SIZE);
gtmsourcelocal_ptr = &jnlpool.gtmsource_local_array[0];
gtmsrclcl_ptr = &jnlpool.gtmsrc_lcl_array[0];
for (index = 0; index < NUM_GTMSRC_LCL; index++, gtmsrclcl_ptr++, gtmsourcelocal_ptr++)
{
COPY_GTMSRCLCL_TO_GTMSOURCELOCAL(gtmsrclcl_ptr, gtmsourcelocal_ptr);
gtmsourcelocal_ptr->gtmsource_state = GTMSOURCE_DUMMY_STATE;
gtmsourcelocal_ptr->gtmsrc_lcl_array_index = index;
/* since we are setting up the journal pool for the first time, use this time to initialize the
* gtmsource_srv_latch as well
*/
SET_LATCH_GLOBAL(&gtmsourcelocal_ptr->gtmsource_srv_latch, LOCK_AVAILABLE);
}
} else if (!jnlpool.jnlpool_ctl->pool_initialized)
{ /* Source server that created the journal pool died before completing initialization. */
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(10) ERR_REPLREQRUNDOWN, 4, DB_LEN_STR(reg), LEN_AND_STR(machine_name),
ERR_TEXT, 2, RTS_ERROR_TEXT("Journal pool is incompletely initialized. Run MUPIP RUNDOWN first."));
}
slot_needs_init = FALSE;
/* Do not release ftok semaphore in the following cases as each of them involve the callers writing to the instance file
* which requires the ftok semaphore to be held. The callers will take care of releasing the semaphore.
* a) MUPIP REPLIC -SOURCE -START
* Invoke the function "gtmsource_rootprimary_init"
* b) MUPIP REPLIC -SOURCE -SHUTDOWN
* Invoke the function "gtmsource_flush_jnlpool" from the function "repl_ipc_cleanup"
* c) MUPIP REPLIC -SOURCE -ACTIVATE -ROOTPRIMARY (or -UPDOK) on a journal pool that has updates disabled.
* Invoke the function "gtmsource_rootprimary_init"
*/
hold_onto_ftok_sem = is_src_srvr && (gtmsource_options.start || gtmsource_options.shut_down);
/* Determine "gtmsourcelocal_ptr" to later initialize jnlpool.gtmsource_local */
if (!is_src_srvr || !gtmsource_options.instsecondary)
{ /* GT.M or Update process or receiver server or a source server command that did not specify INSTSECONDARY */
gtmsourcelocal_ptr = NULL;
} else
{ /* In jnlpool.gtmsource_local_array, find the structure which corresponds to the input secondary instance name.
* Each gtmsource_local structure in the array is termed a slot. A slot is used if the "secondary_instname"
* (the secondary instance name) member has a non-zero value. A slot is unused otherwise. Below is a tabulation
* of the possible cases and actions for each of the source server commands.
*
* -------------------------------------------------------------------------------------------------------
* Used Slot found Unused slot found No slot found
* -------------------------------------------------------------------------------------------------------
* activate Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* deactivate Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* showbacklog Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* checkhealth Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* statslog Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* changelog Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* stopsourcefilter Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* start AAAA BBBB CCCC
* shutdown Set gtmsource_local REPLINSTSECNONE REPLINSTSECNONE
* needrestart Set gtmsource_local DDDD DDDD
*
* AAAA : if the slot has a gtmsource_pid value that is still alive, then issue error message SRCSRVEXISTS
* Else initialize the slot for the new source server.
*
* BBBB : Set gtmsource_local to this slot. Initialize slot.
*
* CCCC : Slot reuse algorithm. Scan through all the slots again looking for those with a
* gtmsource_state field value 0 (GTMSOURCE_DUMMY_STATE) or those with a gtmsource_pid that does
* not exist. Note down the connect_jnl_seqno field of all these slots. Set gtmsource_local to that
* slot with the least value for connect_jnl_seqno. Initialize that slot. If no slot is found issue
* SRCSRVTOOMANY error
*
* DDDD : Set gtmsource_local to NULL. The caller gtmsource_needrestart.c will issue the appropriate
* message.
*
* Slot Initialization : Set "read_jnl_seqno" to 1.
*/
reuse_slot_ptr = NULL;
gtmsourcelocal_ptr = &jnlpool.gtmsource_local_array[0];
for ( index = 0; index < NUM_GTMSRC_LCL; index++, gtmsourcelocal_ptr++)
{
if ((NULL == reuse_slot_ptr) && ('\0' == gtmsourcelocal_ptr->secondary_instname[0]))
reuse_slot_ptr = gtmsourcelocal_ptr;
if (0 == STRCMP(gtmsource_options.secondary_instname, gtmsourcelocal_ptr->secondary_instname))
{ /* Found matching slot */
gtmsource_state = gtmsourcelocal_ptr->gtmsource_state;
gtmsource_pid = gtmsourcelocal_ptr->gtmsource_pid;
/* Check if source server is already running for this secondary instance */
if ((0 != gtmsource_pid) && is_proc_alive(gtmsource_pid, 0))
{ /* Source server is already up and running for this secondary instance */
if (gtmsource_options.start)
{
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
/* Assert we did not create shm or sem so no need to remove any */
assert(!new_ipc);
rts_error(VARLSTCNT(5) ERR_SRCSRVEXISTS, 3,
LEN_AND_STR(gtmsource_options.secondary_instname), gtmsource_pid);
}
} else
{ /* Source server is not running for this secondary instance */
if (gtmsource_options.start)
{ /* We want to reinitialize the source server related fields in "gtmsource_local"
* but do NOT want to reinitialize any fields that intersect with "gtmsrc_lcl"
*/
CHECK_SLOT(gtmsourcelocal_ptr);
slot_needs_init = TRUE;
reset_gtmsrclcl_info = FALSE;
} else if (!gtmsource_options.needrestart && !gtmsource_options.showbacklog
&& !gtmsource_options.checkhealth)
{ /* If NEEDRESTART, we dont care if the source server is alive or not. All that
* we care about is if the primary and secondary did communicate or not. That
* will be determined in gtmsource_needrestart.c. Do not trigger an error here.
* If SHOWBACKLOG or CHECKHEALTH, do not trigger an error as slot was found
* even though the source server is not alive. We can generate backlog/checkhealth
* information using values from the matched slot.
*/
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
/* Assert we did not create shm or sem so no need to remove any */
assert(!new_ipc);
rts_error(VARLSTCNT(4) ERR_SRCSRVNOTEXIST, 2,
LEN_AND_STR(gtmsource_options.secondary_instname));
}
}
break;
}
}
if (NUM_GTMSRC_LCL == index)
{ /* Did not find a matching slot. Use the unused slot if it was already found and if appropriate. */
if (gtmsource_options.needrestart)
{ /* If -NEEDRESTART is specified, set gtmsource_local to NULL. The caller function
* "gtmsource_needrestart" will print the appropriate message.
*/
gtmsourcelocal_ptr = NULL;
} else if (NULL == reuse_slot_ptr)
{ /* No used or unused slot found. Issue REPLINSTSECNONE error except for -start */
if (!gtmsource_options.start)
{
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
/* Assert we did not create shm or sem so no need to remove any */
assert(!new_ipc);
rts_error(VARLSTCNT(6) ERR_REPLINSTSECNONE, 4,
LEN_AND_STR(gtmsource_options.secondary_instname), full_len, udi->fn);
} else
{ /* Find a used slot that can be reused. Find one with least value of "connect_jnl_seqno". */
reuse_slot_seqnum = MAX_SEQNO;
gtmsourcelocal_ptr = &jnlpool.gtmsource_local_array[0];
for (index = 0; index < NUM_GTMSRC_LCL; index++, gtmsourcelocal_ptr++)
{
gtmsource_state = gtmsourcelocal_ptr->gtmsource_state;
gtmsource_pid = gtmsourcelocal_ptr->gtmsource_pid;
if ((0 == gtmsource_pid) || !is_proc_alive(gtmsource_pid, 0))
{ /* Slot can be reused */
if (gtmsourcelocal_ptr->connect_jnl_seqno < reuse_slot_seqnum)
{
reuse_slot_seqnum = gtmsourcelocal_ptr->connect_jnl_seqno;
reuse_slot_ptr = gtmsourcelocal_ptr;
}
}
}
if (NULL == reuse_slot_ptr)
{
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
/* Assert we did not create shm or sem so no need to remove any */
assert(!new_ipc);
rts_error(VARLSTCNT(5) ERR_SRCSRVTOOMANY, 3, NUM_GTMSRC_LCL, full_len, udi->fn);
} else
{ /* We want to reinitialize the source server related fields in "gtmsource_local"
* as well as reinitialize any fields that intersect with "gtmsrc_lcl" as this
* slot is being reused for a different secondary instance than is currently stored.
*/
gtmsourcelocal_ptr = reuse_slot_ptr;
CHECK_SLOT(gtmsourcelocal_ptr);
slot_needs_init = TRUE;
reset_gtmsrclcl_info = TRUE;
}
}
} else
{ /* No used slot found. But an unused slot was found. */
if (gtmsource_options.start)
{ /* Initialize the unused slot. We want to reinitialize the source server related fields
* in "gtmsource_local" as well as reinitialize any fields that intersect with "gtmsrc_lcl"
* as this slot is being reused for the first time for any secondary instance name.
*/
gtmsourcelocal_ptr = reuse_slot_ptr;
CHECK_SLOT(gtmsourcelocal_ptr);
slot_needs_init = TRUE;
reset_gtmsrclcl_info = TRUE;
} else
{ /* One of the following qualifiers has been specified. Issue error.
* ACTIVATE, CHANGELOG, CHECKHEALTH, DEACTIVATE, SHOWBACKLOG,
* STATSLOG, SHUTDOWN or STOPSOURCEFILTER
*/
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
/* Assert we did not create shm or sem so no need to remove any */
assert(!new_ipc);
rts_error(VARLSTCNT(6) ERR_REPLINSTSECNONE, 4,
LEN_AND_STR(gtmsource_options.secondary_instname), full_len, udi->fn);
}
}
}
}
assert((ROOTPRIMARY_UNSPECIFIED == gtmsource_options.rootprimary)
|| (PROPAGATEPRIMARY_SPECIFIED == gtmsource_options.rootprimary)
|| (ROOTPRIMARY_SPECIFIED == gtmsource_options.rootprimary));
if (!new_ipc)
{ /* We did not create shm or sem so no need to remove any of them for any "rts_error" within this IF */
assert(!STRCMP(repl_instance.inst_info.this_instname, jnlpool.repl_inst_filehdr->inst_info.this_instname));
/* Check compatibility of caller source server or receiver server command with the current state of journal pool */
if (!jnlpool.jnlpool_ctl->upd_disabled)
{
if (((is_src_srvr && (PROPAGATEPRIMARY_SPECIFIED == gtmsource_options.rootprimary))
|| ((GTMRECEIVE == pool_user) && !jnlpool.repl_inst_filehdr->is_supplementary)))
{ /* Journal pool was created as -ROOTPRIMARY (or -UPDOK) and a source server command has
* specified -PROPAGATEPRIMARY (or -UPDNOTOK) or a receiver server command is being attempted
* on a non-supplementary instance. Issue error.
*/
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(4) ERR_PRIMARYISROOT, 2,
LEN_AND_STR((char *)repl_instance.inst_info.this_instname));
}
} else if (is_src_srvr)
{ /* Source server command issued on a propagating primary */
if (ROOTPRIMARY_SPECIFIED == gtmsource_options.rootprimary)
{ /* Journal pool was created with a -PROPAGATEPRIMARY command and current source server command
* has specified -ROOTPRIMARY (or -UPDOK).
*/
if (!gtmsource_options.activate)
{ /* START or DEACTIVATE was specified. Issue incompatibility error right away */
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(4) ERR_PRIMARYNOTROOT, 2,
LEN_AND_STR((char *)repl_instance.inst_info.this_instname));
} else
{ /* ACTIVATE was specified. Check if there is only one process attached to the journal
* pool. If yes, we are guaranteed that it is indeed the passive source server process
* that we are trying to activate (we can assert this because we know for sure the source
* server corresponding to this slot "gtmsourcelocal_ptr" is alive and running or else we
* would have issued a SRCSRVNOTEXIST error earlier). and that this is a case of a
* transition from propagating primary to root primary. If not, issue ACTIVATEFAIL error.
*/
assert(NULL != gtmsourcelocal_ptr);
if (1 != shmstat.shm_nattch)
{
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(4) ERR_ACTIVATEFAIL, 2,
LEN_AND_STR(gtmsource_options.secondary_instname));
} else
hold_onto_ftok_sem = TRUE;
}
}
}
}
if (!csa->nl->glob_sec_init)
{
assert(new_ipc);
assert(slot_needs_init);
if (!is_src_srvr || !gtmsource_options.start)
{
assert(FALSE);
rel_sem_immediate(SOURCE, JNL_POOL_ACCESS_SEM);
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(6) ERR_JNLPOOLSETUP, 0,
ERR_TEXT, 2, RTS_ERROR_LITERAL("Journal pool has not been initialized"));
}
/* Initialize the shared memory fields. */
/* Start_jnl_seqno (and jnl_seqno, read_jnl_seqno) need region shared mem to be properly setup. For now set to 0. */
jnlpool_ctl->start_jnl_seqno = 0;
jnlpool_ctl->jnl_seqno = 0;
jnlpool_ctl->max_zqgblmod_seqno = 0;
jnlpool_ctl->jnldata_base_off = JNLDATA_BASE_OFF;
jnlpool_ctl->jnlpool_size = gtmsource_options.buffsize - jnlpool_ctl->jnldata_base_off;
assert((jnlpool_ctl->jnlpool_size & ~JNL_WRT_END_MASK) == 0);
jnlpool_ctl->write = 0;
jnlpool_ctl->lastwrite_len = 0;
QWASSIGNDW(jnlpool_ctl->early_write_addr, 0);
QWASSIGNDW(jnlpool_ctl->write_addr, 0);
if (0 < jnlpool.repl_inst_filehdr->num_histinfo)
{
GRAB_LOCK(jnlpool.jnlpool_dummy_reg, ASSERT_NO_ONLINE_ROLLBACK);
status = repl_inst_histinfo_get(jnlpool.repl_inst_filehdr->num_histinfo - 1, &last_histinfo);
rel_lock(jnlpool.jnlpool_dummy_reg);
assert(0 == status);
if (0 != status)
{
assert(ERR_REPLINSTNOHIST == status); /* the only error returned by repl_inst_histinfo_get() */
GRAB_LOCK(jnlpool.jnlpool_dummy_reg, ASSERT_NO_ONLINE_ROLLBACK);
repl_inst_flush_jnlpool(TRUE, TRUE); /* to reset "crash" field in instance file header to FALSE */
rel_lock(jnlpool.jnlpool_dummy_reg);
DETACH_AND_REMOVE_SHM_AND_SEM; /* remove any sem/shm we had created */
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(6) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
LEN_AND_LIT("Error reading last history record in replication instance file"));
}
instfilehdr_seqno = jnlpool.repl_inst_filehdr->jnl_seqno;
assert(last_histinfo.start_seqno);
assert(instfilehdr_seqno);
if (instfilehdr_seqno < last_histinfo.start_seqno)
{ /* The jnl seqno in the instance file header is not greater than the last histinfo's start seqno */
GRAB_LOCK(jnlpool.jnlpool_dummy_reg, ASSERT_NO_ONLINE_ROLLBACK);
repl_inst_flush_jnlpool(TRUE, TRUE); /* to reset "crash" field in instance file header to FALSE */
rel_lock(jnlpool.jnlpool_dummy_reg);
DETACH_AND_REMOVE_SHM_AND_SEM; /* remove any sem/shm we had created */
udi->grabbed_access_sem = FALSE;
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
rts_error(VARLSTCNT(8) ERR_REPLINSTSEQORD, 6, LEN_AND_LIT("Instance file header"),
&instfilehdr_seqno, &last_histinfo.start_seqno, LEN_AND_STR(udi->fn));
}
jnlpool_ctl->last_histinfo_seqno = last_histinfo.start_seqno;
} else
jnlpool_ctl->last_histinfo_seqno = 0;
assert(ROOTPRIMARY_UNSPECIFIED != gtmsource_options.rootprimary);
jnlpool_ctl->upd_disabled = TRUE; /* truly initialized later by a call to "gtmsource_rootprimary_init" */
jnlpool_ctl->primary_instname[0] = '\0';
jnlpool_ctl->send_losttn_complete = FALSE;
memcpy(jnlpool_ctl->jnlpool_id.instfilename, seg->fname, seg->fname_len);
jnlpool_ctl->jnlpool_id.instfilename[seg->fname_len] = '\0';
memcpy(jnlpool_ctl->jnlpool_id.label, GDS_RPL_LABEL, GDS_LABEL_SZ);
memcpy(jnlpool_ctl->jnlpool_id.now_running, gtm_release_name, gtm_release_name_len + 1);
assert(0 == (offsetof(jnlpool_ctl_struct, start_jnl_seqno) % 8));
/* ensure that start_jnl_seqno starts at an 8 byte boundary */
assert(0 == offsetof(jnlpool_ctl_struct, jnlpool_id));
/* ensure that the pool identifier is at the top of the pool */
jnlpool_ctl->jnlpool_id.pool_type = JNLPOOL_SEGMENT;
DEBUG_ONLY(locknl = csa->nl;) /* for DEBUG_ONLY LOCK_HIST macro */
gtm_mutex_init(reg, NUM_CRIT_ENTRY, FALSE);
DEBUG_ONLY(locknl = NULL;) /* restore "locknl" to default value */
jnlpool_mutex_spin_parms->mutex_hard_spin_count = MUTEX_HARD_SPIN_COUNT;
jnlpool_mutex_spin_parms->mutex_sleep_spin_count = MUTEX_SLEEP_SPIN_COUNT;
jnlpool_mutex_spin_parms->mutex_spin_sleep_mask = MUTEX_SPIN_SLEEP_MASK;
csa->nl->glob_sec_init = TRUE;
assert(NULL != jnlpool_creator);
*jnlpool_creator = TRUE;
} else if (NULL != jnlpool_creator)
*jnlpool_creator = FALSE;
/* If this is a supplementary instance, initialize strm_index to a non-default value.
* In case of an update process and a root primary supplementary instance, "strm_index"
* will be initialized to a non-zero value later in updproc.c.
*/
if (jnlpool.repl_inst_filehdr->is_supplementary)
{ /* The value of 0 is possible in rare cases, if a process does jnlpool_init more than once
* (possible if the first jnlpool_init failed say due to a REPLINSTUNDEF error). In that
* case, we are anyways going to set it to the exact same value so allow that in the assert.
*/
assert((INVALID_SUPPL_STRM == strm_index) || (0 == strm_index));
strm_index = 0;
}
assert(!(is_src_srvr && gtmsource_options.start) || slot_needs_init);
jnlpool.gtmsource_local = gtmsourcelocal_ptr;
reg->open = TRUE; /* this is used by t_commit_cleanup/tp_restart/mutex_deadlock_check */
reg->read_only = FALSE; /* maintain csa->read_write simultaneously */
csa->read_write = TRUE; /* maintain reg->read_only simultaneously */
if (slot_needs_init)
{
assert(is_src_srvr);
assert(NULL != gtmsourcelocal_ptr);
assert(gtmsource_options.start || gtmsource_options.showbacklog);
gtmsourcelocal_ptr->gtmsource_pid = 0;
gtmsourcelocal_ptr->gtmsource_state = GTMSOURCE_DUMMY_STATE;
if (gtmsource_options.start)
{ /* Source server startup needs to initialize source server specific fields in the journal pool */
assert(NULL != gtmsourcelocal_ptr);
QWASSIGNDW(gtmsourcelocal_ptr->read_addr, 0);
gtmsourcelocal_ptr->read = 0;
gtmsourcelocal_ptr->read_state = READ_POOL;
gtmsourcelocal_ptr->mode = gtmsource_options.mode;
assert(gtmsourcelocal_ptr->gtmsrc_lcl_array_index == (gtmsourcelocal_ptr - jnlpool.gtmsource_local_array));
gtmsourcelocal_ptr->statslog = FALSE;
gtmsourcelocal_ptr->shutdown = NO_SHUTDOWN;
gtmsourcelocal_ptr->shutdown_time = -1;
gtmsourcelocal_ptr->secondary_inet_addr = gtmsource_options.sec_inet_addr;
gtmsourcelocal_ptr->secondary_port = gtmsource_options.secondary_port;
STRCPY(gtmsourcelocal_ptr->secondary_host, gtmsource_options.secondary_host);
STRCPY(gtmsourcelocal_ptr->filter_cmd, gtmsource_options.filter_cmd);
STRCPY(gtmsourcelocal_ptr->log_file, gtmsource_options.log_file);
gtmsourcelocal_ptr->log_interval = log_interval = gtmsource_options.src_log_interval;
gtmsourcelocal_ptr->statslog_file[0] = '\0';
gtmsourcelocal_ptr->last_flush_resync_seqno = 0;
gtmsourcelocal_ptr->next_histinfo_seqno = 0;/* fully initialized when source server connects to receiver */
gtmsourcelocal_ptr->next_histinfo_num = 0; /* fully initialized when source server connects to receiver */
gtmsourcelocal_ptr->num_histinfo = 0; /* fully initialized when source server connects to receiver */
if (GTMSOURCE_MODE_ACTIVE == gtmsource_options.mode)
{
gtmsourcelocal_ptr->connect_parms[GTMSOURCE_CONN_HARD_TRIES_COUNT] =
gtmsource_options.connect_parms[GTMSOURCE_CONN_HARD_TRIES_COUNT];
gtmsourcelocal_ptr->connect_parms[GTMSOURCE_CONN_HARD_TRIES_PERIOD] =
gtmsource_options.connect_parms[GTMSOURCE_CONN_HARD_TRIES_PERIOD];
gtmsourcelocal_ptr->connect_parms[GTMSOURCE_CONN_SOFT_TRIES_PERIOD] =
gtmsource_options.connect_parms[GTMSOURCE_CONN_SOFT_TRIES_PERIOD];
gtmsourcelocal_ptr->connect_parms[GTMSOURCE_CONN_ALERT_PERIOD] =
gtmsource_options.connect_parms[GTMSOURCE_CONN_ALERT_PERIOD];
gtmsourcelocal_ptr->connect_parms[GTMSOURCE_CONN_HEARTBEAT_PERIOD] =
gtmsource_options.connect_parms[GTMSOURCE_CONN_HEARTBEAT_PERIOD];
gtmsourcelocal_ptr->connect_parms[GTMSOURCE_CONN_HEARTBEAT_MAX_WAIT] =
gtmsource_options.connect_parms[GTMSOURCE_CONN_HEARTBEAT_MAX_WAIT];
}
/* At this point online rollback cannot be concurrently running because we hold the journal pool access
* control semaphore. So, go ahead and initialize the gtmsource_srv_latch for this source server.
*/
SET_LATCH_GLOBAL(&gtmsourcelocal_ptr->gtmsource_srv_latch, LOCK_AVAILABLE);
}
if (reset_gtmsrclcl_info)
{ /* Initialize all fields of "gtmsource_local" that are also present in the corresponding "gtmsrc_lcl" */
gtmsourcelocal_ptr->read_jnl_seqno = 1; /* fully initialized when source server connects to receiver */
memcpy(gtmsourcelocal_ptr->secondary_instname, gtmsource_options.secondary_instname, MAX_INSTNAME_LEN - 1);
gtmsourcelocal_ptr->connect_jnl_seqno = 0; /* fully initialized when source server connects to receiver */
gtmsourcelocal_ptr->send_losttn_complete = FALSE;
/* Now make the corresponding changes from gtmsource_local to the gtmsrc_lcl structure and flush to disk.
* This assumes "jnlpool.gtmsource_local" is set appropriately.
*/
GRAB_LOCK(jnlpool.jnlpool_dummy_reg, ASSERT_NO_ONLINE_ROLLBACK);
repl_inst_flush_gtmsrc_lcl();
rel_lock(jnlpool.jnlpool_dummy_reg);
}
}
/* Assert that gtmsource_local is set to non-NULL value for all those qualifiers that care about it */
assert(!(gtmsource_options.start || gtmsource_options.activate || gtmsource_options.deactivate
|| gtmsource_options.stopsourcefilter || gtmsource_options.changelog || gtmsource_options.statslog)
|| (NULL != jnlpool.gtmsource_local));
assert((NULL == jnlpool.gtmsource_local)
|| !STRCMP(jnlpool.gtmsource_local->secondary_instname, gtmsource_options.secondary_instname));
temp_jnlpool_ctl->jnlpool_size = jnlpool_ctl->jnlpool_size;
/* Release control lockout now that this process has attached to the journal pool except if caller is source server.
* Source Server will release the control lockout only after it is done with
* a) initializing other fields in the pool (in case of source server startup) or
* b) the actual command (in case of any other source server command like checkhealth, shutdown, showbacklog etc.)
*/
if (!is_src_srvr)
{
if (0 != (save_errno = rel_sem(SOURCE, JNL_POOL_ACCESS_SEM)))
{
ftok_sem_release(jnlpool.jnlpool_dummy_reg, TRUE, TRUE);
/* Assert we did not create shm or sem so no need to remove any */
assert(!new_ipc);
rts_error(VARLSTCNT(7) ERR_JNLPOOLSETUP, 0, ERR_TEXT, 2,
RTS_ERROR_LITERAL("Error in rel_sem"), save_errno);
}
udi->grabbed_access_sem = FALSE;
} else
{
this_side = &jnlpool_ctl->this_side;
remote_side = &gtmsourcelocal_ptr->remote_side; /* Set global variable now. Structure will be initialized
* later when source server connects to receiver */
}
if (!hold_onto_ftok_sem && !ftok_sem_release(jnlpool.jnlpool_dummy_reg, FALSE, FALSE))
rts_error(VARLSTCNT(1) ERR_JNLPOOLSETUP);
pool_init = TRUE;
return;
}
void jnlpool_detach(void)
{
if (TRUE == pool_init)
{
if (-1 == shmdt((caddr_t)jnlpool_ctl))
rts_error(VARLSTCNT(5) ERR_REPLWARN, 2, RTS_ERROR_LITERAL("Could not detach from journal pool"), errno);
jnlpool_ctl = NULL;
jnlpool.jnlpool_ctl = NULL;
jnlpool.repl_inst_filehdr = NULL;
jnlpool.gtmsrc_lcl_array = NULL;
jnlpool.gtmsource_local_array = NULL;
jnlpool.jnldata_base = NULL;
pool_init = FALSE;
}
}