675 lines
26 KiB
C
675 lines
26 KiB
C
/****************************************************************
|
|
* *
|
|
* Copyright 2001, 2011 Fidelity Information Services, Inc *
|
|
* *
|
|
* This source code contains the intellectual property *
|
|
* of its copyright holder(s), and is made available *
|
|
* under a license. If you do not know the terms of *
|
|
* the license, please stop and do not read further. *
|
|
* *
|
|
****************************************************************/
|
|
|
|
#include "mdef.h"
|
|
|
|
#include "gtm_ipc.h"
|
|
#include "gtm_inet.h"
|
|
#include "gtm_fcntl.h"
|
|
#include "gtm_string.h"
|
|
#include "gtm_unistd.h"
|
|
#include "gtm_time.h"
|
|
|
|
#include <sys/sem.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/shm.h>
|
|
#include <errno.h>
|
|
#include <signal.h> /* for VSIG_ATOMIC_T type */
|
|
|
|
#include "gdsroot.h"
|
|
#include "gtm_facility.h"
|
|
#include "fileinfo.h"
|
|
#include "gdsbt.h"
|
|
#include "gdsfhead.h"
|
|
#include "filestruct.h"
|
|
#include "gdsblk.h"
|
|
#include "gt_timer.h"
|
|
#include "jnl.h"
|
|
#include "interlock.h"
|
|
#include "error.h"
|
|
#include "iosp.h"
|
|
#include "gdsbgtr.h"
|
|
#include "repl_msg.h"
|
|
#include "gtmsource.h"
|
|
#include "aswp.h"
|
|
#include "gtm_c_stack_trace.h"
|
|
#include "eintr_wrappers.h"
|
|
#include "eintr_wrapper_semop.h"
|
|
#include "util.h"
|
|
#include "send_msg.h"
|
|
#include "change_reg.h"
|
|
#include "compswap.h"
|
|
#ifdef UNIX
|
|
#include "mutex.h"
|
|
#endif
|
|
#include "gds_rundown.h"
|
|
#include "gvusr.h"
|
|
#include "do_semop.h"
|
|
#include "mmseg.h"
|
|
#include "ipcrmid.h"
|
|
#include "gtmmsg.h"
|
|
#include "wcs_recover.h"
|
|
#include "wcs_mm_recover.h"
|
|
#include "tp_change_reg.h"
|
|
#include "wcs_flu.h"
|
|
#include "add_inter.h"
|
|
#include "io.h"
|
|
#include "gtmsecshr.h"
|
|
#include "ftok_sems.h"
|
|
#include "gtmimagename.h"
|
|
#include "gtmio.h"
|
|
#include "have_crit.h"
|
|
#include "wcs_clean_dbsync.h"
|
|
#include "is_proc_alive.h"
|
|
#include "shmpool.h"
|
|
#include "db_snapshot.h"
|
|
#include "tp_grab_crit.h"
|
|
#include "ss_lock_facility.h"
|
|
|
|
#ifndef GTM_SNAPSHOT
|
|
# error "Snapshot facility not available on this platform"
|
|
#endif
|
|
|
|
#define CANCEL_DB_TIMERS(region, csa, cancelled_timer, cancelled_dbsync_timer) \
|
|
{ \
|
|
if (csa->timer) \
|
|
{ \
|
|
cancel_timer((TID)region); \
|
|
if (NULL != csa->nl) \
|
|
DECR_CNT(&csa->nl->wcs_timers, &csa->nl->wc_var_lock); \
|
|
cancelled_timer = TRUE; \
|
|
csa->timer = FALSE; \
|
|
} \
|
|
if (csa->dbsync_timer) \
|
|
{ \
|
|
CANCEL_DBSYNC_TIMER(csa, FALSE); \
|
|
cancelled_dbsync_timer = TRUE; \
|
|
} \
|
|
}
|
|
|
|
GBLREF VSIG_ATOMIC_T forced_exit;
|
|
GBLREF boolean_t mupip_jnl_recover;
|
|
GBLREF boolean_t have_standalone_access;
|
|
GBLREF boolean_t created_core, need_core, dont_want_core, is_src_server, is_updproc;
|
|
GBLREF gd_region *gv_cur_region;
|
|
GBLREF sgmnt_addrs *cs_addrs;
|
|
GBLREF sgmnt_data_ptr_t cs_data;
|
|
GBLREF uint4 process_id;
|
|
GBLREF ipcs_mesg db_ipcs;
|
|
GBLREF jnl_process_vector *prc_vec;
|
|
GBLREF jnl_process_vector *originator_prc_vec;
|
|
GBLREF jnl_gbls_t jgbl;
|
|
GBLREF boolean_t dse_running;
|
|
|
|
error_def(ERR_ASSERT);
|
|
error_def(ERR_CRITSEMFAIL);
|
|
error_def(ERR_DBFILERR);
|
|
error_def(ERR_DBRNDWN);
|
|
error_def(ERR_DBRNDWNWRN);
|
|
error_def(ERR_GTMASSERT);
|
|
error_def(ERR_GTMCHECK);
|
|
error_def(ERR_IPCNOTDEL);
|
|
error_def(ERR_JNLFLUSH);
|
|
error_def(ERR_MEMORY);
|
|
error_def(ERR_OUTOFSPACE);
|
|
error_def(ERR_RNDWNSEMFAIL);
|
|
error_def(ERR_TEXT);
|
|
error_def(ERR_WCBLOCKED);
|
|
error_def(ERR_STACKOFLOW);
|
|
|
|
LITREF char gtm_release_name[];
|
|
LITREF int4 gtm_release_name_len;
|
|
|
|
static boolean_t grabbed_access_sem;
|
|
|
|
void gds_rundown(void)
|
|
{
|
|
bool is_mm, we_are_last_user, we_are_last_writer;
|
|
boolean_t ipc_deleted, remove_shm, cancelled_timer, cancelled_dbsync_timer, vermismatch;
|
|
now_t now; /* for GET_CUR_TIME macro */
|
|
char *time_ptr, time_str[CTIME_BEFORE_NL + 2]; /* for GET_CUR_TIME macro */
|
|
gd_region *reg;
|
|
int save_errno, status, rc;
|
|
int4 semval, ftok_semval, sopcnt, ftok_sopcnt;
|
|
short crash_count;
|
|
sm_long_t munmap_len;
|
|
sgmnt_addrs *csa;
|
|
sgmnt_data_ptr_t csd;
|
|
struct shmid_ds shm_buf;
|
|
struct sembuf sop[2], ftok_sop[2];
|
|
uint4 jnl_status;
|
|
unix_db_info *udi;
|
|
jnl_private_control *jpc;
|
|
jnl_buffer_ptr_t jbp;
|
|
shm_snapshot_t *ss_shm_ptr;
|
|
boolean_t is_cur_process_ss_initiator;
|
|
uint4 ss_pid;
|
|
|
|
grabbed_access_sem = FALSE;
|
|
jnl_status = 0;
|
|
reg = gv_cur_region; /* Local copy */
|
|
|
|
/*
|
|
* early out for cluster regions
|
|
* to avoid tripping the assert below.
|
|
* Note:
|
|
* This early out is consistent with VMS. It has been
|
|
* noted that all of the gtcm assignments
|
|
* to gv_cur_region should use the TP_CHANGE_REG
|
|
* macro. This would also avoid the assert problem
|
|
* and should be done eventually.
|
|
*/
|
|
if (dba_cm == reg->dyn.addr->acc_meth)
|
|
return;
|
|
|
|
udi = FILE_INFO(reg);
|
|
csa = &udi->s_addrs;
|
|
csd = csa->hdr;
|
|
assert(csa == cs_addrs && csd == cs_data);
|
|
if ((reg->open) && (dba_usr == csd->acc_meth))
|
|
{
|
|
change_reg();
|
|
gvusr_rundown();
|
|
return;
|
|
}
|
|
ESTABLISH(gds_rundown_ch);
|
|
assert(!csa->hold_onto_crit); /* because of this, we dont need to do "was_crit" checks before doing "rel_crit" below */
|
|
if (!reg->open) /* Not open, no point to rundown */
|
|
{
|
|
if (reg->opening) /* Died partway open, kill rest of way */
|
|
{
|
|
rel_crit(reg);
|
|
mutex_cleanup(reg);
|
|
/* revist this to handle MM properly SMW 98/12/16
|
|
if (NULL != csa->nl)
|
|
{
|
|
status = shmdt((caddr_t)csa->nl);
|
|
if (-1 == status)
|
|
send_msg(VARLSTCNT(9) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, LEN_AND_LIT("Error during shmdt"), errno);
|
|
}
|
|
*/
|
|
shmdt((caddr_t)csa->nl);
|
|
csa->nl = NULL;
|
|
REMOVE_CSA_FROM_CSADDRSLIST(csa); /* remove "csa" from list of open regions (cs_addrs_list) */
|
|
}
|
|
REVERT;
|
|
return;
|
|
}
|
|
switch(csd->acc_meth)
|
|
{ /* Pass mm and bg through */
|
|
case dba_bg:
|
|
is_mm = FALSE;
|
|
break;
|
|
case dba_mm:
|
|
is_mm = TRUE;
|
|
break;
|
|
case dba_usr:
|
|
assert(FALSE);
|
|
default:
|
|
REVERT;
|
|
return;
|
|
}
|
|
/* Cancel any pending flush timer for this region by this task */
|
|
cancelled_timer = FALSE;
|
|
cancelled_dbsync_timer = FALSE;
|
|
CANCEL_DB_TIMERS(reg, csa, cancelled_timer, cancelled_dbsync_timer);
|
|
we_are_last_user = FALSE;
|
|
if (!csa->persistent_freeze)
|
|
region_freeze(reg, FALSE, FALSE, FALSE);
|
|
rel_crit(reg); /* get locks to known state */
|
|
mutex_cleanup(reg);
|
|
/*
|
|
* We need to guarantee that none else access database file header when semid/shmid fields are reset.
|
|
* We already have created ftok semaphore in db_init or, mu_rndwn_file and did not remove it.
|
|
* So just lock it. We do it in blocking mode.
|
|
*/
|
|
DEFER_INTERRUPTS(INTRPT_IN_GDS_RUNDOWN);
|
|
if (!ftok_sem_lock(reg, FALSE, FALSE))
|
|
rts_error(VARLSTCNT(4) ERR_DBFILERR, 2, DB_LEN_STR(reg));
|
|
FTOK_TRACE(csa, csa->ti->curr_tn, ftok_ops_lock, process_id);
|
|
/*
|
|
* If we already have database access control semaphore, we do not release it now. We will release it
|
|
* in mupip_exit_handler.c
|
|
*/
|
|
if (!have_standalone_access)
|
|
{
|
|
sop[0].sem_num = 0; sop[0].sem_op = 0; /* Wait for 0 */
|
|
sop[1].sem_num = 0; sop[1].sem_op = 1; /* Lock */
|
|
sopcnt = 2;
|
|
sop[0].sem_flg = sop[1].sem_flg = SEM_UNDO | IPC_NOWAIT; /* Don't wait the first time thru */
|
|
SEMOP(udi->semid, sop, sopcnt, status, NO_WAIT);
|
|
if (-1 == status) /* We couldn't get it in one shot -- see if we already have it */
|
|
{
|
|
save_errno = errno;
|
|
if (semctl(udi->semid, 0, GETPID) == process_id)
|
|
{
|
|
send_msg(VARLSTCNT(5) MAKE_MSG_INFO(ERR_CRITSEMFAIL), 2,
|
|
DB_LEN_STR(reg),
|
|
ERR_RNDWNSEMFAIL);
|
|
REVERT;
|
|
ENABLE_INTERRUPTS(INTRPT_IN_GDS_RUNDOWN);
|
|
return; /* Already in rundown for this region */
|
|
}
|
|
if (EAGAIN != save_errno)
|
|
{
|
|
assert(FALSE);
|
|
rts_error(VARLSTCNT(9) ERR_CRITSEMFAIL, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("gds_rundown first semop/semctl"), save_errno);
|
|
}
|
|
sop[0].sem_flg = sop[1].sem_flg = SEM_UNDO; /* Try again - blocking this time */
|
|
SEMOP(udi->semid, sop, 2, status, FORCED_WAIT);
|
|
if (-1 == status) /* We couldn't get it at all.. */
|
|
rts_error(VARLSTCNT(5) ERR_CRITSEMFAIL, 2, DB_LEN_STR(reg), errno);
|
|
}
|
|
}
|
|
grabbed_access_sem = TRUE;
|
|
/*
|
|
* We now have the dbinit/rundown lock, so we are alone in this code for this region
|
|
* and nobody else can attach.
|
|
* See if we are all alone in accessing this database shared memory.
|
|
*/
|
|
assert(csa->ref_cnt); /* decrement private ref_cnt before shared ref_cnt decrement. */
|
|
csa->ref_cnt--; /* Currently journaling logic in gds_rundown() in VMS relies on this order to detect last writer */
|
|
assert(!csa->ref_cnt);
|
|
--csa->nl->ref_cnt;
|
|
if (memcmp(csa->nl->now_running, gtm_release_name, gtm_release_name_len + 1))
|
|
{ /* VERMISMATCH condition. Possible only if DSE */
|
|
assert(dse_running);
|
|
vermismatch = TRUE;
|
|
} else
|
|
vermismatch = FALSE;
|
|
if (-1 == shmctl(udi->shmid, IPC_STAT, &shm_buf))
|
|
{
|
|
save_errno = errno;
|
|
rts_error(VARLSTCNT(9) ERR_CRITSEMFAIL, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("gds_rundown shmctl"), save_errno);
|
|
} else
|
|
we_are_last_user = (1 == shm_buf.shm_nattch) && !vermismatch;
|
|
assert(!have_standalone_access || we_are_last_user); /* recover => one user */
|
|
if (-1 == (semval = semctl(udi->semid, 1, GETVAL)))
|
|
rts_error(VARLSTCNT(5) ERR_CRITSEMFAIL, 2, DB_LEN_STR(reg), errno);
|
|
we_are_last_writer = (1 == semval) && (FALSE == reg->read_only) && !vermismatch;/* There's one writer left and I am it */
|
|
assert(!(have_standalone_access && !reg->read_only) || we_are_last_writer); /* recover + R/W region => one writer */
|
|
if (-1 == (ftok_semval = semctl(udi->ftok_semid, 1, GETVAL)))
|
|
rts_error(VARLSTCNT(5) ERR_CRITSEMFAIL, 2, DB_LEN_STR(reg), errno);
|
|
if (NULL != csa->ss_ctx)
|
|
ss_destroy_context(csa->ss_ctx);
|
|
/* SS_MULTI: If multiple snapshots are supported, then we have to run through each of the snapshots */
|
|
assert(1 == MAX_SNAPSHOTS);
|
|
ss_shm_ptr = (shm_snapshot_ptr_t)SS_GETSTARTPTR(csa);
|
|
ss_pid = ss_shm_ptr->ss_info.ss_pid;
|
|
is_cur_process_ss_initiator = (process_id == ss_pid);
|
|
if (ss_pid && (is_cur_process_ss_initiator || we_are_last_user))
|
|
{
|
|
/* Try getting snapshot crit latch. If we don't get latch, we won't hang for eternity and will skip
|
|
* doing the orphaned snapshot cleanup. It will be cleaned up eventually either by subsequent MUPIP
|
|
* INTEG or by a MUPIP RUNDOWN.
|
|
*/
|
|
if (ss_get_lock_nowait(reg) && (ss_pid == ss_shm_ptr->ss_info.ss_pid)
|
|
&& (is_cur_process_ss_initiator || !is_proc_alive(ss_pid, 0)))
|
|
{
|
|
ss_release(NULL);
|
|
ss_release_lock(reg);
|
|
}
|
|
}
|
|
/* If csa->nl->donotflush_dbjnl is set, it means mupip recover/rollback was interrupted and therefore we should
|
|
* not flush shared memory contents to disk as they might be in an inconsistent state.
|
|
* In this case, we will go ahead and remove shared memory (without flushing the contents) in this routine.
|
|
* A reissue of the recover/rollback command will restore the database to a consistent state.
|
|
* Otherwise, if we have write access to this region, let us perform a few writing tasks.
|
|
*/
|
|
if (csa->nl->donotflush_dbjnl)
|
|
csa->wbuf_dqd = 0; /* ignore csa->wbuf_dqd status as we do not care about the cache contents */
|
|
else if (!reg->read_only && !vermismatch)
|
|
{ /* If we had an orphaned block and were interrupted, set wc_blocked so we can invoke wcs_recover */
|
|
if (csa->wbuf_dqd)
|
|
{
|
|
grab_crit(reg);
|
|
SET_TRACEABLE_VAR(csd->wc_blocked, TRUE);
|
|
BG_TRACE_PRO_ANY(csa, wcb_gds_rundown);
|
|
send_msg(VARLSTCNT(8) ERR_WCBLOCKED, 6, LEN_AND_LIT("wcb_gds_rundown"),
|
|
process_id, &csa->ti->curr_tn, DB_LEN_STR(reg));
|
|
csa->wbuf_dqd = 0;
|
|
wcs_recover(reg);
|
|
if (is_mm)
|
|
{
|
|
assert(FALSE);
|
|
csd = csa->hdr;
|
|
}
|
|
BG_TRACE_PRO_ANY(csa, lost_block_recovery);
|
|
rel_crit(reg);
|
|
}
|
|
if (JNL_ENABLED(csd) && IS_GTCM_GNP_SERVER_IMAGE)
|
|
originator_prc_vec = NULL;
|
|
/* If we are the last writing user, then everything must be flushed */
|
|
if (we_are_last_writer)
|
|
{ /* Time to flush out all of our buffers */
|
|
if (is_mm)
|
|
{
|
|
if (csa->total_blks != csa->ti->total_blks) /* do remap if file had been extended */
|
|
{
|
|
grab_crit(reg);
|
|
wcs_mm_recover(reg);
|
|
csd = csa->hdr;
|
|
rel_crit(reg);
|
|
}
|
|
csa->nl->remove_shm = TRUE;
|
|
}
|
|
/* Note WCSFLU_SYNC_EPOCH ensures the epoch is synced to the journal and indirectly
|
|
* also ensures that the db is fsynced. We don't want to use it in the calls to
|
|
* wcs_flu() from t_end() and tp_tend() since we can defer it to out-of-crit there.
|
|
* In this case, since we are running down, we don't have any such option.
|
|
*/
|
|
csa->nl->remove_shm = wcs_flu(WCSFLU_FLUSH_HDR | WCSFLU_WRITE_EPOCH | WCSFLU_SYNC_EPOCH);
|
|
/* Since we_are_last_writer, we should be guaranteed that wcs_flu() did not change csd, (in
|
|
* case of MM for potential file extension), even if it did a grab_crit(). Therefore, make
|
|
* sure that's true.
|
|
*/
|
|
assert(csd == csa->hdr);
|
|
assert(0 == memcmp(csd->label, GDS_LABEL, GDS_LABEL_SZ - 1));
|
|
} else if ((cancelled_timer && (0 > csa->nl->wcs_timers)) || cancelled_dbsync_timer)
|
|
{ /* cancelled pending db or jnl flush timers - flush database and journal buffers to disk */
|
|
grab_crit(reg);
|
|
/* we need to sync the epoch as the fact that there is no active pending flush timer implies
|
|
* there will be noone else who will flush the dirty buffers and EPOCH to disk in a timely fashion
|
|
*/
|
|
wcs_flu(WCSFLU_FLUSH_HDR | WCSFLU_WRITE_EPOCH | WCSFLU_SYNC_EPOCH);
|
|
rel_crit(reg);
|
|
assert((dba_mm == cs_data->acc_meth) || (csd == cs_data));
|
|
csd = cs_data; /* In case this is MM and wcs_flu() remapped an extended database, reset csd */
|
|
}
|
|
/* Do rundown journal processing after buffer flushes since they require jnl to be open */
|
|
if (JNL_ENABLED(csd))
|
|
{ /* the following tp_change_reg() is not needed due to the assert csa == cs_addrs at the beginning
|
|
* of gds_rundown(), but just to be safe. To be removed by 2002!! --- nars -- 2001/04/25.
|
|
*/
|
|
tp_change_reg(); /* call this because jnl_ensure_open checks cs_addrs rather than gv_cur_region */
|
|
jpc = csa->jnl;
|
|
jbp = jpc->jnl_buff;
|
|
if (jbp->fsync_in_prog_latch.u.parts.latch_pid == process_id)
|
|
{
|
|
assert(FALSE);
|
|
COMPSWAP_UNLOCK(&jbp->fsync_in_prog_latch, process_id, 0, LOCK_AVAILABLE, 0);
|
|
}
|
|
if (jbp->io_in_prog_latch.u.parts.latch_pid == process_id)
|
|
{
|
|
assert(FALSE);
|
|
COMPSWAP_UNLOCK(&jbp->io_in_prog_latch, process_id, 0, LOCK_AVAILABLE, 0);
|
|
}
|
|
if (((NOJNL != jpc->channel) && !JNL_FILE_SWITCHED(jpc))
|
|
|| we_are_last_writer && (0 != csa->nl->jnl_file.u.inode))
|
|
{ /* We need to close the journal file cleanly if we have the latest generation journal file open
|
|
* or if we are the last writer and the journal file is open in shared memory (not necessarily
|
|
* by ourselves e.g. the only process that opened the journal got shot abnormally)
|
|
* Note: we should not infer anything from the shared memory value of csa->nl->jnl_file.u.inode
|
|
* if we are not the last writer as it can be concurrently updated.
|
|
*/
|
|
grab_crit(reg);
|
|
if (JNL_ENABLED(csd))
|
|
{
|
|
SET_GBL_JREC_TIME; /* jnl_ensure_open/jnl_put_jrt_pini/pfin/jnl_file_close all need it */
|
|
/* Before writing to jnlfile, adjust jgbl.gbl_jrec_time if needed to maintain time order
|
|
* of jnl records. This needs to be done BEFORE the jnl_ensure_open as that could write
|
|
* journal records (if it decides to switch to a new journal file).
|
|
*/
|
|
ADJUST_GBL_JREC_TIME(jgbl, jbp);
|
|
jnl_status = jnl_ensure_open();
|
|
if (0 == jnl_status)
|
|
{ /* If we_are_last_writer, we would have already done a wcs_flu() which would
|
|
* have written an epoch record and we are guaranteed no further updates
|
|
* since we are the last writer. So, just close the journal.
|
|
* Although we assert pini_addr should be non-zero for last_writer, we
|
|
* play it safe in PRO and write a PINI record if not written already.
|
|
*/
|
|
assert(!jbp->before_images || is_mm
|
|
|| !we_are_last_writer || 0 != jpc->pini_addr);
|
|
if (we_are_last_writer && 0 == jpc->pini_addr)
|
|
jnl_put_jrt_pini(csa);
|
|
if (0 != jpc->pini_addr)
|
|
jnl_put_jrt_pfin(csa);
|
|
/* If not the last writer and no pending flush timer left, do jnl flush now */
|
|
if (!we_are_last_writer && (0 > csa->nl->wcs_timers))
|
|
{
|
|
if (SS_NORMAL == (jnl_status = jnl_flush(reg)))
|
|
{
|
|
assert(jbp->freeaddr == jbp->dskaddr);
|
|
jnl_fsync(reg, jbp->dskaddr);
|
|
assert(jbp->fsync_dskaddr == jbp->dskaddr);
|
|
} else
|
|
{
|
|
send_msg(VARLSTCNT(9) ERR_JNLFLUSH, 2, JNL_LEN_STR(csd),
|
|
ERR_TEXT, 2,
|
|
RTS_ERROR_TEXT("Error with journal flush in gds_rundown"),
|
|
jnl_status);
|
|
assert(NOJNL == jpc->channel);/* jnl file lost has been triggered */
|
|
/* In this routine, all code that follows from here on does not
|
|
* assume anything about the journaling characteristics of this
|
|
* database so it is safe to continue execution even though
|
|
* journaling got closed in the middle.
|
|
*/
|
|
}
|
|
}
|
|
jnl_file_close(reg, we_are_last_writer, FALSE);
|
|
} else
|
|
send_msg(VARLSTCNT(6) jnl_status, 4, JNL_LEN_STR(csd), DB_LEN_STR(reg));
|
|
}
|
|
rel_crit(reg);
|
|
}
|
|
}
|
|
if (we_are_last_writer) /* Flush the fileheader last and harden the file to disk */
|
|
{
|
|
grab_crit(reg); /* To satisfy crit requirement in fileheader_sync() */
|
|
memset(csd->machine_name, 0, MAX_MCNAMELEN); /* clear the machine_name field */
|
|
if (!have_standalone_access && we_are_last_user)
|
|
{ /* mupip_exit_handler will do this after mur_close_file */
|
|
csd->semid = INVALID_SEMID;
|
|
csd->shmid = INVALID_SHMID;
|
|
csd->gt_sem_ctime.ctime = 0;
|
|
csd->gt_shm_ctime.ctime = 0;
|
|
}
|
|
fileheader_sync(reg);
|
|
rel_crit(reg);
|
|
if (FALSE == is_mm)
|
|
{
|
|
if (-1 == fsync(udi->fd)) /* Sync it all */
|
|
{
|
|
rts_error(VARLSTCNT(9) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("Error during file sync at close"), errno);
|
|
}
|
|
} else
|
|
{ /* Now do final MM file sync before exit */
|
|
#if !defined(TARGETED_MSYNC) && !defined(NO_MSYNC)
|
|
if (-1 == fsync(udi->fd)) /* Sync it all */
|
|
{
|
|
rts_error(VARLSTCNT(9) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("Error during file sync at close"), errno);
|
|
}
|
|
#else
|
|
if (-1 == msync((caddr_t)csa->db_addrs[0], (size_t)(csa->db_addrs[1] - csa->db_addrs[0]), MS_SYNC))
|
|
{
|
|
rts_error(VARLSTCNT(9) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("Error during file msync at close"), errno);
|
|
}
|
|
#endif
|
|
}
|
|
}
|
|
} /* end if (!reg->read_only && !csa->nl->donotflush_dbjnl) */
|
|
/* We had cancelled all db timers at start of rundown. In case as part of rundown (wcs_flu above), we had started
|
|
* any timers, cancel them BEFORE setting reg->open to FALSE (assert in wcs_clean_dbsync relies on this).
|
|
*/
|
|
CANCEL_DB_TIMERS(reg, csa, cancelled_timer, cancelled_dbsync_timer);
|
|
if (reg->read_only && we_are_last_user && !have_standalone_access)
|
|
{ /* mupip_exit_handler will do this after mur_close_file */
|
|
db_ipcs.semid = INVALID_SEMID;
|
|
db_ipcs.shmid = INVALID_SHMID;
|
|
db_ipcs.gt_sem_ctime = 0;
|
|
db_ipcs.gt_shm_ctime = 0;
|
|
db_ipcs.fn_len = reg->dyn.addr->fname_len;
|
|
memcpy(db_ipcs.fn, reg->dyn.addr->fname, reg->dyn.addr->fname_len);
|
|
db_ipcs.fn[reg->dyn.addr->fname_len] = 0;
|
|
/* request gtmsecshr to flush. read_only cannot flush itself */
|
|
if (0 != send_mesg2gtmsecshr(FLUSH_DB_IPCS_INFO, 0, (char *)NULL, 0))
|
|
rts_error(VARLSTCNT(8) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("gtmsecshr failed to update database file header"));
|
|
}
|
|
/* Done with file now, close it */
|
|
CLOSEFILE_RESET(udi->fd, rc); /* resets "udi->fd" to FD_INVALID */
|
|
if (-1 == rc)
|
|
{
|
|
rts_error(VARLSTCNT(9) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, LEN_AND_LIT("Error during file close"), errno);
|
|
}
|
|
/* Unmap storage if mm mode but only the part that is not the fileheader (so shows up in dumps) */
|
|
if (is_mm)
|
|
{
|
|
munmap_len = (sm_long_t)((csa->db_addrs[1] - csa->db_addrs[0]) - ROUND_UP(SIZEOF_FILE_HDR(csa->hdr),
|
|
MSYNC_ADDR_INCS));
|
|
if (munmap_len > 0)
|
|
{
|
|
munmap((caddr_t)(csa->db_addrs[0] + ROUND_UP(SIZEOF_FILE_HDR(csa->hdr), MSYNC_ADDR_INCS)),
|
|
(size_t)(munmap_len));
|
|
# ifdef DEBUG_DB64
|
|
rel_mmseg((caddr_t)csa->db_addrs[0]);
|
|
# endif
|
|
}
|
|
}
|
|
/* Detach our shared memory while still under lock so reference counts will be
|
|
* correct for the next process to run down this region.
|
|
* In the process also get the remove_shm status from node_local before detaching.
|
|
* If csa->nl->donotflush_dbjnl is TRUE, it means we can safely remove shared memory without compromising data
|
|
* integrity as a reissue of recover will restore the database to a consistent state.
|
|
*/
|
|
remove_shm = !vermismatch && (csa->nl->remove_shm || csa->nl->donotflush_dbjnl);
|
|
status = shmdt((caddr_t)csa->nl);
|
|
csa->nl = NULL; /* dereferencing nl after detach is not right, so we set it to NULL so that we can test before dereference*/
|
|
if (-1 == status)
|
|
send_msg(VARLSTCNT(9) ERR_DBFILERR, 2, DB_LEN_STR(reg), ERR_TEXT, 2, LEN_AND_LIT("Error during shmdt"), errno);
|
|
REMOVE_CSA_FROM_CSADDRSLIST(csa); /* remove "csa" from list of open regions (cs_addrs_list) */
|
|
reg->open = FALSE;
|
|
/* If file is still not in good shape, die here and now before we get rid of our storage */
|
|
if (csa->wbuf_dqd)
|
|
GTMASSERT;
|
|
ipc_deleted = FALSE;
|
|
/* If we are the very last user, remove shared storage id and the semaphores */
|
|
if (we_are_last_user)
|
|
{ /* remove shared storage, only if last writer to rundown did a successful wcs_flu() */
|
|
assert(!vermismatch);
|
|
if (remove_shm)
|
|
{
|
|
ipc_deleted = TRUE;
|
|
if (0 != shm_rmid(udi->shmid))
|
|
rts_error(VARLSTCNT(8) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("Unable to remove shared memory"));
|
|
} else if (is_src_server || is_updproc)
|
|
{
|
|
gtm_putmsg(VARLSTCNT(6) ERR_DBRNDWNWRN, 4, DB_LEN_STR(reg), process_id, process_id);
|
|
send_msg(VARLSTCNT(6) ERR_DBRNDWNWRN, 4, DB_LEN_STR(reg), process_id, process_id);
|
|
} else
|
|
send_msg(VARLSTCNT(6) ERR_DBRNDWNWRN, 4, DB_LEN_STR(reg), process_id, process_id);
|
|
/*
|
|
* Don't release semaphore in case of mupip recover/rollback; since it has standalone access.
|
|
* It will release the semaphore in mupip_exit_handler.
|
|
*/
|
|
if (!have_standalone_access)
|
|
{
|
|
if (0 != sem_rmid(udi->semid))
|
|
rts_error(VARLSTCNT(8) ERR_DBFILERR, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("Unable to remove semaphore"));
|
|
grabbed_access_sem = FALSE;
|
|
}
|
|
} else
|
|
{
|
|
assert(!have_standalone_access);
|
|
/* If we were writing, get rid of our writer access count semaphore */
|
|
if (!reg->read_only)
|
|
if (0 != (save_errno = do_semop(udi->semid, 1, -1, SEM_UNDO)))
|
|
rts_error(VARLSTCNT(9) ERR_CRITSEMFAIL, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("gds_rundown write semaphore release"), save_errno);
|
|
/* Now remove the rundown lock */
|
|
if (0 != (save_errno = do_semop(udi->semid, 0, -1, SEM_UNDO)))
|
|
rts_error(VARLSTCNT(9) ERR_CRITSEMFAIL, 2, DB_LEN_STR(reg),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("gds_rundown rundown semaphore release"), save_errno);
|
|
grabbed_access_sem = FALSE;
|
|
}
|
|
if (!ftok_sem_release(reg, !have_standalone_access, FALSE))
|
|
rts_error(VARLSTCNT(4) ERR_DBFILERR, 2, DB_LEN_STR(reg));
|
|
FTOK_TRACE(csa, csa->ti->curr_tn, ftok_ops_release, process_id);
|
|
ENABLE_INTERRUPTS(INTRPT_IN_GDS_RUNDOWN);
|
|
if (!ipc_deleted)
|
|
{
|
|
GET_CUR_TIME;
|
|
if (is_src_server)
|
|
gtm_putmsg(VARLSTCNT(8) ERR_IPCNOTDEL, 6, CTIME_BEFORE_NL, time_ptr,
|
|
LEN_AND_LIT("Source server"), REG_LEN_STR(reg));
|
|
if (is_updproc)
|
|
gtm_putmsg(VARLSTCNT(8) ERR_IPCNOTDEL, 6, CTIME_BEFORE_NL, time_ptr,
|
|
LEN_AND_LIT("Update process"), REG_LEN_STR(reg));
|
|
if (mupip_jnl_recover)
|
|
{
|
|
gtm_putmsg(VARLSTCNT(8) ERR_IPCNOTDEL, 6, CTIME_BEFORE_NL, time_ptr,
|
|
LEN_AND_LIT("Mupip journal process"), REG_LEN_STR(reg));
|
|
send_msg(VARLSTCNT(8) ERR_IPCNOTDEL, 6, CTIME_BEFORE_NL, time_ptr,
|
|
LEN_AND_LIT("Mupip journal process"), REG_LEN_STR(reg));
|
|
}
|
|
}
|
|
REVERT;
|
|
}
|
|
|
|
CONDITION_HANDLER(gds_rundown_ch)
|
|
{
|
|
pid_t sem_pid;
|
|
int semop_res;
|
|
unix_db_info *udi;
|
|
sgmnt_addrs *csa;
|
|
boolean_t cancelled_timer, cancelled_dbsync_timer;
|
|
|
|
START_CH;
|
|
/* To get as virgin a state as possible in the core, take the core now if we
|
|
would be doing so anyway. This will set created_core so it doesn't happen again.
|
|
*/
|
|
if (DUMPABLE && !SUPPRESS_DUMP)
|
|
{
|
|
need_core = TRUE;
|
|
gtm_fork_n_core();
|
|
}
|
|
udi = FILE_INFO(gv_cur_region);
|
|
csa = &udi->s_addrs;
|
|
/* We got here on an error and are going to close the region. Cancel any pending flush timer for this region by this task*/
|
|
CANCEL_DB_TIMERS(gv_cur_region, csa, cancelled_timer, cancelled_dbsync_timer);
|
|
/* release the access control semaphore, if you hold it */
|
|
if (grabbed_access_sem)
|
|
{
|
|
if (csa->now_crit) /* Might hold crit if wcs_flu or other failure */
|
|
{
|
|
assert(!csa->hold_onto_crit);
|
|
if (NULL != csa->nl)
|
|
rel_crit(gv_cur_region); /* also sets csa->now_crit to FALSE */
|
|
else
|
|
csa->now_crit = FALSE;
|
|
}
|
|
sem_pid = semctl(udi->semid, 0, GETPID);
|
|
assert(sem_pid == process_id);
|
|
if (0 != (semop_res = do_semop(udi->semid, 0, -1, SEM_UNDO | IPC_NOWAIT)))
|
|
gtm_putmsg(VARLSTCNT(9) ERR_CRITSEMFAIL, 2, DB_LEN_STR(gv_cur_region),
|
|
ERR_TEXT, 2, RTS_ERROR_TEXT("Error releasing access semaphore"), semop_res);
|
|
|
|
}
|
|
if (udi->grabbed_ftok_sem)
|
|
ftok_sem_release(gv_cur_region, !have_standalone_access, TRUE);
|
|
gv_cur_region->open = FALSE;
|
|
csa->nl = NULL;
|
|
REMOVE_CSA_FROM_CSADDRSLIST(csa); /* remove "csa" from list of open regions (cs_addrs_list) */
|
|
PRN_ERROR;
|
|
gtm_putmsg(VARLSTCNT(4) ERR_DBRNDWN, 2, REG_LEN_STR(gv_cur_region));
|
|
UNWIND(NULL, NULL);
|
|
}
|