2012-02-05 11:35:58 -05:00
|
|
|
/****************************************************************
|
|
|
|
* *
|
2012-03-24 14:06:46 -04:00
|
|
|
* Copyright 2001, 2012 Fidelity Information Services, Inc *
|
2012-02-05 11:35:58 -05:00
|
|
|
* *
|
|
|
|
* This source code contains the intellectual property *
|
|
|
|
* of its copyright holder(s), and is made available *
|
|
|
|
* under a license. If you do not know the terms of *
|
|
|
|
* the license, please stop and do not read further. *
|
|
|
|
* *
|
|
|
|
****************************************************************/
|
|
|
|
|
|
|
|
#include "mdef.h"
|
|
|
|
|
|
|
|
#include "gtm_string.h"
|
|
|
|
|
|
|
|
#ifdef VMS
|
|
|
|
#include <descrip.h>
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#include "gdsroot.h"
|
|
|
|
#include "gdsblk.h"
|
|
|
|
#include "gdskill.h"
|
|
|
|
#include "gtm_facility.h"
|
|
|
|
#include "fileinfo.h"
|
|
|
|
#include "gdsbt.h"
|
|
|
|
#include "gdsfhead.h"
|
|
|
|
#include "filestruct.h"
|
|
|
|
#include "gdscc.h"
|
|
|
|
#include "cdb_sc.h"
|
|
|
|
#include "error.h"
|
|
|
|
#include "iosp.h" /* for declaration of SS_NORMAL */
|
|
|
|
#include "jnl.h"
|
2012-10-29 18:54:31 -04:00
|
|
|
#include <rtnhdr.h>
|
2012-02-05 11:35:58 -05:00
|
|
|
#include "mv_stent.h"
|
|
|
|
#include "stack_frame.h"
|
|
|
|
#include "hashtab_int4.h" /* needed for tp.h */
|
|
|
|
#include "buddy_list.h" /* needed for tp.h */
|
|
|
|
#include "tp.h"
|
|
|
|
#include "tp_frame.h"
|
|
|
|
#include "gtm_stdio.h"
|
|
|
|
#include "gtm_stdlib.h" /* for ATOI */
|
|
|
|
#include "send_msg.h"
|
|
|
|
#include "op.h"
|
|
|
|
#include "io.h"
|
|
|
|
#include "targ_alloc.h"
|
|
|
|
#include "getzposition.h"
|
|
|
|
#include "wcs_recover.h"
|
|
|
|
#include "tp_unwind.h"
|
|
|
|
#include "wcs_backoff.h"
|
|
|
|
#include "rel_quant.h"
|
|
|
|
#include "wcs_mm_recover.h"
|
|
|
|
#include "tp_restart.h"
|
|
|
|
#include "repl_msg.h" /* for gtmsource.h */
|
|
|
|
#include "gtmsource.h" /* for jnlpool_addrs structure definition */
|
|
|
|
#include "wbox_test_init.h"
|
|
|
|
#include "gtmimagename.h"
|
2012-03-24 14:06:46 -04:00
|
|
|
#include "have_crit.h"
|
2012-10-29 18:54:31 -04:00
|
|
|
#include "anticipatory_freeze.h"
|
2012-02-05 11:35:58 -05:00
|
|
|
#ifdef GTM_TRIGGER
|
|
|
|
#include "gv_trigger.h"
|
|
|
|
#include "gtm_trigger.h"
|
|
|
|
#endif
|
2012-03-24 14:06:46 -04:00
|
|
|
#ifdef ENABLE_EXTENDED_RESTART_TRACE_HIST
|
|
|
|
#include "caller_id.h"
|
|
|
|
#endif
|
2012-02-05 11:35:58 -05:00
|
|
|
|
|
|
|
GBLDEF trans_num tp_fail_histtn[CDB_MAX_TRIES], tp_fail_bttn[CDB_MAX_TRIES];
|
|
|
|
GBLDEF int4 tp_fail_n, tp_fail_level;
|
|
|
|
GBLDEF int4 n_pvtmods, n_blkmods;
|
|
|
|
GBLDEF gv_namehead *tp_fail_hist[CDB_MAX_TRIES];
|
|
|
|
GBLDEF block_id t_fail_hist_blk[CDB_MAX_TRIES];
|
|
|
|
GBLDEF gd_region *tp_fail_hist_reg[CDB_MAX_TRIES];
|
|
|
|
|
|
|
|
GBLREF uint4 dollar_tlevel;
|
|
|
|
GBLREF uint4 dollar_trestart;
|
|
|
|
GBLREF int dollar_truth;
|
|
|
|
GBLREF mval dollar_zgbldir;
|
|
|
|
GBLREF gd_addr *gd_header;
|
|
|
|
GBLREF gv_key *gv_currkey;
|
|
|
|
GBLREF gv_namehead *gv_target;
|
|
|
|
GBLREF stack_frame *frame_pointer;
|
|
|
|
GBLREF tp_frame *tp_pointer;
|
|
|
|
GBLREF sgm_info *sgm_info_ptr;
|
|
|
|
GBLREF tp_region *tp_reg_list; /* Chained list of regions used in this transaction not cleared on tp_restart */
|
|
|
|
GBLREF mv_stent *mv_chain;
|
|
|
|
GBLREF unsigned char *msp, *stackbase, *stacktop, t_fail_hist[CDB_MAX_TRIES];
|
|
|
|
GBLREF sgm_info *first_sgm_info;
|
|
|
|
GBLREF unsigned int t_tries;
|
|
|
|
GBLREF int process_id;
|
|
|
|
GBLREF gd_region *gv_cur_region;
|
|
|
|
GBLREF jnlpool_addrs jnlpool;
|
|
|
|
GBLREF bool caller_id_flag;
|
|
|
|
GBLREF unsigned char *tpstackbase, *tpstacktop;
|
|
|
|
GBLREF trans_num local_tn; /* transaction number for THIS PROCESS */
|
|
|
|
GBLREF sgmnt_addrs *cs_addrs;
|
2012-03-24 14:06:46 -04:00
|
|
|
GBLREF sgmnt_data *cs_data;
|
2012-02-05 11:35:58 -05:00
|
|
|
GBLREF symval *curr_symval;
|
|
|
|
GBLREF trans_num tstart_local_tn; /* copy of global variable "local_tn" at op_tstart time */
|
2012-10-29 18:54:31 -04:00
|
|
|
GBLREF boolean_t mupip_jnl_recover;
|
2012-02-05 11:35:58 -05:00
|
|
|
#ifdef VMS
|
|
|
|
GBLREF struct chf$signal_array *tp_restart_fail_sig;
|
2012-03-24 14:06:46 -04:00
|
|
|
GBLREF boolean_t tp_restart_fail_sig_used;
|
|
|
|
#else
|
|
|
|
GBLREF jnl_gbls_t jgbl;
|
|
|
|
GBLREF sgmnt_addrs *cs_addrs_list;
|
|
|
|
GBLREF boolean_t is_updproc;
|
2012-02-05 11:35:58 -05:00
|
|
|
#endif
|
|
|
|
#ifdef GTM_TRIGGER
|
2012-03-24 14:06:46 -04:00
|
|
|
GBLREF int tprestart_state; /* When triggers restart, multiple states possible. See tp_restart.h */
|
|
|
|
GBLREF mval dollar_ztwormhole; /* Previous value (mval) restored on restart */
|
|
|
|
GBLREF mval dollar_ztslate;
|
|
|
|
LITREF mval literal_null;
|
2012-02-05 11:35:58 -05:00
|
|
|
#endif
|
|
|
|
|
2012-03-24 14:06:46 -04:00
|
|
|
#ifdef UNIX
|
|
|
|
error_def(ERR_GVFAILCORE);
|
|
|
|
error_def(ERR_REPLONLNRLBK);
|
|
|
|
#endif
|
2012-02-05 11:35:58 -05:00
|
|
|
error_def(ERR_TLVLZERO);
|
|
|
|
error_def(ERR_TPFAIL);
|
|
|
|
error_def(ERR_TPLOCKRESTMAX);
|
|
|
|
error_def(ERR_TPRESTART);
|
|
|
|
error_def(ERR_TPRETRY);
|
|
|
|
error_def(ERR_TRESTLOC);
|
|
|
|
error_def(ERR_TRESTNOT);
|
|
|
|
|
|
|
|
#define GVNAME_UNKNOWN "*UNKNOWN"
|
|
|
|
|
|
|
|
static readonly char gvname_unknown[] = GVNAME_UNKNOWN;
|
|
|
|
static readonly int4 gvname_unknown_len = STR_LIT_LEN(GVNAME_UNKNOWN);
|
|
|
|
|
|
|
|
CONDITION_HANDLER(tp_restart_ch)
|
|
|
|
{
|
|
|
|
START_CH;
|
|
|
|
/* On Unix, there is only one set of the signal info and this error will handily replace it. For VMS,
|
2012-03-24 14:06:46 -04:00
|
|
|
* far more subterfuge is required. We will save the signal information and paramters and overlay the
|
|
|
|
* TPRETRY signal information with it so that the signal will be handled properly. Note also that since
|
|
|
|
* VMS does not support triggers, no special avoidance of the below needs to occur when we are dealing with
|
|
|
|
* a trigger unwind initiated rethrow.
|
|
|
|
*/
|
2012-02-05 11:35:58 -05:00
|
|
|
# ifdef VMS
|
|
|
|
assert(TPRESTART_ARG_CNT >= sig->chf$is_sig_args);
|
|
|
|
if (NULL == tp_restart_fail_sig)
|
|
|
|
tp_restart_fail_sig = (struct chf$signal_array *)malloc((TPRESTART_ARG_CNT + 1) * SIZEOF(int));
|
|
|
|
memcpy(tp_restart_fail_sig, sig, (sig->chf$is_sig_args + 1) * SIZEOF(int));
|
|
|
|
assert(FALSE == tp_restart_fail_sig_used);
|
|
|
|
tp_restart_fail_sig_used = TRUE;
|
|
|
|
# endif
|
|
|
|
GTMTRIG_ONLY(DBGTRIGR((stderr, "tp_restart_ch: ERROR!! unwinding C frame to return. Error is %d, tprestart_state is %d\n",
|
|
|
|
arg, tprestart_state)));
|
|
|
|
UNWIND(NULL, NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Note that adding a new rts_error in "tp_restart" might need a change to the INVOKE_RESTART macro in tp.h and
|
|
|
|
* TPRESTART_ARG_CNT in errorsp.h (sl_vvms). See comment in tp.h for INVOKE_RESTART macro for the details.
|
|
|
|
*/
|
|
|
|
int tp_restart(int newlevel, boolean_t handle_errors_internally)
|
|
|
|
{
|
|
|
|
unsigned char *cp;
|
|
|
|
unsigned int hist_index;
|
|
|
|
tp_frame *tf;
|
|
|
|
mv_stent *mvc;
|
|
|
|
tp_region *tr;
|
|
|
|
mval beganHere;
|
|
|
|
sgmnt_addrs *csa;
|
|
|
|
int4 num_closed = 0;
|
|
|
|
boolean_t tp_tend_status;
|
2012-10-29 18:54:31 -04:00
|
|
|
boolean_t reset_clues_done = FALSE;
|
2012-02-05 11:35:58 -05:00
|
|
|
mstr gvname_mstr, reg_mstr;
|
|
|
|
gd_region *restart_reg, *reg;
|
|
|
|
int tprestart_rc;
|
|
|
|
enum cdb_sc status;
|
|
|
|
DCL_THREADGBL_ACCESS;
|
|
|
|
|
|
|
|
SETUP_THREADGBL_ACCESS;
|
|
|
|
tprestart_rc = 0;
|
|
|
|
/* Some callers of "tp_restart" want this function to return with an error code instead of issue an rts_error
|
|
|
|
* if there is an error inside tp_restart. The SIGNAL macro is supposed to reflect the error code in that
|
|
|
|
* case and the caller handles this error code accordingly after tp_restart returns. For those callers,
|
|
|
|
* establish the "tp_restart_ch" condition handler to catch all errors. For the remaining callers, any errors
|
|
|
|
* inside tp_restart will invoke whatever parent condition handler is active at that point.
|
|
|
|
*
|
|
|
|
* The reason why a few callers prefer this inside-tprestart error handling is because they are already a
|
|
|
|
* condition/error handler (e.g. mdb_condition_handler) when they invoke tp_restart and do not want another
|
|
|
|
* rts_error to happen inside tp_restart and trigger any other condition/error handlers that can alter the
|
|
|
|
* flow of control elsewhere until this condition handler returns.
|
|
|
|
*/
|
|
|
|
if (handle_errors_internally)
|
|
|
|
{ /* Currently, the only callers of tp_restart with handle_errors_internally set to TRUE are
|
|
|
|
* "mdb_condition_handler", "updproc.c" and "mupip_recover.c". All of those have SIGNAL set
|
|
|
|
* to ERR_TPRETRY so assert that. This is one way of protecting against a new caller of tp_restart
|
|
|
|
* inadvertently using a TRUE value for "handle_errors_internally". One reason for being paranoid
|
|
|
|
* about the TRUE usage is for example in gv_trigger.c, if tp_restart is incorrectly invoked with
|
|
|
|
* TRUE as the second paramter, it will result in indefinite number of cores on a broken database.
|
|
|
|
* In VMS, SIGNAL can be used only if we are inside a condition handler so do the assert only in Unix.
|
|
|
|
*/
|
|
|
|
UNIX_ONLY(assert(ERR_TPRETRY == SIGNAL));
|
|
|
|
ESTABLISH_RET(tp_restart_ch, tprestart_rc);
|
|
|
|
}
|
|
|
|
assert(1 == newlevel);
|
|
|
|
if (!dollar_tlevel)
|
|
|
|
{
|
|
|
|
rts_error(VARLSTCNT(1) ERR_TLVLZERO);
|
|
|
|
return 0; /* for the compiler only -- never executed */
|
|
|
|
}
|
|
|
|
# ifdef GTM_TRIGGER
|
|
|
|
DBGTRIGR((stderr, "tp_restart: Entry state: %d\n", tprestart_state));
|
|
|
|
if (TPRESTART_STATE_NORMAL == tprestart_state)
|
|
|
|
{ /* Only do if a normal invocation - otherwise we've already done this code for this TP restart */
|
|
|
|
# endif
|
|
|
|
/* Increment restart counts for each region in this transaction */
|
2012-03-24 14:06:46 -04:00
|
|
|
for (tr = tp_reg_list; NULL != tr; tr = tr->fPtr)
|
2012-02-05 11:35:58 -05:00
|
|
|
{
|
|
|
|
reg = tr->reg;
|
|
|
|
if (reg->open)
|
|
|
|
{
|
|
|
|
csa = &FILE_INFO(reg)->s_addrs;
|
|
|
|
switch (dollar_trestart)
|
|
|
|
{
|
|
|
|
case 0:
|
|
|
|
INCR_GVSTATS_COUNTER(csa, csa->nl, n_tp_tot_retries_0, 1);
|
|
|
|
break;
|
|
|
|
case 1:
|
|
|
|
INCR_GVSTATS_COUNTER(csa, csa->nl, n_tp_tot_retries_1, 1);
|
|
|
|
break;
|
|
|
|
case 2:
|
|
|
|
INCR_GVSTATS_COUNTER(csa, csa->nl, n_tp_tot_retries_2, 1);
|
|
|
|
break;
|
|
|
|
case 3:
|
|
|
|
INCR_GVSTATS_COUNTER(csa, csa->nl, n_tp_tot_retries_3, 1);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
INCR_GVSTATS_COUNTER(csa, csa->nl, n_tp_tot_retries_4, 1);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
} else
|
|
|
|
{
|
|
|
|
assert(cdb_sc_needcrit == t_fail_hist[t_tries]);
|
|
|
|
assert(!num_closed); /* we can have at the most 1 region not opened in the whole tp_reg_list */
|
|
|
|
num_closed++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
status = t_fail_hist[t_tries];
|
2012-10-29 18:54:31 -04:00
|
|
|
TREF(prev_t_tries) = t_tries;
|
|
|
|
/* Even though rollback and recover operate standalone, there are certain kind of restarts that can still happen due
|
|
|
|
* to white box test cases. Assert accordingly.
|
2012-03-24 14:06:46 -04:00
|
|
|
*/
|
2012-10-29 18:54:31 -04:00
|
|
|
assert(!mupip_jnl_recover || WB_COMMIT_ERR_ENABLED ||
|
|
|
|
(WBTEST_TP_HIST_CDB_SC_BLKMOD == gtm_white_box_test_case_number));
|
2012-03-24 14:06:46 -04:00
|
|
|
if (TREF(tprestart_syslog_delta) && (((TREF(tp_restart_count))++ < TREF(tprestart_syslog_limit))
|
|
|
|
|| (0 == ((TREF(tp_restart_count) - TREF(tprestart_syslog_limit)) % TREF(tprestart_syslog_delta)))))
|
2012-02-05 11:35:58 -05:00
|
|
|
{
|
|
|
|
if (NULL != tp_fail_hist[t_tries])
|
|
|
|
gvname_mstr = tp_fail_hist[t_tries]->gvname.var_name;
|
|
|
|
else
|
|
|
|
{
|
|
|
|
gvname_mstr.addr = (char *)gvname_unknown;
|
|
|
|
gvname_mstr.len = gvname_unknown_len;
|
|
|
|
}
|
|
|
|
caller_id_flag = FALSE; /* don't want caller_id in the operator log */
|
|
|
|
assert(0 == cdb_sc_normal);
|
|
|
|
if (cdb_sc_normal == status)
|
|
|
|
t_fail_hist[t_tries] = '0'; /* temporarily reset just for pretty printing */
|
|
|
|
restart_reg = tp_fail_hist_reg[t_tries];
|
|
|
|
if (NULL != restart_reg)
|
|
|
|
{
|
|
|
|
reg_mstr.len = restart_reg->dyn.addr->fname_len;
|
|
|
|
reg_mstr.addr = (char *)restart_reg->dyn.addr->fname;
|
|
|
|
} else
|
|
|
|
{
|
|
|
|
reg_mstr.len = 0;
|
|
|
|
reg_mstr.addr = NULL;
|
|
|
|
}
|
|
|
|
if (cdb_sc_blkmod != status)
|
|
|
|
{
|
|
|
|
send_msg(VARLSTCNT(16) ERR_TPRESTART, 14, reg_mstr.len, reg_mstr.addr,
|
|
|
|
t_tries + 1, t_fail_hist, t_fail_hist_blk[t_tries], gvname_mstr.len, gvname_mstr.addr,
|
|
|
|
0, 0, 0, tp_blkmod_nomod,
|
|
|
|
(NULL != sgm_info_ptr) ? sgm_info_ptr->num_of_blks : 0,
|
|
|
|
(NULL != sgm_info_ptr) ? sgm_info_ptr->cw_set_depth : 0, &local_tn);
|
|
|
|
} else
|
|
|
|
{
|
|
|
|
send_msg(VARLSTCNT(16) ERR_TPRESTART, 14, reg_mstr.len, reg_mstr.addr,
|
|
|
|
t_tries + 1, t_fail_hist, t_fail_hist_blk[t_tries], gvname_mstr.len, gvname_mstr.addr,
|
|
|
|
n_pvtmods, n_blkmods, tp_fail_level, tp_fail_n,
|
|
|
|
sgm_info_ptr->num_of_blks,
|
|
|
|
sgm_info_ptr->cw_set_depth, &local_tn);
|
|
|
|
}
|
|
|
|
tp_fail_hist_reg[t_tries] = NULL;
|
|
|
|
tp_fail_hist[t_tries] = NULL;
|
|
|
|
if ('0' == t_fail_hist[t_tries])
|
|
|
|
t_fail_hist[t_tries] = cdb_sc_normal; /* get back to where it was */
|
|
|
|
caller_id_flag = TRUE;
|
|
|
|
n_pvtmods = n_blkmods = 0;
|
|
|
|
}
|
|
|
|
/* Should never come here with a normal restart code */
|
|
|
|
assert(cdb_sc_normal != status);
|
|
|
|
# ifdef DEBUG
|
|
|
|
TAREF1(tp_restart_failhist_arry, (TREF(tp_restart_failhist_indx))++) = status;
|
|
|
|
if (FAIL_HIST_ARRAY_SIZE <= TREF(tp_restart_failhist_indx))
|
|
|
|
TREF(tp_restart_failhist_indx) = 0;
|
2012-03-24 14:06:46 -04:00
|
|
|
TRACE_TRANS_RESTART(status);
|
2012-02-05 11:35:58 -05:00
|
|
|
# endif
|
|
|
|
/* the following code is similar, but not identical, to code in t_retry,
|
|
|
|
* which should also be maintained in parallel
|
|
|
|
*/
|
|
|
|
switch (status)
|
|
|
|
{
|
|
|
|
case cdb_sc_helpedout:
|
|
|
|
csa = sgm_info_ptr->tp_csa;
|
|
|
|
if (dba_bg == csa->hdr->acc_meth)
|
|
|
|
{
|
|
|
|
if (!csa->now_crit)
|
|
|
|
{ /* The following grab/rel crit logic is purely to ensure that wcs_recover
|
|
|
|
* gets called if needed. This is because we saw wc_blocked to be TRUE in
|
|
|
|
* tp_tend and decided to restart.
|
|
|
|
*/
|
|
|
|
assert(!csa->hold_onto_crit);
|
|
|
|
grab_crit(sgm_info_ptr->gv_cur_region);
|
|
|
|
rel_crit(sgm_info_ptr->gv_cur_region);
|
|
|
|
} else
|
|
|
|
{ /* Some non-crit holding process set wc_blocked to TRUE causing us to
|
|
|
|
* restart even though we held crit. Most likely phase2 commit or a
|
|
|
|
* process in wcs_wtstart encountered an error. In any case, need to run
|
|
|
|
* cache-recovery to fix the shared memory structures. Since we hold crit,
|
|
|
|
* so no need to grab/rel crit. Call wcs_recover right away.
|
|
|
|
*/
|
2012-03-24 14:06:46 -04:00
|
|
|
assert(!csa->hold_onto_crit UNIX_ONLY(|| jgbl.onlnrlbk));
|
2012-02-05 11:35:58 -05:00
|
|
|
DEBUG_ONLY(TREF(ok_to_call_wcs_recover) = TRUE);
|
|
|
|
wcs_recover(sgm_info_ptr->gv_cur_region);
|
|
|
|
DEBUG_ONLY(TREF(ok_to_call_wcs_recover) = FALSE);
|
|
|
|
}
|
|
|
|
} else
|
|
|
|
{
|
|
|
|
assert(dba_mm == csa->hdr->acc_meth);
|
|
|
|
wcs_recover(sgm_info_ptr->gv_cur_region);
|
|
|
|
}
|
|
|
|
if (CDB_STAGNATE > t_tries)
|
|
|
|
{
|
|
|
|
t_tries++;
|
|
|
|
break;
|
|
|
|
}
|
2012-10-29 18:54:31 -04:00
|
|
|
# ifdef DEBUG
|
|
|
|
if (0 <= TREF(tp_restart_dont_counts)) /* skip increment on negative from TPNOTACID */
|
|
|
|
(TREF(tp_restart_dont_counts))++;
|
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
/* WARNING - fallthrough !!! */
|
|
|
|
case cdb_sc_needcrit:
|
|
|
|
/* Here when a final (4th) attempt has failed with a need for crit in some routine. The
|
|
|
|
* assumption is that the previous attempt failed somewhere before transaction end
|
|
|
|
* therefore tp_reg_list did not have a complete list of regions necessary to complete the
|
|
|
|
* transaction and therefore not all the regions have been locked down. The new region (by
|
|
|
|
* virtue of it having now been referenced) has been added to tp_reg_list so all we need
|
|
|
|
* now is a retry.
|
|
|
|
*/
|
|
|
|
assert(CDB_STAGNATE == t_tries);
|
2012-03-24 14:06:46 -04:00
|
|
|
for (tr = tp_reg_list; NULL != tr; tr = tr->fPtr)
|
2012-02-05 11:35:58 -05:00
|
|
|
{ /* regions might not have been opened if we t_retried in gvcst_init(). dont
|
|
|
|
* rel_crit in that case.
|
|
|
|
*/
|
|
|
|
reg = tr->reg;
|
|
|
|
if (reg->open)
|
|
|
|
{
|
|
|
|
DEBUG_ONLY(csa = &FILE_INFO(reg)->s_addrs;)
|
|
|
|
assert(!csa->hold_onto_crit);
|
|
|
|
rel_crit(reg); /* to ensure deadlock safe order, release all regions
|
2012-03-24 14:06:46 -04:00
|
|
|
* before retry */
|
2012-02-05 11:35:58 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
DEBUG_ONLY(
|
|
|
|
/* The journal pool crit lock is currently obtained only inside commit logic at
|
2012-10-29 18:54:31 -04:00
|
|
|
* which point we will never signal a cdb_sc_needcrit restart code.
|
|
|
|
* So no need to verify if we need to release crit there. Assert this though.
|
2012-02-05 11:35:58 -05:00
|
|
|
*/
|
|
|
|
if ((NULL != jnlpool.jnlpool_dummy_reg) && jnlpool.jnlpool_dummy_reg->open)
|
|
|
|
{
|
|
|
|
csa = &FILE_INFO(jnlpool.jnlpool_dummy_reg)->s_addrs;
|
|
|
|
assert(!csa->now_crit);
|
|
|
|
}
|
|
|
|
)
|
|
|
|
/* If retry due to M-locks, sleep so needed locks have a chance to get released */
|
|
|
|
break;
|
|
|
|
/* Journaling might get turned off in the final retry INSIDE crit while trying to flush journal buffer or
|
|
|
|
* during extending the journal file (due to possible disk issues) in which case we will come here with
|
|
|
|
* t_tries = CDB_STAGNATE and failure status set to cdb_sc_jnlclose
|
|
|
|
*/
|
|
|
|
case cdb_sc_jnlclose:
|
|
|
|
/* cdb_sc_jnlstatemod is expected in final retry because csa->jnl_state is noted from csd->jnl_state only
|
|
|
|
* if they are different INSIDE crit. Therefore it is possible that in the final retry one might start with
|
|
|
|
* a stale value of csa->jnl_state which will be noticed only in t_end just before commit as a result of
|
|
|
|
* which we would restart. Such a restart is okay (instead of the checking for jnl state change during the
|
|
|
|
* beginning of final retry) since jnl state changes are considered infrequent that too in the final retry
|
|
|
|
*/
|
|
|
|
case cdb_sc_jnlstatemod:
|
2012-03-24 14:06:46 -04:00
|
|
|
# ifdef UNIX
|
|
|
|
/* cdb_sc_onln_rlbk[1,2] are possible in the final retry. See comment below that explains why. So, decrement
|
|
|
|
* t_tries to account for later increment
|
|
|
|
*/
|
|
|
|
case cdb_sc_onln_rlbk1:
|
|
|
|
case cdb_sc_onln_rlbk2:
|
2012-10-29 18:54:31 -04:00
|
|
|
case cdb_sc_instancefreeze:
|
|
|
|
case cdb_sc_gvtrootmod2:
|
2012-03-24 14:06:46 -04:00
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
case cdb_sc_optrestart:
|
|
|
|
if (CDB_STAGNATE <= t_tries)
|
|
|
|
{
|
|
|
|
t_tries--;
|
2012-10-29 18:54:31 -04:00
|
|
|
# ifdef DEBUG
|
|
|
|
if (0 <= TREF(tp_restart_dont_counts)) /* skip increment on negative from TPNOTACID */
|
|
|
|
(TREF(tp_restart_dont_counts))++;
|
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
}
|
2012-10-29 18:54:31 -04:00
|
|
|
# ifdef UNIX
|
|
|
|
if (cdb_sc_instancefreeze == status)
|
|
|
|
WAIT_FOR_REPL_INST_UNFREEZE_NOCSA;
|
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
/* fall through */
|
|
|
|
default:
|
|
|
|
if (CDB_STAGNATE < ++t_tries)
|
|
|
|
{
|
|
|
|
hist_index = t_tries;
|
|
|
|
t_tries = 0;
|
2012-03-24 14:06:46 -04:00
|
|
|
assert(0 != have_crit(CRIT_HAVE_ANY_REG)); /* we should still be holding crit */
|
2012-02-05 11:35:58 -05:00
|
|
|
assert(gtm_white_box_test_case_enabled
|
2012-03-24 14:06:46 -04:00
|
|
|
&& (WBTEST_TP_HIST_CDB_SC_BLKMOD == gtm_white_box_test_case_number));
|
|
|
|
# ifdef UNIX
|
2012-02-05 11:35:58 -05:00
|
|
|
send_msg(VARLSTCNT(5) ERR_TPFAIL, 2, hist_index, t_fail_hist, ERR_GVFAILCORE);
|
|
|
|
/* Generate core only if not triggering this codepath using white-box tests */
|
2012-03-24 14:06:46 -04:00
|
|
|
DEBUG_ONLY(
|
|
|
|
if (!gtm_white_box_test_case_enabled
|
|
|
|
|| (WBTEST_TP_HIST_CDB_SC_BLKMOD != gtm_white_box_test_case_number))
|
|
|
|
)
|
|
|
|
gtm_fork_n_core();
|
2012-02-05 11:35:58 -05:00
|
|
|
# endif
|
|
|
|
VMS_ONLY(send_msg(VARLSTCNT(4) ERR_TPFAIL, 2, hist_index, t_fail_hist));
|
|
|
|
rts_error(VARLSTCNT(4) ERR_TPFAIL, 2, hist_index, t_fail_hist);
|
|
|
|
return 0; /* for the compiler only -- never executed */
|
|
|
|
} else
|
|
|
|
{
|
|
|
|
/* Yield the CPU so that the restarting process does not block the crit holder
|
|
|
|
* and/or other processes (referencing potentially non-intersecting database
|
|
|
|
* regions) in case they are waiting for the CPU. This is done using the
|
|
|
|
* rel_quant function.
|
|
|
|
* As of this writing, this operates only between the 2nd and 3rd tries;
|
|
|
|
* The 2nd is fast with the assumption of coincidental conflict in an attempt
|
|
|
|
* to take advantage of the buffer state created by the 1st try.
|
|
|
|
* The next to last try is not followed by a rel_quant as it may leave the buffers
|
|
|
|
* locked, to reduce live lock and deadlock issues.
|
|
|
|
* With only 4 tries that leaves only the "middle" for rel_quant.
|
|
|
|
*/
|
|
|
|
if ((CDB_STAGNATE - 1) == t_tries)
|
|
|
|
rel_quant();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if ((CDB_STAGNATE <= t_tries))
|
|
|
|
{ /* If there are any regions that haven't yet been opened, open them before attempting for crit on
|
|
|
|
* all. Since we don't hold any crit locks now, we can rest assured this cannot cause a deadlock.
|
|
|
|
* The only exception (to holding crit) currently is mupip journal rollback/recovery if online when
|
2012-03-24 14:06:46 -04:00
|
|
|
* we will be holding crit on all regions but in that case we should NOT have restarted in the first
|
|
|
|
* place as all the concurrent GT.M processes will be hung waiting for crit. The only exception is
|
|
|
|
* if some process set wc_blocked outside crit and we restarted to do cache recovery (which is asserted
|
|
|
|
* above)
|
2012-02-05 11:35:58 -05:00
|
|
|
*/
|
|
|
|
if (num_closed)
|
|
|
|
{
|
2012-03-24 14:06:46 -04:00
|
|
|
for (tr = tp_reg_list; NULL != tr; tr = tr->fPtr)
|
2012-02-05 11:35:58 -05:00
|
|
|
{ /* to open region use gv_init_reg() instead of gvcst_init() since that does extra
|
|
|
|
* manipulations with gv_keysize, gv_currkey and gv_altkey.
|
|
|
|
*/
|
|
|
|
reg = tr->reg;
|
|
|
|
if (!reg->open)
|
|
|
|
{
|
|
|
|
gv_init_reg(reg);
|
|
|
|
assert(reg->open);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
DBG_CHECK_TP_REG_LIST_SORTING(tp_reg_list);
|
|
|
|
}
|
|
|
|
DEBUG_ONLY(TREF(ok_to_call_wcs_recover) = TRUE);
|
2012-03-24 14:06:46 -04:00
|
|
|
UNIX_ONLY(if (!jgbl.onlnrlbk))
|
|
|
|
tp_tend_status = tp_crit_all_regions(); /* grab crits on all regions */
|
|
|
|
/* else online rollback in which case we already hold crit on all regions */
|
|
|
|
/* It is possible we came into tp_restart to handle a different restart code but as part of the final
|
|
|
|
* retry detected an Online Rollback. If so, we shouldn't go into the final retry with this out-of-sync
|
|
|
|
* state as tp_tend will detect this out-of-sync during validation causing restart in the final
|
|
|
|
* retry resulting in TPFAIL. So, check for a concurrent Online Rollback and handle it if needed. It is
|
|
|
|
* enough to check any region as Online Rollback will increment the cycle field ON all the regions
|
|
|
|
* unconditionally
|
|
|
|
*/
|
|
|
|
# ifdef UNIX
|
2012-10-29 18:54:31 -04:00
|
|
|
if (cs_addrs_list && MISMATCH_ONLN_RLBK_CYCLES(cs_addrs_list, cs_addrs_list->nl))
|
2012-03-24 14:06:46 -04:00
|
|
|
{ /* We came in to handle a different restart code in the penultimate retry and grab_crit before going
|
|
|
|
* to final retry. As part of grabbing crit, we detected an online rollback. Although we could treat
|
|
|
|
* this as just an online rollback restart and handle it by syncing cycles, but by doing so, we will
|
|
|
|
* loose the information that an online rollback happened when we go back to gvcst_{put,kill}. This
|
|
|
|
* is usually fine except when we are in implicit TP (due to triggers). In case of implicit TP,
|
|
|
|
* gvcst_{put,kill} has specific code to handle online rollback differently than other restart codes
|
|
|
|
* Because of this reason, we don't want to sync cycles but instead continue with the final retry.
|
|
|
|
* t_end/tp_tend/tp_hist will notice the cycle mismatch and will restart (once more) in final retry
|
|
|
|
* with the appropriate cdb_sc code which gvcst_put/gvcst_kill will intercept and act accordingly.
|
|
|
|
* Even if we are not syncing cycles, we need to reset the clues and root block numbers so that we
|
|
|
|
* proceeds smoothly in the final retry.
|
|
|
|
*/
|
|
|
|
RESET_ALL_GVT_CLUES;
|
2012-10-29 18:54:31 -04:00
|
|
|
reset_clues_done = TRUE;
|
2012-03-24 14:06:46 -04:00
|
|
|
}
|
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
DEBUG_ONLY(TREF(ok_to_call_wcs_recover) = FALSE);
|
|
|
|
assert(FALSE != tp_tend_status);
|
|
|
|
/* pick up all MM extension information */
|
|
|
|
for (tr = tp_reg_list; NULL != tr; tr = tr->fPtr)
|
|
|
|
{
|
|
|
|
reg = tr->reg;
|
|
|
|
if (dba_mm == reg->dyn.addr->acc_meth)
|
|
|
|
{
|
|
|
|
TP_CHANGE_REG_IF_NEEDED(reg);
|
|
|
|
MM_DBFILEXT_REMAP_IF_NEEDED(cs_addrs, gv_cur_region);
|
|
|
|
}
|
2012-10-29 18:54:31 -04:00
|
|
|
# ifdef UNIX
|
|
|
|
csa = (sgmnt_addrs *)&FILE_INFO(reg)->s_addrs;
|
|
|
|
if (MISMATCH_ROOT_CYCLES(csa, csa->nl) && !reset_clues_done)
|
|
|
|
{
|
|
|
|
RESET_ALL_GVT_CLUES;
|
|
|
|
reset_clues_done = TRUE;
|
|
|
|
}
|
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
# ifdef GTM_TRIGGER
|
2012-10-29 18:54:31 -04:00
|
|
|
} else
|
|
|
|
status = LAST_RESTART_CODE;
|
2012-02-05 11:35:58 -05:00
|
|
|
DBGTRIGR((stderr, "tp_restart: past initial normal state processing\n"));
|
|
|
|
# endif
|
|
|
|
/* The below code to determine the roll-back point depends on tp_frame sized blocks being pushed on the TP
|
|
|
|
* stack. If ever other sized blocks are pushed on, a different method will need to be found.
|
|
|
|
*/
|
|
|
|
assert(0 == ((tpstackbase - (unsigned char *)tp_pointer) % SIZEOF(tp_frame))); /* Simple check for above condition */
|
|
|
|
tf = (tp_frame *)(tpstackbase - (newlevel * SIZEOF(tp_frame)));
|
|
|
|
assert(NULL != tf);
|
|
|
|
assert(tpstacktop < (unsigned char *)tf);
|
|
|
|
# ifdef GTM_TRIGGER
|
|
|
|
if (TPRESTART_STATE_NORMAL == tprestart_state)
|
|
|
|
{ /* Only if normal tp_restart call - else we've already done this for this tp_restart */
|
|
|
|
# endif
|
|
|
|
/* Before we get too far unwound here, if this is a nonrestartable transaction,
|
2012-03-24 14:06:46 -04:00
|
|
|
* let's record where we are for the message later on.
|
|
|
|
*/
|
2012-02-05 11:35:58 -05:00
|
|
|
if (FALSE == tf->restartable && IS_MCODE_RUNNING)
|
|
|
|
getzposition(TADR(tp_restart_entryref));
|
|
|
|
/* Do a rollback type cleanup (invalidate gv_target clues of read as well as updated blocks) */
|
|
|
|
tp_clean_up(TRUE);
|
2012-03-24 14:06:46 -04:00
|
|
|
/* Note: At this point, we are ready to begin the next retry. While we can sync the trigger cycles now to avoid
|
|
|
|
* further restarts, we don't need to because tp_set_sgm (done for each region that is updated in a TP transaction)
|
|
|
|
* does the syncing anyways.
|
|
|
|
*/
|
|
|
|
# ifdef UNIX
|
|
|
|
assert(cdb_sc_normal != status);
|
2012-10-29 18:54:31 -04:00
|
|
|
switch (status)
|
|
|
|
{
|
|
|
|
case cdb_sc_gvtrootmod2: /* restarted due to MUPIP REORG moving root blocks */
|
|
|
|
RESET_ALL_GVT_CLUES;
|
|
|
|
break;
|
|
|
|
case cdb_sc_onln_rlbk1: /* restarted due to online rollback */
|
|
|
|
case cdb_sc_onln_rlbk2:
|
|
|
|
RESET_ALL_GVT_CLUES;
|
|
|
|
if (IS_MCODE_RUNNING && (cdb_sc_onln_rlbk2 == status))
|
|
|
|
(TREF(dollar_zonlnrlbk))++;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
2012-03-24 14:06:46 -04:00
|
|
|
}
|
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
# ifdef GTM_TRIGGER
|
|
|
|
}
|
|
|
|
if (TPRESTART_STATE_TPUNW >= tprestart_state)
|
|
|
|
{ /* Either this is a normal tp_restart call or we ran into a trigger base frame while "tp_unwind"
|
|
|
|
* was running which required M and C stack unwinding before we could proceed so this call is
|
|
|
|
* being restarted.
|
|
|
|
*/
|
|
|
|
# endif
|
|
|
|
/* Note that this form of "tp_unwind" will not only unwind the TP stack but also most if not all of
|
|
|
|
* the M stackframe and mv_stent chain as well.
|
|
|
|
*/
|
|
|
|
GTMTRIG_ONLY(DBGTRIGR((stderr, "tp_restart: Beginning state 0/1 processing (state %d)\n", tprestart_state)));
|
|
|
|
tp_unwind(newlevel, RESTART_INVOCATION, &tprestart_rc);
|
|
|
|
assert(tf == tp_pointer); /* Needs to be true for now. Revisit when can restart to other than newlevel == 1 */
|
|
|
|
gd_header = tp_pointer->gd_header;
|
|
|
|
gv_target = tp_pointer->orig_gv_target;
|
|
|
|
gv_cur_region = tp_pointer->gd_reg;
|
|
|
|
TP_CHANGE_REG(gv_cur_region);
|
|
|
|
DBG_CHECK_GVTARGET_CSADDRS_IN_SYNC;
|
|
|
|
dollar_tlevel = newlevel;
|
2012-10-29 18:54:31 -04:00
|
|
|
COPY_KEY(gv_currkey, tp_pointer->orig_key);
|
2012-02-05 11:35:58 -05:00
|
|
|
DBG_CHECK_GVTARGET_GVCURRKEY_IN_SYNC;
|
|
|
|
# ifdef GTM_TRIGGER
|
2012-10-29 18:54:31 -04:00
|
|
|
/* Maintenance of SFF_IMPLTSTART_CALLD stack frame flag:
|
2012-02-05 11:35:58 -05:00
|
|
|
* - Set by gtm_trigger when trigger base frame is created. Purpose to prevent MUM_TSTART from restarting
|
|
|
|
* a frame making a call-in to a trigger (flag is checked in MUM_TSTART macro) because the mpc in the
|
|
|
|
* stack frame is not the return point to the frame, which is only available in the C stack.
|
|
|
|
* - Both TP restart and error handling unwinds can use MUM_TSTART to restart frame.
|
|
|
|
* - TP restart changes the mpc to the proper address (where TSTART was done) before invoking MUM_TSTART. We allow
|
2012-10-29 18:54:31 -04:00
|
|
|
* this by shutting the SFF_IMPLTSTART_CALLD flag off when mpc is changed.
|
2012-02-05 11:35:58 -05:00
|
|
|
* - For TSTARTs done implcitly by triggers, MUM_TSTART would break things so we do not turn off the flag
|
|
|
|
* for that type.
|
|
|
|
*/
|
|
|
|
if (!tp_pointer->implicit_tstart)
|
2012-10-29 18:54:31 -04:00
|
|
|
{ /* SFF_IMPLTSTART_CALLD validation:
|
2012-02-05 11:35:58 -05:00
|
|
|
* - This is not a trigger-initiated implicit TSTART.
|
|
|
|
* - If the flag is is not on, no further checks. Turning off flag is unconditional for best performance.
|
|
|
|
* - If flag is on, verify the address in the stack frame is in fact being modified so it points to
|
|
|
|
* a TSTART instead of the (currently) trigger call point.
|
|
|
|
*/
|
2012-10-29 18:54:31 -04:00
|
|
|
assert(!(tp_pointer->fp->flags & SFF_IMPLTSTART_CALLD) || (tp_pointer->fp->mpc != tp_pointer->restart_pc));
|
|
|
|
tp_pointer->fp->flags &= SFF_IMPLTSTART_CALLD_OFF;
|
|
|
|
DBGTRIGR((stderr, "tp_restart: Removing SFF_IMPLTSTART_CALLD in frame 0x"lvaddr"\n", tp_pointer->fp));
|
2012-02-05 11:35:58 -05:00
|
|
|
}
|
|
|
|
# endif
|
|
|
|
tp_pointer->fp->mpc = tp_pointer->restart_pc;
|
|
|
|
tp_pointer->fp->ctxt = tp_pointer->restart_ctxt;
|
|
|
|
# ifdef GTM_TRIGGER
|
|
|
|
} else
|
|
|
|
assert(TPRESTART_STATE_MSTKUNW == tprestart_state);
|
|
|
|
/* From here on, all states run */
|
|
|
|
# endif
|
|
|
|
/* Make sure everything else added to the stack since the transaction started is unwound. Note this loop only
|
|
|
|
* has work to do if there were NO local vars to restore. Otherwise tp_unwind would have already unwound the
|
|
|
|
* stack for us.
|
|
|
|
*/
|
|
|
|
GTMTRIG_ONLY(DBGTRIGR((stderr, "tp_restart: beginning frame unwinds (state %d)\n", tprestart_state)));
|
|
|
|
while (frame_pointer < tf->fp)
|
|
|
|
{
|
|
|
|
# ifdef GTM_TRIGGER
|
|
|
|
if (SFT_TRIGR & frame_pointer->type)
|
|
|
|
{ /* We have encountered a trigger base frame. We cannot unroll it because there are C frames
|
|
|
|
* associated with it so we must interrupt this tp_restart and return to gtm_trigger() so
|
|
|
|
* it can unroll the base frame and rethrow the error to properly unroll the C stack.
|
|
|
|
*/
|
|
|
|
tprestart_rc = ERR_TPRETRY;
|
|
|
|
tprestart_state = TPRESTART_STATE_MSTKUNW;
|
|
|
|
DBGTRIGR((stderr, "tp_restart: Encountered trigger base frame in M-stack unwind - rethrowing\n"));
|
|
|
|
INVOKE_RESTART;
|
|
|
|
}
|
|
|
|
# endif
|
|
|
|
op_unwind();
|
|
|
|
}
|
|
|
|
/* From here on, no further rethrows of tp_restart() - the final finishing touches */
|
|
|
|
assert((msp <= stackbase) && (msp > stacktop));
|
|
|
|
assert((mv_chain <= (mv_stent *)stackbase) && (mv_chain > (mv_stent *)stacktop));
|
|
|
|
assert(MVST_TPHOLD == tf->mvc->mv_st_type);
|
|
|
|
/* Our stack frames are unwound to the correct frame but there could be mv_stents pushed on after we last (re)started
|
|
|
|
* this transaction. We need to get rid of them to get back to the correct restart state.
|
|
|
|
*/
|
2012-03-24 14:06:46 -04:00
|
|
|
for (mvc = mv_chain; mvc < tf->mvc;)
|
2012-02-05 11:35:58 -05:00
|
|
|
{ /* Make sure we don't unwind the MVST_TPHOLD for our target level */
|
|
|
|
assert((MVST_TPHOLD != mvc->mv_st_type) || ((newlevel - 1) != mvc->mv_st_cont.mvs_tp_holder.tphold_tlevel));
|
|
|
|
DBGEHND((stderr, "tp_restart: unwinding mv_stent addr 0x"lvaddr" type %d\n", mvc, mvc->mv_st_type));
|
|
|
|
unw_mv_ent(mvc);
|
|
|
|
mvc = (mv_stent *)(mvc->mv_st_next + (char *)mvc);
|
|
|
|
}
|
|
|
|
assert((void *)mvc < (void *)frame_pointer);
|
|
|
|
assert(mvc == tf->mvc);
|
|
|
|
assert(mvc->mv_st_cont.mvs_tp_holder.tphold_tlevel == (dollar_tlevel - 1));
|
|
|
|
DBGEHND((stderr, "tp_restart: Resetting msp from 0x"lvaddr", to 0x"lvaddr" (diff=%d)\n",
|
|
|
|
msp, mvc, INTCAST((unsigned char *)mvc - msp)));
|
|
|
|
mv_chain = mvc;
|
|
|
|
msp = (unsigned char *)mvc;
|
|
|
|
# ifdef GTM_TRIGGER
|
|
|
|
/* Revert $ZTWormhole to its previous value */
|
|
|
|
DBGTRIGR((stderr, "tp_restart: Restoring $ZTWORMHOLE and NULLifying $ZTSLATE (state %d)\n", tprestart_state));
|
|
|
|
memcpy(&dollar_ztwormhole, &mvc->mv_st_cont.mvs_tp_holder.ztwormhole_save, SIZEOF(mval));
|
|
|
|
if (1 == newlevel)
|
|
|
|
memcpy(&dollar_ztslate, &literal_null, SIZEOF(mval)); /* Zap $ZTSLate at (re)start of lvl 1 transaction */
|
|
|
|
# endif
|
|
|
|
assert(curr_symval == tf->sym);
|
|
|
|
if (frame_pointer->flags & SFF_UNW_SYMVAL)
|
|
|
|
{ /* A symval was popped in THIS stackframe by one of our last mv_stent unwinds which means
|
|
|
|
* l_symtab is fairly borked.
|
|
|
|
*/
|
|
|
|
assert(frame_pointer->l_symtab); /* Would be NULL in replication processor */
|
|
|
|
if ((unsigned char *)frame_pointer->l_symtab < msp)
|
|
|
|
{ /* This condition is set up when a local routine is called which, since it is using the
|
|
|
|
* same code, uses the same l_symtab as the caller. But when an exclusive new is done in
|
|
|
|
* this frame, op_xnew creates a NEW symtab just for this frame. But when this code
|
|
|
|
* unwound back to the TSTART, we also unwound the l_symtab this frame was using. So here
|
|
|
|
* we verify this frame is a simple call frame from the previous and restore the use of its
|
|
|
|
* l_symtab if so. If not, GTMASSERT. Note the outer SFF_UWN_SYMVAL check keeps us from having
|
|
|
|
* non-existant l_symtab issues which is possible when we are MUPIP.
|
|
|
|
*/
|
|
|
|
if ((frame_pointer->rvector == frame_pointer->old_frame_pointer->rvector)
|
|
|
|
&& (frame_pointer->vartab_ptr == frame_pointer->old_frame_pointer->vartab_ptr))
|
|
|
|
{
|
|
|
|
frame_pointer->l_symtab = frame_pointer->old_frame_pointer->l_symtab;
|
|
|
|
frame_pointer->flags &= SFF_UNW_SYMVAL_OFF; /* No need to clear symtab now */
|
|
|
|
} else
|
|
|
|
GTMASSERT;
|
|
|
|
} else
|
|
|
|
{ /* Otherwise the l_symtab needs to be cleared so its references get re-resolved to *this* symtab */
|
|
|
|
memset(frame_pointer->l_symtab, 0, frame_pointer->vartab_len * SIZEOF(ht_ent_mname *));
|
|
|
|
frame_pointer->flags &= SFF_UNW_SYMVAL_OFF;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
dollar_truth = tp_pointer->dlr_t;
|
|
|
|
dollar_zgbldir = tp_pointer->zgbldir;
|
|
|
|
GTMTRIG_ONLY(tprestart_state = TPRESTART_STATE_NORMAL);
|
|
|
|
if (FALSE == tf->restartable)
|
|
|
|
{ /* Transation is not restartable. Be sure to leave things in a state that the transaction could
|
|
|
|
* be continued if an error handler has a mind to do that.
|
|
|
|
*/
|
|
|
|
GTMTRIG_ONLY(DBGTRIGR((stderr, "tp_restart: Leaving tp_restart via TRESTNOT error - state reset to 0\n")));
|
|
|
|
if (IS_MCODE_RUNNING)
|
|
|
|
{
|
|
|
|
getzposition(&beganHere);
|
|
|
|
send_msg(VARLSTCNT(1) ERR_TRESTNOT); /* Separate msgs so we get both */
|
|
|
|
send_msg(VARLSTCNT(6) ERR_TRESTLOC, 4, beganHere.str.len, beganHere.str.addr,
|
|
|
|
(TREF(tp_restart_entryref)).str.len, (TREF(tp_restart_entryref)).str.addr);
|
|
|
|
rts_error(VARLSTCNT(8) ERR_TRESTNOT, 0, ERR_TRESTLOC, 4, beganHere.str.len, beganHere.str.addr,
|
|
|
|
(TREF(tp_restart_entryref)).str.len, (TREF(tp_restart_entryref)).str.addr);
|
|
|
|
} else
|
|
|
|
rts_error(VARLSTCNT(1) ERR_TRESTNOT);
|
|
|
|
return 0; /* for the compiler only -- never executed */
|
|
|
|
}
|
|
|
|
++dollar_trestart;
|
2012-10-29 18:54:31 -04:00
|
|
|
# ifdef DEBUG
|
|
|
|
if (0 > TREF(tp_restart_dont_counts)) /* negative is a flag from TPNOTACID - do increment here */
|
|
|
|
{
|
|
|
|
TREF(tp_restart_dont_counts) = -TREF(tp_restart_dont_counts);
|
|
|
|
(TREF(tp_restart_dont_counts))++;
|
|
|
|
}
|
|
|
|
assert(0 <= TREF(tp_restart_dont_counts)); /* NOTE: should tp_restart_dont_counts wrap, asserts will fail */
|
2012-02-05 11:35:58 -05:00
|
|
|
assert(dollar_trestart >= TREF(tp_restart_dont_counts));
|
|
|
|
assert(MAX_TRESTARTS > (dollar_trestart - TREF(tp_restart_dont_counts))); /* a magic number limit for restarts */
|
2012-10-29 18:54:31 -04:00
|
|
|
# endif
|
2012-02-05 11:35:58 -05:00
|
|
|
if (!dollar_trestart) /* in case of a wrap */
|
|
|
|
dollar_trestart--;
|
2012-03-24 14:06:46 -04:00
|
|
|
UNIX_ONLY(
|
|
|
|
/* Now that we are done with all the cleanup related to this restart, issue rts_error if we are update process.
|
|
|
|
* updproc_ch knows to handle this SIGNAL.
|
|
|
|
*/
|
2012-10-29 18:54:31 -04:00
|
|
|
assert(cdb_sc_normal != status);
|
|
|
|
if (is_updproc && ((cdb_sc_onln_rlbk1 == status) || (cdb_sc_onln_rlbk2 == status)))
|
2012-03-24 14:06:46 -04:00
|
|
|
rts_error(VARLSTCNT(1) ERR_REPLONLNRLBK);
|
|
|
|
)
|
2012-02-05 11:35:58 -05:00
|
|
|
if (handle_errors_internally)
|
|
|
|
REVERT;
|
|
|
|
GTMTRIG_ONLY(DBGTRIGR((stderr, "tp_restart: completed\n")));
|
|
|
|
return 0;
|
|
|
|
}
|