about summary refs log tree commit diff
path: root/db2/hash/hash.c
diff options
context:
space:
mode:
Diffstat (limited to 'db2/hash/hash.c')
-rw-r--r--db2/hash/hash.c135
1 files changed, 83 insertions, 52 deletions
diff --git a/db2/hash/hash.c b/db2/hash/hash.c
index d986e08087..c08495378e 100644
--- a/db2/hash/hash.c
+++ b/db2/hash/hash.c
@@ -47,7 +47,7 @@
 #include "config.h"
 
 #ifndef lint
-static const char sccsid[] = "@(#)hash.c	10.27 (Sleepycat) 9/15/97";
+static const char sccsid[] = "@(#)hash.c	10.33 (Sleepycat) 11/2/97";
 #endif /* not lint */
 
 #ifndef NO_SYSTEM_INCLUDES
@@ -79,7 +79,7 @@ static int  __ham_cursor __P((DB *, DB_TXN *, DBC **));
 static int  __ham_delete __P((DB *, DB_TXN *, DBT *, int));
 static int  __ham_dup_return __P((HTAB *, HASH_CURSOR *, DBT *, int));
 static int  __ham_get __P((DB *, DB_TXN *, DBT *, DBT *, int));
-static void __ham_init_htab __P((HTAB *));
+static void __ham_init_htab __P((HTAB *, u_int));
 static int  __ham_lookup __P((HTAB *,
 		HASH_CURSOR *, const DBT *, u_int32_t, db_lockmode_t));
 static int  __ham_overwrite __P((HTAB *, HASH_CURSOR *, DBT *));
@@ -106,7 +106,7 @@ __ham_open(dbp, dbinfo)
 
 	dbenv = dbp->dbenv;
 
-	if ((hashp = (HTAB *)calloc(1, sizeof(HTAB))) == NULL)
+	if ((hashp = (HTAB *)__db_calloc(1, sizeof(HTAB))) == NULL)
 		return (ENOMEM);
 	hashp->dbp = dbp;
 
@@ -175,10 +175,9 @@ __ham_open(dbp, dbinfo)
 			goto out;
 		}
 
-		hashp->hdr->nelem = dbinfo != NULL ? dbinfo->h_nelem : 0;
 		hashp->hdr->ffactor =
 		    dbinfo != NULL && dbinfo->h_ffactor ? dbinfo->h_ffactor : 0;
-		__ham_init_htab(hashp);
+		__ham_init_htab(hashp, dbinfo != NULL ? dbinfo->h_nelem : 0);
 		if (F_ISSET(dbp, DB_AM_DUP))
 			F_SET(hashp->hdr, DB_HASH_DUP);
 		if ((ret = __ham_dirty_page(hashp, (PAGE *)hashp->hdr)) != 0)
@@ -190,7 +189,7 @@ __ham_open(dbp, dbinfo)
 	TAILQ_INSERT_TAIL(&dbp->curs_queue, curs, links);
 
 	/* Allocate memory for our split buffer. */
-	if ((hashp->split_buf = (PAGE *)malloc(dbp->pgsize)) == NULL) {
+	if ((hashp->split_buf = (PAGE *)__db_malloc(dbp->pgsize)) == NULL) {
 		ret = ENOMEM;
 		goto out;
 	}
@@ -265,13 +264,13 @@ __ham_close(dbp)
  * Returns 0 on No Error
  */
 static void
-__ham_init_htab(hashp)
+__ham_init_htab(hashp, nelem)
 	HTAB *hashp;
+	u_int nelem;
 {
-	u_int32_t nelem;
 	int32_t l2, nbuckets;
 
-	nelem = hashp->hdr->nelem;
+	hashp->hdr->nelem = 0;
 	hashp->hdr->pagesize = hashp->dbp->pgsize;
 	ZERO_LSN(hashp->hdr->lsn);
 	hashp->hdr->magic = DB_HASHMAGIC;
@@ -502,11 +501,11 @@ __ham_c_init(dbp, txnid, dbcp)
 	DBC *db_curs;
 	HASH_CURSOR *new_curs;
 
-	if ((db_curs = (DBC *)calloc(sizeof(DBC), 1)) == NULL)
+	if ((db_curs = (DBC *)__db_calloc(sizeof(DBC), 1)) == NULL)
 		return (ENOMEM);
 
 	if ((new_curs =
-	    (HASH_CURSOR *)calloc(sizeof(struct cursor_t), 1)) == NULL) {
+	    (HASH_CURSOR *)__db_calloc(sizeof(struct cursor_t), 1)) == NULL) {
 		FREE(db_curs, sizeof(DBC));
 		return (ENOMEM);
 	}
@@ -555,7 +554,7 @@ __ham_delete(dbp, txn, key, flags)
 	hashp->hash_accesses++;
 	if ((ret = __ham_lookup(hashp, hcp, key, 0, DB_LOCK_WRITE)) == 0)
 		if (F_ISSET(hcp, H_OK))
-			ret = __ham_del_pair(hashp, hcp);
+			ret = __ham_del_pair(hashp, hcp, 1);
 		else
 			ret = DB_NOTFOUND;
 
@@ -669,30 +668,41 @@ __ham_c_del(cursor, flags)
 	if ((ret = __ham_get_cpage(hashp, hcp, DB_LOCK_WRITE)) != 0)
 		goto out;
 	if (F_ISSET(hcp, H_ISDUP) && hcp->dpgno != PGNO_INVALID) {
-		ppgno = PREV_PGNO(hcp->dpagep);
-
-		/* Remove item from duplicate page. */
-		chg_pgno = hcp->dpgno;
-		if ((ret = __db_drem(hashp->dbp,
-		    &hcp->dpagep, hcp->dndx, __ham_del_page)) != 0)
-			goto out;
-
 		/*
+		 * We are about to remove a duplicate from offpage.
+		 *
 		 * There are 4 cases.
-		 * 1. We removed an item on a page, but nothing else changed.
-		 * 2. We removed the last item on a page, but there is a
+		 * 1. We will remove an item on a page, but there are more
+		 *    items on that page.
+		 * 2. We will remove the last item on a page, but there is a
 		 *    following page of duplicates.
-		 * 3. We removed the last item on a page, this page was the
+		 * 3. We will remove the last item on a page, this page was the
 		 *    last page in a duplicate set, but there were dups before
 		 *    it.
-		 * 4. We removed the last item on a page, removing the last
+		 * 4. We will remove the last item on a page, removing the last
 		 *    duplicate.
 		 * In case 1 hcp->dpagep is unchanged.
 		 * In case 2 hcp->dpagep comes back pointing to the next dup
 		 *     page.
 		 * In case 3 hcp->dpagep comes back NULL.
 		 * In case 4 hcp->dpagep comes back NULL.
+		 *
+		 * Case 4 results in deleting the pair off the master page.
+		 * The normal code for doing this knows how to delete the
+		 * duplicates, so we will handle this case in the normal code.
 		 */
+		ppgno = PREV_PGNO(hcp->dpagep);
+		if (ppgno == PGNO_INVALID &&
+		    NEXT_PGNO(hcp->dpagep) == PGNO_INVALID &&
+		    NUM_ENT(hcp->dpagep) == 1)
+			goto normal;
+
+		/* Remove item from duplicate page. */
+		chg_pgno = hcp->dpgno;
+		if ((ret = __db_drem(hashp->dbp,
+		    &hcp->dpagep, hcp->dndx, __ham_del_page)) != 0)
+			goto out;
+
 		if (hcp->dpagep == NULL) {
 			if (ppgno != PGNO_INVALID) {		/* Case 3 */
 				hcp->dpgno = ppgno;
@@ -702,7 +712,7 @@ __ham_c_del(cursor, flags)
 				hcp->dndx = NUM_ENT(hcp->dpagep);
 				F_SET(hcp, H_DELETED);
 			} else {				/* Case 4 */
-				ret = __ham_del_pair(hashp, hcp);
+				ret = __ham_del_pair(hashp, hcp, 1);
 				hcp->dpgno = PGNO_INVALID;
 				/*
 				 * Delpair updated the cursor queue, so we
@@ -718,14 +728,14 @@ __ham_c_del(cursor, flags)
 				    H_DATAINDEX(hcp->bndx))),
 				    &hcp->dpgno, sizeof(db_pgno_t));
 			F_SET(hcp, H_DELETED);
-		} else					/* Case 1 */
+		} else						/* Case 1 */
 			F_SET(hcp, H_DELETED);
 		if (chg_pgno != PGNO_INVALID)
 			__ham_c_update(hashp, hcp, chg_pgno, 0, 0, 1);
 	} else if (F_ISSET(hcp, H_ISDUP)) {			/* on page */
 		if (hcp->dup_off == 0 && DUP_SIZE(hcp->dup_len) ==
 		    LEN_HDATA(hcp->pagep, hashp->hdr->pagesize, hcp->bndx))
-			ret = __ham_del_pair(hashp, hcp);
+			ret = __ham_del_pair(hashp, hcp, 1);
 		else {
 			DBT repldbt;
 
@@ -736,14 +746,14 @@ __ham_c_del(cursor, flags)
 			repldbt.size = 0;
 			ret = __ham_replpair(hashp, hcp, &repldbt, 0);
 			hcp->dup_tlen -= DUP_SIZE(hcp->dup_len);
+			F_SET(hcp, H_DELETED);
 			__ham_c_update(hashp, hcp, hcp->pgno,
 			    DUP_SIZE(hcp->dup_len), 0, 1);
-			F_SET(hcp, H_DELETED);
 		}
 
 	} else
 		/* Not a duplicate */
-		ret = __ham_del_pair(hashp, hcp);
+normal:		ret = __ham_del_pair(hashp, hcp, 1);
 
 out:	if ((t_ret = __ham_item_done(hashp, hcp, ret == 0)) != 0 && ret == 0)
 		t_ret = ret;
@@ -975,8 +985,8 @@ int
 __ham_expand_table(hashp)
 	HTAB *hashp;
 {
-	u_int32_t old_bucket, new_bucket;
-	u_int32_t spare_ndx;
+	DB_LSN new_lsn;
+	u_int32_t old_bucket, new_bucket, spare_ndx;
 	int ret;
 
 	ret = 0;
@@ -984,9 +994,30 @@ __ham_expand_table(hashp)
 	if (ret)
 		return (ret);
 
-	if (DB_LOGGING(hashp->dbp)) {
-		DB_LSN new_lsn;
+	/*
+	 * If the split point is about to increase, make sure that we
+	 * have enough extra pages.  The calculation here is weird.
+	 * We'd like to do this after we've upped max_bucket, but it's
+	 * too late then because we've logged the meta-data split.  What
+	 * we'll do between then and now is increment max bucket and then
+	 * see what the log of one greater than that is; here we have to
+	 * look at the log of max + 2.  VERY NASTY STUFF.
+	 */
+	if (__db_log2(hashp->hdr->max_bucket + 2) > hashp->hdr->ovfl_point) {
+		/*
+		 * We are about to shift the split point.  Make sure that
+		 * if the next doubling is going to be big (more than 8
+		 * pages), we have some extra pages around.
+		 */
+		if (hashp->hdr->max_bucket + 1 >= 8 && 
+		    hashp->hdr->spares[hashp->hdr->ovfl_point] <
+		    hashp->hdr->spares[hashp->hdr->ovfl_point - 1] + 
+		    hashp->hdr->ovfl_point + 1)
+			__ham_init_ovflpages(hashp);
+	}
 
+	/* Now we can log the meta-data split. */
+	if (DB_LOGGING(hashp->dbp)) {
 		if ((ret = __ham_splitmeta_log(hashp->dbp->dbenv->lg_info,
 		    (DB_TXN *)hashp->dbp->txn, &new_lsn, 0,
 		    hashp->dbp->log_fileid,
@@ -1003,22 +1034,11 @@ __ham_expand_table(hashp)
 	old_bucket = (hashp->hdr->max_bucket & hashp->hdr->low_mask);
 
 	/*
-	 * If the split point is increasing (hdr.max_bucket's log base 2
-	 * increases), max sure that we have enough extra pages, then
-	 * copy the current contents of the spare split bucket to the
-	 * next bucket.
+	 * If the split point is increasing, copy the current contents
+	 * of the spare split bucket to the next bucket.
 	 */
 	spare_ndx = __db_log2(hashp->hdr->max_bucket + 1);
 	if (spare_ndx > hashp->hdr->ovfl_point) {
-		/*
-		 * We are about to shift the split point.  Make sure that
-		 * if the next doubling is going to be big (more than 8
-		 * pages), we have some extra pages around.
-		 */
-		if (hashp->hdr->spares[hashp->hdr->ovfl_point] == 0 &&
-		    new_bucket >= 8)
-			__ham_init_ovflpages(hashp);
-
 		hashp->hdr->spares[spare_ndx] =
 		    hashp->hdr->spares[hashp->hdr->ovfl_point];
 		hashp->hdr->ovfl_point = spare_ndx;
@@ -1306,7 +1326,7 @@ __ham_init_dbt(dbt, size, bufp, sizep)
 	memset(dbt, 0, sizeof(*dbt));
 	if (*sizep < size) {
 		if ((*bufp = (void *)(*bufp == NULL ?
-		    malloc(size) : realloc(*bufp, size))) == NULL) {
+		    __db_malloc(size) : __db_realloc(*bufp, size))) == NULL) {
 			*sizep = 0;
 			return (ENOMEM);
 		}
@@ -1352,9 +1372,20 @@ __ham_c_update(hashp, hcp, chg_pgno, len, add, dup)
 	if (!dup && add)
 		return;
 
-	page_deleted = chg_pgno != PGNO_INVALID &&
-	    ((!dup && chg_pgno != hcp->pgno) ||
-	    (dup && chg_pgno != hcp->dpgno));
+	/*
+	 * Determine if a page was deleted.    If this is a regular update
+	 * (i.e., not dup) then the deleted page's number will be that in
+	 * chg_pgno, and the pgno in the cursor will be different.  If this
+	 * was an onpage-duplicate, then the same conditions apply.  If this
+	 * was an off-page duplicate, then we need to verify if hcp->dpgno
+	 * is the same (no delete) or different (delete) than chg_pgno.
+	 */
+	if (!dup || hcp->dpgno == PGNO_INVALID)
+		page_deleted =
+		    chg_pgno != PGNO_INVALID && chg_pgno != hcp->pgno;
+	else
+		page_deleted =
+		    chg_pgno != PGNO_INVALID && chg_pgno != hcp->dpgno;
 
 	hp = hcp->db_cursor->dbp->master->internal;
 	DB_THREAD_LOCK(hp->dbp);
@@ -1432,7 +1463,7 @@ __ham_hdup(orig, new)
 	DBC *curs;
 	int ret;
 
-	if ((hashp = (HTAB *)malloc(sizeof(HTAB))) == NULL)
+	if ((hashp = (HTAB *)__db_malloc(sizeof(HTAB))) == NULL)
 		return (ENOMEM);
 
 	new->internal = hashp;
@@ -1441,7 +1472,7 @@ __ham_hdup(orig, new)
 	hashp->hlock = 0;
 	hashp->hdr = NULL;
 	hashp->hash = ((HTAB *)orig->internal)->hash;
-	if ((hashp->split_buf = (PAGE *)malloc(orig->pgsize)) == NULL)
+	if ((hashp->split_buf = (PAGE *)__db_malloc(orig->pgsize)) == NULL)
 		return (ENOMEM);
 	hashp->local_errno = 0;
 	hashp->hash_accesses = 0;