|
|
|
/*
|
|
|
|
* Hash Array Mapped Trie (HAMT) implementation
|
|
|
|
*
|
|
|
|
* Copyright (C) 2001-2007 Peter Johnson
|
|
|
|
*
|
|
|
|
* Based on the paper "Ideal Hash Tries" by Phil Bagwell [2000].
|
|
|
|
* One algorithmic change from that described in the paper: we use the LSB's
|
|
|
|
* of the key to index the root table and move upward in the key rather than
|
|
|
|
* use the MSBs as described in the paper. The LSBs have more entropy.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND OTHER CONTRIBUTORS ``AS IS''
|
|
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR OTHER CONTRIBUTORS BE
|
|
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
#include "util.h"
|
|
|
|
/*@unused@*/ RCSID("$Id$");
|
|
|
|
|
|
|
|
#include <ctype.h>
|
|
|
|
|
|
|
|
#include "libyasm-stdint.h"
|
|
|
|
#include "coretype.h"
|
|
|
|
#include "hamt.h"
|
|
|
|
|
|
|
|
struct HAMTEntry {
|
|
|
|
STAILQ_ENTRY(HAMTEntry) next; /* next hash table entry */
|
|
|
|
/*@dependent@*/ const char *str; /* string being hashed */
|
|
|
|
/*@owned@*/ void *data; /* data pointer being stored */
|
|
|
|
};
|
|
|
|
|
|
|
|
typedef struct HAMTNode {
|
|
|
|
unsigned long BitMapKey; /* 32 bits, bitmap or hash key */
|
|
|
|
uintptr_t BaseValue; /* Base of HAMTNode list or value */
|
|
|
|
} HAMTNode;
|
|
|
|
|
|
|
|
struct HAMT {
|
Add some abilities to base symrec so that it can be more effectively used by
object formats without creating duplicate lists of symbols. XDF and COFF
were updated; ELF needs to reorder the symbols on its own, so for now it's
not been updated to use the common implementation.
* hamt.c (HAMTEntry, HAMT, HAMT_destroy)
(HAMT_insert): Change SLIST to STAILQ.
* symrec.c (sym_type): Add SYM_SPECIAL.
(yasm_symtab_define_special): New.
(yasm_symrec_declare): New; includes all functionality from symtab_declare.
(yasm_symtab_declare): Call yasm_symrec_declare now.
(yasm_symrec_is_special): New.
* symrec.h: Add prototypes for above.
* xdf-objfmt.c: Use symrec_data instead of declaring xdf_symtab.
* xdflong.hex, xdfprotect.hex, xdfother.hex: Update due to symbol reordering.
* coff-objfmt.c: Use symrec_data instead of declaring coff_symtab.
* elftimes.hex, elfso.hex, elfabssect.hex: Update due to symbol reordering.
* elfglobext.hex, elf-x86id.hex, elftest.hex, elfso64.hex: Likewise.
* stabs-elf.hex: Likewise.
svn path=/trunk/yasm/; revision=1188
20 years ago
|
|
|
STAILQ_HEAD(HAMTEntryHead, HAMTEntry) entries;
|
|
|
|
HAMTNode *root;
|
|
|
|
/*@exits@*/ void (*error_func) (const char *file, unsigned int line,
|
|
|
|
const char *message);
|
|
|
|
unsigned long (*HashKey) (const char *key);
|
|
|
|
unsigned long (*ReHashKey) (const char *key, int Level);
|
|
|
|
int (*CmpKey) (const char *s1, const char *s2);
|
|
|
|
};
|
|
|
|
|
|
|
|
/* XXX make a portable version of this. This depends on the pointer being
|
|
|
|
* 4 or 2-byte aligned (as it uses the LSB of the pointer variable to store
|
|
|
|
* the subtrie flag!
|
|
|
|
*/
|
|
|
|
#define IsSubTrie(n) ((n)->BaseValue & 1)
|
|
|
|
#define SetSubTrie(h, n, v) do { \
|
|
|
|
if ((uintptr_t)(v) & 1) \
|
|
|
|
h->error_func(__FILE__, __LINE__, \
|
|
|
|
N_("Subtrie is seen as subtrie before flag is set (misaligned?)")); \
|
|
|
|
(n)->BaseValue = (uintptr_t)(v) | 1; \
|
|
|
|
} while (0)
|
|
|
|
#define SetValue(h, n, v) do { \
|
|
|
|
if ((uintptr_t)(v) & 1) \
|
|
|
|
h->error_func(__FILE__, __LINE__, \
|
|
|
|
N_("Value is seen as subtrie (misaligned?)")); \
|
|
|
|
(n)->BaseValue = (uintptr_t)(v); \
|
|
|
|
} while (0)
|
|
|
|
#define GetSubTrie(n) (HAMTNode *)(((n)->BaseValue | 1) ^ 1)
|
|
|
|
|
|
|
|
static unsigned long
|
|
|
|
HashKey(const char *key)
|
|
|
|
{
|
|
|
|
unsigned long a=31415, b=27183, vHash;
|
|
|
|
for (vHash=0; *key; key++, a*=b)
|
|
|
|
vHash = a*vHash + *key;
|
|
|
|
return vHash;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned long
|
|
|
|
ReHashKey(const char *key, int Level)
|
|
|
|
{
|
|
|
|
unsigned long a=31415, b=27183, vHash;
|
|
|
|
for (vHash=0; *key; key++, a*=b)
|
|
|
|
vHash = a*vHash*(unsigned long)Level + *key;
|
|
|
|
return vHash;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned long
|
|
|
|
HashKey_nocase(const char *key)
|
|
|
|
{
|
|
|
|
unsigned long a=31415, b=27183, vHash;
|
|
|
|
for (vHash=0; *key; key++, a*=b)
|
|
|
|
vHash = a*vHash + tolower(*key);
|
|
|
|
return vHash;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned long
|
|
|
|
ReHashKey_nocase(const char *key, int Level)
|
|
|
|
{
|
|
|
|
unsigned long a=31415, b=27183, vHash;
|
|
|
|
for (vHash=0; *key; key++, a*=b)
|
|
|
|
vHash = a*vHash*(unsigned long)Level + tolower(*key);
|
|
|
|
return vHash;
|
|
|
|
}
|
|
|
|
|
|
|
|
HAMT *
|
|
|
|
HAMT_create(int nocase, /*@exits@*/ void (*error_func)
|
Massive libyasm / module interface update - Phase 1
As yasm has evolved, various minor additions have been made to libyasm to
support the new features. These minor additions have accumulated, and
some contain significant redundancies. In addition, the core focus of
yasm has begun to move away from the front-end commandline program "yasm"
to focusing on libyasm, a collection of reusable routines for use in all
sorts of programs dealing with code at the assembly level, and the modules
that provide specific features for parsing such code.
This libyasm/module update focuses on cleaning up much of the cruft that
has accumulated in libyasm, standardizing function names, eliminating
redundancies, making many of the core objects more reusable for future
extensions, and starting to make libyasm and the modules thread-safe by
eliminating static variables.
Specific changes include:
- Making a symbol table data structure (no longer global). It follows a
factory model for creating symrecs.
- Label symbols now refer only to bytecodes; bytecodes have a pointer to
their containing section.
- Standardizing on *_create() and *_destroy() for allocation/deallocation.
- Adding a standardized callback mechanism for all data structures that
allow associated data. Allowed the removal of objfmt and
dbgfmt-specific data callbacks in their interfaces.
- Unmodularizing linemgr, but allowing multiple linemap instances (linemgr
is now renamed linemap).
- Remove references to lindex; all virtual lines (from linemap) are now
just "line"s.
- Eliminating the bytecode "type" enum, instead adding a standardized
callback mechanism for custom (and standard internal) bytecode types.
This will make it much easier to add new bytecodes, and eliminate the
possibility of type collisions. This also allowed the removal of the
of_data and df_data bytecodes, as objfmts and dbgfmts can now easily
implement their own bytecodes, and the cleanup of arch's bytecode usage.
- Remove the bytecodehead and sectionhead pseudo-containers, instead
making true containers: section now implements all the functions of
bytecodehead, and the new object data structure implements all the
functions of sectionhead.
- Add object data structure: it's a container that contains sections, a
symbol table, and a line mapping for a single object. Every former use
of sectionhead now takes an object.
- Make arch interface and all standard architectures thread-safe:
yasm_arch_module is the module interface; it contains a create()
function that returns a yasm_arch * to store local yasm_arch data; all
yasm_arch_module functions take the yasm_arch *.
- Make nasm parser thread-safe.
To be done in phase 2: making other module interfaces thread-safe. Note
that while the module interface may be thread-safe, not all modules may be
written in such a fashion (hopefully all the "standard" ones will be, but
this is yet to be determined).
svn path=/trunk/yasm/; revision=1058
21 years ago
|
|
|
(const char *file, unsigned int line, const char *message))
|
|
|
|
{
|
|
|
|
/*@out@*/ HAMT *hamt = yasm_xmalloc(sizeof(HAMT));
|
|
|
|
int i;
|
|
|
|
|
Add some abilities to base symrec so that it can be more effectively used by
object formats without creating duplicate lists of symbols. XDF and COFF
were updated; ELF needs to reorder the symbols on its own, so for now it's
not been updated to use the common implementation.
* hamt.c (HAMTEntry, HAMT, HAMT_destroy)
(HAMT_insert): Change SLIST to STAILQ.
* symrec.c (sym_type): Add SYM_SPECIAL.
(yasm_symtab_define_special): New.
(yasm_symrec_declare): New; includes all functionality from symtab_declare.
(yasm_symtab_declare): Call yasm_symrec_declare now.
(yasm_symrec_is_special): New.
* symrec.h: Add prototypes for above.
* xdf-objfmt.c: Use symrec_data instead of declaring xdf_symtab.
* xdflong.hex, xdfprotect.hex, xdfother.hex: Update due to symbol reordering.
* coff-objfmt.c: Use symrec_data instead of declaring coff_symtab.
* elftimes.hex, elfso.hex, elfabssect.hex: Update due to symbol reordering.
* elfglobext.hex, elf-x86id.hex, elftest.hex, elfso64.hex: Likewise.
* stabs-elf.hex: Likewise.
svn path=/trunk/yasm/; revision=1188
20 years ago
|
|
|
STAILQ_INIT(&hamt->entries);
|
|
|
|
hamt->root = yasm_xmalloc(32*sizeof(HAMTNode));
|
|
|
|
|
|
|
|
for (i=0; i<32; i++) {
|
|
|
|
hamt->root[i].BitMapKey = 0;
|
|
|
|
hamt->root[i].BaseValue = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
hamt->error_func = error_func;
|
|
|
|
if (nocase) {
|
|
|
|
hamt->HashKey = HashKey_nocase;
|
|
|
|
hamt->ReHashKey = ReHashKey_nocase;
|
|
|
|
hamt->CmpKey = yasm__strcasecmp;
|
|
|
|
} else {
|
|
|
|
hamt->HashKey = HashKey;
|
|
|
|
hamt->ReHashKey = ReHashKey;
|
|
|
|
hamt->CmpKey = strcmp;
|
|
|
|
}
|
|
|
|
|
|
|
|
return hamt;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
HAMT_delete_trie(HAMTNode *node)
|
|
|
|
{
|
|
|
|
if (IsSubTrie(node)) {
|
|
|
|
unsigned long i, Size;
|
|
|
|
|
|
|
|
/* Count total number of bits in bitmap to determine size */
|
|
|
|
BitCount(Size, node->BitMapKey);
|
|
|
|
Size &= 0x1F;
|
|
|
|
if (Size == 0)
|
|
|
|
Size = 32;
|
|
|
|
|
|
|
|
for (i=0; i<Size; i++)
|
|
|
|
HAMT_delete_trie(&(GetSubTrie(node))[i]);
|
|
|
|
yasm_xfree(GetSubTrie(node));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
Massive libyasm / module interface update - Phase 1
As yasm has evolved, various minor additions have been made to libyasm to
support the new features. These minor additions have accumulated, and
some contain significant redundancies. In addition, the core focus of
yasm has begun to move away from the front-end commandline program "yasm"
to focusing on libyasm, a collection of reusable routines for use in all
sorts of programs dealing with code at the assembly level, and the modules
that provide specific features for parsing such code.
This libyasm/module update focuses on cleaning up much of the cruft that
has accumulated in libyasm, standardizing function names, eliminating
redundancies, making many of the core objects more reusable for future
extensions, and starting to make libyasm and the modules thread-safe by
eliminating static variables.
Specific changes include:
- Making a symbol table data structure (no longer global). It follows a
factory model for creating symrecs.
- Label symbols now refer only to bytecodes; bytecodes have a pointer to
their containing section.
- Standardizing on *_create() and *_destroy() for allocation/deallocation.
- Adding a standardized callback mechanism for all data structures that
allow associated data. Allowed the removal of objfmt and
dbgfmt-specific data callbacks in their interfaces.
- Unmodularizing linemgr, but allowing multiple linemap instances (linemgr
is now renamed linemap).
- Remove references to lindex; all virtual lines (from linemap) are now
just "line"s.
- Eliminating the bytecode "type" enum, instead adding a standardized
callback mechanism for custom (and standard internal) bytecode types.
This will make it much easier to add new bytecodes, and eliminate the
possibility of type collisions. This also allowed the removal of the
of_data and df_data bytecodes, as objfmts and dbgfmts can now easily
implement their own bytecodes, and the cleanup of arch's bytecode usage.
- Remove the bytecodehead and sectionhead pseudo-containers, instead
making true containers: section now implements all the functions of
bytecodehead, and the new object data structure implements all the
functions of sectionhead.
- Add object data structure: it's a container that contains sections, a
symbol table, and a line mapping for a single object. Every former use
of sectionhead now takes an object.
- Make arch interface and all standard architectures thread-safe:
yasm_arch_module is the module interface; it contains a create()
function that returns a yasm_arch * to store local yasm_arch data; all
yasm_arch_module functions take the yasm_arch *.
- Make nasm parser thread-safe.
To be done in phase 2: making other module interfaces thread-safe. Note
that while the module interface may be thread-safe, not all modules may be
written in such a fashion (hopefully all the "standard" ones will be, but
this is yet to be determined).
svn path=/trunk/yasm/; revision=1058
21 years ago
|
|
|
HAMT_destroy(HAMT *hamt, void (*deletefunc) (/*@only@*/ void *data))
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
/* delete entries */
|
Add some abilities to base symrec so that it can be more effectively used by
object formats without creating duplicate lists of symbols. XDF and COFF
were updated; ELF needs to reorder the symbols on its own, so for now it's
not been updated to use the common implementation.
* hamt.c (HAMTEntry, HAMT, HAMT_destroy)
(HAMT_insert): Change SLIST to STAILQ.
* symrec.c (sym_type): Add SYM_SPECIAL.
(yasm_symtab_define_special): New.
(yasm_symrec_declare): New; includes all functionality from symtab_declare.
(yasm_symtab_declare): Call yasm_symrec_declare now.
(yasm_symrec_is_special): New.
* symrec.h: Add prototypes for above.
* xdf-objfmt.c: Use symrec_data instead of declaring xdf_symtab.
* xdflong.hex, xdfprotect.hex, xdfother.hex: Update due to symbol reordering.
* coff-objfmt.c: Use symrec_data instead of declaring coff_symtab.
* elftimes.hex, elfso.hex, elfabssect.hex: Update due to symbol reordering.
* elfglobext.hex, elf-x86id.hex, elftest.hex, elfso64.hex: Likewise.
* stabs-elf.hex: Likewise.
svn path=/trunk/yasm/; revision=1188
20 years ago
|
|
|
while (!STAILQ_EMPTY(&hamt->entries)) {
|
|
|
|
HAMTEntry *entry;
|
|
|
|
entry = STAILQ_FIRST(&hamt->entries);
|
|
|
|
STAILQ_REMOVE_HEAD(&hamt->entries, next);
|
|
|
|
deletefunc(entry->data);
|
|
|
|
yasm_xfree(entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* delete trie */
|
|
|
|
for (i=0; i<32; i++)
|
|
|
|
HAMT_delete_trie(&hamt->root[i]);
|
|
|
|
|
|
|
|
yasm_xfree(hamt->root);
|
|
|
|
yasm_xfree(hamt);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
HAMT_traverse(HAMT *hamt, void *d,
|
|
|
|
int (*func) (/*@dependent@*/ /*@null@*/ void *node,
|
|
|
|
/*@null@*/ void *d))
|
|
|
|
{
|
|
|
|
HAMTEntry *entry;
|
|
|
|
STAILQ_FOREACH(entry, &hamt->entries, next) {
|
|
|
|
int retval = func(entry->data, d);
|
|
|
|
if (retval != 0)
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
const HAMTEntry *
|
|
|
|
HAMT_first(const HAMT *hamt)
|
|
|
|
{
|
|
|
|
return STAILQ_FIRST(&hamt->entries);
|
|
|
|
}
|
|
|
|
|
|
|
|
const HAMTEntry *
|
|
|
|
HAMT_next(const HAMTEntry *prev)
|
|
|
|
{
|
|
|
|
return STAILQ_NEXT(prev, next);
|
|
|
|
}
|
|
|
|
|
|
|
|
void *
|
|
|
|
HAMTEntry_get_data(const HAMTEntry *entry)
|
|
|
|
{
|
|
|
|
return entry->data;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*@-temptrans -kepttrans -mustfree@*/
|
|
|
|
void *
|
|
|
|
HAMT_insert(HAMT *hamt, const char *str, void *data, int *replace,
|
|
|
|
void (*deletefunc) (/*@only@*/ void *data))
|
|
|
|
{
|
|
|
|
HAMTNode *node, *newnodes;
|
|
|
|
HAMTEntry *entry;
|
|
|
|
unsigned long key, keypart, Map;
|
|
|
|
int keypartbits = 0;
|
|
|
|
int level = 0;
|
|
|
|
|
|
|
|
key = hamt->HashKey(str);
|
|
|
|
keypart = key & 0x1F;
|
|
|
|
node = &hamt->root[keypart];
|
|
|
|
|
|
|
|
if (!node->BaseValue) {
|
|
|
|
node->BitMapKey = key;
|
|
|
|
entry = yasm_xmalloc(sizeof(HAMTEntry));
|
|
|
|
entry->str = str;
|
|
|
|
entry->data = data;
|
|
|
|
STAILQ_INSERT_TAIL(&hamt->entries, entry, next);
|
|
|
|
SetValue(hamt, node, entry);
|
|
|
|
if (IsSubTrie(node))
|
|
|
|
hamt->error_func(__FILE__, __LINE__,
|
|
|
|
N_("Data is seen as subtrie (misaligned?)"));
|
|
|
|
*replace = 1;
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (;;) {
|
|
|
|
if (!(IsSubTrie(node))) {
|
|
|
|
if (node->BitMapKey == key
|
|
|
|
&& hamt->CmpKey(((HAMTEntry *)(node->BaseValue))->str,
|
|
|
|
str) == 0) {
|
|
|
|
/*@-branchstate@*/
|
|
|
|
if (*replace) {
|
|
|
|
deletefunc(((HAMTEntry *)(node->BaseValue))->data);
|
|
|
|
((HAMTEntry *)(node->BaseValue))->str = str;
|
|
|
|
((HAMTEntry *)(node->BaseValue))->data = data;
|
|
|
|
} else
|
|
|
|
deletefunc(data);
|
|
|
|
/*@=branchstate@*/
|
|
|
|
return ((HAMTEntry *)(node->BaseValue))->data;
|
|
|
|
} else {
|
|
|
|
unsigned long key2 = node->BitMapKey;
|
|
|
|
/* build tree downward until keys differ */
|
|
|
|
for (;;) {
|
|
|
|
unsigned long keypart2;
|
|
|
|
|
|
|
|
/* replace node with subtrie */
|
|
|
|
keypartbits += 5;
|
|
|
|
if (keypartbits > 30) {
|
|
|
|
/* Exceeded 32 bits: rehash */
|
|
|
|
key = hamt->ReHashKey(str, level);
|
|
|
|
key2 = hamt->ReHashKey(
|
|
|
|
((HAMTEntry *)(node->BaseValue))->str, level);
|
|
|
|
keypartbits = 0;
|
|
|
|
}
|
|
|
|
keypart = (key >> keypartbits) & 0x1F;
|
|
|
|
keypart2 = (key2 >> keypartbits) & 0x1F;
|
|
|
|
|
|
|
|
if (keypart == keypart2) {
|
|
|
|
/* Still equal, build one-node subtrie and continue
|
|
|
|
* downward.
|
|
|
|
*/
|
|
|
|
newnodes = yasm_xmalloc(sizeof(HAMTNode));
|
|
|
|
newnodes[0].BitMapKey = key2;
|
|
|
|
newnodes[0].BaseValue = node->BaseValue;
|
|
|
|
node->BitMapKey = 1<<keypart;
|
|
|
|
SetSubTrie(hamt, node, newnodes);
|
|
|
|
node = &newnodes[0];
|
|
|
|
level++;
|
|
|
|
} else {
|
|
|
|
/* partitioned: allocate two-node subtrie */
|
|
|
|
newnodes = yasm_xmalloc(2*sizeof(HAMTNode));
|
|
|
|
|
|
|
|
entry = yasm_xmalloc(sizeof(HAMTEntry));
|
|
|
|
entry->str = str;
|
|
|
|
entry->data = data;
|
|
|
|
STAILQ_INSERT_TAIL(&hamt->entries, entry, next);
|
|
|
|
|
|
|
|
/* Copy nodes into subtrie based on order */
|
|
|
|
if (keypart2 < keypart) {
|
|
|
|
newnodes[0].BitMapKey = key2;
|
|
|
|
newnodes[0].BaseValue = node->BaseValue;
|
|
|
|
newnodes[1].BitMapKey = key;
|
|
|
|
SetValue(hamt, &newnodes[1], entry);
|
|
|
|
} else {
|
|
|
|
newnodes[0].BitMapKey = key;
|
|
|
|
SetValue(hamt, &newnodes[0], entry);
|
|
|
|
newnodes[1].BitMapKey = key2;
|
|
|
|
newnodes[1].BaseValue = node->BaseValue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Set bits in bitmap corresponding to keys */
|
|
|
|
node->BitMapKey = (1UL<<keypart) | (1UL<<keypart2);
|
|
|
|
SetSubTrie(hamt, node, newnodes);
|
|
|
|
*replace = 1;
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Subtrie: look up in bitmap */
|
|
|
|
keypartbits += 5;
|
|
|
|
if (keypartbits > 30) {
|
|
|
|
/* Exceeded 32 bits of current key: rehash */
|
|
|
|
key = hamt->ReHashKey(str, level);
|
|
|
|
keypartbits = 0;
|
|
|
|
}
|
|
|
|
keypart = (key >> keypartbits) & 0x1F;
|
|
|
|
if (!(node->BitMapKey & (1<<keypart))) {
|
|
|
|
/* bit is 0 in bitmap -> add node to table */
|
|
|
|
unsigned long Size;
|
|
|
|
|
|
|
|
/* set bit to 1 */
|
|
|
|
node->BitMapKey |= 1<<keypart;
|
|
|
|
|
|
|
|
/* Count total number of bits in bitmap to determine new size */
|
|
|
|
BitCount(Size, node->BitMapKey);
|
|
|
|
Size &= 0x1F;
|
|
|
|
if (Size == 0)
|
|
|
|
Size = 32;
|
|
|
|
newnodes = yasm_xmalloc(Size*sizeof(HAMTNode));
|
|
|
|
|
|
|
|
/* Count bits below to find where to insert new node at */
|
|
|
|
BitCount(Map, node->BitMapKey & ~((~0UL)<<keypart));
|
|
|
|
Map &= 0x1F; /* Clamp to <32 */
|
|
|
|
/* Copy existing nodes leaving gap for new node */
|
|
|
|
memcpy(newnodes, GetSubTrie(node), Map*sizeof(HAMTNode));
|
|
|
|
memcpy(&newnodes[Map+1], &(GetSubTrie(node))[Map],
|
|
|
|
(Size-Map-1)*sizeof(HAMTNode));
|
|
|
|
/* Delete old subtrie */
|
|
|
|
yasm_xfree(GetSubTrie(node));
|
|
|
|
/* Set up new node */
|
|
|
|
newnodes[Map].BitMapKey = key;
|
|
|
|
entry = yasm_xmalloc(sizeof(HAMTEntry));
|
|
|
|
entry->str = str;
|
|
|
|
entry->data = data;
|
|
|
|
STAILQ_INSERT_TAIL(&hamt->entries, entry, next);
|
|
|
|
SetValue(hamt, &newnodes[Map], entry);
|
|
|
|
SetSubTrie(hamt, node, newnodes);
|
|
|
|
|
|
|
|
*replace = 1;
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Count bits below */
|
|
|
|
BitCount(Map, node->BitMapKey & ~((~0UL)<<keypart));
|
|
|
|
Map &= 0x1F; /* Clamp to <32 */
|
|
|
|
|
|
|
|
/* Go down a level */
|
|
|
|
level++;
|
|
|
|
node = &(GetSubTrie(node))[Map];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*@=temptrans =kepttrans =mustfree@*/
|
|
|
|
|
|
|
|
void *
|
|
|
|
HAMT_search(HAMT *hamt, const char *str)
|
|
|
|
{
|
|
|
|
HAMTNode *node;
|
|
|
|
unsigned long key, keypart, Map;
|
|
|
|
int keypartbits = 0;
|
|
|
|
int level = 0;
|
|
|
|
|
|
|
|
key = hamt->HashKey(str);
|
|
|
|
keypart = key & 0x1F;
|
|
|
|
node = &hamt->root[keypart];
|
|
|
|
|
|
|
|
if (!node->BaseValue)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
for (;;) {
|
|
|
|
if (!(IsSubTrie(node))) {
|
|
|
|
if (node->BitMapKey == key
|
|
|
|
&& hamt->CmpKey(((HAMTEntry *)(node->BaseValue))->str,
|
|
|
|
str) == 0)
|
|
|
|
return ((HAMTEntry *)(node->BaseValue))->data;
|
|
|
|
else
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Subtree: look up in bitmap */
|
|
|
|
keypartbits += 5;
|
|
|
|
if (keypartbits > 30) {
|
|
|
|
/* Exceeded 32 bits of current key: rehash */
|
|
|
|
key = hamt->ReHashKey(str, level);
|
|
|
|
keypartbits = 0;
|
|
|
|
}
|
|
|
|
keypart = (key >> keypartbits) & 0x1F;
|
|
|
|
if (!(node->BitMapKey & (1<<keypart)))
|
|
|
|
return NULL; /* bit is 0 in bitmap -> no match */
|
|
|
|
|
|
|
|
/* Count bits below */
|
|
|
|
BitCount(Map, node->BitMapKey & ~((~0UL)<<keypart));
|
|
|
|
Map &= 0x1F; /* Clamp to <32 */
|
|
|
|
|
|
|
|
/* Go down a level */
|
|
|
|
level++;
|
|
|
|
node = &(GetSubTrie(node))[Map];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|