libnl 3.7.0
hash.c
1/* SPDX-License-Identifier: LGPL-2.1-only */
2/*
3 * This code was taken from http://ccodearchive.net/info/hash.html
4 * The original file was modified to remove unwanted code
5 * and some changes to fit the current build environment
6 */
7/*
8-------------------------------------------------------------------------------
9lookup3.c, by Bob Jenkins, May 2006, Public Domain.
10
11These are functions for producing 32-bit hashes for hash table lookup.
12hash_word(), hashlittle(), hashlittle2(), hashbig(), mix(), and final()
13are externally useful functions. Routines to test the hash are included
14if SELF_TEST is defined. You can use this free for any purpose. It's in
15the public domain. It has no warranty.
16
17You probably want to use hashlittle(). hashlittle() and hashbig()
18hash byte arrays. hashlittle() is is faster than hashbig() on
19little-endian machines. Intel and AMD are little-endian machines.
20On second thought, you probably want hashlittle2(), which is identical to
21hashlittle() except it returns two 32-bit hashes for the price of one.
22You could implement hashbig2() if you wanted but I haven't bothered here.
23
24If you want to find a hash of, say, exactly 7 integers, do
25 a = i1; b = i2; c = i3;
26 mix(a,b,c);
27 a += i4; b += i5; c += i6;
28 mix(a,b,c);
29 a += i7;
30 final(a,b,c);
31then use c as the hash value. If you have a variable length array of
324-byte integers to hash, use hash_word(). If you have a byte array (like
33a character string), use hashlittle(). If you have several byte arrays, or
34a mix of things, see the comments above hashlittle().
35
36Why is this so big? I read 12 bytes at a time into 3 4-byte integers,
37then mix those integers. This is fast (you can do a lot more thorough
38mixing with 12*3 instructions on 3 integers than you can with 3 instructions
39on 1 byte), but shoehorning those bytes into integers efficiently is messy.
40-------------------------------------------------------------------------------
41*/
42#include <netlink/hash.h>
43
44#if HAVE_LITTLE_ENDIAN
45#define HASH_LITTLE_ENDIAN 1
46#define HASH_BIG_ENDIAN 0
47#elif HAVE_BIG_ENDIAN
48#define HASH_LITTLE_ENDIAN 0
49#define HASH_BIG_ENDIAN 1
50#else
51#error Unknown endian
52#endif
53
54#define hashsize(n) ((uint32_t)1<<(n))
55#define hashmask(n) (hashsize(n)-1)
56#define rot(x,k) (((x)<<(k)) | ((x)>>(32-(k))))
57
58/*
59-------------------------------------------------------------------------------
60mix -- mix 3 32-bit values reversibly.
61
62This is reversible, so any information in (a,b,c) before mix() is
63still in (a,b,c) after mix().
64
65If four pairs of (a,b,c) inputs are run through mix(), or through
66mix() in reverse, there are at least 32 bits of the output that
67are sometimes the same for one pair and different for another pair.
68This was tested for:
69* pairs that differed by one bit, by two bits, in any combination
70 of top bits of (a,b,c), or in any combination of bottom bits of
71 (a,b,c).
72* "differ" is defined as +, -, ^, or ~^. For + and -, I transformed
73 the output delta to a Gray code (a^(a>>1)) so a string of 1's (as
74 is commonly produced by subtraction) look like a single 1-bit
75 difference.
76* the base values were pseudorandom, all zero but one bit set, or
77 all zero plus a counter that starts at zero.
78
79Some k values for my "a-=c; a^=rot(c,k); c+=b;" arrangement that
80satisfy this are
81 4 6 8 16 19 4
82 9 15 3 18 27 15
83 14 9 3 7 17 3
84Well, "9 15 3 18 27 15" didn't quite get 32 bits diffing
85for "differ" defined as + with a one-bit base and a two-bit delta. I
86used http://burtleburtle.net/bob/hash/avalanche.html to choose
87the operations, constants, and arrangements of the variables.
88
89This does not achieve avalanche. There are input bits of (a,b,c)
90that fail to affect some output bits of (a,b,c), especially of a. The
91most thoroughly mixed value is c, but it doesn't really even achieve
92avalanche in c.
93
94This allows some parallelism. Read-after-writes are good at doubling
95the number of bits affected, so the goal of mixing pulls in the opposite
96direction as the goal of parallelism. I did what I could. Rotates
97seem to cost as much as shifts on every machine I could lay my hands
98on, and rotates are much kinder to the top and bottom bits, so I used
99rotates.
100-------------------------------------------------------------------------------
101*/
102#define mix(a,b,c) \
103{ \
104 a -= c; a ^= rot(c, 4); c += b; \
105 b -= a; b ^= rot(a, 6); a += c; \
106 c -= b; c ^= rot(b, 8); b += a; \
107 a -= c; a ^= rot(c,16); c += b; \
108 b -= a; b ^= rot(a,19); a += c; \
109 c -= b; c ^= rot(b, 4); b += a; \
110}
111
112/*
113-------------------------------------------------------------------------------
114final -- final mixing of 3 32-bit values (a,b,c) into c
115
116Pairs of (a,b,c) values differing in only a few bits will usually
117produce values of c that look totally different. This was tested for
118* pairs that differed by one bit, by two bits, in any combination
119 of top bits of (a,b,c), or in any combination of bottom bits of
120 (a,b,c).
121* "differ" is defined as +, -, ^, or ~^. For + and -, I transformed
122 the output delta to a Gray code (a^(a>>1)) so a string of 1's (as
123 is commonly produced by subtraction) look like a single 1-bit
124 difference.
125* the base values were pseudorandom, all zero but one bit set, or
126 all zero plus a counter that starts at zero.
127
128These constants passed:
129 14 11 25 16 4 14 24
130 12 14 25 16 4 14 24
131and these came close:
132 4 8 15 26 3 22 24
133 10 8 15 26 3 22 24
134 11 8 15 26 3 22 24
135-------------------------------------------------------------------------------
136*/
137#define final(a,b,c) \
138{ \
139 c ^= b; c -= rot(b,14); \
140 a ^= c; a -= rot(c,11); \
141 b ^= a; b -= rot(a,25); \
142 c ^= b; c -= rot(b,16); \
143 a ^= c; a -= rot(c,4); \
144 b ^= a; b -= rot(a,14); \
145 c ^= b; c -= rot(b,24); \
146}
147
148/*
149-------------------------------------------------------------------------------
150hashlittle() -- hash a variable-length key into a 32-bit value
151 k : the key (the unaligned variable-length array of bytes)
152 length : the length of the key, counting by bytes
153 val2 : IN: can be any 4-byte value OUT: second 32 bit hash.
154Returns a 32-bit value. Every bit of the key affects every bit of
155the return value. Two keys differing by one or two bits will have
156totally different hash values. Note that the return value is better
157mixed than val2, so use that first.
158
159The best hash table sizes are powers of 2. There is no need to do
160mod a prime (mod is sooo slow!). If you need less than 32 bits,
161use a bitmask. For example, if you need only 10 bits, do
162 h = (h & hashmask(10));
163In which case, the hash table should have hashsize(10) elements.
164
165If you are hashing n strings (uint8_t **)k, do it like this:
166 for (i=0, h=0; i<n; ++i) h = hashlittle( k[i], len[i], h);
167
168By Bob Jenkins, 2006. bob_jenkins@burtleburtle.net. You may use this
169code any way you wish, private, educational, or commercial. It's free.
170
171Use for hash table lookup, or anything where one collision in 2^^32 is
172acceptable. Do NOT use for cryptographic purposes.
173-------------------------------------------------------------------------------
174*/
175
176static uint32_t hashlittle( const void *key, size_t length, uint32_t *val2 )
177{
178 uint32_t a,b,c; /* internal state */
179 union { const void *ptr; size_t i; } u; /* needed for Mac Powerbook G4 */
180
181 /* Set up the internal state */
182 a = b = c = 0xdeadbeef + ((uint32_t)length) + *val2;
183
184 u.ptr = key;
185 if (HASH_LITTLE_ENDIAN && ((u.i & 0x3) == 0)) {
186 const uint32_t *k = (const uint32_t *)key; /* read 32-bit chunks */
187 const uint8_t *k8;
188
189 /*------ all but last block: aligned reads and affect 32 bits of (a,b,c) */
190 while (length > 12)
191 {
192 a += k[0];
193 b += k[1];
194 c += k[2];
195 mix(a,b,c);
196 length -= 12;
197 k += 3;
198 }
199
200 /*----------------------------- handle the last (probably partial) block */
201 /*
202 * "k[2]&0xffffff" actually reads beyond the end of the string, but
203 * then masks off the part it's not allowed to read. Because the
204 * string is aligned, the masked-off tail is in the same word as the
205 * rest of the string. Every machine with memory protection I've seen
206 * does it on word boundaries, so is OK with this. But VALGRIND will
207 * still catch it and complain. The masking trick does make the hash
208 * noticably faster for short strings (like English words).
209 *
210 * Not on my testing with gcc 4.5 on an intel i5 CPU, at least --RR.
211 */
212#if 0
213 switch(length)
214 {
215 case 12: c+=k[2]; b+=k[1]; a+=k[0]; break;
216 case 11: c+=k[2]&0xffffff; b+=k[1]; a+=k[0]; break;
217 case 10: c+=k[2]&0xffff; b+=k[1]; a+=k[0]; break;
218 case 9 : c+=k[2]&0xff; b+=k[1]; a+=k[0]; break;
219 case 8 : b+=k[1]; a+=k[0]; break;
220 case 7 : b+=k[1]&0xffffff; a+=k[0]; break;
221 case 6 : b+=k[1]&0xffff; a+=k[0]; break;
222 case 5 : b+=k[1]&0xff; a+=k[0]; break;
223 case 4 : a+=k[0]; break;
224 case 3 : a+=k[0]&0xffffff; break;
225 case 2 : a+=k[0]&0xffff; break;
226 case 1 : a+=k[0]&0xff; break;
227 case 0 : return c; /* zero length strings require no mixing */
228 }
229
230#else /* make valgrind happy */
231
232 k8 = (const uint8_t *)k;
233 switch(length)
234 {
235 case 12: c+=k[2]; b+=k[1]; a+=k[0]; break;
236 case 11: c+=((uint32_t)k8[10])<<16; /* fall through */
237 case 10: c+=((uint32_t)k8[9])<<8; /* fall through */
238 case 9 : c+=k8[8]; /* fall through */
239 case 8 : b+=k[1]; a+=k[0]; break;
240 case 7 : b+=((uint32_t)k8[6])<<16; /* fall through */
241 case 6 : b+=((uint32_t)k8[5])<<8; /* fall through */
242 case 5 : b+=k8[4]; /* fall through */
243 case 4 : a+=k[0]; break;
244 case 3 : a+=((uint32_t)k8[2])<<16; /* fall through */
245 case 2 : a+=((uint32_t)k8[1])<<8; /* fall through */
246 case 1 : a+=k8[0]; break;
247 case 0 : return c;
248 }
249
250#endif /* !valgrind */
251
252 } else if (HASH_LITTLE_ENDIAN && ((u.i & 0x1) == 0)) {
253 const uint16_t *k = (const uint16_t *)key; /* read 16-bit chunks */
254 const uint8_t *k8;
255
256 /*--------------- all but last block: aligned reads and different mixing */
257 while (length > 12)
258 {
259 a += k[0] + (((uint32_t)k[1])<<16);
260 b += k[2] + (((uint32_t)k[3])<<16);
261 c += k[4] + (((uint32_t)k[5])<<16);
262 mix(a,b,c);
263 length -= 12;
264 k += 6;
265 }
266
267 /*----------------------------- handle the last (probably partial) block */
268 k8 = (const uint8_t *)k;
269 switch(length)
270 {
271 case 12: c+=k[4]+(((uint32_t)k[5])<<16);
272 b+=k[2]+(((uint32_t)k[3])<<16);
273 a+=k[0]+(((uint32_t)k[1])<<16);
274 break;
275 case 11: c+=((uint32_t)k8[10])<<16; /* fall through */
276 case 10: c+=k[4];
277 b+=k[2]+(((uint32_t)k[3])<<16);
278 a+=k[0]+(((uint32_t)k[1])<<16);
279 break;
280 case 9 : c+=k8[8]; /* fall through */
281 case 8 : b+=k[2]+(((uint32_t)k[3])<<16);
282 a+=k[0]+(((uint32_t)k[1])<<16);
283 break;
284 case 7 : b+=((uint32_t)k8[6])<<16; /* fall through */
285 case 6 : b+=k[2];
286 a+=k[0]+(((uint32_t)k[1])<<16);
287 break;
288 case 5 : b+=k8[4]; /* fall through */
289 case 4 : a+=k[0]+(((uint32_t)k[1])<<16);
290 break;
291 case 3 : a+=((uint32_t)k8[2])<<16; /* fall through */
292 case 2 : a+=k[0];
293 break;
294 case 1 : a+=k8[0];
295 break;
296 case 0 : return c; /* zero length requires no mixing */
297 }
298
299 } else { /* need to read the key one byte at a time */
300 const uint8_t *k = (const uint8_t *)key;
301
302 /*--------------- all but the last block: affect some 32 bits of (a,b,c) */
303 while (length > 12)
304 {
305 a += k[0];
306 a += ((uint32_t)k[1])<<8;
307 a += ((uint32_t)k[2])<<16;
308 a += ((uint32_t)k[3])<<24;
309 b += k[4];
310 b += ((uint32_t)k[5])<<8;
311 b += ((uint32_t)k[6])<<16;
312 b += ((uint32_t)k[7])<<24;
313 c += k[8];
314 c += ((uint32_t)k[9])<<8;
315 c += ((uint32_t)k[10])<<16;
316 c += ((uint32_t)k[11])<<24;
317 mix(a,b,c);
318 length -= 12;
319 k += 12;
320 }
321
322 /*-------------------------------- last block: affect all 32 bits of (c) */
323 switch(length) /* all the case statements fall through */
324 {
325 case 12: c+=((uint32_t)k[11])<<24; /* fall through */
326 case 11: c+=((uint32_t)k[10])<<16; /* fall through */
327 case 10: c+=((uint32_t)k[9])<<8; /* fall through */
328 case 9 : c+=k[8]; /* fall through */
329 case 8 : b+=((uint32_t)k[7])<<24; /* fall through */
330 case 7 : b+=((uint32_t)k[6])<<16; /* fall through */
331 case 6 : b+=((uint32_t)k[5])<<8; /* fall through */
332 case 5 : b+=k[4]; /* fall through */
333 case 4 : a+=((uint32_t)k[3])<<24; /* fall through */
334 case 3 : a+=((uint32_t)k[2])<<16; /* fall through */
335 case 2 : a+=((uint32_t)k[1])<<8; /* fall through */
336 case 1 : a+=k[0];
337 break;
338 case 0 : return c;
339 }
340 }
341
342 final(a,b,c);
343 *val2 = b;
344 return c;
345}
346
347/*
348 * hashbig():
349 * This is the same as hash_word() on big-endian machines. It is different
350 * from hashlittle() on all machines. hashbig() takes advantage of
351 * big-endian byte ordering.
352 */
353static uint32_t hashbig( const void *key, size_t length, uint32_t *val2)
354{
355 uint32_t a,b,c;
356 union { const void *ptr; size_t i; } u; /* to cast key to (size_t) happily */
357
358 /* Set up the internal state */
359 a = b = c = 0xdeadbeef + ((uint32_t)length) + *val2;
360
361 u.ptr = key;
362 if (HASH_BIG_ENDIAN && ((u.i & 0x3) == 0)) {
363 const uint32_t *k = (const uint32_t *)key; /* read 32-bit chunks */
364 const uint8_t *k8;
365
366 /*------ all but last block: aligned reads and affect 32 bits of (a,b,c) */
367 while (length > 12)
368 {
369 a += k[0];
370 b += k[1];
371 c += k[2];
372 mix(a,b,c);
373 length -= 12;
374 k += 3;
375 }
376
377 /*----------------------------- handle the last (probably partial) block */
378 /*
379 * "k[2]<<8" actually reads beyond the end of the string, but
380 * then shifts out the part it's not allowed to read. Because the
381 * string is aligned, the illegal read is in the same word as the
382 * rest of the string. Every machine with memory protection I've seen
383 * does it on word boundaries, so is OK with this. But VALGRIND will
384 * still catch it and complain. The masking trick does make the hash
385 * noticably faster for short strings (like English words).
386 *
387 * Not on my testing with gcc 4.5 on an intel i5 CPU, at least --RR.
388 */
389#if 0
390 switch(length)
391 {
392 case 12: c+=k[2]; b+=k[1]; a+=k[0]; break;
393 case 11: c+=k[2]&0xffffff00; b+=k[1]; a+=k[0]; break;
394 case 10: c+=k[2]&0xffff0000; b+=k[1]; a+=k[0]; break;
395 case 9 : c+=k[2]&0xff000000; b+=k[1]; a+=k[0]; break;
396 case 8 : b+=k[1]; a+=k[0]; break;
397 case 7 : b+=k[1]&0xffffff00; a+=k[0]; break;
398 case 6 : b+=k[1]&0xffff0000; a+=k[0]; break;
399 case 5 : b+=k[1]&0xff000000; a+=k[0]; break;
400 case 4 : a+=k[0]; break;
401 case 3 : a+=k[0]&0xffffff00; break;
402 case 2 : a+=k[0]&0xffff0000; break;
403 case 1 : a+=k[0]&0xff000000; break;
404 case 0 : return c; /* zero length strings require no mixing */
405 }
406
407#else /* make valgrind happy */
408
409 k8 = (const uint8_t *)k;
410 switch(length) /* all the case statements fall through */
411 {
412 case 12: c+=k[2]; b+=k[1]; a+=k[0]; break;
413 case 11: c+=((uint32_t)k8[10])<<8; /* fall through */
414 case 10: c+=((uint32_t)k8[9])<<16; /* fall through */
415 case 9 : c+=((uint32_t)k8[8])<<24; /* fall through */
416 case 8 : b+=k[1]; a+=k[0]; break;
417 case 7 : b+=((uint32_t)k8[6])<<8; /* fall through */
418 case 6 : b+=((uint32_t)k8[5])<<16; /* fall through */
419 case 5 : b+=((uint32_t)k8[4])<<24; /* fall through */
420 case 4 : a+=k[0]; break;
421 case 3 : a+=((uint32_t)k8[2])<<8; /* fall through */
422 case 2 : a+=((uint32_t)k8[1])<<16; /* fall through */
423 case 1 : a+=((uint32_t)k8[0])<<24; break;
424 case 0 : return c;
425 }
426
427#endif /* !VALGRIND */
428
429 } else { /* need to read the key one byte at a time */
430 const uint8_t *k = (const uint8_t *)key;
431
432 /*--------------- all but the last block: affect some 32 bits of (a,b,c) */
433 while (length > 12)
434 {
435 a += ((uint32_t)k[0])<<24;
436 a += ((uint32_t)k[1])<<16;
437 a += ((uint32_t)k[2])<<8;
438 a += ((uint32_t)k[3]);
439 b += ((uint32_t)k[4])<<24;
440 b += ((uint32_t)k[5])<<16;
441 b += ((uint32_t)k[6])<<8;
442 b += ((uint32_t)k[7]);
443 c += ((uint32_t)k[8])<<24;
444 c += ((uint32_t)k[9])<<16;
445 c += ((uint32_t)k[10])<<8;
446 c += ((uint32_t)k[11]);
447 mix(a,b,c);
448 length -= 12;
449 k += 12;
450 }
451
452 /*-------------------------------- last block: affect all 32 bits of (c) */
453 switch(length) /* all the case statements fall through */
454 {
455 case 12: c+=k[11]; /* fall through */
456 case 11: c+=((uint32_t)k[10])<<8; /* fall through */
457 case 10: c+=((uint32_t)k[9])<<16; /* fall through */
458 case 9 : c+=((uint32_t)k[8])<<24; /* fall through */
459 case 8 : b+=k[7]; /* fall through */
460 case 7 : b+=((uint32_t)k[6])<<8; /* fall through */
461 case 6 : b+=((uint32_t)k[5])<<16; /* fall through */
462 case 5 : b+=((uint32_t)k[4])<<24; /* fall through */
463 case 4 : a+=k[3]; /* fall through */
464 case 3 : a+=((uint32_t)k[2])<<8; /* fall through */
465 case 2 : a+=((uint32_t)k[1])<<16; /* fall through */
466 case 1 : a+=((uint32_t)k[0])<<24; /* fall through */
467 break;
468 case 0 : return c;
469 }
470 }
471
472 final(a,b,c);
473 *val2 = b;
474 return c;
475}
476
477uint32_t nl_hash_any(const void *key, size_t length, uint32_t base)
478{
479 if (HASH_BIG_ENDIAN)
480 return hashbig(key, length, &base);
481 else
482 return hashlittle(key, length, &base);
483}