]> git.zerfleddert.de Git - proxmark3-svn/blob - client/hardnested/hardnested_bitarray_core.c
fix iso14a_set_timeout() adjustment
[proxmark3-svn] / client / hardnested / hardnested_bitarray_core.c
1 //-----------------------------------------------------------------------------
2 // Copyright (C) 2016, 2017 by piwi
3 //
4 // This code is licensed to you under the terms of the GNU GPL, version 2 or,
5 // at your option, any later version. See the LICENSE.txt file for the text of
6 // the license.ch b
7 //-----------------------------------------------------------------------------
8 // Implements a card only attack based on crypto text (encrypted nonces
9 // received during a nested authentication) only. Unlike other card only
10 // attacks this doesn't rely on implementation errors but only on the
11 // inherent weaknesses of the crypto1 cypher. Described in
12 // Carlo Meijer, Roel Verdult, "Ciphertext-only Cryptanalysis on Hardened
13 // Mifare Classic Cards" in Proceedings of the 22nd ACM SIGSAC Conference on
14 // Computer and Communications Security, 2015
15 //-----------------------------------------------------------------------------
16 // some helper functions which can benefit from SIMD instructions or other special instructions
17 //
18
19 #include "hardnested_bitarray_core.h"
20
21 #include <stdint.h>
22 #include <stdio.h>
23 #include <stdlib.h>
24 #ifndef __APPLE__
25 #include <malloc.h>
26 #endif
27
28 // this needs to be compiled several times for each instruction set.
29 // For each instruction set, define a dedicated function name:
30 #if defined (__AVX512F__)
31 #define MALLOC_BITARRAY malloc_bitarray_AVX512
32 #define FREE_BITARRAY free_bitarray_AVX512
33 #define BITCOUNT bitcount_AVX512
34 #define COUNT_STATES count_states_AVX512
35 #define BITARRAY_AND bitarray_AND_AVX512
36 #define BITARRAY_LOW20_AND bitarray_low20_AND_AVX512
37 #define COUNT_BITARRAY_AND count_bitarray_AND_AVX512
38 #define COUNT_BITARRAY_LOW20_AND count_bitarray_low20_AND_AVX512
39 #define BITARRAY_AND4 bitarray_AND4_AVX512
40 #define BITARRAY_OR bitarray_OR_AVX512
41 #define COUNT_BITARRAY_AND2 count_bitarray_AND2_AVX512
42 #define COUNT_BITARRAY_AND3 count_bitarray_AND3_AVX512
43 #define COUNT_BITARRAY_AND4 count_bitarray_AND4_AVX512
44 #elif defined (__AVX2__)
45 #define MALLOC_BITARRAY malloc_bitarray_AVX2
46 #define FREE_BITARRAY free_bitarray_AVX2
47 #define BITCOUNT bitcount_AVX2
48 #define COUNT_STATES count_states_AVX2
49 #define BITARRAY_AND bitarray_AND_AVX2
50 #define BITARRAY_LOW20_AND bitarray_low20_AND_AVX2
51 #define COUNT_BITARRAY_AND count_bitarray_AND_AVX2
52 #define COUNT_BITARRAY_LOW20_AND count_bitarray_low20_AND_AVX2
53 #define BITARRAY_AND4 bitarray_AND4_AVX2
54 #define BITARRAY_OR bitarray_OR_AVX2
55 #define COUNT_BITARRAY_AND2 count_bitarray_AND2_AVX2
56 #define COUNT_BITARRAY_AND3 count_bitarray_AND3_AVX2
57 #define COUNT_BITARRAY_AND4 count_bitarray_AND4_AVX2
58 #elif defined (__AVX__)
59 #define MALLOC_BITARRAY malloc_bitarray_AVX
60 #define FREE_BITARRAY free_bitarray_AVX
61 #define BITCOUNT bitcount_AVX
62 #define COUNT_STATES count_states_AVX
63 #define BITARRAY_AND bitarray_AND_AVX
64 #define BITARRAY_LOW20_AND bitarray_low20_AND_AVX
65 #define COUNT_BITARRAY_AND count_bitarray_AND_AVX
66 #define COUNT_BITARRAY_LOW20_AND count_bitarray_low20_AND_AVX
67 #define BITARRAY_AND4 bitarray_AND4_AVX
68 #define BITARRAY_OR bitarray_OR_AVX
69 #define COUNT_BITARRAY_AND2 count_bitarray_AND2_AVX
70 #define COUNT_BITARRAY_AND3 count_bitarray_AND3_AVX
71 #define COUNT_BITARRAY_AND4 count_bitarray_AND4_AVX
72 #elif defined (__SSE2__)
73 #define MALLOC_BITARRAY malloc_bitarray_SSE2
74 #define FREE_BITARRAY free_bitarray_SSE2
75 #define BITCOUNT bitcount_SSE2
76 #define COUNT_STATES count_states_SSE2
77 #define BITARRAY_AND bitarray_AND_SSE2
78 #define BITARRAY_LOW20_AND bitarray_low20_AND_SSE2
79 #define COUNT_BITARRAY_AND count_bitarray_AND_SSE2
80 #define COUNT_BITARRAY_LOW20_AND count_bitarray_low20_AND_SSE2
81 #define BITARRAY_AND4 bitarray_AND4_SSE2
82 #define BITARRAY_OR bitarray_OR_SSE2
83 #define COUNT_BITARRAY_AND2 count_bitarray_AND2_SSE2
84 #define COUNT_BITARRAY_AND3 count_bitarray_AND3_SSE2
85 #define COUNT_BITARRAY_AND4 count_bitarray_AND4_SSE2
86 #elif defined (__MMX__)
87 #define MALLOC_BITARRAY malloc_bitarray_MMX
88 #define FREE_BITARRAY free_bitarray_MMX
89 #define BITCOUNT bitcount_MMX
90 #define COUNT_STATES count_states_MMX
91 #define BITARRAY_AND bitarray_AND_MMX
92 #define BITARRAY_LOW20_AND bitarray_low20_AND_MMX
93 #define COUNT_BITARRAY_AND count_bitarray_AND_MMX
94 #define COUNT_BITARRAY_LOW20_AND count_bitarray_low20_AND_MMX
95 #define BITARRAY_AND4 bitarray_AND4_MMX
96 #define BITARRAY_OR bitarray_OR_MMX
97 #define COUNT_BITARRAY_AND2 count_bitarray_AND2_MMX
98 #define COUNT_BITARRAY_AND3 count_bitarray_AND3_MMX
99 #define COUNT_BITARRAY_AND4 count_bitarray_AND4_MMX
100 #else
101 #define MALLOC_BITARRAY malloc_bitarray_NOSIMD
102 #define FREE_BITARRAY free_bitarray_NOSIMD
103 #define BITCOUNT bitcount_NOSIMD
104 #define COUNT_STATES count_states_NOSIMD
105 #define BITARRAY_AND bitarray_AND_NOSIMD
106 #define BITARRAY_LOW20_AND bitarray_low20_AND_NOSIMD
107 #define COUNT_BITARRAY_AND count_bitarray_AND_NOSIMD
108 #define COUNT_BITARRAY_LOW20_AND count_bitarray_low20_AND_NOSIMD
109 #define BITARRAY_AND4 bitarray_AND4_NOSIMD
110 #define BITARRAY_OR bitarray_OR_NOSIMD
111 #define COUNT_BITARRAY_AND2 count_bitarray_AND2_NOSIMD
112 #define COUNT_BITARRAY_AND3 count_bitarray_AND3_NOSIMD
113 #define COUNT_BITARRAY_AND4 count_bitarray_AND4_NOSIMD
114 #endif
115
116
117 // typedefs and declaration of functions:
118 typedef uint32_t* malloc_bitarray_t(uint32_t);
119 malloc_bitarray_t malloc_bitarray_AVX512, malloc_bitarray_AVX2, malloc_bitarray_AVX, malloc_bitarray_SSE2, malloc_bitarray_MMX, malloc_bitarray_NOSIMD, malloc_bitarray_dispatch;
120 typedef void free_bitarray_t(uint32_t*);
121 free_bitarray_t free_bitarray_AVX512, free_bitarray_AVX2, free_bitarray_AVX, free_bitarray_SSE2, free_bitarray_MMX, free_bitarray_NOSIMD, free_bitarray_dispatch;
122 typedef uint32_t bitcount_t(uint32_t);
123 bitcount_t bitcount_AVX512, bitcount_AVX2, bitcount_AVX, bitcount_SSE2, bitcount_MMX, bitcount_NOSIMD, bitcount_dispatch;
124 typedef uint32_t count_states_t(uint32_t*);
125 count_states_t count_states_AVX512, count_states_AVX2, count_states_AVX, count_states_SSE2, count_states_MMX, count_states_NOSIMD, count_states_dispatch;
126 typedef void bitarray_AND_t(uint32_t[], uint32_t[]);
127 bitarray_AND_t bitarray_AND_AVX512, bitarray_AND_AVX2, bitarray_AND_AVX, bitarray_AND_SSE2, bitarray_AND_MMX, bitarray_AND_NOSIMD, bitarray_AND_dispatch;
128 typedef void bitarray_low20_AND_t(uint32_t*, uint32_t*);
129 bitarray_low20_AND_t bitarray_low20_AND_AVX512, bitarray_low20_AND_AVX2, bitarray_low20_AND_AVX, bitarray_low20_AND_SSE2, bitarray_low20_AND_MMX, bitarray_low20_AND_NOSIMD, bitarray_low20_AND_dispatch;
130 typedef uint32_t count_bitarray_AND_t(uint32_t*, uint32_t*);
131 count_bitarray_AND_t count_bitarray_AND_AVX512, count_bitarray_AND_AVX2, count_bitarray_AND_AVX, count_bitarray_AND_SSE2, count_bitarray_AND_MMX, count_bitarray_AND_NOSIMD, count_bitarray_AND_dispatch;
132 typedef uint32_t count_bitarray_low20_AND_t(uint32_t*, uint32_t*);
133 count_bitarray_low20_AND_t count_bitarray_low20_AND_AVX512, count_bitarray_low20_AND_AVX2, count_bitarray_low20_AND_AVX, count_bitarray_low20_AND_SSE2, count_bitarray_low20_AND_MMX, count_bitarray_low20_AND_NOSIMD, count_bitarray_low20_AND_dispatch;
134 typedef void bitarray_AND4_t(uint32_t*, uint32_t*, uint32_t*, uint32_t*);
135 bitarray_AND4_t bitarray_AND4_AVX512, bitarray_AND4_AVX2, bitarray_AND4_AVX, bitarray_AND4_SSE2, bitarray_AND4_MMX, bitarray_AND4_NOSIMD, bitarray_AND4_dispatch;
136 typedef void bitarray_OR_t(uint32_t[], uint32_t[]);
137 bitarray_OR_t bitarray_OR_AVX512, bitarray_OR_AVX2, bitarray_OR_AVX, bitarray_OR_SSE2, bitarray_OR_MMX, bitarray_OR_NOSIMD, bitarray_OR_dispatch;
138 typedef uint32_t count_bitarray_AND2_t(uint32_t*, uint32_t*);
139 count_bitarray_AND2_t count_bitarray_AND2_AVX512, count_bitarray_AND2_AVX2, count_bitarray_AND2_AVX, count_bitarray_AND2_SSE2, count_bitarray_AND2_MMX, count_bitarray_AND2_NOSIMD, count_bitarray_AND2_dispatch;
140 typedef uint32_t count_bitarray_AND3_t(uint32_t*, uint32_t*, uint32_t*);
141 count_bitarray_AND3_t count_bitarray_AND3_AVX512, count_bitarray_AND3_AVX2, count_bitarray_AND3_AVX, count_bitarray_AND3_SSE2, count_bitarray_AND3_MMX, count_bitarray_AND3_NOSIMD, count_bitarray_AND3_dispatch;
142 typedef uint32_t count_bitarray_AND4_t(uint32_t*, uint32_t*, uint32_t*, uint32_t*);
143 count_bitarray_AND4_t count_bitarray_AND4_AVX512, count_bitarray_AND4_AVX2, count_bitarray_AND4_AVX, count_bitarray_AND4_SSE2, count_bitarray_AND4_MMX, count_bitarray_AND4_NOSIMD, count_bitarray_AND4_dispatch;
144
145
146 inline uint32_t *MALLOC_BITARRAY(uint32_t x)
147 {
148 #if defined (_WIN32)
149 return __builtin_assume_aligned(_aligned_malloc((x), __BIGGEST_ALIGNMENT__), __BIGGEST_ALIGNMENT__);
150 #elif defined (__APPLE__)
151 uint32_t *allocated_memory;
152 if (posix_memalign((void**)&allocated_memory, __BIGGEST_ALIGNMENT__, x)) {
153 return NULL;
154 } else {
155 return __builtin_assume_aligned(allocated_memory, __BIGGEST_ALIGNMENT__);
156 }
157 #else
158 return __builtin_assume_aligned(memalign(__BIGGEST_ALIGNMENT__, (x)), __BIGGEST_ALIGNMENT__);
159 #endif
160 }
161
162
163 inline void FREE_BITARRAY(uint32_t *x)
164 {
165 #ifdef _WIN32
166 _aligned_free(x);
167 #else
168 free(x);
169 #endif
170 }
171
172
173 inline uint32_t BITCOUNT(uint32_t a)
174 {
175 return __builtin_popcountl(a);
176 }
177
178
179 inline uint32_t COUNT_STATES(uint32_t *A)
180 {
181 uint32_t count = 0;
182 for (uint32_t i = 0; i < (1<<19); i++) {
183 count += BITCOUNT(A[i]);
184 }
185 return count;
186 }
187
188
189 inline void BITARRAY_AND(uint32_t *restrict A, uint32_t *restrict B)
190 {
191 A = __builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
192 B = __builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
193 for (uint32_t i = 0; i < (1<<19); i++) {
194 A[i] &= B[i];
195 }
196 }
197
198
199 inline void BITARRAY_LOW20_AND(uint32_t *restrict A, uint32_t *restrict B)
200 {
201 uint16_t *a = (uint16_t *)__builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
202 uint16_t *b = (uint16_t *)__builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
203
204 for (uint32_t i = 0; i < (1<<20); i++) {
205 if (!b[i]) {
206 a[i] = 0;
207 }
208 }
209 }
210
211
212 inline uint32_t COUNT_BITARRAY_AND(uint32_t *restrict A, uint32_t *restrict B)
213 {
214 A = __builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
215 B = __builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
216 uint32_t count = 0;
217 for (uint32_t i = 0; i < (1<<19); i++) {
218 A[i] &= B[i];
219 count += BITCOUNT(A[i]);
220 }
221 return count;
222 }
223
224
225 inline uint32_t COUNT_BITARRAY_LOW20_AND(uint32_t *restrict A, uint32_t *restrict B)
226 {
227 uint16_t *a = (uint16_t *)__builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
228 uint16_t *b = (uint16_t *)__builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
229 uint32_t count = 0;
230
231 for (uint32_t i = 0; i < (1<<20); i++) {
232 if (!b[i]) {
233 a[i] = 0;
234 }
235 count += BITCOUNT(a[i]);
236 }
237 return count;
238 }
239
240
241 inline void BITARRAY_AND4(uint32_t *restrict A, uint32_t *restrict B, uint32_t *restrict C, uint32_t *restrict D)
242 {
243 A = __builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
244 B = __builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
245 C = __builtin_assume_aligned(C, __BIGGEST_ALIGNMENT__);
246 D = __builtin_assume_aligned(D, __BIGGEST_ALIGNMENT__);
247 for (uint32_t i = 0; i < (1<<19); i++) {
248 A[i] = B[i] & C[i] & D[i];
249 }
250 }
251
252
253 inline void BITARRAY_OR(uint32_t *restrict A, uint32_t *restrict B)
254 {
255 A = __builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
256 B = __builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
257 for (uint32_t i = 0; i < (1<<19); i++) {
258 A[i] |= B[i];
259 }
260 }
261
262
263 inline uint32_t COUNT_BITARRAY_AND2(uint32_t *restrict A, uint32_t *restrict B)
264 {
265 A = __builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
266 B = __builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
267 uint32_t count = 0;
268 for (uint32_t i = 0; i < (1<<19); i++) {
269 count += BITCOUNT(A[i] & B[i]);
270 }
271 return count;
272 }
273
274
275 inline uint32_t COUNT_BITARRAY_AND3(uint32_t *restrict A, uint32_t *restrict B, uint32_t *restrict C)
276 {
277 A = __builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
278 B = __builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
279 C = __builtin_assume_aligned(C, __BIGGEST_ALIGNMENT__);
280 uint32_t count = 0;
281 for (uint32_t i = 0; i < (1<<19); i++) {
282 count += BITCOUNT(A[i] & B[i] & C[i]);
283 }
284 return count;
285 }
286
287
288 inline uint32_t COUNT_BITARRAY_AND4(uint32_t *restrict A, uint32_t *restrict B, uint32_t *restrict C, uint32_t *restrict D)
289 {
290 A = __builtin_assume_aligned(A, __BIGGEST_ALIGNMENT__);
291 B = __builtin_assume_aligned(B, __BIGGEST_ALIGNMENT__);
292 C = __builtin_assume_aligned(C, __BIGGEST_ALIGNMENT__);
293 D = __builtin_assume_aligned(D, __BIGGEST_ALIGNMENT__);
294 uint32_t count = 0;
295 for (uint32_t i = 0; i < (1<<19); i++) {
296 count += BITCOUNT(A[i] & B[i] & C[i] & D[i]);
297 }
298 return count;
299 }
300
301
302 #ifndef __MMX__
303
304 // pointers to functions:
305 malloc_bitarray_t *malloc_bitarray_function_p = &malloc_bitarray_dispatch;
306 free_bitarray_t *free_bitarray_function_p = &free_bitarray_dispatch;
307 bitcount_t *bitcount_function_p = &bitcount_dispatch;
308 count_states_t *count_states_function_p = &count_states_dispatch;
309 bitarray_AND_t *bitarray_AND_function_p = &bitarray_AND_dispatch;
310 bitarray_low20_AND_t *bitarray_low20_AND_function_p = &bitarray_low20_AND_dispatch;
311 count_bitarray_AND_t *count_bitarray_AND_function_p = &count_bitarray_AND_dispatch;
312 count_bitarray_low20_AND_t *count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_dispatch;
313 bitarray_AND4_t *bitarray_AND4_function_p = &bitarray_AND4_dispatch;
314 bitarray_OR_t *bitarray_OR_function_p = &bitarray_OR_dispatch;
315 count_bitarray_AND2_t *count_bitarray_AND2_function_p = &count_bitarray_AND2_dispatch;
316 count_bitarray_AND3_t *count_bitarray_AND3_function_p = &count_bitarray_AND3_dispatch;
317 count_bitarray_AND4_t *count_bitarray_AND4_function_p = &count_bitarray_AND4_dispatch;
318
319 // determine the available instruction set at runtime and call the correct function
320 uint32_t *malloc_bitarray_dispatch(uint32_t x) {
321 #if defined (__i386__) || defined (__x86_64__)
322 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
323 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
324 if (__builtin_cpu_supports("avx512f")) malloc_bitarray_function_p = &malloc_bitarray_AVX512;
325 else if (__builtin_cpu_supports("avx2")) malloc_bitarray_function_p = &malloc_bitarray_AVX2;
326 #else
327 if (__builtin_cpu_supports("avx2")) malloc_bitarray_function_p = &malloc_bitarray_AVX2;
328 #endif
329 else if (__builtin_cpu_supports("avx")) malloc_bitarray_function_p = &malloc_bitarray_AVX;
330 else if (__builtin_cpu_supports("sse2")) malloc_bitarray_function_p = &malloc_bitarray_SSE2;
331 else if (__builtin_cpu_supports("mmx")) malloc_bitarray_function_p = &malloc_bitarray_MMX;
332 else
333 #endif
334 #endif
335 malloc_bitarray_function_p = &malloc_bitarray_NOSIMD;
336
337 // call the most optimized function for this CPU
338 return (*malloc_bitarray_function_p)(x);
339 }
340
341 void free_bitarray_dispatch(uint32_t *x) {
342 #if defined (__i386__) || defined (__x86_64__)
343 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
344 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
345 if (__builtin_cpu_supports("avx512f")) free_bitarray_function_p = &free_bitarray_AVX512;
346 else if (__builtin_cpu_supports("avx2")) free_bitarray_function_p = &free_bitarray_AVX2;
347 #else
348 if (__builtin_cpu_supports("avx2")) free_bitarray_function_p = &free_bitarray_AVX2;
349 #endif
350 else if (__builtin_cpu_supports("avx")) free_bitarray_function_p = &free_bitarray_AVX;
351 else if (__builtin_cpu_supports("sse2")) free_bitarray_function_p = &free_bitarray_SSE2;
352 else if (__builtin_cpu_supports("mmx")) free_bitarray_function_p = &free_bitarray_MMX;
353 else
354 #endif
355 #endif
356 free_bitarray_function_p = &free_bitarray_NOSIMD;
357
358 // call the most optimized function for this CPU
359 (*free_bitarray_function_p)(x);
360 }
361
362 uint32_t bitcount_dispatch(uint32_t a) {
363 #if defined (__i386__) || defined (__x86_64__)
364 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
365 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
366 if (__builtin_cpu_supports("avx512f")) bitcount_function_p = &bitcount_AVX512;
367 else if (__builtin_cpu_supports("avx2")) bitcount_function_p = &bitcount_AVX2;
368 #else
369 if (__builtin_cpu_supports("avx2")) bitcount_function_p = &bitcount_AVX2;
370 #endif
371 else if (__builtin_cpu_supports("avx")) bitcount_function_p = &bitcount_AVX;
372 else if (__builtin_cpu_supports("sse2")) bitcount_function_p = &bitcount_SSE2;
373 else if (__builtin_cpu_supports("mmx")) bitcount_function_p = &bitcount_MMX;
374 else
375 #endif
376 #endif
377 bitcount_function_p = &bitcount_NOSIMD;
378
379 // call the most optimized function for this CPU
380 return (*bitcount_function_p)(a);
381 }
382
383 uint32_t count_states_dispatch(uint32_t *bitarray) {
384 #if defined (__i386__) || defined (__x86_64__)
385 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
386 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
387 if (__builtin_cpu_supports("avx512f")) count_states_function_p = &count_states_AVX512;
388 else if (__builtin_cpu_supports("avx2")) count_states_function_p = &count_states_AVX2;
389 #else
390 if (__builtin_cpu_supports("avx2")) count_states_function_p = &count_states_AVX2;
391 #endif
392 else if (__builtin_cpu_supports("avx")) count_states_function_p = &count_states_AVX;
393 else if (__builtin_cpu_supports("sse2")) count_states_function_p = &count_states_SSE2;
394 else if (__builtin_cpu_supports("mmx")) count_states_function_p = &count_states_MMX;
395 else
396 #endif
397 #endif
398 count_states_function_p = &count_states_NOSIMD;
399
400 // call the most optimized function for this CPU
401 return (*count_states_function_p)(bitarray);
402 }
403
404 void bitarray_AND_dispatch(uint32_t *A, uint32_t *B) {
405 #if defined (__i386__) || defined (__x86_64__)
406 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
407 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
408 if (__builtin_cpu_supports("avx512f")) bitarray_AND_function_p = &bitarray_AND_AVX512;
409 else if (__builtin_cpu_supports("avx2")) bitarray_AND_function_p = &bitarray_AND_AVX2;
410 #else
411 if (__builtin_cpu_supports("avx2")) bitarray_AND_function_p = &bitarray_AND_AVX2;
412 #endif
413 else if (__builtin_cpu_supports("avx")) bitarray_AND_function_p = &bitarray_AND_AVX;
414 else if (__builtin_cpu_supports("sse2")) bitarray_AND_function_p = &bitarray_AND_SSE2;
415 else if (__builtin_cpu_supports("mmx")) bitarray_AND_function_p = &bitarray_AND_MMX;
416 else
417 #endif
418 #endif
419 bitarray_AND_function_p = &bitarray_AND_NOSIMD;
420
421 // call the most optimized function for this CPU
422 (*bitarray_AND_function_p)(A,B);
423 }
424
425 void bitarray_low20_AND_dispatch(uint32_t *A, uint32_t *B) {
426 #if defined (__i386__) || defined (__x86_64__)
427 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
428 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
429 if (__builtin_cpu_supports("avx512f")) bitarray_low20_AND_function_p = &bitarray_low20_AND_AVX512;
430 else if (__builtin_cpu_supports("avx2")) bitarray_low20_AND_function_p = &bitarray_low20_AND_AVX2;
431 #else
432 if (__builtin_cpu_supports("avx2")) bitarray_low20_AND_function_p = &bitarray_low20_AND_AVX2;
433 #endif
434 else if (__builtin_cpu_supports("avx")) bitarray_low20_AND_function_p = &bitarray_low20_AND_AVX;
435 else if (__builtin_cpu_supports("sse2")) bitarray_low20_AND_function_p = &bitarray_low20_AND_SSE2;
436 else if (__builtin_cpu_supports("mmx")) bitarray_low20_AND_function_p = &bitarray_low20_AND_MMX;
437 else
438 #endif
439 #endif
440 bitarray_low20_AND_function_p = &bitarray_low20_AND_NOSIMD;
441
442 // call the most optimized function for this CPU
443 (*bitarray_low20_AND_function_p)(A, B);
444 }
445
446 uint32_t count_bitarray_AND_dispatch(uint32_t *A, uint32_t *B) {
447 #if defined (__i386__) || defined (__x86_64__)
448 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
449 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
450 if (__builtin_cpu_supports("avx512f")) count_bitarray_AND_function_p = &count_bitarray_AND_AVX512;
451 else if (__builtin_cpu_supports("avx2")) count_bitarray_AND_function_p = &count_bitarray_AND_AVX2;
452 #else
453 if (__builtin_cpu_supports("avx2")) count_bitarray_AND_function_p = &count_bitarray_AND_AVX2;
454 #endif
455 else if (__builtin_cpu_supports("avx")) count_bitarray_AND_function_p = &count_bitarray_AND_AVX;
456 else if (__builtin_cpu_supports("sse2")) count_bitarray_AND_function_p = &count_bitarray_AND_SSE2;
457 else if (__builtin_cpu_supports("mmx")) count_bitarray_AND_function_p = &count_bitarray_AND_MMX;
458 else
459 #endif
460 #endif
461 count_bitarray_AND_function_p = &count_bitarray_AND_NOSIMD;
462
463 // call the most optimized function for this CPU
464 return (*count_bitarray_AND_function_p)(A, B);
465 }
466
467 uint32_t count_bitarray_low20_AND_dispatch(uint32_t *A, uint32_t *B) {
468 #if defined (__i386__) || defined (__x86_64__)
469 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
470 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
471 if (__builtin_cpu_supports("avx512f")) count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_AVX512;
472 else if (__builtin_cpu_supports("avx2")) count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_AVX2;
473 #else
474 if (__builtin_cpu_supports("avx2")) count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_AVX2;
475 #endif
476 else if (__builtin_cpu_supports("avx")) count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_AVX;
477 else if (__builtin_cpu_supports("sse2")) count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_SSE2;
478 else if (__builtin_cpu_supports("mmx")) count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_MMX;
479 else
480 #endif
481 #endif
482 count_bitarray_low20_AND_function_p = &count_bitarray_low20_AND_NOSIMD;
483
484 // call the most optimized function for this CPU
485 return (*count_bitarray_low20_AND_function_p)(A, B);
486 }
487
488 void bitarray_AND4_dispatch(uint32_t *A, uint32_t *B, uint32_t *C, uint32_t *D) {
489 #if defined (__i386__) || defined (__x86_64__)
490 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
491 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
492 if (__builtin_cpu_supports("avx512f")) bitarray_AND4_function_p = &bitarray_AND4_AVX512;
493 else if (__builtin_cpu_supports("avx2")) bitarray_AND4_function_p = &bitarray_AND4_AVX2;
494 #else
495 if (__builtin_cpu_supports("avx2")) bitarray_AND4_function_p = &bitarray_AND4_AVX2;
496 #endif
497 else if (__builtin_cpu_supports("avx")) bitarray_AND4_function_p = &bitarray_AND4_AVX;
498 else if (__builtin_cpu_supports("sse2")) bitarray_AND4_function_p = &bitarray_AND4_SSE2;
499 else if (__builtin_cpu_supports("mmx")) bitarray_AND4_function_p = &bitarray_AND4_MMX;
500 else
501 #endif
502 #endif
503 bitarray_AND4_function_p = &bitarray_AND4_NOSIMD;
504
505 // call the most optimized function for this CPU
506 (*bitarray_AND4_function_p)(A, B, C, D);
507 }
508
509 void bitarray_OR_dispatch(uint32_t *A, uint32_t *B) {
510 #if defined (__i386__) || defined (__x86_64__)
511 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
512 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
513 if (__builtin_cpu_supports("avx512f")) bitarray_OR_function_p = &bitarray_OR_AVX512;
514 else if (__builtin_cpu_supports("avx2")) bitarray_OR_function_p = &bitarray_OR_AVX2;
515 #else
516 if (__builtin_cpu_supports("avx2")) bitarray_OR_function_p = &bitarray_OR_AVX2;
517 #endif
518 else if (__builtin_cpu_supports("avx")) bitarray_OR_function_p = &bitarray_OR_AVX;
519 else if (__builtin_cpu_supports("sse2")) bitarray_OR_function_p = &bitarray_OR_SSE2;
520 else if (__builtin_cpu_supports("mmx")) bitarray_OR_function_p = &bitarray_OR_MMX;
521 else
522 #endif
523 #endif
524 bitarray_OR_function_p = &bitarray_OR_NOSIMD;
525
526 // call the most optimized function for this CPU
527 (*bitarray_OR_function_p)(A,B);
528 }
529
530 uint32_t count_bitarray_AND2_dispatch(uint32_t *A, uint32_t *B) {
531 #if defined (__i386__) || defined (__x86_64__)
532 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
533 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
534 if (__builtin_cpu_supports("avx512f")) count_bitarray_AND2_function_p = &count_bitarray_AND2_AVX512;
535 else if (__builtin_cpu_supports("avx2")) count_bitarray_AND2_function_p = &count_bitarray_AND2_AVX2;
536 #else
537 if (__builtin_cpu_supports("avx2")) count_bitarray_AND2_function_p = &count_bitarray_AND2_AVX2;
538 #endif
539 else if (__builtin_cpu_supports("avx")) count_bitarray_AND2_function_p = &count_bitarray_AND2_AVX;
540 else if (__builtin_cpu_supports("sse2")) count_bitarray_AND2_function_p = &count_bitarray_AND2_SSE2;
541 else if (__builtin_cpu_supports("mmx")) count_bitarray_AND2_function_p = &count_bitarray_AND2_MMX;
542 else
543 #endif
544 #endif
545 count_bitarray_AND2_function_p = &count_bitarray_AND2_NOSIMD;
546
547 // call the most optimized function for this CPU
548 return (*count_bitarray_AND2_function_p)(A, B);
549 }
550
551 uint32_t count_bitarray_AND3_dispatch(uint32_t *A, uint32_t *B, uint32_t *C) {
552 #if defined (__i386__) || defined (__x86_64__)
553 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
554 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
555 if (__builtin_cpu_supports("avx512f")) count_bitarray_AND3_function_p = &count_bitarray_AND3_AVX512;
556 else if (__builtin_cpu_supports("avx2")) count_bitarray_AND3_function_p = &count_bitarray_AND3_AVX2;
557 #else
558 if (__builtin_cpu_supports("avx2")) count_bitarray_AND3_function_p = &count_bitarray_AND3_AVX2;
559 #endif
560 else if (__builtin_cpu_supports("avx")) count_bitarray_AND3_function_p = &count_bitarray_AND3_AVX;
561 else if (__builtin_cpu_supports("sse2")) count_bitarray_AND3_function_p = &count_bitarray_AND3_SSE2;
562 else if (__builtin_cpu_supports("mmx")) count_bitarray_AND3_function_p = &count_bitarray_AND3_MMX;
563 else
564 #endif
565 #endif
566 count_bitarray_AND3_function_p = &count_bitarray_AND3_NOSIMD;
567
568 // call the most optimized function for this CPU
569 return (*count_bitarray_AND3_function_p)(A, B, C);
570 }
571
572 uint32_t count_bitarray_AND4_dispatch(uint32_t *A, uint32_t *B, uint32_t *C, uint32_t *D) {
573 #if defined (__i386__) || defined (__x86_64__)
574 #if !defined(__APPLE__) || (defined(__APPLE__) && (__clang_major__ > 8 || __clang_major__ == 8 && __clang_minor__ >= 1))
575 #if (__GNUC__ >= 5) && (__GNUC__ > 5 || __GNUC_MINOR__ > 2)
576 if (__builtin_cpu_supports("avx512f")) count_bitarray_AND4_function_p = &count_bitarray_AND4_AVX512;
577 else if (__builtin_cpu_supports("avx2")) count_bitarray_AND4_function_p = &count_bitarray_AND4_AVX2;
578 #else
579 if (__builtin_cpu_supports("avx2")) count_bitarray_AND4_function_p = &count_bitarray_AND4_AVX2;
580 #endif
581 else if (__builtin_cpu_supports("avx")) count_bitarray_AND4_function_p = &count_bitarray_AND4_AVX;
582 else if (__builtin_cpu_supports("sse2")) count_bitarray_AND4_function_p = &count_bitarray_AND4_SSE2;
583 else if (__builtin_cpu_supports("mmx")) count_bitarray_AND4_function_p = &count_bitarray_AND4_MMX;
584 else
585 #endif
586 #endif
587 count_bitarray_AND4_function_p = &count_bitarray_AND4_NOSIMD;
588
589 // call the most optimized function for this CPU
590 return (*count_bitarray_AND4_function_p)(A, B, C, D);
591 }
592
593
594 ///////////////////////////////////////////////77
595 // Entries to dispatched function calls
596
597 uint32_t *malloc_bitarray(uint32_t x) {
598 return (*malloc_bitarray_function_p)(x);
599 }
600
601 void free_bitarray(uint32_t *x) {
602 (*free_bitarray_function_p)(x);
603 }
604
605 uint32_t bitcount(uint32_t a) {
606 return (*bitcount_function_p)(a);
607 }
608
609 uint32_t count_states(uint32_t *bitarray) {
610 return (*count_states_function_p)(bitarray);
611 }
612
613 void bitarray_AND(uint32_t *A, uint32_t *B) {
614 (*bitarray_AND_function_p)(A, B);
615 }
616
617 void bitarray_low20_AND(uint32_t *A, uint32_t *B) {
618 (*bitarray_low20_AND_function_p)(A, B);
619 }
620
621 uint32_t count_bitarray_AND(uint32_t *A, uint32_t *B) {
622 return (*count_bitarray_AND_function_p)(A, B);
623 }
624
625 uint32_t count_bitarray_low20_AND(uint32_t *A, uint32_t *B) {
626 return (*count_bitarray_low20_AND_function_p)(A, B);
627 }
628
629 void bitarray_AND4(uint32_t *A, uint32_t *B, uint32_t *C, uint32_t *D) {
630 (*bitarray_AND4_function_p)(A, B, C, D);
631 }
632
633 void bitarray_OR(uint32_t *A, uint32_t *B) {
634 (*bitarray_OR_function_p)(A, B);
635 }
636
637 uint32_t count_bitarray_AND2(uint32_t *A, uint32_t *B) {
638 return (*count_bitarray_AND2_function_p)(A, B);
639 }
640
641 uint32_t count_bitarray_AND3(uint32_t *A, uint32_t *B, uint32_t *C) {
642 return (*count_bitarray_AND3_function_p)(A, B, C);
643 }
644
645 uint32_t count_bitarray_AND4(uint32_t *A, uint32_t *B, uint32_t *C, uint32_t *D) {
646 return (*count_bitarray_AND4_function_p)(A, B, C, D);
647 }
648
649 #endif
650
Impressum, Datenschutz