forked from vectorclass/version2
-
Notifications
You must be signed in to change notification settings - Fork 0
/
vectori512s.h
2294 lines (2027 loc) · 93.2 KB
/
vectori512s.h
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
/**************************** vectori512s.h ********************************
* Author: Agner Fog
* Date created: 2019-04-20
* Last modified: 2020-02-23
* Version: 2.01.01
* Project: vector classes
* Description:
* Header file defining 512-bit integer vector classes for 8 and 16 bit integers.
* For x86 microprocessors with AVX512BW and later instruction sets.
*
* Instructions: see vcl_manual.pdf
*
* The following vector classes are defined here:
* Vec64c Vector of 64 8-bit signed integers
* Vec64uc Vector of 64 8-bit unsigned integers
* Vec64cb Vector of 64 booleans for use with Vec64c and Vec64uc
* Vec32s Vector of 32 16-bit signed integers
* Vec32us Vector of 32 16-bit unsigned integers
* Vec32sb Vector of 32 booleans for use with Vec32s and Vec32us
* Other 512-bit integer vectors are defined in Vectori512.h
*
* Each vector object is represented internally in the CPU as a 512-bit register.
* This header file defines operators and functions for these vectors.
*
* (c) Copyright 2012-2020 Agner Fog.
* Apache License version 2.0 or later.
******************************************************************************/
#ifndef VECTORI512S_H
#define VECTORI512S_H
#ifndef VECTORCLASS_H
#include "vectorclass.h"
#endif
#if VECTORCLASS_H < 20100
#error Incompatible versions of vector class library mixed
#endif
// check combination of header files
#ifdef VECTORI512SE_H
#error Two different versions of vectorf256.h included
#endif
#ifdef VCL_NAMESPACE
namespace VCL_NAMESPACE {
#endif
/*****************************************************************************
*
* Vector of 64 8-bit signed integers
*
*****************************************************************************/
class Vec64c: public Vec512b {
public:
// Default constructor:
Vec64c() {
}
// Constructor to broadcast the same value into all elements:
Vec64c(int8_t i) {
zmm = _mm512_set1_epi8(i);
}
// Constructor to build from all elements:
Vec64c(int8_t i0, int8_t i1, int8_t i2, int8_t i3, int8_t i4, int8_t i5, int8_t i6, int8_t i7,
int8_t i8, int8_t i9, int8_t i10, int8_t i11, int8_t i12, int8_t i13, int8_t i14, int8_t i15,
int8_t i16, int8_t i17, int8_t i18, int8_t i19, int8_t i20, int8_t i21, int8_t i22, int8_t i23,
int8_t i24, int8_t i25, int8_t i26, int8_t i27, int8_t i28, int8_t i29, int8_t i30, int8_t i31,
int8_t i32, int8_t i33, int8_t i34, int8_t i35, int8_t i36, int8_t i37, int8_t i38, int8_t i39,
int8_t i40, int8_t i41, int8_t i42, int8_t i43, int8_t i44, int8_t i45, int8_t i46, int8_t i47,
int8_t i48, int8_t i49, int8_t i50, int8_t i51, int8_t i52, int8_t i53, int8_t i54, int8_t i55,
int8_t i56, int8_t i57, int8_t i58, int8_t i59, int8_t i60, int8_t i61, int8_t i62, int8_t i63) {
// _mm512_set_epi8 and _mm512_set_epi16 missing in GCC 7.4.0
int8_t aa[64] = {
i0, i1, i2, i3, i4, i5, i6, i7,i8, i9, i10, i11, i12, i13, i14, i15,
i16, i17, i18, i19, i20, i21, i22, i23, i24, i25, i26, i27, i28, i29, i30, i31,
i32, i33, i34, i35, i36, i37, i38, i39, i40, i41, i42, i43, i44, i45, i46, i47,
i48, i49, i50, i51, i52, i53, i54, i55, i56, i57, i58, i59, i60, i61, i62, i63 };
load(aa);
}
// Constructor to build from two Vec32c:
Vec64c(Vec32c const a0, Vec32c const a1) {
zmm = _mm512_inserti64x4(_mm512_castsi256_si512(a0), a1, 1);
}
// Constructor to convert from type __m512i used in intrinsics:
Vec64c(__m512i const x) {
zmm = x;
}
// Assignment operator to convert from type __m512i used in intrinsics:
Vec64c & operator = (__m512i const x) {
zmm = x;
return *this;
}
// Type cast operator to convert to __m512i used in intrinsics
operator __m512i() const {
return zmm;
}
// Member function to load from array (unaligned)
Vec64c & load(void const * p) {
zmm = _mm512_loadu_si512(p);
return *this;
}
// Member function to load from array, aligned by 64
Vec64c & load_a(void const * p) {
zmm = _mm512_load_si512(p);
return *this;
}
// Partial load. Load n elements and set the rest to 0
Vec64c & load_partial(int n, void const * p) {
if (n >= 64) {
zmm = _mm512_loadu_si512(p);
}
else {
zmm = _mm512_maskz_loadu_epi8(__mmask64(((uint64_t)1 << n) - 1), p);
}
return *this;
}
// Partial store. Store n elements
void store_partial(int n, void * p) const {
if (n >= 64) {
// _mm512_storeu_epi8(p, zmm);
_mm512_storeu_si512(p, zmm);
}
else {
_mm512_mask_storeu_epi8(p, __mmask64(((uint64_t)1 << n) - 1), zmm);
}
}
// cut off vector to n elements. The last 64-n elements are set to zero
Vec64c & cutoff(int n) {
if (n < 64) {
zmm = _mm512_maskz_mov_epi8(__mmask64(((uint64_t)1 << n) - 1), zmm);
}
return *this;
}
// Member function to change a single element in vector
Vec64c const insert(int index, int8_t value) {
zmm = _mm512_mask_set1_epi8(zmm, __mmask64((uint64_t)1 << index), value);
return *this;
}
// Member function extract a single element from vector
int8_t extract(int index) const {
#if INSTRSET >= 10 && defined (__AVX512VBMI2__)
__m512i x = _mm512_maskz_compress_epi8(__mmask64((uint64_t)1 << index), zmm);
return (int8_t)_mm_cvtsi128_si32(_mm512_castsi512_si128(x));
#else
int8_t a[64];
store(a);
return a[index & 63];
#endif
}
// Extract a single element. Use store function if extracting more than one element.
// Operator [] can only read an element, not write.
int8_t operator [] (int index) const {
return extract(index);
}
// Member functions to split into two Vec32c:
Vec32c get_low() const {
return _mm512_castsi512_si256(zmm);
}
Vec32c get_high() const {
return _mm512_extracti64x4_epi64(zmm,1);
}
static constexpr int size() {
return 64;
}
static constexpr int elementtype() {
return 4;
}
};
/*****************************************************************************
*
* Vec64b: Vector of 64 Booleans for use with Vec64c and Vec64uc
*
*****************************************************************************/
class Vec64b {
protected:
__mmask64 mm; // Boolean vector
public:
// Default constructor:
Vec64b () {
}
// Constructor to build from all elements:
/*
Vec64b(bool b0, bool b1, bool b2, bool b3, bool b4, bool b5, bool b6, bool b7,
bool b8, bool b9, bool b10, bool b11, bool b12, bool b13, bool b14, bool b15,
bool b16, bool b17, bool b18, bool b19, bool b20, bool b21, bool b22, bool b23,
bool b24, bool b25, bool b26, bool b27, bool b28, bool b29, bool b30, bool b31,
bool b32, bool b33, bool b34, bool b35, bool b36, bool b37, bool b38, bool b39,
bool b40, bool b41, bool b42, bool b43, bool b44, bool b45, bool b46, bool b47,
bool b48, bool b49, bool b50, bool b51, bool b52, bool b53, bool b54, bool b55,
bool b56, bool b57, bool b58, bool b59, bool b60, bool b61, bool b62, bool b63) {
mm = uint64_t(
(uint64_t)b0 | (uint64_t)b1 << 1 | (uint64_t)b2 << 2 | (uint64_t)b3 << 3 |
(uint64_t)b4 << 4 | (uint64_t)b5 << 5 | (uint64_t)b6 << 6 | (uint64_t)b7 << 7 |
(uint64_t)b8 << 8 | (uint64_t)b9 << 9 | (uint64_t)b10 << 10 | (uint64_t)b11 << 11 |
(uint64_t)b12 << 12 | (uint64_t)b13 << 13 | (uint64_t)b14 << 14 | (uint64_t)b15 << 15 |
(uint64_t)b16 << 16 | (uint64_t)b17 << 17 | (uint64_t)b18 << 18 | (uint64_t)b19 << 19 |
(uint64_t)b20 << 20 | (uint64_t)b21 << 21 | (uint64_t)b22 << 22 | (uint64_t)b23 << 23 |
(uint64_t)b24 << 24 | (uint64_t)b25 << 25 | (uint64_t)b26 << 26 | (uint64_t)b27 << 27 |
(uint64_t)b28 << 28 | (uint64_t)b29 << 29 | (uint64_t)b30 << 30 | (uint64_t)b31 << 31 |
(uint64_t)b32 << 32 | (uint64_t)b33 << 33 | (uint64_t)b34 << 34 | (uint64_t)b35 << 35 |
(uint64_t)b36 << 36 | (uint64_t)b37 << 37 | (uint64_t)b38 << 38 | (uint64_t)b39 << 39 |
(uint64_t)b40 << 40 | (uint64_t)b41 << 41 | (uint64_t)b42 << 42 | (uint64_t)b43 << 43 |
(uint64_t)b44 << 44 | (uint64_t)b45 << 45 | (uint64_t)b46 << 46 | (uint64_t)b47 << 47 |
(uint64_t)b48 << 48 | (uint64_t)b49 << 49 | (uint64_t)b50 << 50 | (uint64_t)b51 << 51 |
(uint64_t)b52 << 52 | (uint64_t)b53 << 53 | (uint64_t)b54 << 54 | (uint64_t)b55 << 55 |
(uint64_t)b56 << 56 | (uint64_t)b57 << 57 | (uint64_t)b58 << 58 | (uint64_t)b59 << 59 |
(uint64_t)b60 << 60 | (uint64_t)b61 << 61 | (uint64_t)b62 << 62 | (uint64_t)b63 << 63);
} */
// Constructor to convert from type __mmask64 used in intrinsics:
Vec64b (__mmask64 x) {
mm = x;
}
// Constructor to broadcast single value:
Vec64b(bool b) {
mm = __mmask64(-int64_t(b));
}
// Constructor to make from two halves
Vec64b(Vec32b const x0, Vec32b const x1) {
mm = uint32_t(__mmask32(x0)) | uint64_t(__mmask32(x1)) << 32;
}
// Assignment operator to convert from type __mmask64 used in intrinsics:
Vec64b & operator = (__mmask64 x) {
mm = x;
return *this;
}
// Assignment operator to broadcast scalar value:
Vec64b & operator = (bool b) {
mm = Vec64b(b);
return *this;
}
// split into two halves
Vec32b get_low() const {
return Vec32b(__mmask32(mm));
}
Vec32b get_high() const {
return Vec32b(__mmask32(mm >> 32));
}
// Member function to change a single element in vector
Vec64b & insert (uint32_t index, bool a) {
uint64_t mask = uint64_t(1) << index;
mm = (mm & ~mask) | uint64_t(a) << index;
return *this;
}
// Member function extract a single element from vector
bool extract(int index) const {
return ((mm >> index) & 1) != 0;
}
// Extract a single element. Use store function if extracting more than one element.
// Operator [] can only read an element, not write.
bool operator [] (int index) const {
return extract(index);
}
// Type cast operator to convert to __mmask64 used in intrinsics
operator __mmask64() const {
return mm;
}
// Member function to change a bitfield to a boolean vector
Vec64b & load_bits(uint64_t a) {
mm = __mmask64(a);
return *this;
}
static constexpr int size() {
return 64;
}
static constexpr int elementtype() {
return 2;
}
};
typedef Vec64b Vec64cb; // compact boolean vector
typedef Vec64b Vec64ucb; // compact boolean vector
/*****************************************************************************
*
* Define operators and functions for Vec64cb
*
*****************************************************************************/
// vector operator & : bitwise and
static inline Vec64cb operator & (Vec64cb const a, Vec64cb const b) {
//return _kand_mask64(a, b);
return __mmask64(a) & __mmask64(b);
}
static inline Vec64cb operator && (Vec64cb const a, Vec64cb const b) {
return a & b;
}
// vector operator &= : bitwise and
static inline Vec64cb & operator &= (Vec64cb & a, Vec64cb const b) {
a = a & b;
return a;
}
// vector operator | : bitwise or
static inline Vec64cb operator | (Vec64cb const a, Vec64cb const b) {
//return _kor_mask64(a, b);
return __mmask64(a) | __mmask64(b);
}
static inline Vec64cb operator || (Vec64cb const a, Vec64cb const b) {
return a | b;
}
// vector operator |= : bitwise or
static inline Vec64cb & operator |= (Vec64cb & a, Vec64cb const b) {
a = a | b;
return a;
}
// vector operator ^ : bitwise xor
static inline Vec64cb operator ^ (Vec64cb const a, Vec64cb const b) {
//return _kxor_mask64(a, b);
return __mmask64(a) ^ __mmask64(b);
}
// vector operator ^= : bitwise xor
static inline Vec64cb & operator ^= (Vec64cb & a, Vec64cb const b) {
a = a ^ b;
return a;
}
// vector operator == : xnor
static inline Vec64cb operator == (Vec64cb const a, Vec64cb const b) {
return __mmask64(a) ^ ~ __mmask64(b);
//return _kxnor_mask64(a, b); // not all compilers have this intrinsic
}
// vector operator != : xor
static inline Vec64cb operator != (Vec64cb const a, Vec64cb const b) {
//return _kxor_mask64(a, b);
return __mmask64(a) ^ __mmask64(b);
}
// vector operator ~ : bitwise not
static inline Vec64cb operator ~ (Vec64cb const a) {
//return _knot_mask64(a);
return ~ __mmask64(a);
}
// vector operator ! : element not
static inline Vec64cb operator ! (Vec64cb const a) {
return ~a;
}
// vector function andnot
static inline Vec64cb andnot (Vec64cb const a, Vec64cb const b) {
//return _kxnor_mask64(b, a);
return __mmask64(a) & ~ __mmask64(b);
}
// horizontal_and. Returns true if all bits are 1
static inline bool horizontal_and (Vec64cb const a) {
return int64_t(__mmask64(a)) == -(int64_t)(1);
}
// horizontal_or. Returns true if at least one bit is 1
static inline bool horizontal_or (Vec64cb const a) {
return int64_t(__mmask64(a)) != 0;
}
// to_bits: convert boolean vector to integer bitfield
static inline uint64_t to_bits(Vec64cb x) {
return uint64_t(__mmask64(x));
}
/*****************************************************************************
*
* Define operators for Vec64c
*
*****************************************************************************/
// vector operator + : add element by element
static inline Vec64c operator + (Vec64c const a, Vec64c const b) {
return _mm512_add_epi8(a, b);
}
// vector operator += : add
static inline Vec64c & operator += (Vec64c & a, Vec64c const b) {
a = a + b;
return a;
}
// postfix operator ++
static inline Vec64c operator ++ (Vec64c & a, int) {
Vec64c a0 = a;
a = a + 1;
return a0;
}
// prefix operator ++
static inline Vec64c & operator ++ (Vec64c & a) {
a = a + 1;
return a;
}
// vector operator - : subtract element by element
static inline Vec64c operator - (Vec64c const a, Vec64c const b) {
return _mm512_sub_epi8(a, b);
}
// vector operator - : unary minus
static inline Vec64c operator - (Vec64c const a) {
return _mm512_sub_epi8(_mm512_setzero_epi32(), a);
}
// vector operator -= : subtract
static inline Vec64c & operator -= (Vec64c & a, Vec64c const b) {
a = a - b;
return a;
}
// postfix operator --
static inline Vec64c operator -- (Vec64c & a, int) {
Vec64c a0 = a;
a = a - 1;
return a0;
}
// prefix operator --
static inline Vec64c & operator -- (Vec64c & a) {
a = a - 1;
return a;
}
// vector operator * : multiply element by element
static inline Vec64c operator * (Vec64c const a, Vec64c const b) {
// There is no 8-bit multiply. Split into two 16-bit multiplies
__m512i aodd = _mm512_srli_epi16(a,8); // odd numbered elements of a
__m512i bodd = _mm512_srli_epi16(b,8); // odd numbered elements of b
__m512i muleven = _mm512_mullo_epi16(a,b); // product of even numbered elements
__m512i mulodd = _mm512_mullo_epi16(aodd,bodd); // product of odd numbered elements
mulodd = _mm512_slli_epi16(mulodd,8); // put odd numbered elements back in place
__m512i product = _mm512_mask_mov_epi8(muleven, 0xAAAAAAAAAAAAAAAA, mulodd); // interleave even and odd
return product;
}
// vector operator *= : multiply
static inline Vec64c & operator *= (Vec64c & a, Vec64c const b) {
a = a * b;
return a;
}
// vector operator / : divide all elements by same integer. See bottom of file
// vector operator << : shift left
static inline Vec64c operator << (Vec64c const a, int32_t b) {
uint32_t mask = (uint32_t)0xFF >> (uint32_t)b; // mask to remove bits that are shifted out
__m512i am = _mm512_and_si512(a,_mm512_set1_epi8((char)mask));// remove bits that will overflow
__m512i res = _mm512_sll_epi16(am,_mm_cvtsi32_si128(b)); // 16-bit shifts
return res;
}
// vector operator <<= : shift left
static inline Vec64c & operator <<= (Vec64c & a, int32_t b) {
a = a << b;
return a;
}
// vector operator >> : shift right arithmetic
static inline Vec64c operator >> (Vec64c const a, int32_t b) {
__m512i aeven = _mm512_slli_epi16(a, 8); // even numbered elements of a. get sign bit in position
aeven = _mm512_sra_epi16(aeven, _mm_cvtsi32_si128(b + 8)); // shift arithmetic, back to position
__m512i aodd = _mm512_sra_epi16(a, _mm_cvtsi32_si128(b)); // shift odd numbered elements arithmetic
__m512i res = _mm512_mask_mov_epi8(aeven, 0xAAAAAAAAAAAAAAAA, aodd);// interleave even and odd
return res;
}
// vector operator >>= : shift right arithmetic
static inline Vec64c & operator >>= (Vec64c & a, int32_t b) {
a = a >> b;
return a;
}
// vector operator == : returns true for elements for which a == b
static inline Vec64cb operator == (Vec64c const a, Vec64c const b) {
return _mm512_cmpeq_epi8_mask(a, b);
}
// vector operator != : returns true for elements for which a != b
static inline Vec64cb operator != (Vec64c const a, Vec64c const b) {
return _mm512_cmpneq_epi8_mask(a, b);
}
// vector operator > : returns true for elements for which a > b
static inline Vec64cb operator > (Vec64c const a, Vec64c const b) {
return _mm512_cmp_epi8_mask(a, b, 6);
}
// vector operator < : returns true for elements for which a < b
static inline Vec64cb operator < (Vec64c const a, Vec64c const b) {
return _mm512_cmp_epi8_mask(a, b, 1);
}
// vector operator >= : returns true for elements for which a >= b (signed)
static inline Vec64cb operator >= (Vec64c const a, Vec64c const b) {
return _mm512_cmp_epi8_mask(a, b, 5);
}
// vector operator <= : returns true for elements for which a <= b (signed)
static inline Vec64cb operator <= (Vec64c const a, Vec64c const b) {
return _mm512_cmp_epi8_mask(a, b, 2);
}
// vector operator & : bitwise and
static inline Vec64c operator & (Vec64c const a, Vec64c const b) {
return _mm512_and_epi32(a, b);
}
// vector operator &= : bitwise and
static inline Vec64c & operator &= (Vec64c & a, Vec64c const b) {
a = a & b;
return a;
}
// vector operator | : bitwise or
static inline Vec64c operator | (Vec64c const a, Vec64c const b) {
return _mm512_or_epi32(a, b);
}
// vector operator |= : bitwise or
static inline Vec64c & operator |= (Vec64c & a, Vec64c const b) {
a = a | b;
return a;
}
// vector operator ^ : bitwise xor
static inline Vec64c operator ^ (Vec64c const a, Vec64c const b) {
return _mm512_xor_epi32(a, b);
}
// vector operator ^= : bitwise xor
static inline Vec64c & operator ^= (Vec64c & a, Vec64c const b) {
a = a ^ b;
return a;
}
// vector operator ~ : bitwise not
static inline Vec64c operator ~ (Vec64c const a) {
return Vec64c(~ Vec16i(a));
}
// Functions for this class
// Select between two operands. Corresponds to this pseudocode:
// for (int i = 0; i < 16; i++) result[i] = s[i] ? a[i] : b[i];
static inline Vec64c select (Vec64cb const s, Vec64c const a, Vec64c const b) {
return _mm512_mask_mov_epi8(b, s, a); // conditional move may be optimized better by the compiler than blend
}
// Conditional add: For all vector elements i: result[i] = f[i] ? (a[i] + b[i]) : a[i]
static inline Vec64c if_add (Vec64cb const f, Vec64c const a, Vec64c const b) {
return _mm512_mask_add_epi8(a, f, a, b);
}
// Conditional subtract
static inline Vec64c if_sub (Vec64cb const f, Vec64c const a, Vec64c const b) {
return _mm512_mask_sub_epi8(a, f, a, b);
}
// Conditional multiply
static inline Vec64c if_mul (Vec64cb const f, Vec64c const a, Vec64c const b) {
Vec64c m = a * b;
return select(f, m, a);
}
// Horizontal add: Calculates the sum of all vector elements. Overflow will wrap around
static inline int8_t horizontal_add (Vec64c const a) {
__m512i sum1 = _mm512_sad_epu8(a,_mm512_setzero_si512());
return (int8_t)horizontal_add(Vec8q(sum1));
}
// Horizontal add extended: Calculates the sum of all vector elements.
// Each element is sign-extended before addition to avoid overflow
static inline int32_t horizontal_add_x (Vec64c const a) {
return horizontal_add_x(a.get_low()) + horizontal_add_x(a.get_high());
}
// function add_saturated: add element by element, signed with saturation
static inline Vec64c add_saturated(Vec64c const a, Vec64c const b) {
return _mm512_adds_epi8(a, b);
}
// function sub_saturated: subtract element by element, signed with saturation
static inline Vec64c sub_saturated(Vec64c const a, Vec64c const b) {
return _mm512_subs_epi8(a, b);
}
// function max: a > b ? a : b
static inline Vec64c max(Vec64c const a, Vec64c const b) {
return _mm512_max_epi8(a,b);
}
// function min: a < b ? a : b
static inline Vec64c min(Vec64c const a, Vec64c const b) {
return _mm512_min_epi8(a,b);
}
// function abs: a >= 0 ? a : -a
static inline Vec64c abs(Vec64c const a) {
return _mm512_abs_epi8(a);
}
// function abs_saturated: same as abs, saturate if overflow
static inline Vec64c abs_saturated(Vec64c const a) {
return _mm512_min_epu8(abs(a), Vec64c(0x7F));
}
// function rotate_left all elements
// Use negative count to rotate right
static inline Vec64c rotate_left(Vec64c const a, int b) {
uint8_t mask = 0xFFu << b; // mask off overflow bits
__m512i m = _mm512_set1_epi8(mask);
__m128i bb = _mm_cvtsi32_si128(b & 7); // b modulo 8
__m128i mbb = _mm_cvtsi32_si128((- b) & 7); // 8-b modulo 8
__m512i left = _mm512_sll_epi16(a, bb); // a << b
__m512i right = _mm512_srl_epi16(a, mbb); // a >> 8-b
left = _mm512_and_si512(m, left); // mask off overflow bits
right = _mm512_andnot_si512(m, right);
return _mm512_or_si512(left, right); // combine left and right shifted bits
}
/*****************************************************************************
*
* Vector of 64 8-bit unsigned integers
*
*****************************************************************************/
class Vec64uc : public Vec64c {
public:
// Default constructor:
Vec64uc() {
}
// Constructor to broadcast the same value into all elements:
Vec64uc(uint8_t i) {
zmm = _mm512_set1_epi8((int8_t)i);
}
// Constructor to build from all elements:
Vec64uc(uint8_t i0, uint8_t i1, uint8_t i2, uint8_t i3, uint8_t i4, uint8_t i5, uint8_t i6, uint8_t i7,
uint8_t i8, uint8_t i9, uint8_t i10, uint8_t i11, uint8_t i12, uint8_t i13, uint8_t i14, uint8_t i15,
uint8_t i16, uint8_t i17, uint8_t i18, uint8_t i19, uint8_t i20, uint8_t i21, uint8_t i22, uint8_t i23,
uint8_t i24, uint8_t i25, uint8_t i26, uint8_t i27, uint8_t i28, uint8_t i29, uint8_t i30, uint8_t i31,
uint8_t i32, uint8_t i33, uint8_t i34, uint8_t i35, uint8_t i36, uint8_t i37, uint8_t i38, uint8_t i39,
uint8_t i40, uint8_t i41, uint8_t i42, uint8_t i43, uint8_t i44, uint8_t i45, uint8_t i46, uint8_t i47,
uint8_t i48, uint8_t i49, uint8_t i50, uint8_t i51, uint8_t i52, uint8_t i53, uint8_t i54, uint8_t i55,
uint8_t i56, uint8_t i57, uint8_t i58, uint8_t i59, uint8_t i60, uint8_t i61, uint8_t i62, uint8_t i63)
: Vec64c(i0, i1, i2, i3, i4, i5, i6, i7, i8, i9, i10, i11, i12, i13, i14, i15,
i16, i17, i18, i19, i20, i21, i22, i23, i24, i25, i26, i27, i28, i29, i30, i31,
i32, i33, i34, i35, i36, i37, i38, i39, i40, i41, i42, i43, i44, i45, i46, i47,
i48, i49, i50, i51, i52, i53, i54, i55, i56, i57, i58, i59, i60, i61, i62, i63) {}
// Constructor to build from two Vec32uc:
Vec64uc(Vec32uc const a0, Vec32uc const a1) {
zmm = _mm512_inserti64x4(_mm512_castsi256_si512(a0), a1, 1);
}
// Constructor to convert from type __m512i used in intrinsics:
Vec64uc(__m512i const x) {
zmm = x;
}
// Assignment operator to convert from type __m512i used in intrinsics:
Vec64uc & operator = (__m512i const x) {
zmm = x;
return *this;
}
// Member function to load from array (unaligned)
Vec64uc & load(void const * p) {
Vec64c::load(p);
return *this;
}
// Member function to load from array, aligned by 64
Vec64uc & load_a(void const * p) {
Vec64c::load_a(p);
return *this;
}
// Member function to change a single element in vector
Vec64uc const insert(int index, uint8_t value) {
Vec64c::insert(index, (int8_t)value);
return *this;
}
// Member function extract a single element from vector
uint8_t extract(int index) const {
return (uint8_t)Vec64c::extract(index);
}
// Extract a single element. Use store function if extracting more than one element.
// Operator [] can only read an element, not write.
uint8_t operator [] (int index) const {
return (uint8_t)Vec64c::extract(index);
}
// Member functions to split into two Vec32uc:
Vec32uc get_low() const {
return Vec32uc(Vec64c::get_low());
}
Vec32uc get_high() const {
return Vec32uc(Vec64c::get_high());
}
static constexpr int elementtype() {
return 5;
}
};
// Define operators for this class
// vector operator + : add element by element
static inline Vec64uc operator + (Vec64uc const a, Vec64uc const b) {
return _mm512_add_epi8(a, b);
}
// vector operator - : subtract element by element
static inline Vec64uc operator - (Vec64uc const a, Vec64uc const b) {
return _mm512_sub_epi8(a, b);
}
// vector operator ' : multiply element by element
static inline Vec64uc operator * (Vec64uc const a, Vec64uc const b) {
return Vec64uc(Vec64c(a) * Vec64c(b));
}
// vector operator / : divide. See bottom of file
// vector operator >> : shift right logical all elements
static inline Vec64uc operator >> (Vec64uc const a, uint32_t b) {
uint32_t mask = (uint32_t)0xFF << (uint32_t)b; // mask to remove bits that are shifted out
__m512i am = _mm512_and_si512(a,_mm512_set1_epi8((char)mask)); // remove bits that will overflow
__m512i res = _mm512_srl_epi16(am,_mm_cvtsi32_si128((int32_t)b));// 16-bit shifts
return res;
}
static inline Vec64uc operator >> (Vec64uc const a, int b) {
return a >> (uint32_t)b;
}
// vector operator >>= : shift right logical
static inline Vec64uc & operator >>= (Vec64uc & a, uint32_t b) {
a = a >> b;
return a;
}
// vector operator >>= : shift right logical (signed b)
static inline Vec64uc & operator >>= (Vec64uc & a, int32_t b) {
a = a >> uint32_t(b);
return a;
}
// vector operator << : shift left all elements
static inline Vec64uc operator << (Vec64uc const a, uint32_t b) {
return Vec64uc(Vec64c(a) << int32_t(b));
}
static inline Vec64uc operator << (Vec64uc const a, int b) {
return a << (uint32_t)b;
}
// vector operator < : returns true for elements for which a < b (unsigned)
static inline Vec64cb operator < (Vec64uc const a, Vec64uc const b) {
return _mm512_cmp_epu8_mask(a, b, 1);
}
// vector operator > : returns true for elements for which a > b (unsigned)
static inline Vec64cb operator > (Vec64uc const a, Vec64uc const b) {
return _mm512_cmp_epu8_mask(a, b, 6);
}
// vector operator >= : returns true for elements for which a >= b (unsigned)
static inline Vec64cb operator >= (Vec64uc const a, Vec64uc const b) {
return _mm512_cmp_epu8_mask(a, b, 5);
}
// vector operator <= : returns true for elements for which a <= b (unsigned)
static inline Vec64cb operator <= (Vec64uc const a, Vec64uc const b) {
return _mm512_cmp_epu8_mask(a, b, 2);
}
// vector operator & : bitwise and
static inline Vec64uc operator & (Vec64uc const a, Vec64uc const b) {
return Vec64uc(Vec64c(a) & Vec64c(b));
}
// vector operator | : bitwise or
static inline Vec64uc operator | (Vec64uc const a, Vec64uc const b) {
return Vec64uc(Vec64c(a) | Vec64c(b));
}
// vector operator ^ : bitwise xor
static inline Vec64uc operator ^ (Vec64uc const a, Vec64uc const b) {
return Vec64uc(Vec64c(a) ^ Vec64c(b));
}
// vector operator ~ : bitwise not
static inline Vec64uc operator ~ (Vec64uc const a) {
return Vec64uc( ~ Vec64c(a));
}
// Functions for this class
// Select between two operands. Corresponds to this pseudocode:
// for (int i = 0; i < 16; i++) result[i] = s[i] ? a[i] : b[i];
static inline Vec64uc select (Vec64cb const s, Vec64uc const a, Vec64uc const b) {
return Vec64uc(select(s, Vec64c(a), Vec64c(b)));
}
// Conditional add: For all vector elements i: result[i] = f[i] ? (a[i] + b[i]) : a[i]
static inline Vec64uc if_add (Vec64cb const f, Vec64uc const a, Vec64uc const b) {
return _mm512_mask_add_epi8(a, f, a, b);
}
// Conditional subtract
static inline Vec64uc if_sub (Vec64cb const f, Vec64uc const a, Vec64uc const b) {
return _mm512_mask_sub_epi8(a, f, a, b);
}
// Conditional multiply
static inline Vec64uc if_mul (Vec64cb const f, Vec64uc const a, Vec64uc const b) {
Vec64uc m = a * b;
return select(f, m, a);
}
// function add_saturated: add element by element, unsigned with saturation
static inline Vec64uc add_saturated(Vec64uc const a, Vec64uc const b) {
return _mm512_adds_epu8(a, b);
}
// function sub_saturated: subtract element by element, unsigned with saturation
static inline Vec64uc sub_saturated(Vec64uc const a, Vec64uc const b) {
return _mm512_subs_epu8(a, b);
}
// function max: a > b ? a : b
static inline Vec64uc max(Vec64uc const a, Vec64uc const b) {
return _mm512_max_epu8(a,b);
}
// function min: a < b ? a : b
static inline Vec64uc min(Vec64uc const a, Vec64uc const b) {
return _mm512_min_epu8(a,b);
}
/*****************************************************************************
*
* Vector of 32 16-bit signed integers
*
*****************************************************************************/
class Vec32s: public Vec512b {
public:
// Default constructor:
Vec32s() {
}
// Constructor to broadcast the same value into all elements:
Vec32s(int16_t i) {
zmm = _mm512_set1_epi16(i);
}
// Constructor to build from all elements:
Vec32s(int16_t i0, int16_t i1, int16_t i2, int16_t i3, int16_t i4, int16_t i5, int16_t i6, int16_t i7,
int16_t i8, int16_t i9, int16_t i10, int16_t i11, int16_t i12, int16_t i13, int16_t i14, int16_t i15,
int16_t i16, int16_t i17, int16_t i18, int16_t i19, int16_t i20, int16_t i21, int16_t i22, int16_t i23,
int16_t i24, int16_t i25, int16_t i26, int16_t i27, int16_t i28, int16_t i29, int16_t i30, int16_t i31) {
#if true
// _mm512_set_epi16 missing in GCC 7.4.0. This may be more efficient after all:
int16_t aa[32] = {
i0, i1, i2, i3, i4, i5, i6, i7, i8, i9, i10, i11, i12, i13, i14, i15,
i16, i17, i18, i19, i20, i21, i22, i23, i24, i25, i26, i27, i28, i29, i30, i31 };
load(aa);
#else
zmm = _mm512_set_epi16(
i31, i30, i29, i28, i27, i26, i25, i24, i23, i22, i21, i20, i19, i18, i17, i16,
i15, i14, i13, i12, i11, i10, i9, i8, i7, i6, i5, i4, i3, i2, i1, i0);
#endif
}
// Constructor to build from two Vec16s:
Vec32s(Vec16s const a0, Vec16s const a1) {
zmm = _mm512_inserti64x4(_mm512_castsi256_si512(a0), a1, 1);
}
// Constructor to convert from type __m512i used in intrinsics:
Vec32s(__m512i const x) {
zmm = x;
}
// Assignment operator to convert from type __m512i used in intrinsics:
Vec32s & operator = (__m512i const x) {
zmm = x;
return *this;
}
// Type cast operator to convert to __m512i used in intrinsics
operator __m512i() const {
return zmm;
}
// Member function to load from array (unaligned)
Vec32s & load(void const * p) {
zmm = _mm512_loadu_si512(p);
return *this;
}
// Member function to load from array, aligned by 64
Vec32s & load_a(void const * p) {
zmm = _mm512_load_si512(p);
return *this;
}
// Partial load. Load n elements and set the rest to 0
Vec32s & load_partial(int n, void const * p) {
zmm = _mm512_maskz_loadu_epi16(__mmask32(((uint64_t)1 << n) - 1), p);
return *this;
}
// Partial store. Store n elements
void store_partial(int n, void * p) const {
_mm512_mask_storeu_epi16(p, __mmask32(((uint64_t)1 << n) - 1), zmm);
}
// cut off vector to n elements. The last 32-n elements are set to zero
Vec32s & cutoff(int n) {
zmm = _mm512_maskz_mov_epi16(__mmask32(((uint64_t)1 << n) - 1), zmm);
return *this;
}
// Member function to change a single element in vector
Vec32s const insert(int index, int16_t value) {
zmm = _mm512_mask_set1_epi16(zmm, __mmask64((uint64_t)1 << index), value);
return *this;
}
// Member function extract a single element from vector
int16_t extract(int index) const {
#if INSTRSET >= 10 && defined (__AVX512VBMI2__)
__m512i x = _mm512_maskz_compress_epi16(__mmask32(1u << index), zmm);
return (int16_t)_mm_cvtsi128_si32(_mm512_castsi512_si128(x));
#else
int16_t a[32];
store(a);
return a[index & 31];
#endif
}
// Extract a single element. Use store function if extracting more than one element.
// Operator [] can only read an element, not write.
int16_t operator [] (int index) const {
return extract(index);
}
// Member functions to split into two Vec16s:
Vec16s get_low() const {
return _mm512_castsi512_si256(zmm);
}
Vec16s get_high() const {
return _mm512_extracti64x4_epi64(zmm,1);
}
static constexpr int size() {
return 32;
}
static constexpr int elementtype() {
return 6;
}
};
/*****************************************************************************
*
* Vec32sb: Vector of 64 Booleans for use with Vec32s and Vec32us
*
*****************************************************************************/
typedef Vec32b Vec32sb; // compact boolean vector
/*****************************************************************************
*
* Define operators for Vec32s
*
*****************************************************************************/
// vector operator + : add element by element
static inline Vec32s operator + (Vec32s const a, Vec32s const b) {
return _mm512_add_epi16(a, b);
}
// vector operator += : add
static inline Vec32s & operator += (Vec32s & a, Vec32s const b) {
a = a + b;
return a;
}
// postfix operator ++
static inline Vec32s operator ++ (Vec32s & a, int) {
Vec32s a0 = a;
a = a + 1;
return a0;
}
// prefix operator ++
static inline Vec32s & operator ++ (Vec32s & a) {
a = a + 1;
return a;
}
// vector operator - : subtract element by element
static inline Vec32s operator - (Vec32s const a, Vec32s const b) {
return _mm512_sub_epi16(a, b);
}
// vector operator - : unary minus
static inline Vec32s operator - (Vec32s const a) {
return _mm512_sub_epi16(_mm512_setzero_epi32(), a);
}
// vector operator -= : subtract
static inline Vec32s & operator -= (Vec32s & a, Vec32s const b) {
a = a - b;
return a;
}
// postfix operator --
static inline Vec32s operator -- (Vec32s & a, int) {
Vec32s a0 = a;
a = a - 1;