summaryrefslogtreecommitdiff
path: root/nacl/crypto_stream/aes128ctr/core2/afternm.s
diff options
context:
space:
mode:
Diffstat (limited to 'nacl/crypto_stream/aes128ctr/core2/afternm.s')
-rw-r--r--nacl/crypto_stream/aes128ctr/core2/afternm.s12308
1 files changed, 12308 insertions, 0 deletions
diff --git a/nacl/crypto_stream/aes128ctr/core2/afternm.s b/nacl/crypto_stream/aes128ctr/core2/afternm.s
new file mode 100644
index 00000000..c1ba79ef
--- /dev/null
+++ b/nacl/crypto_stream/aes128ctr/core2/afternm.s
@@ -0,0 +1,12308 @@
1# Author: Emilia Käsper and Peter Schwabe
2# Date: 2009-03-19
3# +2010.01.31: minor namespace modifications
4# Public domain
5
6.data
7.p2align 6
8
9RCON: .int 0x00000000, 0x00000000, 0x00000000, 0xffffffff
10ROTB: .int 0x0c000000, 0x00000000, 0x04000000, 0x08000000
11EXPB0: .int 0x03030303, 0x07070707, 0x0b0b0b0b, 0x0f0f0f0f
12CTRINC1: .int 0x00000001, 0x00000000, 0x00000000, 0x00000000
13CTRINC2: .int 0x00000002, 0x00000000, 0x00000000, 0x00000000
14CTRINC3: .int 0x00000003, 0x00000000, 0x00000000, 0x00000000
15CTRINC4: .int 0x00000004, 0x00000000, 0x00000000, 0x00000000
16CTRINC5: .int 0x00000005, 0x00000000, 0x00000000, 0x00000000
17CTRINC6: .int 0x00000006, 0x00000000, 0x00000000, 0x00000000
18CTRINC7: .int 0x00000007, 0x00000000, 0x00000000, 0x00000000
19RCTRINC1: .int 0x00000000, 0x00000000, 0x00000000, 0x00000001
20RCTRINC2: .int 0x00000000, 0x00000000, 0x00000000, 0x00000002
21RCTRINC3: .int 0x00000000, 0x00000000, 0x00000000, 0x00000003
22RCTRINC4: .int 0x00000000, 0x00000000, 0x00000000, 0x00000004
23RCTRINC5: .int 0x00000000, 0x00000000, 0x00000000, 0x00000005
24RCTRINC6: .int 0x00000000, 0x00000000, 0x00000000, 0x00000006
25RCTRINC7: .int 0x00000000, 0x00000000, 0x00000000, 0x00000007
26
27SWAP32: .int 0x00010203, 0x04050607, 0x08090a0b, 0x0c0d0e0f
28M0SWAP: .quad 0x0105090d0004080c , 0x03070b0f02060a0e
29
30BS0: .quad 0x5555555555555555, 0x5555555555555555
31BS1: .quad 0x3333333333333333, 0x3333333333333333
32BS2: .quad 0x0f0f0f0f0f0f0f0f, 0x0f0f0f0f0f0f0f0f
33ONE: .quad 0xffffffffffffffff, 0xffffffffffffffff
34M0: .quad 0x02060a0e03070b0f, 0x0004080c0105090d
35SRM0: .quad 0x0304090e00050a0f, 0x01060b0c0207080d
36SR: .quad 0x0504070600030201, 0x0f0e0d0c0a09080b
37
38# qhasm: int64 outp
39
40# qhasm: int64 len
41
42# qhasm: int64 np
43
44# qhasm: int64 c
45
46# qhasm: input outp
47
48# qhasm: input len
49
50# qhasm: input np
51
52# qhasm: input c
53
54# qhasm: int64 lensav
55
56# qhasm: int6464 xmm0
57
58# qhasm: int6464 xmm1
59
60# qhasm: int6464 xmm2
61
62# qhasm: int6464 xmm3
63
64# qhasm: int6464 xmm4
65
66# qhasm: int6464 xmm5
67
68# qhasm: int6464 xmm6
69
70# qhasm: int6464 xmm7
71
72# qhasm: int6464 xmm8
73
74# qhasm: int6464 xmm9
75
76# qhasm: int6464 xmm10
77
78# qhasm: int6464 xmm11
79
80# qhasm: int6464 xmm12
81
82# qhasm: int6464 xmm13
83
84# qhasm: int6464 xmm14
85
86# qhasm: int6464 xmm15
87
88# qhasm: int6464 t
89
90# qhasm: stack1024 bl
91
92# qhasm: stack128 nonce_stack
93
94# qhasm: int64 blp
95
96# qhasm: int64 b
97
98# qhasm: int64 tmp
99
100# qhasm: enter crypto_stream_aes128ctr_core2_afternm
101.text
102.p2align 5
103.globl _crypto_stream_aes128ctr_core2_afternm
104.globl crypto_stream_aes128ctr_core2_afternm
105_crypto_stream_aes128ctr_core2_afternm:
106crypto_stream_aes128ctr_core2_afternm:
107mov %rsp,%r11
108and $31,%r11
109add $160,%r11
110sub %r11,%rsp
111
112# qhasm: xmm0 = *(int128 *) (np + 0)
113# asm 1: movdqa 0(<np=int64#3),>xmm0=int6464#1
114# asm 2: movdqa 0(<np=%rdx),>xmm0=%xmm0
115movdqa 0(%rdx),%xmm0
116
117# qhasm: nonce_stack = xmm0
118# asm 1: movdqa <xmm0=int6464#1,>nonce_stack=stack128#1
119# asm 2: movdqa <xmm0=%xmm0,>nonce_stack=0(%rsp)
120movdqa %xmm0,0(%rsp)
121
122# qhasm: np = &nonce_stack
123# asm 1: leaq <nonce_stack=stack128#1,>np=int64#3
124# asm 2: leaq <nonce_stack=0(%rsp),>np=%rdx
125leaq 0(%rsp),%rdx
126
127# qhasm: enc_block:
128._enc_block:
129
130# qhasm: xmm0 = *(int128 *) (np + 0)
131# asm 1: movdqa 0(<np=int64#3),>xmm0=int6464#1
132# asm 2: movdqa 0(<np=%rdx),>xmm0=%xmm0
133movdqa 0(%rdx),%xmm0
134
135# qhasm: xmm1 = xmm0
136# asm 1: movdqa <xmm0=int6464#1,>xmm1=int6464#2
137# asm 2: movdqa <xmm0=%xmm0,>xmm1=%xmm1
138movdqa %xmm0,%xmm1
139
140# qhasm: shuffle bytes of xmm1 by SWAP32
141# asm 1: pshufb SWAP32,<xmm1=int6464#2
142# asm 2: pshufb SWAP32,<xmm1=%xmm1
143pshufb SWAP32,%xmm1
144
145# qhasm: xmm2 = xmm1
146# asm 1: movdqa <xmm1=int6464#2,>xmm2=int6464#3
147# asm 2: movdqa <xmm1=%xmm1,>xmm2=%xmm2
148movdqa %xmm1,%xmm2
149
150# qhasm: xmm3 = xmm1
151# asm 1: movdqa <xmm1=int6464#2,>xmm3=int6464#4
152# asm 2: movdqa <xmm1=%xmm1,>xmm3=%xmm3
153movdqa %xmm1,%xmm3
154
155# qhasm: xmm4 = xmm1
156# asm 1: movdqa <xmm1=int6464#2,>xmm4=int6464#5
157# asm 2: movdqa <xmm1=%xmm1,>xmm4=%xmm4
158movdqa %xmm1,%xmm4
159
160# qhasm: xmm5 = xmm1
161# asm 1: movdqa <xmm1=int6464#2,>xmm5=int6464#6
162# asm 2: movdqa <xmm1=%xmm1,>xmm5=%xmm5
163movdqa %xmm1,%xmm5
164
165# qhasm: xmm6 = xmm1
166# asm 1: movdqa <xmm1=int6464#2,>xmm6=int6464#7
167# asm 2: movdqa <xmm1=%xmm1,>xmm6=%xmm6
168movdqa %xmm1,%xmm6
169
170# qhasm: xmm7 = xmm1
171# asm 1: movdqa <xmm1=int6464#2,>xmm7=int6464#8
172# asm 2: movdqa <xmm1=%xmm1,>xmm7=%xmm7
173movdqa %xmm1,%xmm7
174
175# qhasm: int32323232 xmm1 += RCTRINC1
176# asm 1: paddd RCTRINC1,<xmm1=int6464#2
177# asm 2: paddd RCTRINC1,<xmm1=%xmm1
178paddd RCTRINC1,%xmm1
179
180# qhasm: int32323232 xmm2 += RCTRINC2
181# asm 1: paddd RCTRINC2,<xmm2=int6464#3
182# asm 2: paddd RCTRINC2,<xmm2=%xmm2
183paddd RCTRINC2,%xmm2
184
185# qhasm: int32323232 xmm3 += RCTRINC3
186# asm 1: paddd RCTRINC3,<xmm3=int6464#4
187# asm 2: paddd RCTRINC3,<xmm3=%xmm3
188paddd RCTRINC3,%xmm3
189
190# qhasm: int32323232 xmm4 += RCTRINC4
191# asm 1: paddd RCTRINC4,<xmm4=int6464#5
192# asm 2: paddd RCTRINC4,<xmm4=%xmm4
193paddd RCTRINC4,%xmm4
194
195# qhasm: int32323232 xmm5 += RCTRINC5
196# asm 1: paddd RCTRINC5,<xmm5=int6464#6
197# asm 2: paddd RCTRINC5,<xmm5=%xmm5
198paddd RCTRINC5,%xmm5
199
200# qhasm: int32323232 xmm6 += RCTRINC6
201# asm 1: paddd RCTRINC6,<xmm6=int6464#7
202# asm 2: paddd RCTRINC6,<xmm6=%xmm6
203paddd RCTRINC6,%xmm6
204
205# qhasm: int32323232 xmm7 += RCTRINC7
206# asm 1: paddd RCTRINC7,<xmm7=int6464#8
207# asm 2: paddd RCTRINC7,<xmm7=%xmm7
208paddd RCTRINC7,%xmm7
209
210# qhasm: shuffle bytes of xmm0 by M0
211# asm 1: pshufb M0,<xmm0=int6464#1
212# asm 2: pshufb M0,<xmm0=%xmm0
213pshufb M0,%xmm0
214
215# qhasm: shuffle bytes of xmm1 by M0SWAP
216# asm 1: pshufb M0SWAP,<xmm1=int6464#2
217# asm 2: pshufb M0SWAP,<xmm1=%xmm1
218pshufb M0SWAP,%xmm1
219
220# qhasm: shuffle bytes of xmm2 by M0SWAP
221# asm 1: pshufb M0SWAP,<xmm2=int6464#3
222# asm 2: pshufb M0SWAP,<xmm2=%xmm2
223pshufb M0SWAP,%xmm2
224
225# qhasm: shuffle bytes of xmm3 by M0SWAP
226# asm 1: pshufb M0SWAP,<xmm3=int6464#4
227# asm 2: pshufb M0SWAP,<xmm3=%xmm3
228pshufb M0SWAP,%xmm3
229
230# qhasm: shuffle bytes of xmm4 by M0SWAP
231# asm 1: pshufb M0SWAP,<xmm4=int6464#5
232# asm 2: pshufb M0SWAP,<xmm4=%xmm4
233pshufb M0SWAP,%xmm4
234
235# qhasm: shuffle bytes of xmm5 by M0SWAP
236# asm 1: pshufb M0SWAP,<xmm5=int6464#6
237# asm 2: pshufb M0SWAP,<xmm5=%xmm5
238pshufb M0SWAP,%xmm5
239
240# qhasm: shuffle bytes of xmm6 by M0SWAP
241# asm 1: pshufb M0SWAP,<xmm6=int6464#7
242# asm 2: pshufb M0SWAP,<xmm6=%xmm6
243pshufb M0SWAP,%xmm6
244
245# qhasm: shuffle bytes of xmm7 by M0SWAP
246# asm 1: pshufb M0SWAP,<xmm7=int6464#8
247# asm 2: pshufb M0SWAP,<xmm7=%xmm7
248pshufb M0SWAP,%xmm7
249
250# qhasm: xmm8 = xmm6
251# asm 1: movdqa <xmm6=int6464#7,>xmm8=int6464#9
252# asm 2: movdqa <xmm6=%xmm6,>xmm8=%xmm8
253movdqa %xmm6,%xmm8
254
255# qhasm: uint6464 xmm8 >>= 1
256# asm 1: psrlq $1,<xmm8=int6464#9
257# asm 2: psrlq $1,<xmm8=%xmm8
258psrlq $1,%xmm8
259
260# qhasm: xmm8 ^= xmm7
261# asm 1: pxor <xmm7=int6464#8,<xmm8=int6464#9
262# asm 2: pxor <xmm7=%xmm7,<xmm8=%xmm8
263pxor %xmm7,%xmm8
264
265# qhasm: xmm8 &= BS0
266# asm 1: pand BS0,<xmm8=int6464#9
267# asm 2: pand BS0,<xmm8=%xmm8
268pand BS0,%xmm8
269
270# qhasm: xmm7 ^= xmm8
271# asm 1: pxor <xmm8=int6464#9,<xmm7=int6464#8
272# asm 2: pxor <xmm8=%xmm8,<xmm7=%xmm7
273pxor %xmm8,%xmm7
274
275# qhasm: uint6464 xmm8 <<= 1
276# asm 1: psllq $1,<xmm8=int6464#9
277# asm 2: psllq $1,<xmm8=%xmm8
278psllq $1,%xmm8
279
280# qhasm: xmm6 ^= xmm8
281# asm 1: pxor <xmm8=int6464#9,<xmm6=int6464#7
282# asm 2: pxor <xmm8=%xmm8,<xmm6=%xmm6
283pxor %xmm8,%xmm6
284
285# qhasm: xmm8 = xmm4
286# asm 1: movdqa <xmm4=int6464#5,>xmm8=int6464#9
287# asm 2: movdqa <xmm4=%xmm4,>xmm8=%xmm8
288movdqa %xmm4,%xmm8
289
290# qhasm: uint6464 xmm8 >>= 1
291# asm 1: psrlq $1,<xmm8=int6464#9
292# asm 2: psrlq $1,<xmm8=%xmm8
293psrlq $1,%xmm8
294
295# qhasm: xmm8 ^= xmm5
296# asm 1: pxor <xmm5=int6464#6,<xmm8=int6464#9
297# asm 2: pxor <xmm5=%xmm5,<xmm8=%xmm8
298pxor %xmm5,%xmm8
299
300# qhasm: xmm8 &= BS0
301# asm 1: pand BS0,<xmm8=int6464#9
302# asm 2: pand BS0,<xmm8=%xmm8
303pand BS0,%xmm8
304
305# qhasm: xmm5 ^= xmm8
306# asm 1: pxor <xmm8=int6464#9,<xmm5=int6464#6
307# asm 2: pxor <xmm8=%xmm8,<xmm5=%xmm5
308pxor %xmm8,%xmm5
309
310# qhasm: uint6464 xmm8 <<= 1
311# asm 1: psllq $1,<xmm8=int6464#9
312# asm 2: psllq $1,<xmm8=%xmm8
313psllq $1,%xmm8
314
315# qhasm: xmm4 ^= xmm8
316# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
317# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
318pxor %xmm8,%xmm4
319
320# qhasm: xmm8 = xmm2
321# asm 1: movdqa <xmm2=int6464#3,>xmm8=int6464#9
322# asm 2: movdqa <xmm2=%xmm2,>xmm8=%xmm8
323movdqa %xmm2,%xmm8
324
325# qhasm: uint6464 xmm8 >>= 1
326# asm 1: psrlq $1,<xmm8=int6464#9
327# asm 2: psrlq $1,<xmm8=%xmm8
328psrlq $1,%xmm8
329
330# qhasm: xmm8 ^= xmm3
331# asm 1: pxor <xmm3=int6464#4,<xmm8=int6464#9
332# asm 2: pxor <xmm3=%xmm3,<xmm8=%xmm8
333pxor %xmm3,%xmm8
334
335# qhasm: xmm8 &= BS0
336# asm 1: pand BS0,<xmm8=int6464#9
337# asm 2: pand BS0,<xmm8=%xmm8
338pand BS0,%xmm8
339
340# qhasm: xmm3 ^= xmm8
341# asm 1: pxor <xmm8=int6464#9,<xmm3=int6464#4
342# asm 2: pxor <xmm8=%xmm8,<xmm3=%xmm3
343pxor %xmm8,%xmm3
344
345# qhasm: uint6464 xmm8 <<= 1
346# asm 1: psllq $1,<xmm8=int6464#9
347# asm 2: psllq $1,<xmm8=%xmm8
348psllq $1,%xmm8
349
350# qhasm: xmm2 ^= xmm8
351# asm 1: pxor <xmm8=int6464#9,<xmm2=int6464#3
352# asm 2: pxor <xmm8=%xmm8,<xmm2=%xmm2
353pxor %xmm8,%xmm2
354
355# qhasm: xmm8 = xmm0
356# asm 1: movdqa <xmm0=int6464#1,>xmm8=int6464#9
357# asm 2: movdqa <xmm0=%xmm0,>xmm8=%xmm8
358movdqa %xmm0,%xmm8
359
360# qhasm: uint6464 xmm8 >>= 1
361# asm 1: psrlq $1,<xmm8=int6464#9
362# asm 2: psrlq $1,<xmm8=%xmm8
363psrlq $1,%xmm8
364
365# qhasm: xmm8 ^= xmm1
366# asm 1: pxor <xmm1=int6464#2,<xmm8=int6464#9
367# asm 2: pxor <xmm1=%xmm1,<xmm8=%xmm8
368pxor %xmm1,%xmm8
369
370# qhasm: xmm8 &= BS0
371# asm 1: pand BS0,<xmm8=int6464#9
372# asm 2: pand BS0,<xmm8=%xmm8
373pand BS0,%xmm8
374
375# qhasm: xmm1 ^= xmm8
376# asm 1: pxor <xmm8=int6464#9,<xmm1=int6464#2
377# asm 2: pxor <xmm8=%xmm8,<xmm1=%xmm1
378pxor %xmm8,%xmm1
379
380# qhasm: uint6464 xmm8 <<= 1
381# asm 1: psllq $1,<xmm8=int6464#9
382# asm 2: psllq $1,<xmm8=%xmm8
383psllq $1,%xmm8
384
385# qhasm: xmm0 ^= xmm8
386# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
387# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
388pxor %xmm8,%xmm0
389
390# qhasm: xmm8 = xmm5
391# asm 1: movdqa <xmm5=int6464#6,>xmm8=int6464#9
392# asm 2: movdqa <xmm5=%xmm5,>xmm8=%xmm8
393movdqa %xmm5,%xmm8
394
395# qhasm: uint6464 xmm8 >>= 2
396# asm 1: psrlq $2,<xmm8=int6464#9
397# asm 2: psrlq $2,<xmm8=%xmm8
398psrlq $2,%xmm8
399
400# qhasm: xmm8 ^= xmm7
401# asm 1: pxor <xmm7=int6464#8,<xmm8=int6464#9
402# asm 2: pxor <xmm7=%xmm7,<xmm8=%xmm8
403pxor %xmm7,%xmm8
404
405# qhasm: xmm8 &= BS1
406# asm 1: pand BS1,<xmm8=int6464#9
407# asm 2: pand BS1,<xmm8=%xmm8
408pand BS1,%xmm8
409
410# qhasm: xmm7 ^= xmm8
411# asm 1: pxor <xmm8=int6464#9,<xmm7=int6464#8
412# asm 2: pxor <xmm8=%xmm8,<xmm7=%xmm7
413pxor %xmm8,%xmm7
414
415# qhasm: uint6464 xmm8 <<= 2
416# asm 1: psllq $2,<xmm8=int6464#9
417# asm 2: psllq $2,<xmm8=%xmm8
418psllq $2,%xmm8
419
420# qhasm: xmm5 ^= xmm8
421# asm 1: pxor <xmm8=int6464#9,<xmm5=int6464#6
422# asm 2: pxor <xmm8=%xmm8,<xmm5=%xmm5
423pxor %xmm8,%xmm5
424
425# qhasm: xmm8 = xmm4
426# asm 1: movdqa <xmm4=int6464#5,>xmm8=int6464#9
427# asm 2: movdqa <xmm4=%xmm4,>xmm8=%xmm8
428movdqa %xmm4,%xmm8
429
430# qhasm: uint6464 xmm8 >>= 2
431# asm 1: psrlq $2,<xmm8=int6464#9
432# asm 2: psrlq $2,<xmm8=%xmm8
433psrlq $2,%xmm8
434
435# qhasm: xmm8 ^= xmm6
436# asm 1: pxor <xmm6=int6464#7,<xmm8=int6464#9
437# asm 2: pxor <xmm6=%xmm6,<xmm8=%xmm8
438pxor %xmm6,%xmm8
439
440# qhasm: xmm8 &= BS1
441# asm 1: pand BS1,<xmm8=int6464#9
442# asm 2: pand BS1,<xmm8=%xmm8
443pand BS1,%xmm8
444
445# qhasm: xmm6 ^= xmm8
446# asm 1: pxor <xmm8=int6464#9,<xmm6=int6464#7
447# asm 2: pxor <xmm8=%xmm8,<xmm6=%xmm6
448pxor %xmm8,%xmm6
449
450# qhasm: uint6464 xmm8 <<= 2
451# asm 1: psllq $2,<xmm8=int6464#9
452# asm 2: psllq $2,<xmm8=%xmm8
453psllq $2,%xmm8
454
455# qhasm: xmm4 ^= xmm8
456# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
457# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
458pxor %xmm8,%xmm4
459
460# qhasm: xmm8 = xmm1
461# asm 1: movdqa <xmm1=int6464#2,>xmm8=int6464#9
462# asm 2: movdqa <xmm1=%xmm1,>xmm8=%xmm8
463movdqa %xmm1,%xmm8
464
465# qhasm: uint6464 xmm8 >>= 2
466# asm 1: psrlq $2,<xmm8=int6464#9
467# asm 2: psrlq $2,<xmm8=%xmm8
468psrlq $2,%xmm8
469
470# qhasm: xmm8 ^= xmm3
471# asm 1: pxor <xmm3=int6464#4,<xmm8=int6464#9
472# asm 2: pxor <xmm3=%xmm3,<xmm8=%xmm8
473pxor %xmm3,%xmm8
474
475# qhasm: xmm8 &= BS1
476# asm 1: pand BS1,<xmm8=int6464#9
477# asm 2: pand BS1,<xmm8=%xmm8
478pand BS1,%xmm8
479
480# qhasm: xmm3 ^= xmm8
481# asm 1: pxor <xmm8=int6464#9,<xmm3=int6464#4
482# asm 2: pxor <xmm8=%xmm8,<xmm3=%xmm3
483pxor %xmm8,%xmm3
484
485# qhasm: uint6464 xmm8 <<= 2
486# asm 1: psllq $2,<xmm8=int6464#9
487# asm 2: psllq $2,<xmm8=%xmm8
488psllq $2,%xmm8
489
490# qhasm: xmm1 ^= xmm8
491# asm 1: pxor <xmm8=int6464#9,<xmm1=int6464#2
492# asm 2: pxor <xmm8=%xmm8,<xmm1=%xmm1
493pxor %xmm8,%xmm1
494
495# qhasm: xmm8 = xmm0
496# asm 1: movdqa <xmm0=int6464#1,>xmm8=int6464#9
497# asm 2: movdqa <xmm0=%xmm0,>xmm8=%xmm8
498movdqa %xmm0,%xmm8
499
500# qhasm: uint6464 xmm8 >>= 2
501# asm 1: psrlq $2,<xmm8=int6464#9
502# asm 2: psrlq $2,<xmm8=%xmm8
503psrlq $2,%xmm8
504
505# qhasm: xmm8 ^= xmm2
506# asm 1: pxor <xmm2=int6464#3,<xmm8=int6464#9
507# asm 2: pxor <xmm2=%xmm2,<xmm8=%xmm8
508pxor %xmm2,%xmm8
509
510# qhasm: xmm8 &= BS1
511# asm 1: pand BS1,<xmm8=int6464#9
512# asm 2: pand BS1,<xmm8=%xmm8
513pand BS1,%xmm8
514
515# qhasm: xmm2 ^= xmm8
516# asm 1: pxor <xmm8=int6464#9,<xmm2=int6464#3
517# asm 2: pxor <xmm8=%xmm8,<xmm2=%xmm2
518pxor %xmm8,%xmm2
519
520# qhasm: uint6464 xmm8 <<= 2
521# asm 1: psllq $2,<xmm8=int6464#9
522# asm 2: psllq $2,<xmm8=%xmm8
523psllq $2,%xmm8
524
525# qhasm: xmm0 ^= xmm8
526# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
527# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
528pxor %xmm8,%xmm0
529
530# qhasm: xmm8 = xmm3
531# asm 1: movdqa <xmm3=int6464#4,>xmm8=int6464#9
532# asm 2: movdqa <xmm3=%xmm3,>xmm8=%xmm8
533movdqa %xmm3,%xmm8
534
535# qhasm: uint6464 xmm8 >>= 4
536# asm 1: psrlq $4,<xmm8=int6464#9
537# asm 2: psrlq $4,<xmm8=%xmm8
538psrlq $4,%xmm8
539
540# qhasm: xmm8 ^= xmm7
541# asm 1: pxor <xmm7=int6464#8,<xmm8=int6464#9
542# asm 2: pxor <xmm7=%xmm7,<xmm8=%xmm8
543pxor %xmm7,%xmm8
544
545# qhasm: xmm8 &= BS2
546# asm 1: pand BS2,<xmm8=int6464#9
547# asm 2: pand BS2,<xmm8=%xmm8
548pand BS2,%xmm8
549
550# qhasm: xmm7 ^= xmm8
551# asm 1: pxor <xmm8=int6464#9,<xmm7=int6464#8
552# asm 2: pxor <xmm8=%xmm8,<xmm7=%xmm7
553pxor %xmm8,%xmm7
554
555# qhasm: uint6464 xmm8 <<= 4
556# asm 1: psllq $4,<xmm8=int6464#9
557# asm 2: psllq $4,<xmm8=%xmm8
558psllq $4,%xmm8
559
560# qhasm: xmm3 ^= xmm8
561# asm 1: pxor <xmm8=int6464#9,<xmm3=int6464#4
562# asm 2: pxor <xmm8=%xmm8,<xmm3=%xmm3
563pxor %xmm8,%xmm3
564
565# qhasm: xmm8 = xmm2
566# asm 1: movdqa <xmm2=int6464#3,>xmm8=int6464#9
567# asm 2: movdqa <xmm2=%xmm2,>xmm8=%xmm8
568movdqa %xmm2,%xmm8
569
570# qhasm: uint6464 xmm8 >>= 4
571# asm 1: psrlq $4,<xmm8=int6464#9
572# asm 2: psrlq $4,<xmm8=%xmm8
573psrlq $4,%xmm8
574
575# qhasm: xmm8 ^= xmm6
576# asm 1: pxor <xmm6=int6464#7,<xmm8=int6464#9
577# asm 2: pxor <xmm6=%xmm6,<xmm8=%xmm8
578pxor %xmm6,%xmm8
579
580# qhasm: xmm8 &= BS2
581# asm 1: pand BS2,<xmm8=int6464#9
582# asm 2: pand BS2,<xmm8=%xmm8
583pand BS2,%xmm8
584
585# qhasm: xmm6 ^= xmm8
586# asm 1: pxor <xmm8=int6464#9,<xmm6=int6464#7
587# asm 2: pxor <xmm8=%xmm8,<xmm6=%xmm6
588pxor %xmm8,%xmm6
589
590# qhasm: uint6464 xmm8 <<= 4
591# asm 1: psllq $4,<xmm8=int6464#9
592# asm 2: psllq $4,<xmm8=%xmm8
593psllq $4,%xmm8
594
595# qhasm: xmm2 ^= xmm8
596# asm 1: pxor <xmm8=int6464#9,<xmm2=int6464#3
597# asm 2: pxor <xmm8=%xmm8,<xmm2=%xmm2
598pxor %xmm8,%xmm2
599
600# qhasm: xmm8 = xmm1
601# asm 1: movdqa <xmm1=int6464#2,>xmm8=int6464#9
602# asm 2: movdqa <xmm1=%xmm1,>xmm8=%xmm8
603movdqa %xmm1,%xmm8
604
605# qhasm: uint6464 xmm8 >>= 4
606# asm 1: psrlq $4,<xmm8=int6464#9
607# asm 2: psrlq $4,<xmm8=%xmm8
608psrlq $4,%xmm8
609
610# qhasm: xmm8 ^= xmm5
611# asm 1: pxor <xmm5=int6464#6,<xmm8=int6464#9
612# asm 2: pxor <xmm5=%xmm5,<xmm8=%xmm8
613pxor %xmm5,%xmm8
614
615# qhasm: xmm8 &= BS2
616# asm 1: pand BS2,<xmm8=int6464#9
617# asm 2: pand BS2,<xmm8=%xmm8
618pand BS2,%xmm8
619
620# qhasm: xmm5 ^= xmm8
621# asm 1: pxor <xmm8=int6464#9,<xmm5=int6464#6
622# asm 2: pxor <xmm8=%xmm8,<xmm5=%xmm5
623pxor %xmm8,%xmm5
624
625# qhasm: uint6464 xmm8 <<= 4
626# asm 1: psllq $4,<xmm8=int6464#9
627# asm 2: psllq $4,<xmm8=%xmm8
628psllq $4,%xmm8
629
630# qhasm: xmm1 ^= xmm8
631# asm 1: pxor <xmm8=int6464#9,<xmm1=int6464#2
632# asm 2: pxor <xmm8=%xmm8,<xmm1=%xmm1
633pxor %xmm8,%xmm1
634
635# qhasm: xmm8 = xmm0
636# asm 1: movdqa <xmm0=int6464#1,>xmm8=int6464#9
637# asm 2: movdqa <xmm0=%xmm0,>xmm8=%xmm8
638movdqa %xmm0,%xmm8
639
640# qhasm: uint6464 xmm8 >>= 4
641# asm 1: psrlq $4,<xmm8=int6464#9
642# asm 2: psrlq $4,<xmm8=%xmm8
643psrlq $4,%xmm8
644
645# qhasm: xmm8 ^= xmm4
646# asm 1: pxor <xmm4=int6464#5,<xmm8=int6464#9
647# asm 2: pxor <xmm4=%xmm4,<xmm8=%xmm8
648pxor %xmm4,%xmm8
649
650# qhasm: xmm8 &= BS2
651# asm 1: pand BS2,<xmm8=int6464#9
652# asm 2: pand BS2,<xmm8=%xmm8
653pand BS2,%xmm8
654
655# qhasm: xmm4 ^= xmm8
656# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
657# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
658pxor %xmm8,%xmm4
659
660# qhasm: uint6464 xmm8 <<= 4
661# asm 1: psllq $4,<xmm8=int6464#9
662# asm 2: psllq $4,<xmm8=%xmm8
663psllq $4,%xmm8
664
665# qhasm: xmm0 ^= xmm8
666# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
667# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
668pxor %xmm8,%xmm0
669
670# qhasm: xmm0 ^= *(int128 *)(c + 0)
671# asm 1: pxor 0(<c=int64#4),<xmm0=int6464#1
672# asm 2: pxor 0(<c=%rcx),<xmm0=%xmm0
673pxor 0(%rcx),%xmm0
674
675# qhasm: shuffle bytes of xmm0 by SR
676# asm 1: pshufb SR,<xmm0=int6464#1
677# asm 2: pshufb SR,<xmm0=%xmm0
678pshufb SR,%xmm0
679
680# qhasm: xmm1 ^= *(int128 *)(c + 16)
681# asm 1: pxor 16(<c=int64#4),<xmm1=int6464#2
682# asm 2: pxor 16(<c=%rcx),<xmm1=%xmm1
683pxor 16(%rcx),%xmm1
684
685# qhasm: shuffle bytes of xmm1 by SR
686# asm 1: pshufb SR,<xmm1=int6464#2
687# asm 2: pshufb SR,<xmm1=%xmm1
688pshufb SR,%xmm1
689
690# qhasm: xmm2 ^= *(int128 *)(c + 32)
691# asm 1: pxor 32(<c=int64#4),<xmm2=int6464#3
692# asm 2: pxor 32(<c=%rcx),<xmm2=%xmm2
693pxor 32(%rcx),%xmm2
694
695# qhasm: shuffle bytes of xmm2 by SR
696# asm 1: pshufb SR,<xmm2=int6464#3
697# asm 2: pshufb SR,<xmm2=%xmm2
698pshufb SR,%xmm2
699
700# qhasm: xmm3 ^= *(int128 *)(c + 48)
701# asm 1: pxor 48(<c=int64#4),<xmm3=int6464#4
702# asm 2: pxor 48(<c=%rcx),<xmm3=%xmm3
703pxor 48(%rcx),%xmm3
704
705# qhasm: shuffle bytes of xmm3 by SR
706# asm 1: pshufb SR,<xmm3=int6464#4
707# asm 2: pshufb SR,<xmm3=%xmm3
708pshufb SR,%xmm3
709
710# qhasm: xmm4 ^= *(int128 *)(c + 64)
711# asm 1: pxor 64(<c=int64#4),<xmm4=int6464#5
712# asm 2: pxor 64(<c=%rcx),<xmm4=%xmm4
713pxor 64(%rcx),%xmm4
714
715# qhasm: shuffle bytes of xmm4 by SR
716# asm 1: pshufb SR,<xmm4=int6464#5
717# asm 2: pshufb SR,<xmm4=%xmm4
718pshufb SR,%xmm4
719
720# qhasm: xmm5 ^= *(int128 *)(c + 80)
721# asm 1: pxor 80(<c=int64#4),<xmm5=int6464#6
722# asm 2: pxor 80(<c=%rcx),<xmm5=%xmm5
723pxor 80(%rcx),%xmm5
724
725# qhasm: shuffle bytes of xmm5 by SR
726# asm 1: pshufb SR,<xmm5=int6464#6
727# asm 2: pshufb SR,<xmm5=%xmm5
728pshufb SR,%xmm5
729
730# qhasm: xmm6 ^= *(int128 *)(c + 96)
731# asm 1: pxor 96(<c=int64#4),<xmm6=int6464#7
732# asm 2: pxor 96(<c=%rcx),<xmm6=%xmm6
733pxor 96(%rcx),%xmm6
734
735# qhasm: shuffle bytes of xmm6 by SR
736# asm 1: pshufb SR,<xmm6=int6464#7
737# asm 2: pshufb SR,<xmm6=%xmm6
738pshufb SR,%xmm6
739
740# qhasm: xmm7 ^= *(int128 *)(c + 112)
741# asm 1: pxor 112(<c=int64#4),<xmm7=int6464#8
742# asm 2: pxor 112(<c=%rcx),<xmm7=%xmm7
743pxor 112(%rcx),%xmm7
744
745# qhasm: shuffle bytes of xmm7 by SR
746# asm 1: pshufb SR,<xmm7=int6464#8
747# asm 2: pshufb SR,<xmm7=%xmm7
748pshufb SR,%xmm7
749
750# qhasm: xmm5 ^= xmm6
751# asm 1: pxor <xmm6=int6464#7,<xmm5=int6464#6
752# asm 2: pxor <xmm6=%xmm6,<xmm5=%xmm5
753pxor %xmm6,%xmm5
754
755# qhasm: xmm2 ^= xmm1
756# asm 1: pxor <xmm1=int6464#2,<xmm2=int6464#3
757# asm 2: pxor <xmm1=%xmm1,<xmm2=%xmm2
758pxor %xmm1,%xmm2
759
760# qhasm: xmm5 ^= xmm0
761# asm 1: pxor <xmm0=int6464#1,<xmm5=int6464#6
762# asm 2: pxor <xmm0=%xmm0,<xmm5=%xmm5
763pxor %xmm0,%xmm5
764
765# qhasm: xmm6 ^= xmm2
766# asm 1: pxor <xmm2=int6464#3,<xmm6=int6464#7
767# asm 2: pxor <xmm2=%xmm2,<xmm6=%xmm6
768pxor %xmm2,%xmm6
769
770# qhasm: xmm3 ^= xmm0
771# asm 1: pxor <xmm0=int6464#1,<xmm3=int6464#4
772# asm 2: pxor <xmm0=%xmm0,<xmm3=%xmm3
773pxor %xmm0,%xmm3
774
775# qhasm: xmm6 ^= xmm3
776# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
777# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
778pxor %xmm3,%xmm6
779
780# qhasm: xmm3 ^= xmm7
781# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#4
782# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm3
783pxor %xmm7,%xmm3
784
785# qhasm: xmm3 ^= xmm4
786# asm 1: pxor <xmm4=int6464#5,<xmm3=int6464#4
787# asm 2: pxor <xmm4=%xmm4,<xmm3=%xmm3
788pxor %xmm4,%xmm3
789
790# qhasm: xmm7 ^= xmm5
791# asm 1: pxor <xmm5=int6464#6,<xmm7=int6464#8
792# asm 2: pxor <xmm5=%xmm5,<xmm7=%xmm7
793pxor %xmm5,%xmm7
794
795# qhasm: xmm3 ^= xmm1
796# asm 1: pxor <xmm1=int6464#2,<xmm3=int6464#4
797# asm 2: pxor <xmm1=%xmm1,<xmm3=%xmm3
798pxor %xmm1,%xmm3
799
800# qhasm: xmm4 ^= xmm5
801# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
802# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
803pxor %xmm5,%xmm4
804
805# qhasm: xmm2 ^= xmm7
806# asm 1: pxor <xmm7=int6464#8,<xmm2=int6464#3
807# asm 2: pxor <xmm7=%xmm7,<xmm2=%xmm2
808pxor %xmm7,%xmm2
809
810# qhasm: xmm1 ^= xmm5
811# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
812# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
813pxor %xmm5,%xmm1
814
815# qhasm: xmm11 = xmm7
816# asm 1: movdqa <xmm7=int6464#8,>xmm11=int6464#9
817# asm 2: movdqa <xmm7=%xmm7,>xmm11=%xmm8
818movdqa %xmm7,%xmm8
819
820# qhasm: xmm10 = xmm1
821# asm 1: movdqa <xmm1=int6464#2,>xmm10=int6464#10
822# asm 2: movdqa <xmm1=%xmm1,>xmm10=%xmm9
823movdqa %xmm1,%xmm9
824
825# qhasm: xmm9 = xmm5
826# asm 1: movdqa <xmm5=int6464#6,>xmm9=int6464#11
827# asm 2: movdqa <xmm5=%xmm5,>xmm9=%xmm10
828movdqa %xmm5,%xmm10
829
830# qhasm: xmm13 = xmm2
831# asm 1: movdqa <xmm2=int6464#3,>xmm13=int6464#12
832# asm 2: movdqa <xmm2=%xmm2,>xmm13=%xmm11
833movdqa %xmm2,%xmm11
834
835# qhasm: xmm12 = xmm6
836# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#13
837# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm12
838movdqa %xmm6,%xmm12
839
840# qhasm: xmm11 ^= xmm4
841# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#9
842# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm8
843pxor %xmm4,%xmm8
844
845# qhasm: xmm10 ^= xmm2
846# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#10
847# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm9
848pxor %xmm2,%xmm9
849
850# qhasm: xmm9 ^= xmm3
851# asm 1: pxor <xmm3=int6464#4,<xmm9=int6464#11
852# asm 2: pxor <xmm3=%xmm3,<xmm9=%xmm10
853pxor %xmm3,%xmm10
854
855# qhasm: xmm13 ^= xmm4
856# asm 1: pxor <xmm4=int6464#5,<xmm13=int6464#12
857# asm 2: pxor <xmm4=%xmm4,<xmm13=%xmm11
858pxor %xmm4,%xmm11
859
860# qhasm: xmm12 ^= xmm0
861# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
862# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
863pxor %xmm0,%xmm12
864
865# qhasm: xmm14 = xmm11
866# asm 1: movdqa <xmm11=int6464#9,>xmm14=int6464#14
867# asm 2: movdqa <xmm11=%xmm8,>xmm14=%xmm13
868movdqa %xmm8,%xmm13
869
870# qhasm: xmm8 = xmm10
871# asm 1: movdqa <xmm10=int6464#10,>xmm8=int6464#15
872# asm 2: movdqa <xmm10=%xmm9,>xmm8=%xmm14
873movdqa %xmm9,%xmm14
874
875# qhasm: xmm15 = xmm11
876# asm 1: movdqa <xmm11=int6464#9,>xmm15=int6464#16
877# asm 2: movdqa <xmm11=%xmm8,>xmm15=%xmm15
878movdqa %xmm8,%xmm15
879
880# qhasm: xmm10 |= xmm9
881# asm 1: por <xmm9=int6464#11,<xmm10=int6464#10
882# asm 2: por <xmm9=%xmm10,<xmm10=%xmm9
883por %xmm10,%xmm9
884
885# qhasm: xmm11 |= xmm12
886# asm 1: por <xmm12=int6464#13,<xmm11=int6464#9
887# asm 2: por <xmm12=%xmm12,<xmm11=%xmm8
888por %xmm12,%xmm8
889
890# qhasm: xmm15 ^= xmm8
891# asm 1: pxor <xmm8=int6464#15,<xmm15=int6464#16
892# asm 2: pxor <xmm8=%xmm14,<xmm15=%xmm15
893pxor %xmm14,%xmm15
894
895# qhasm: xmm14 &= xmm12
896# asm 1: pand <xmm12=int6464#13,<xmm14=int6464#14
897# asm 2: pand <xmm12=%xmm12,<xmm14=%xmm13
898pand %xmm12,%xmm13
899
900# qhasm: xmm8 &= xmm9
901# asm 1: pand <xmm9=int6464#11,<xmm8=int6464#15
902# asm 2: pand <xmm9=%xmm10,<xmm8=%xmm14
903pand %xmm10,%xmm14
904
905# qhasm: xmm12 ^= xmm9
906# asm 1: pxor <xmm9=int6464#11,<xmm12=int6464#13
907# asm 2: pxor <xmm9=%xmm10,<xmm12=%xmm12
908pxor %xmm10,%xmm12
909
910# qhasm: xmm15 &= xmm12
911# asm 1: pand <xmm12=int6464#13,<xmm15=int6464#16
912# asm 2: pand <xmm12=%xmm12,<xmm15=%xmm15
913pand %xmm12,%xmm15
914
915# qhasm: xmm12 = xmm3
916# asm 1: movdqa <xmm3=int6464#4,>xmm12=int6464#11
917# asm 2: movdqa <xmm3=%xmm3,>xmm12=%xmm10
918movdqa %xmm3,%xmm10
919
920# qhasm: xmm12 ^= xmm0
921# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#11
922# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm10
923pxor %xmm0,%xmm10
924
925# qhasm: xmm13 &= xmm12
926# asm 1: pand <xmm12=int6464#11,<xmm13=int6464#12
927# asm 2: pand <xmm12=%xmm10,<xmm13=%xmm11
928pand %xmm10,%xmm11
929
930# qhasm: xmm11 ^= xmm13
931# asm 1: pxor <xmm13=int6464#12,<xmm11=int6464#9
932# asm 2: pxor <xmm13=%xmm11,<xmm11=%xmm8
933pxor %xmm11,%xmm8
934
935# qhasm: xmm10 ^= xmm13
936# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
937# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
938pxor %xmm11,%xmm9
939
940# qhasm: xmm13 = xmm7
941# asm 1: movdqa <xmm7=int6464#8,>xmm13=int6464#11
942# asm 2: movdqa <xmm7=%xmm7,>xmm13=%xmm10
943movdqa %xmm7,%xmm10
944
945# qhasm: xmm13 ^= xmm1
946# asm 1: pxor <xmm1=int6464#2,<xmm13=int6464#11
947# asm 2: pxor <xmm1=%xmm1,<xmm13=%xmm10
948pxor %xmm1,%xmm10
949
950# qhasm: xmm12 = xmm5
951# asm 1: movdqa <xmm5=int6464#6,>xmm12=int6464#12
952# asm 2: movdqa <xmm5=%xmm5,>xmm12=%xmm11
953movdqa %xmm5,%xmm11
954
955# qhasm: xmm9 = xmm13
956# asm 1: movdqa <xmm13=int6464#11,>xmm9=int6464#13
957# asm 2: movdqa <xmm13=%xmm10,>xmm9=%xmm12
958movdqa %xmm10,%xmm12
959
960# qhasm: xmm12 ^= xmm6
961# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#12
962# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm11
963pxor %xmm6,%xmm11
964
965# qhasm: xmm9 |= xmm12
966# asm 1: por <xmm12=int6464#12,<xmm9=int6464#13
967# asm 2: por <xmm12=%xmm11,<xmm9=%xmm12
968por %xmm11,%xmm12
969
970# qhasm: xmm13 &= xmm12
971# asm 1: pand <xmm12=int6464#12,<xmm13=int6464#11
972# asm 2: pand <xmm12=%xmm11,<xmm13=%xmm10
973pand %xmm11,%xmm10
974
975# qhasm: xmm8 ^= xmm13
976# asm 1: pxor <xmm13=int6464#11,<xmm8=int6464#15
977# asm 2: pxor <xmm13=%xmm10,<xmm8=%xmm14
978pxor %xmm10,%xmm14
979
980# qhasm: xmm11 ^= xmm15
981# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#9
982# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm8
983pxor %xmm15,%xmm8
984
985# qhasm: xmm10 ^= xmm14
986# asm 1: pxor <xmm14=int6464#14,<xmm10=int6464#10
987# asm 2: pxor <xmm14=%xmm13,<xmm10=%xmm9
988pxor %xmm13,%xmm9
989
990# qhasm: xmm9 ^= xmm15
991# asm 1: pxor <xmm15=int6464#16,<xmm9=int6464#13
992# asm 2: pxor <xmm15=%xmm15,<xmm9=%xmm12
993pxor %xmm15,%xmm12
994
995# qhasm: xmm8 ^= xmm14
996# asm 1: pxor <xmm14=int6464#14,<xmm8=int6464#15
997# asm 2: pxor <xmm14=%xmm13,<xmm8=%xmm14
998pxor %xmm13,%xmm14
999
1000# qhasm: xmm9 ^= xmm14
1001# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
1002# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
1003pxor %xmm13,%xmm12
1004
1005# qhasm: xmm12 = xmm2
1006# asm 1: movdqa <xmm2=int6464#3,>xmm12=int6464#11
1007# asm 2: movdqa <xmm2=%xmm2,>xmm12=%xmm10
1008movdqa %xmm2,%xmm10
1009
1010# qhasm: xmm13 = xmm4
1011# asm 1: movdqa <xmm4=int6464#5,>xmm13=int6464#12
1012# asm 2: movdqa <xmm4=%xmm4,>xmm13=%xmm11
1013movdqa %xmm4,%xmm11
1014
1015# qhasm: xmm14 = xmm1
1016# asm 1: movdqa <xmm1=int6464#2,>xmm14=int6464#14
1017# asm 2: movdqa <xmm1=%xmm1,>xmm14=%xmm13
1018movdqa %xmm1,%xmm13
1019
1020# qhasm: xmm15 = xmm7
1021# asm 1: movdqa <xmm7=int6464#8,>xmm15=int6464#16
1022# asm 2: movdqa <xmm7=%xmm7,>xmm15=%xmm15
1023movdqa %xmm7,%xmm15
1024
1025# qhasm: xmm12 &= xmm3
1026# asm 1: pand <xmm3=int6464#4,<xmm12=int6464#11
1027# asm 2: pand <xmm3=%xmm3,<xmm12=%xmm10
1028pand %xmm3,%xmm10
1029
1030# qhasm: xmm13 &= xmm0
1031# asm 1: pand <xmm0=int6464#1,<xmm13=int6464#12
1032# asm 2: pand <xmm0=%xmm0,<xmm13=%xmm11
1033pand %xmm0,%xmm11
1034
1035# qhasm: xmm14 &= xmm5
1036# asm 1: pand <xmm5=int6464#6,<xmm14=int6464#14
1037# asm 2: pand <xmm5=%xmm5,<xmm14=%xmm13
1038pand %xmm5,%xmm13
1039
1040# qhasm: xmm15 |= xmm6
1041# asm 1: por <xmm6=int6464#7,<xmm15=int6464#16
1042# asm 2: por <xmm6=%xmm6,<xmm15=%xmm15
1043por %xmm6,%xmm15
1044
1045# qhasm: xmm11 ^= xmm12
1046# asm 1: pxor <xmm12=int6464#11,<xmm11=int6464#9
1047# asm 2: pxor <xmm12=%xmm10,<xmm11=%xmm8
1048pxor %xmm10,%xmm8
1049
1050# qhasm: xmm10 ^= xmm13
1051# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
1052# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
1053pxor %xmm11,%xmm9
1054
1055# qhasm: xmm9 ^= xmm14
1056# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
1057# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
1058pxor %xmm13,%xmm12
1059
1060# qhasm: xmm8 ^= xmm15
1061# asm 1: pxor <xmm15=int6464#16,<xmm8=int6464#15
1062# asm 2: pxor <xmm15=%xmm15,<xmm8=%xmm14
1063pxor %xmm15,%xmm14
1064
1065# qhasm: xmm12 = xmm11
1066# asm 1: movdqa <xmm11=int6464#9,>xmm12=int6464#11
1067# asm 2: movdqa <xmm11=%xmm8,>xmm12=%xmm10
1068movdqa %xmm8,%xmm10
1069
1070# qhasm: xmm12 ^= xmm10
1071# asm 1: pxor <xmm10=int6464#10,<xmm12=int6464#11
1072# asm 2: pxor <xmm10=%xmm9,<xmm12=%xmm10
1073pxor %xmm9,%xmm10
1074
1075# qhasm: xmm11 &= xmm9
1076# asm 1: pand <xmm9=int6464#13,<xmm11=int6464#9
1077# asm 2: pand <xmm9=%xmm12,<xmm11=%xmm8
1078pand %xmm12,%xmm8
1079
1080# qhasm: xmm14 = xmm8
1081# asm 1: movdqa <xmm8=int6464#15,>xmm14=int6464#12
1082# asm 2: movdqa <xmm8=%xmm14,>xmm14=%xmm11
1083movdqa %xmm14,%xmm11
1084
1085# qhasm: xmm14 ^= xmm11
1086# asm 1: pxor <xmm11=int6464#9,<xmm14=int6464#12
1087# asm 2: pxor <xmm11=%xmm8,<xmm14=%xmm11
1088pxor %xmm8,%xmm11
1089
1090# qhasm: xmm15 = xmm12
1091# asm 1: movdqa <xmm12=int6464#11,>xmm15=int6464#14
1092# asm 2: movdqa <xmm12=%xmm10,>xmm15=%xmm13
1093movdqa %xmm10,%xmm13
1094
1095# qhasm: xmm15 &= xmm14
1096# asm 1: pand <xmm14=int6464#12,<xmm15=int6464#14
1097# asm 2: pand <xmm14=%xmm11,<xmm15=%xmm13
1098pand %xmm11,%xmm13
1099
1100# qhasm: xmm15 ^= xmm10
1101# asm 1: pxor <xmm10=int6464#10,<xmm15=int6464#14
1102# asm 2: pxor <xmm10=%xmm9,<xmm15=%xmm13
1103pxor %xmm9,%xmm13
1104
1105# qhasm: xmm13 = xmm9
1106# asm 1: movdqa <xmm9=int6464#13,>xmm13=int6464#16
1107# asm 2: movdqa <xmm9=%xmm12,>xmm13=%xmm15
1108movdqa %xmm12,%xmm15
1109
1110# qhasm: xmm13 ^= xmm8
1111# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
1112# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
1113pxor %xmm14,%xmm15
1114
1115# qhasm: xmm11 ^= xmm10
1116# asm 1: pxor <xmm10=int6464#10,<xmm11=int6464#9
1117# asm 2: pxor <xmm10=%xmm9,<xmm11=%xmm8
1118pxor %xmm9,%xmm8
1119
1120# qhasm: xmm13 &= xmm11
1121# asm 1: pand <xmm11=int6464#9,<xmm13=int6464#16
1122# asm 2: pand <xmm11=%xmm8,<xmm13=%xmm15
1123pand %xmm8,%xmm15
1124
1125# qhasm: xmm13 ^= xmm8
1126# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
1127# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
1128pxor %xmm14,%xmm15
1129
1130# qhasm: xmm9 ^= xmm13
1131# asm 1: pxor <xmm13=int6464#16,<xmm9=int6464#13
1132# asm 2: pxor <xmm13=%xmm15,<xmm9=%xmm12
1133pxor %xmm15,%xmm12
1134
1135# qhasm: xmm10 = xmm14
1136# asm 1: movdqa <xmm14=int6464#12,>xmm10=int6464#9
1137# asm 2: movdqa <xmm14=%xmm11,>xmm10=%xmm8
1138movdqa %xmm11,%xmm8
1139
1140# qhasm: xmm10 ^= xmm13
1141# asm 1: pxor <xmm13=int6464#16,<xmm10=int6464#9
1142# asm 2: pxor <xmm13=%xmm15,<xmm10=%xmm8
1143pxor %xmm15,%xmm8
1144
1145# qhasm: xmm10 &= xmm8
1146# asm 1: pand <xmm8=int6464#15,<xmm10=int6464#9
1147# asm 2: pand <xmm8=%xmm14,<xmm10=%xmm8
1148pand %xmm14,%xmm8
1149
1150# qhasm: xmm9 ^= xmm10
1151# asm 1: pxor <xmm10=int6464#9,<xmm9=int6464#13
1152# asm 2: pxor <xmm10=%xmm8,<xmm9=%xmm12
1153pxor %xmm8,%xmm12
1154
1155# qhasm: xmm14 ^= xmm10
1156# asm 1: pxor <xmm10=int6464#9,<xmm14=int6464#12
1157# asm 2: pxor <xmm10=%xmm8,<xmm14=%xmm11
1158pxor %xmm8,%xmm11
1159
1160# qhasm: xmm14 &= xmm15
1161# asm 1: pand <xmm15=int6464#14,<xmm14=int6464#12
1162# asm 2: pand <xmm15=%xmm13,<xmm14=%xmm11
1163pand %xmm13,%xmm11
1164
1165# qhasm: xmm14 ^= xmm12
1166# asm 1: pxor <xmm12=int6464#11,<xmm14=int6464#12
1167# asm 2: pxor <xmm12=%xmm10,<xmm14=%xmm11
1168pxor %xmm10,%xmm11
1169
1170# qhasm: xmm12 = xmm6
1171# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#9
1172# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm8
1173movdqa %xmm6,%xmm8
1174
1175# qhasm: xmm8 = xmm5
1176# asm 1: movdqa <xmm5=int6464#6,>xmm8=int6464#10
1177# asm 2: movdqa <xmm5=%xmm5,>xmm8=%xmm9
1178movdqa %xmm5,%xmm9
1179
1180# qhasm: xmm10 = xmm15
1181# asm 1: movdqa <xmm15=int6464#14,>xmm10=int6464#11
1182# asm 2: movdqa <xmm15=%xmm13,>xmm10=%xmm10
1183movdqa %xmm13,%xmm10
1184
1185# qhasm: xmm10 ^= xmm14
1186# asm 1: pxor <xmm14=int6464#12,<xmm10=int6464#11
1187# asm 2: pxor <xmm14=%xmm11,<xmm10=%xmm10
1188pxor %xmm11,%xmm10
1189
1190# qhasm: xmm10 &= xmm6
1191# asm 1: pand <xmm6=int6464#7,<xmm10=int6464#11
1192# asm 2: pand <xmm6=%xmm6,<xmm10=%xmm10
1193pand %xmm6,%xmm10
1194
1195# qhasm: xmm6 ^= xmm5
1196# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
1197# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
1198pxor %xmm5,%xmm6
1199
1200# qhasm: xmm6 &= xmm14
1201# asm 1: pand <xmm14=int6464#12,<xmm6=int6464#7
1202# asm 2: pand <xmm14=%xmm11,<xmm6=%xmm6
1203pand %xmm11,%xmm6
1204
1205# qhasm: xmm5 &= xmm15
1206# asm 1: pand <xmm15=int6464#14,<xmm5=int6464#6
1207# asm 2: pand <xmm15=%xmm13,<xmm5=%xmm5
1208pand %xmm13,%xmm5
1209
1210# qhasm: xmm6 ^= xmm5
1211# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
1212# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
1213pxor %xmm5,%xmm6
1214
1215# qhasm: xmm5 ^= xmm10
1216# asm 1: pxor <xmm10=int6464#11,<xmm5=int6464#6
1217# asm 2: pxor <xmm10=%xmm10,<xmm5=%xmm5
1218pxor %xmm10,%xmm5
1219
1220# qhasm: xmm12 ^= xmm0
1221# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#9
1222# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm8
1223pxor %xmm0,%xmm8
1224
1225# qhasm: xmm8 ^= xmm3
1226# asm 1: pxor <xmm3=int6464#4,<xmm8=int6464#10
1227# asm 2: pxor <xmm3=%xmm3,<xmm8=%xmm9
1228pxor %xmm3,%xmm9
1229
1230# qhasm: xmm15 ^= xmm13
1231# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
1232# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
1233pxor %xmm15,%xmm13
1234
1235# qhasm: xmm14 ^= xmm9
1236# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
1237# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
1238pxor %xmm12,%xmm11
1239
1240# qhasm: xmm11 = xmm15
1241# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
1242# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
1243movdqa %xmm13,%xmm10
1244
1245# qhasm: xmm11 ^= xmm14
1246# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
1247# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
1248pxor %xmm11,%xmm10
1249
1250# qhasm: xmm11 &= xmm12
1251# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
1252# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
1253pand %xmm8,%xmm10
1254
1255# qhasm: xmm12 ^= xmm8
1256# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
1257# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
1258pxor %xmm9,%xmm8
1259
1260# qhasm: xmm12 &= xmm14
1261# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
1262# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
1263pand %xmm11,%xmm8
1264
1265# qhasm: xmm8 &= xmm15
1266# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
1267# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
1268pand %xmm13,%xmm9
1269
1270# qhasm: xmm8 ^= xmm12
1271# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
1272# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
1273pxor %xmm8,%xmm9
1274
1275# qhasm: xmm12 ^= xmm11
1276# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
1277# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
1278pxor %xmm10,%xmm8
1279
1280# qhasm: xmm10 = xmm13
1281# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
1282# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
1283movdqa %xmm15,%xmm10
1284
1285# qhasm: xmm10 ^= xmm9
1286# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
1287# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
1288pxor %xmm12,%xmm10
1289
1290# qhasm: xmm10 &= xmm0
1291# asm 1: pand <xmm0=int6464#1,<xmm10=int6464#11
1292# asm 2: pand <xmm0=%xmm0,<xmm10=%xmm10
1293pand %xmm0,%xmm10
1294
1295# qhasm: xmm0 ^= xmm3
1296# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
1297# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
1298pxor %xmm3,%xmm0
1299
1300# qhasm: xmm0 &= xmm9
1301# asm 1: pand <xmm9=int6464#13,<xmm0=int6464#1
1302# asm 2: pand <xmm9=%xmm12,<xmm0=%xmm0
1303pand %xmm12,%xmm0
1304
1305# qhasm: xmm3 &= xmm13
1306# asm 1: pand <xmm13=int6464#16,<xmm3=int6464#4
1307# asm 2: pand <xmm13=%xmm15,<xmm3=%xmm3
1308pand %xmm15,%xmm3
1309
1310# qhasm: xmm0 ^= xmm3
1311# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
1312# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
1313pxor %xmm3,%xmm0
1314
1315# qhasm: xmm3 ^= xmm10
1316# asm 1: pxor <xmm10=int6464#11,<xmm3=int6464#4
1317# asm 2: pxor <xmm10=%xmm10,<xmm3=%xmm3
1318pxor %xmm10,%xmm3
1319
1320# qhasm: xmm6 ^= xmm12
1321# asm 1: pxor <xmm12=int6464#9,<xmm6=int6464#7
1322# asm 2: pxor <xmm12=%xmm8,<xmm6=%xmm6
1323pxor %xmm8,%xmm6
1324
1325# qhasm: xmm0 ^= xmm12
1326# asm 1: pxor <xmm12=int6464#9,<xmm0=int6464#1
1327# asm 2: pxor <xmm12=%xmm8,<xmm0=%xmm0
1328pxor %xmm8,%xmm0
1329
1330# qhasm: xmm5 ^= xmm8
1331# asm 1: pxor <xmm8=int6464#10,<xmm5=int6464#6
1332# asm 2: pxor <xmm8=%xmm9,<xmm5=%xmm5
1333pxor %xmm9,%xmm5
1334
1335# qhasm: xmm3 ^= xmm8
1336# asm 1: pxor <xmm8=int6464#10,<xmm3=int6464#4
1337# asm 2: pxor <xmm8=%xmm9,<xmm3=%xmm3
1338pxor %xmm9,%xmm3
1339
1340# qhasm: xmm12 = xmm7
1341# asm 1: movdqa <xmm7=int6464#8,>xmm12=int6464#9
1342# asm 2: movdqa <xmm7=%xmm7,>xmm12=%xmm8
1343movdqa %xmm7,%xmm8
1344
1345# qhasm: xmm8 = xmm1
1346# asm 1: movdqa <xmm1=int6464#2,>xmm8=int6464#10
1347# asm 2: movdqa <xmm1=%xmm1,>xmm8=%xmm9
1348movdqa %xmm1,%xmm9
1349
1350# qhasm: xmm12 ^= xmm4
1351# asm 1: pxor <xmm4=int6464#5,<xmm12=int6464#9
1352# asm 2: pxor <xmm4=%xmm4,<xmm12=%xmm8
1353pxor %xmm4,%xmm8
1354
1355# qhasm: xmm8 ^= xmm2
1356# asm 1: pxor <xmm2=int6464#3,<xmm8=int6464#10
1357# asm 2: pxor <xmm2=%xmm2,<xmm8=%xmm9
1358pxor %xmm2,%xmm9
1359
1360# qhasm: xmm11 = xmm15
1361# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
1362# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
1363movdqa %xmm13,%xmm10
1364
1365# qhasm: xmm11 ^= xmm14
1366# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
1367# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
1368pxor %xmm11,%xmm10
1369
1370# qhasm: xmm11 &= xmm12
1371# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
1372# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
1373pand %xmm8,%xmm10
1374
1375# qhasm: xmm12 ^= xmm8
1376# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
1377# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
1378pxor %xmm9,%xmm8
1379
1380# qhasm: xmm12 &= xmm14
1381# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
1382# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
1383pand %xmm11,%xmm8
1384
1385# qhasm: xmm8 &= xmm15
1386# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
1387# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
1388pand %xmm13,%xmm9
1389
1390# qhasm: xmm8 ^= xmm12
1391# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
1392# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
1393pxor %xmm8,%xmm9
1394
1395# qhasm: xmm12 ^= xmm11
1396# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
1397# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
1398pxor %xmm10,%xmm8
1399
1400# qhasm: xmm10 = xmm13
1401# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
1402# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
1403movdqa %xmm15,%xmm10
1404
1405# qhasm: xmm10 ^= xmm9
1406# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
1407# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
1408pxor %xmm12,%xmm10
1409
1410# qhasm: xmm10 &= xmm4
1411# asm 1: pand <xmm4=int6464#5,<xmm10=int6464#11
1412# asm 2: pand <xmm4=%xmm4,<xmm10=%xmm10
1413pand %xmm4,%xmm10
1414
1415# qhasm: xmm4 ^= xmm2
1416# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
1417# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
1418pxor %xmm2,%xmm4
1419
1420# qhasm: xmm4 &= xmm9
1421# asm 1: pand <xmm9=int6464#13,<xmm4=int6464#5
1422# asm 2: pand <xmm9=%xmm12,<xmm4=%xmm4
1423pand %xmm12,%xmm4
1424
1425# qhasm: xmm2 &= xmm13
1426# asm 1: pand <xmm13=int6464#16,<xmm2=int6464#3
1427# asm 2: pand <xmm13=%xmm15,<xmm2=%xmm2
1428pand %xmm15,%xmm2
1429
1430# qhasm: xmm4 ^= xmm2
1431# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
1432# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
1433pxor %xmm2,%xmm4
1434
1435# qhasm: xmm2 ^= xmm10
1436# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#3
1437# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm2
1438pxor %xmm10,%xmm2
1439
1440# qhasm: xmm15 ^= xmm13
1441# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
1442# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
1443pxor %xmm15,%xmm13
1444
1445# qhasm: xmm14 ^= xmm9
1446# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
1447# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
1448pxor %xmm12,%xmm11
1449
1450# qhasm: xmm11 = xmm15
1451# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
1452# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
1453movdqa %xmm13,%xmm10
1454
1455# qhasm: xmm11 ^= xmm14
1456# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
1457# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
1458pxor %xmm11,%xmm10
1459
1460# qhasm: xmm11 &= xmm7
1461# asm 1: pand <xmm7=int6464#8,<xmm11=int6464#11
1462# asm 2: pand <xmm7=%xmm7,<xmm11=%xmm10
1463pand %xmm7,%xmm10
1464
1465# qhasm: xmm7 ^= xmm1
1466# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
1467# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
1468pxor %xmm1,%xmm7
1469
1470# qhasm: xmm7 &= xmm14
1471# asm 1: pand <xmm14=int6464#12,<xmm7=int6464#8
1472# asm 2: pand <xmm14=%xmm11,<xmm7=%xmm7
1473pand %xmm11,%xmm7
1474
1475# qhasm: xmm1 &= xmm15
1476# asm 1: pand <xmm15=int6464#14,<xmm1=int6464#2
1477# asm 2: pand <xmm15=%xmm13,<xmm1=%xmm1
1478pand %xmm13,%xmm1
1479
1480# qhasm: xmm7 ^= xmm1
1481# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
1482# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
1483pxor %xmm1,%xmm7
1484
1485# qhasm: xmm1 ^= xmm11
1486# asm 1: pxor <xmm11=int6464#11,<xmm1=int6464#2
1487# asm 2: pxor <xmm11=%xmm10,<xmm1=%xmm1
1488pxor %xmm10,%xmm1
1489
1490# qhasm: xmm7 ^= xmm12
1491# asm 1: pxor <xmm12=int6464#9,<xmm7=int6464#8
1492# asm 2: pxor <xmm12=%xmm8,<xmm7=%xmm7
1493pxor %xmm8,%xmm7
1494
1495# qhasm: xmm4 ^= xmm12
1496# asm 1: pxor <xmm12=int6464#9,<xmm4=int6464#5
1497# asm 2: pxor <xmm12=%xmm8,<xmm4=%xmm4
1498pxor %xmm8,%xmm4
1499
1500# qhasm: xmm1 ^= xmm8
1501# asm 1: pxor <xmm8=int6464#10,<xmm1=int6464#2
1502# asm 2: pxor <xmm8=%xmm9,<xmm1=%xmm1
1503pxor %xmm9,%xmm1
1504
1505# qhasm: xmm2 ^= xmm8
1506# asm 1: pxor <xmm8=int6464#10,<xmm2=int6464#3
1507# asm 2: pxor <xmm8=%xmm9,<xmm2=%xmm2
1508pxor %xmm9,%xmm2
1509
1510# qhasm: xmm7 ^= xmm0
1511# asm 1: pxor <xmm0=int6464#1,<xmm7=int6464#8
1512# asm 2: pxor <xmm0=%xmm0,<xmm7=%xmm7
1513pxor %xmm0,%xmm7
1514
1515# qhasm: xmm1 ^= xmm6
1516# asm 1: pxor <xmm6=int6464#7,<xmm1=int6464#2
1517# asm 2: pxor <xmm6=%xmm6,<xmm1=%xmm1
1518pxor %xmm6,%xmm1
1519
1520# qhasm: xmm4 ^= xmm7
1521# asm 1: pxor <xmm7=int6464#8,<xmm4=int6464#5
1522# asm 2: pxor <xmm7=%xmm7,<xmm4=%xmm4
1523pxor %xmm7,%xmm4
1524
1525# qhasm: xmm6 ^= xmm0
1526# asm 1: pxor <xmm0=int6464#1,<xmm6=int6464#7
1527# asm 2: pxor <xmm0=%xmm0,<xmm6=%xmm6
1528pxor %xmm0,%xmm6
1529
1530# qhasm: xmm0 ^= xmm1
1531# asm 1: pxor <xmm1=int6464#2,<xmm0=int6464#1
1532# asm 2: pxor <xmm1=%xmm1,<xmm0=%xmm0
1533pxor %xmm1,%xmm0
1534
1535# qhasm: xmm1 ^= xmm5
1536# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
1537# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
1538pxor %xmm5,%xmm1
1539
1540# qhasm: xmm5 ^= xmm2
1541# asm 1: pxor <xmm2=int6464#3,<xmm5=int6464#6
1542# asm 2: pxor <xmm2=%xmm2,<xmm5=%xmm5
1543pxor %xmm2,%xmm5
1544
1545# qhasm: xmm4 ^= xmm5
1546# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
1547# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
1548pxor %xmm5,%xmm4
1549
1550# qhasm: xmm2 ^= xmm3
1551# asm 1: pxor <xmm3=int6464#4,<xmm2=int6464#3
1552# asm 2: pxor <xmm3=%xmm3,<xmm2=%xmm2
1553pxor %xmm3,%xmm2
1554
1555# qhasm: xmm3 ^= xmm5
1556# asm 1: pxor <xmm5=int6464#6,<xmm3=int6464#4
1557# asm 2: pxor <xmm5=%xmm5,<xmm3=%xmm3
1558pxor %xmm5,%xmm3
1559
1560# qhasm: xmm6 ^= xmm3
1561# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
1562# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
1563pxor %xmm3,%xmm6
1564
1565# qhasm: xmm8 = shuffle dwords of xmm0 by 0x93
1566# asm 1: pshufd $0x93,<xmm0=int6464#1,>xmm8=int6464#9
1567# asm 2: pshufd $0x93,<xmm0=%xmm0,>xmm8=%xmm8
1568pshufd $0x93,%xmm0,%xmm8
1569
1570# qhasm: xmm9 = shuffle dwords of xmm1 by 0x93
1571# asm 1: pshufd $0x93,<xmm1=int6464#2,>xmm9=int6464#10
1572# asm 2: pshufd $0x93,<xmm1=%xmm1,>xmm9=%xmm9
1573pshufd $0x93,%xmm1,%xmm9
1574
1575# qhasm: xmm10 = shuffle dwords of xmm4 by 0x93
1576# asm 1: pshufd $0x93,<xmm4=int6464#5,>xmm10=int6464#11
1577# asm 2: pshufd $0x93,<xmm4=%xmm4,>xmm10=%xmm10
1578pshufd $0x93,%xmm4,%xmm10
1579
1580# qhasm: xmm11 = shuffle dwords of xmm6 by 0x93
1581# asm 1: pshufd $0x93,<xmm6=int6464#7,>xmm11=int6464#12
1582# asm 2: pshufd $0x93,<xmm6=%xmm6,>xmm11=%xmm11
1583pshufd $0x93,%xmm6,%xmm11
1584
1585# qhasm: xmm12 = shuffle dwords of xmm3 by 0x93
1586# asm 1: pshufd $0x93,<xmm3=int6464#4,>xmm12=int6464#13
1587# asm 2: pshufd $0x93,<xmm3=%xmm3,>xmm12=%xmm12
1588pshufd $0x93,%xmm3,%xmm12
1589
1590# qhasm: xmm13 = shuffle dwords of xmm7 by 0x93
1591# asm 1: pshufd $0x93,<xmm7=int6464#8,>xmm13=int6464#14
1592# asm 2: pshufd $0x93,<xmm7=%xmm7,>xmm13=%xmm13
1593pshufd $0x93,%xmm7,%xmm13
1594
1595# qhasm: xmm14 = shuffle dwords of xmm2 by 0x93
1596# asm 1: pshufd $0x93,<xmm2=int6464#3,>xmm14=int6464#15
1597# asm 2: pshufd $0x93,<xmm2=%xmm2,>xmm14=%xmm14
1598pshufd $0x93,%xmm2,%xmm14
1599
1600# qhasm: xmm15 = shuffle dwords of xmm5 by 0x93
1601# asm 1: pshufd $0x93,<xmm5=int6464#6,>xmm15=int6464#16
1602# asm 2: pshufd $0x93,<xmm5=%xmm5,>xmm15=%xmm15
1603pshufd $0x93,%xmm5,%xmm15
1604
1605# qhasm: xmm0 ^= xmm8
1606# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
1607# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
1608pxor %xmm8,%xmm0
1609
1610# qhasm: xmm1 ^= xmm9
1611# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
1612# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
1613pxor %xmm9,%xmm1
1614
1615# qhasm: xmm4 ^= xmm10
1616# asm 1: pxor <xmm10=int6464#11,<xmm4=int6464#5
1617# asm 2: pxor <xmm10=%xmm10,<xmm4=%xmm4
1618pxor %xmm10,%xmm4
1619
1620# qhasm: xmm6 ^= xmm11
1621# asm 1: pxor <xmm11=int6464#12,<xmm6=int6464#7
1622# asm 2: pxor <xmm11=%xmm11,<xmm6=%xmm6
1623pxor %xmm11,%xmm6
1624
1625# qhasm: xmm3 ^= xmm12
1626# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
1627# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
1628pxor %xmm12,%xmm3
1629
1630# qhasm: xmm7 ^= xmm13
1631# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
1632# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
1633pxor %xmm13,%xmm7
1634
1635# qhasm: xmm2 ^= xmm14
1636# asm 1: pxor <xmm14=int6464#15,<xmm2=int6464#3
1637# asm 2: pxor <xmm14=%xmm14,<xmm2=%xmm2
1638pxor %xmm14,%xmm2
1639
1640# qhasm: xmm5 ^= xmm15
1641# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
1642# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
1643pxor %xmm15,%xmm5
1644
1645# qhasm: xmm8 ^= xmm5
1646# asm 1: pxor <xmm5=int6464#6,<xmm8=int6464#9
1647# asm 2: pxor <xmm5=%xmm5,<xmm8=%xmm8
1648pxor %xmm5,%xmm8
1649
1650# qhasm: xmm9 ^= xmm0
1651# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
1652# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
1653pxor %xmm0,%xmm9
1654
1655# qhasm: xmm10 ^= xmm1
1656# asm 1: pxor <xmm1=int6464#2,<xmm10=int6464#11
1657# asm 2: pxor <xmm1=%xmm1,<xmm10=%xmm10
1658pxor %xmm1,%xmm10
1659
1660# qhasm: xmm9 ^= xmm5
1661# asm 1: pxor <xmm5=int6464#6,<xmm9=int6464#10
1662# asm 2: pxor <xmm5=%xmm5,<xmm9=%xmm9
1663pxor %xmm5,%xmm9
1664
1665# qhasm: xmm11 ^= xmm4
1666# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
1667# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
1668pxor %xmm4,%xmm11
1669
1670# qhasm: xmm12 ^= xmm6
1671# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#13
1672# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm12
1673pxor %xmm6,%xmm12
1674
1675# qhasm: xmm13 ^= xmm3
1676# asm 1: pxor <xmm3=int6464#4,<xmm13=int6464#14
1677# asm 2: pxor <xmm3=%xmm3,<xmm13=%xmm13
1678pxor %xmm3,%xmm13
1679
1680# qhasm: xmm11 ^= xmm5
1681# asm 1: pxor <xmm5=int6464#6,<xmm11=int6464#12
1682# asm 2: pxor <xmm5=%xmm5,<xmm11=%xmm11
1683pxor %xmm5,%xmm11
1684
1685# qhasm: xmm14 ^= xmm7
1686# asm 1: pxor <xmm7=int6464#8,<xmm14=int6464#15
1687# asm 2: pxor <xmm7=%xmm7,<xmm14=%xmm14
1688pxor %xmm7,%xmm14
1689
1690# qhasm: xmm15 ^= xmm2
1691# asm 1: pxor <xmm2=int6464#3,<xmm15=int6464#16
1692# asm 2: pxor <xmm2=%xmm2,<xmm15=%xmm15
1693pxor %xmm2,%xmm15
1694
1695# qhasm: xmm12 ^= xmm5
1696# asm 1: pxor <xmm5=int6464#6,<xmm12=int6464#13
1697# asm 2: pxor <xmm5=%xmm5,<xmm12=%xmm12
1698pxor %xmm5,%xmm12
1699
1700# qhasm: xmm0 = shuffle dwords of xmm0 by 0x4E
1701# asm 1: pshufd $0x4E,<xmm0=int6464#1,>xmm0=int6464#1
1702# asm 2: pshufd $0x4E,<xmm0=%xmm0,>xmm0=%xmm0
1703pshufd $0x4E,%xmm0,%xmm0
1704
1705# qhasm: xmm1 = shuffle dwords of xmm1 by 0x4E
1706# asm 1: pshufd $0x4E,<xmm1=int6464#2,>xmm1=int6464#2
1707# asm 2: pshufd $0x4E,<xmm1=%xmm1,>xmm1=%xmm1
1708pshufd $0x4E,%xmm1,%xmm1
1709
1710# qhasm: xmm4 = shuffle dwords of xmm4 by 0x4E
1711# asm 1: pshufd $0x4E,<xmm4=int6464#5,>xmm4=int6464#5
1712# asm 2: pshufd $0x4E,<xmm4=%xmm4,>xmm4=%xmm4
1713pshufd $0x4E,%xmm4,%xmm4
1714
1715# qhasm: xmm6 = shuffle dwords of xmm6 by 0x4E
1716# asm 1: pshufd $0x4E,<xmm6=int6464#7,>xmm6=int6464#7
1717# asm 2: pshufd $0x4E,<xmm6=%xmm6,>xmm6=%xmm6
1718pshufd $0x4E,%xmm6,%xmm6
1719
1720# qhasm: xmm3 = shuffle dwords of xmm3 by 0x4E
1721# asm 1: pshufd $0x4E,<xmm3=int6464#4,>xmm3=int6464#4
1722# asm 2: pshufd $0x4E,<xmm3=%xmm3,>xmm3=%xmm3
1723pshufd $0x4E,%xmm3,%xmm3
1724
1725# qhasm: xmm7 = shuffle dwords of xmm7 by 0x4E
1726# asm 1: pshufd $0x4E,<xmm7=int6464#8,>xmm7=int6464#8
1727# asm 2: pshufd $0x4E,<xmm7=%xmm7,>xmm7=%xmm7
1728pshufd $0x4E,%xmm7,%xmm7
1729
1730# qhasm: xmm2 = shuffle dwords of xmm2 by 0x4E
1731# asm 1: pshufd $0x4E,<xmm2=int6464#3,>xmm2=int6464#3
1732# asm 2: pshufd $0x4E,<xmm2=%xmm2,>xmm2=%xmm2
1733pshufd $0x4E,%xmm2,%xmm2
1734
1735# qhasm: xmm5 = shuffle dwords of xmm5 by 0x4E
1736# asm 1: pshufd $0x4E,<xmm5=int6464#6,>xmm5=int6464#6
1737# asm 2: pshufd $0x4E,<xmm5=%xmm5,>xmm5=%xmm5
1738pshufd $0x4E,%xmm5,%xmm5
1739
1740# qhasm: xmm8 ^= xmm0
1741# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
1742# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
1743pxor %xmm0,%xmm8
1744
1745# qhasm: xmm9 ^= xmm1
1746# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
1747# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
1748pxor %xmm1,%xmm9
1749
1750# qhasm: xmm10 ^= xmm4
1751# asm 1: pxor <xmm4=int6464#5,<xmm10=int6464#11
1752# asm 2: pxor <xmm4=%xmm4,<xmm10=%xmm10
1753pxor %xmm4,%xmm10
1754
1755# qhasm: xmm11 ^= xmm6
1756# asm 1: pxor <xmm6=int6464#7,<xmm11=int6464#12
1757# asm 2: pxor <xmm6=%xmm6,<xmm11=%xmm11
1758pxor %xmm6,%xmm11
1759
1760# qhasm: xmm12 ^= xmm3
1761# asm 1: pxor <xmm3=int6464#4,<xmm12=int6464#13
1762# asm 2: pxor <xmm3=%xmm3,<xmm12=%xmm12
1763pxor %xmm3,%xmm12
1764
1765# qhasm: xmm13 ^= xmm7
1766# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
1767# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
1768pxor %xmm7,%xmm13
1769
1770# qhasm: xmm14 ^= xmm2
1771# asm 1: pxor <xmm2=int6464#3,<xmm14=int6464#15
1772# asm 2: pxor <xmm2=%xmm2,<xmm14=%xmm14
1773pxor %xmm2,%xmm14
1774
1775# qhasm: xmm15 ^= xmm5
1776# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
1777# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
1778pxor %xmm5,%xmm15
1779
1780# qhasm: xmm8 ^= *(int128 *)(c + 128)
1781# asm 1: pxor 128(<c=int64#4),<xmm8=int6464#9
1782# asm 2: pxor 128(<c=%rcx),<xmm8=%xmm8
1783pxor 128(%rcx),%xmm8
1784
1785# qhasm: shuffle bytes of xmm8 by SR
1786# asm 1: pshufb SR,<xmm8=int6464#9
1787# asm 2: pshufb SR,<xmm8=%xmm8
1788pshufb SR,%xmm8
1789
1790# qhasm: xmm9 ^= *(int128 *)(c + 144)
1791# asm 1: pxor 144(<c=int64#4),<xmm9=int6464#10
1792# asm 2: pxor 144(<c=%rcx),<xmm9=%xmm9
1793pxor 144(%rcx),%xmm9
1794
1795# qhasm: shuffle bytes of xmm9 by SR
1796# asm 1: pshufb SR,<xmm9=int6464#10
1797# asm 2: pshufb SR,<xmm9=%xmm9
1798pshufb SR,%xmm9
1799
1800# qhasm: xmm10 ^= *(int128 *)(c + 160)
1801# asm 1: pxor 160(<c=int64#4),<xmm10=int6464#11
1802# asm 2: pxor 160(<c=%rcx),<xmm10=%xmm10
1803pxor 160(%rcx),%xmm10
1804
1805# qhasm: shuffle bytes of xmm10 by SR
1806# asm 1: pshufb SR,<xmm10=int6464#11
1807# asm 2: pshufb SR,<xmm10=%xmm10
1808pshufb SR,%xmm10
1809
1810# qhasm: xmm11 ^= *(int128 *)(c + 176)
1811# asm 1: pxor 176(<c=int64#4),<xmm11=int6464#12
1812# asm 2: pxor 176(<c=%rcx),<xmm11=%xmm11
1813pxor 176(%rcx),%xmm11
1814
1815# qhasm: shuffle bytes of xmm11 by SR
1816# asm 1: pshufb SR,<xmm11=int6464#12
1817# asm 2: pshufb SR,<xmm11=%xmm11
1818pshufb SR,%xmm11
1819
1820# qhasm: xmm12 ^= *(int128 *)(c + 192)
1821# asm 1: pxor 192(<c=int64#4),<xmm12=int6464#13
1822# asm 2: pxor 192(<c=%rcx),<xmm12=%xmm12
1823pxor 192(%rcx),%xmm12
1824
1825# qhasm: shuffle bytes of xmm12 by SR
1826# asm 1: pshufb SR,<xmm12=int6464#13
1827# asm 2: pshufb SR,<xmm12=%xmm12
1828pshufb SR,%xmm12
1829
1830# qhasm: xmm13 ^= *(int128 *)(c + 208)
1831# asm 1: pxor 208(<c=int64#4),<xmm13=int6464#14
1832# asm 2: pxor 208(<c=%rcx),<xmm13=%xmm13
1833pxor 208(%rcx),%xmm13
1834
1835# qhasm: shuffle bytes of xmm13 by SR
1836# asm 1: pshufb SR,<xmm13=int6464#14
1837# asm 2: pshufb SR,<xmm13=%xmm13
1838pshufb SR,%xmm13
1839
1840# qhasm: xmm14 ^= *(int128 *)(c + 224)
1841# asm 1: pxor 224(<c=int64#4),<xmm14=int6464#15
1842# asm 2: pxor 224(<c=%rcx),<xmm14=%xmm14
1843pxor 224(%rcx),%xmm14
1844
1845# qhasm: shuffle bytes of xmm14 by SR
1846# asm 1: pshufb SR,<xmm14=int6464#15
1847# asm 2: pshufb SR,<xmm14=%xmm14
1848pshufb SR,%xmm14
1849
1850# qhasm: xmm15 ^= *(int128 *)(c + 240)
1851# asm 1: pxor 240(<c=int64#4),<xmm15=int6464#16
1852# asm 2: pxor 240(<c=%rcx),<xmm15=%xmm15
1853pxor 240(%rcx),%xmm15
1854
1855# qhasm: shuffle bytes of xmm15 by SR
1856# asm 1: pshufb SR,<xmm15=int6464#16
1857# asm 2: pshufb SR,<xmm15=%xmm15
1858pshufb SR,%xmm15
1859
1860# qhasm: xmm13 ^= xmm14
1861# asm 1: pxor <xmm14=int6464#15,<xmm13=int6464#14
1862# asm 2: pxor <xmm14=%xmm14,<xmm13=%xmm13
1863pxor %xmm14,%xmm13
1864
1865# qhasm: xmm10 ^= xmm9
1866# asm 1: pxor <xmm9=int6464#10,<xmm10=int6464#11
1867# asm 2: pxor <xmm9=%xmm9,<xmm10=%xmm10
1868pxor %xmm9,%xmm10
1869
1870# qhasm: xmm13 ^= xmm8
1871# asm 1: pxor <xmm8=int6464#9,<xmm13=int6464#14
1872# asm 2: pxor <xmm8=%xmm8,<xmm13=%xmm13
1873pxor %xmm8,%xmm13
1874
1875# qhasm: xmm14 ^= xmm10
1876# asm 1: pxor <xmm10=int6464#11,<xmm14=int6464#15
1877# asm 2: pxor <xmm10=%xmm10,<xmm14=%xmm14
1878pxor %xmm10,%xmm14
1879
1880# qhasm: xmm11 ^= xmm8
1881# asm 1: pxor <xmm8=int6464#9,<xmm11=int6464#12
1882# asm 2: pxor <xmm8=%xmm8,<xmm11=%xmm11
1883pxor %xmm8,%xmm11
1884
1885# qhasm: xmm14 ^= xmm11
1886# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
1887# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
1888pxor %xmm11,%xmm14
1889
1890# qhasm: xmm11 ^= xmm15
1891# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#12
1892# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm11
1893pxor %xmm15,%xmm11
1894
1895# qhasm: xmm11 ^= xmm12
1896# asm 1: pxor <xmm12=int6464#13,<xmm11=int6464#12
1897# asm 2: pxor <xmm12=%xmm12,<xmm11=%xmm11
1898pxor %xmm12,%xmm11
1899
1900# qhasm: xmm15 ^= xmm13
1901# asm 1: pxor <xmm13=int6464#14,<xmm15=int6464#16
1902# asm 2: pxor <xmm13=%xmm13,<xmm15=%xmm15
1903pxor %xmm13,%xmm15
1904
1905# qhasm: xmm11 ^= xmm9
1906# asm 1: pxor <xmm9=int6464#10,<xmm11=int6464#12
1907# asm 2: pxor <xmm9=%xmm9,<xmm11=%xmm11
1908pxor %xmm9,%xmm11
1909
1910# qhasm: xmm12 ^= xmm13
1911# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
1912# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
1913pxor %xmm13,%xmm12
1914
1915# qhasm: xmm10 ^= xmm15
1916# asm 1: pxor <xmm15=int6464#16,<xmm10=int6464#11
1917# asm 2: pxor <xmm15=%xmm15,<xmm10=%xmm10
1918pxor %xmm15,%xmm10
1919
1920# qhasm: xmm9 ^= xmm13
1921# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
1922# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
1923pxor %xmm13,%xmm9
1924
1925# qhasm: xmm3 = xmm15
1926# asm 1: movdqa <xmm15=int6464#16,>xmm3=int6464#1
1927# asm 2: movdqa <xmm15=%xmm15,>xmm3=%xmm0
1928movdqa %xmm15,%xmm0
1929
1930# qhasm: xmm2 = xmm9
1931# asm 1: movdqa <xmm9=int6464#10,>xmm2=int6464#2
1932# asm 2: movdqa <xmm9=%xmm9,>xmm2=%xmm1
1933movdqa %xmm9,%xmm1
1934
1935# qhasm: xmm1 = xmm13
1936# asm 1: movdqa <xmm13=int6464#14,>xmm1=int6464#3
1937# asm 2: movdqa <xmm13=%xmm13,>xmm1=%xmm2
1938movdqa %xmm13,%xmm2
1939
1940# qhasm: xmm5 = xmm10
1941# asm 1: movdqa <xmm10=int6464#11,>xmm5=int6464#4
1942# asm 2: movdqa <xmm10=%xmm10,>xmm5=%xmm3
1943movdqa %xmm10,%xmm3
1944
1945# qhasm: xmm4 = xmm14
1946# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#5
1947# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm4
1948movdqa %xmm14,%xmm4
1949
1950# qhasm: xmm3 ^= xmm12
1951# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#1
1952# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm0
1953pxor %xmm12,%xmm0
1954
1955# qhasm: xmm2 ^= xmm10
1956# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#2
1957# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm1
1958pxor %xmm10,%xmm1
1959
1960# qhasm: xmm1 ^= xmm11
1961# asm 1: pxor <xmm11=int6464#12,<xmm1=int6464#3
1962# asm 2: pxor <xmm11=%xmm11,<xmm1=%xmm2
1963pxor %xmm11,%xmm2
1964
1965# qhasm: xmm5 ^= xmm12
1966# asm 1: pxor <xmm12=int6464#13,<xmm5=int6464#4
1967# asm 2: pxor <xmm12=%xmm12,<xmm5=%xmm3
1968pxor %xmm12,%xmm3
1969
1970# qhasm: xmm4 ^= xmm8
1971# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
1972# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
1973pxor %xmm8,%xmm4
1974
1975# qhasm: xmm6 = xmm3
1976# asm 1: movdqa <xmm3=int6464#1,>xmm6=int6464#6
1977# asm 2: movdqa <xmm3=%xmm0,>xmm6=%xmm5
1978movdqa %xmm0,%xmm5
1979
1980# qhasm: xmm0 = xmm2
1981# asm 1: movdqa <xmm2=int6464#2,>xmm0=int6464#7
1982# asm 2: movdqa <xmm2=%xmm1,>xmm0=%xmm6
1983movdqa %xmm1,%xmm6
1984
1985# qhasm: xmm7 = xmm3
1986# asm 1: movdqa <xmm3=int6464#1,>xmm7=int6464#8
1987# asm 2: movdqa <xmm3=%xmm0,>xmm7=%xmm7
1988movdqa %xmm0,%xmm7
1989
1990# qhasm: xmm2 |= xmm1
1991# asm 1: por <xmm1=int6464#3,<xmm2=int6464#2
1992# asm 2: por <xmm1=%xmm2,<xmm2=%xmm1
1993por %xmm2,%xmm1
1994
1995# qhasm: xmm3 |= xmm4
1996# asm 1: por <xmm4=int6464#5,<xmm3=int6464#1
1997# asm 2: por <xmm4=%xmm4,<xmm3=%xmm0
1998por %xmm4,%xmm0
1999
2000# qhasm: xmm7 ^= xmm0
2001# asm 1: pxor <xmm0=int6464#7,<xmm7=int6464#8
2002# asm 2: pxor <xmm0=%xmm6,<xmm7=%xmm7
2003pxor %xmm6,%xmm7
2004
2005# qhasm: xmm6 &= xmm4
2006# asm 1: pand <xmm4=int6464#5,<xmm6=int6464#6
2007# asm 2: pand <xmm4=%xmm4,<xmm6=%xmm5
2008pand %xmm4,%xmm5
2009
2010# qhasm: xmm0 &= xmm1
2011# asm 1: pand <xmm1=int6464#3,<xmm0=int6464#7
2012# asm 2: pand <xmm1=%xmm2,<xmm0=%xmm6
2013pand %xmm2,%xmm6
2014
2015# qhasm: xmm4 ^= xmm1
2016# asm 1: pxor <xmm1=int6464#3,<xmm4=int6464#5
2017# asm 2: pxor <xmm1=%xmm2,<xmm4=%xmm4
2018pxor %xmm2,%xmm4
2019
2020# qhasm: xmm7 &= xmm4
2021# asm 1: pand <xmm4=int6464#5,<xmm7=int6464#8
2022# asm 2: pand <xmm4=%xmm4,<xmm7=%xmm7
2023pand %xmm4,%xmm7
2024
2025# qhasm: xmm4 = xmm11
2026# asm 1: movdqa <xmm11=int6464#12,>xmm4=int6464#3
2027# asm 2: movdqa <xmm11=%xmm11,>xmm4=%xmm2
2028movdqa %xmm11,%xmm2
2029
2030# qhasm: xmm4 ^= xmm8
2031# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#3
2032# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm2
2033pxor %xmm8,%xmm2
2034
2035# qhasm: xmm5 &= xmm4
2036# asm 1: pand <xmm4=int6464#3,<xmm5=int6464#4
2037# asm 2: pand <xmm4=%xmm2,<xmm5=%xmm3
2038pand %xmm2,%xmm3
2039
2040# qhasm: xmm3 ^= xmm5
2041# asm 1: pxor <xmm5=int6464#4,<xmm3=int6464#1
2042# asm 2: pxor <xmm5=%xmm3,<xmm3=%xmm0
2043pxor %xmm3,%xmm0
2044
2045# qhasm: xmm2 ^= xmm5
2046# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
2047# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
2048pxor %xmm3,%xmm1
2049
2050# qhasm: xmm5 = xmm15
2051# asm 1: movdqa <xmm15=int6464#16,>xmm5=int6464#3
2052# asm 2: movdqa <xmm15=%xmm15,>xmm5=%xmm2
2053movdqa %xmm15,%xmm2
2054
2055# qhasm: xmm5 ^= xmm9
2056# asm 1: pxor <xmm9=int6464#10,<xmm5=int6464#3
2057# asm 2: pxor <xmm9=%xmm9,<xmm5=%xmm2
2058pxor %xmm9,%xmm2
2059
2060# qhasm: xmm4 = xmm13
2061# asm 1: movdqa <xmm13=int6464#14,>xmm4=int6464#4
2062# asm 2: movdqa <xmm13=%xmm13,>xmm4=%xmm3
2063movdqa %xmm13,%xmm3
2064
2065# qhasm: xmm1 = xmm5
2066# asm 1: movdqa <xmm5=int6464#3,>xmm1=int6464#5
2067# asm 2: movdqa <xmm5=%xmm2,>xmm1=%xmm4
2068movdqa %xmm2,%xmm4
2069
2070# qhasm: xmm4 ^= xmm14
2071# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#4
2072# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm3
2073pxor %xmm14,%xmm3
2074
2075# qhasm: xmm1 |= xmm4
2076# asm 1: por <xmm4=int6464#4,<xmm1=int6464#5
2077# asm 2: por <xmm4=%xmm3,<xmm1=%xmm4
2078por %xmm3,%xmm4
2079
2080# qhasm: xmm5 &= xmm4
2081# asm 1: pand <xmm4=int6464#4,<xmm5=int6464#3
2082# asm 2: pand <xmm4=%xmm3,<xmm5=%xmm2
2083pand %xmm3,%xmm2
2084
2085# qhasm: xmm0 ^= xmm5
2086# asm 1: pxor <xmm5=int6464#3,<xmm0=int6464#7
2087# asm 2: pxor <xmm5=%xmm2,<xmm0=%xmm6
2088pxor %xmm2,%xmm6
2089
2090# qhasm: xmm3 ^= xmm7
2091# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#1
2092# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm0
2093pxor %xmm7,%xmm0
2094
2095# qhasm: xmm2 ^= xmm6
2096# asm 1: pxor <xmm6=int6464#6,<xmm2=int6464#2
2097# asm 2: pxor <xmm6=%xmm5,<xmm2=%xmm1
2098pxor %xmm5,%xmm1
2099
2100# qhasm: xmm1 ^= xmm7
2101# asm 1: pxor <xmm7=int6464#8,<xmm1=int6464#5
2102# asm 2: pxor <xmm7=%xmm7,<xmm1=%xmm4
2103pxor %xmm7,%xmm4
2104
2105# qhasm: xmm0 ^= xmm6
2106# asm 1: pxor <xmm6=int6464#6,<xmm0=int6464#7
2107# asm 2: pxor <xmm6=%xmm5,<xmm0=%xmm6
2108pxor %xmm5,%xmm6
2109
2110# qhasm: xmm1 ^= xmm6
2111# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
2112# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
2113pxor %xmm5,%xmm4
2114
2115# qhasm: xmm4 = xmm10
2116# asm 1: movdqa <xmm10=int6464#11,>xmm4=int6464#3
2117# asm 2: movdqa <xmm10=%xmm10,>xmm4=%xmm2
2118movdqa %xmm10,%xmm2
2119
2120# qhasm: xmm5 = xmm12
2121# asm 1: movdqa <xmm12=int6464#13,>xmm5=int6464#4
2122# asm 2: movdqa <xmm12=%xmm12,>xmm5=%xmm3
2123movdqa %xmm12,%xmm3
2124
2125# qhasm: xmm6 = xmm9
2126# asm 1: movdqa <xmm9=int6464#10,>xmm6=int6464#6
2127# asm 2: movdqa <xmm9=%xmm9,>xmm6=%xmm5
2128movdqa %xmm9,%xmm5
2129
2130# qhasm: xmm7 = xmm15
2131# asm 1: movdqa <xmm15=int6464#16,>xmm7=int6464#8
2132# asm 2: movdqa <xmm15=%xmm15,>xmm7=%xmm7
2133movdqa %xmm15,%xmm7
2134
2135# qhasm: xmm4 &= xmm11
2136# asm 1: pand <xmm11=int6464#12,<xmm4=int6464#3
2137# asm 2: pand <xmm11=%xmm11,<xmm4=%xmm2
2138pand %xmm11,%xmm2
2139
2140# qhasm: xmm5 &= xmm8
2141# asm 1: pand <xmm8=int6464#9,<xmm5=int6464#4
2142# asm 2: pand <xmm8=%xmm8,<xmm5=%xmm3
2143pand %xmm8,%xmm3
2144
2145# qhasm: xmm6 &= xmm13
2146# asm 1: pand <xmm13=int6464#14,<xmm6=int6464#6
2147# asm 2: pand <xmm13=%xmm13,<xmm6=%xmm5
2148pand %xmm13,%xmm5
2149
2150# qhasm: xmm7 |= xmm14
2151# asm 1: por <xmm14=int6464#15,<xmm7=int6464#8
2152# asm 2: por <xmm14=%xmm14,<xmm7=%xmm7
2153por %xmm14,%xmm7
2154
2155# qhasm: xmm3 ^= xmm4
2156# asm 1: pxor <xmm4=int6464#3,<xmm3=int6464#1
2157# asm 2: pxor <xmm4=%xmm2,<xmm3=%xmm0
2158pxor %xmm2,%xmm0
2159
2160# qhasm: xmm2 ^= xmm5
2161# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
2162# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
2163pxor %xmm3,%xmm1
2164
2165# qhasm: xmm1 ^= xmm6
2166# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
2167# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
2168pxor %xmm5,%xmm4
2169
2170# qhasm: xmm0 ^= xmm7
2171# asm 1: pxor <xmm7=int6464#8,<xmm0=int6464#7
2172# asm 2: pxor <xmm7=%xmm7,<xmm0=%xmm6
2173pxor %xmm7,%xmm6
2174
2175# qhasm: xmm4 = xmm3
2176# asm 1: movdqa <xmm3=int6464#1,>xmm4=int6464#3
2177# asm 2: movdqa <xmm3=%xmm0,>xmm4=%xmm2
2178movdqa %xmm0,%xmm2
2179
2180# qhasm: xmm4 ^= xmm2
2181# asm 1: pxor <xmm2=int6464#2,<xmm4=int6464#3
2182# asm 2: pxor <xmm2=%xmm1,<xmm4=%xmm2
2183pxor %xmm1,%xmm2
2184
2185# qhasm: xmm3 &= xmm1
2186# asm 1: pand <xmm1=int6464#5,<xmm3=int6464#1
2187# asm 2: pand <xmm1=%xmm4,<xmm3=%xmm0
2188pand %xmm4,%xmm0
2189
2190# qhasm: xmm6 = xmm0
2191# asm 1: movdqa <xmm0=int6464#7,>xmm6=int6464#4
2192# asm 2: movdqa <xmm0=%xmm6,>xmm6=%xmm3
2193movdqa %xmm6,%xmm3
2194
2195# qhasm: xmm6 ^= xmm3
2196# asm 1: pxor <xmm3=int6464#1,<xmm6=int6464#4
2197# asm 2: pxor <xmm3=%xmm0,<xmm6=%xmm3
2198pxor %xmm0,%xmm3
2199
2200# qhasm: xmm7 = xmm4
2201# asm 1: movdqa <xmm4=int6464#3,>xmm7=int6464#6
2202# asm 2: movdqa <xmm4=%xmm2,>xmm7=%xmm5
2203movdqa %xmm2,%xmm5
2204
2205# qhasm: xmm7 &= xmm6
2206# asm 1: pand <xmm6=int6464#4,<xmm7=int6464#6
2207# asm 2: pand <xmm6=%xmm3,<xmm7=%xmm5
2208pand %xmm3,%xmm5
2209
2210# qhasm: xmm7 ^= xmm2
2211# asm 1: pxor <xmm2=int6464#2,<xmm7=int6464#6
2212# asm 2: pxor <xmm2=%xmm1,<xmm7=%xmm5
2213pxor %xmm1,%xmm5
2214
2215# qhasm: xmm5 = xmm1
2216# asm 1: movdqa <xmm1=int6464#5,>xmm5=int6464#8
2217# asm 2: movdqa <xmm1=%xmm4,>xmm5=%xmm7
2218movdqa %xmm4,%xmm7
2219
2220# qhasm: xmm5 ^= xmm0
2221# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
2222# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
2223pxor %xmm6,%xmm7
2224
2225# qhasm: xmm3 ^= xmm2
2226# asm 1: pxor <xmm2=int6464#2,<xmm3=int6464#1
2227# asm 2: pxor <xmm2=%xmm1,<xmm3=%xmm0
2228pxor %xmm1,%xmm0
2229
2230# qhasm: xmm5 &= xmm3
2231# asm 1: pand <xmm3=int6464#1,<xmm5=int6464#8
2232# asm 2: pand <xmm3=%xmm0,<xmm5=%xmm7
2233pand %xmm0,%xmm7
2234
2235# qhasm: xmm5 ^= xmm0
2236# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
2237# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
2238pxor %xmm6,%xmm7
2239
2240# qhasm: xmm1 ^= xmm5
2241# asm 1: pxor <xmm5=int6464#8,<xmm1=int6464#5
2242# asm 2: pxor <xmm5=%xmm7,<xmm1=%xmm4
2243pxor %xmm7,%xmm4
2244
2245# qhasm: xmm2 = xmm6
2246# asm 1: movdqa <xmm6=int6464#4,>xmm2=int6464#1
2247# asm 2: movdqa <xmm6=%xmm3,>xmm2=%xmm0
2248movdqa %xmm3,%xmm0
2249
2250# qhasm: xmm2 ^= xmm5
2251# asm 1: pxor <xmm5=int6464#8,<xmm2=int6464#1
2252# asm 2: pxor <xmm5=%xmm7,<xmm2=%xmm0
2253pxor %xmm7,%xmm0
2254
2255# qhasm: xmm2 &= xmm0
2256# asm 1: pand <xmm0=int6464#7,<xmm2=int6464#1
2257# asm 2: pand <xmm0=%xmm6,<xmm2=%xmm0
2258pand %xmm6,%xmm0
2259
2260# qhasm: xmm1 ^= xmm2
2261# asm 1: pxor <xmm2=int6464#1,<xmm1=int6464#5
2262# asm 2: pxor <xmm2=%xmm0,<xmm1=%xmm4
2263pxor %xmm0,%xmm4
2264
2265# qhasm: xmm6 ^= xmm2
2266# asm 1: pxor <xmm2=int6464#1,<xmm6=int6464#4
2267# asm 2: pxor <xmm2=%xmm0,<xmm6=%xmm3
2268pxor %xmm0,%xmm3
2269
2270# qhasm: xmm6 &= xmm7
2271# asm 1: pand <xmm7=int6464#6,<xmm6=int6464#4
2272# asm 2: pand <xmm7=%xmm5,<xmm6=%xmm3
2273pand %xmm5,%xmm3
2274
2275# qhasm: xmm6 ^= xmm4
2276# asm 1: pxor <xmm4=int6464#3,<xmm6=int6464#4
2277# asm 2: pxor <xmm4=%xmm2,<xmm6=%xmm3
2278pxor %xmm2,%xmm3
2279
2280# qhasm: xmm4 = xmm14
2281# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#1
2282# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm0
2283movdqa %xmm14,%xmm0
2284
2285# qhasm: xmm0 = xmm13
2286# asm 1: movdqa <xmm13=int6464#14,>xmm0=int6464#2
2287# asm 2: movdqa <xmm13=%xmm13,>xmm0=%xmm1
2288movdqa %xmm13,%xmm1
2289
2290# qhasm: xmm2 = xmm7
2291# asm 1: movdqa <xmm7=int6464#6,>xmm2=int6464#3
2292# asm 2: movdqa <xmm7=%xmm5,>xmm2=%xmm2
2293movdqa %xmm5,%xmm2
2294
2295# qhasm: xmm2 ^= xmm6
2296# asm 1: pxor <xmm6=int6464#4,<xmm2=int6464#3
2297# asm 2: pxor <xmm6=%xmm3,<xmm2=%xmm2
2298pxor %xmm3,%xmm2
2299
2300# qhasm: xmm2 &= xmm14
2301# asm 1: pand <xmm14=int6464#15,<xmm2=int6464#3
2302# asm 2: pand <xmm14=%xmm14,<xmm2=%xmm2
2303pand %xmm14,%xmm2
2304
2305# qhasm: xmm14 ^= xmm13
2306# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
2307# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
2308pxor %xmm13,%xmm14
2309
2310# qhasm: xmm14 &= xmm6
2311# asm 1: pand <xmm6=int6464#4,<xmm14=int6464#15
2312# asm 2: pand <xmm6=%xmm3,<xmm14=%xmm14
2313pand %xmm3,%xmm14
2314
2315# qhasm: xmm13 &= xmm7
2316# asm 1: pand <xmm7=int6464#6,<xmm13=int6464#14
2317# asm 2: pand <xmm7=%xmm5,<xmm13=%xmm13
2318pand %xmm5,%xmm13
2319
2320# qhasm: xmm14 ^= xmm13
2321# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
2322# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
2323pxor %xmm13,%xmm14
2324
2325# qhasm: xmm13 ^= xmm2
2326# asm 1: pxor <xmm2=int6464#3,<xmm13=int6464#14
2327# asm 2: pxor <xmm2=%xmm2,<xmm13=%xmm13
2328pxor %xmm2,%xmm13
2329
2330# qhasm: xmm4 ^= xmm8
2331# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#1
2332# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm0
2333pxor %xmm8,%xmm0
2334
2335# qhasm: xmm0 ^= xmm11
2336# asm 1: pxor <xmm11=int6464#12,<xmm0=int6464#2
2337# asm 2: pxor <xmm11=%xmm11,<xmm0=%xmm1
2338pxor %xmm11,%xmm1
2339
2340# qhasm: xmm7 ^= xmm5
2341# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
2342# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
2343pxor %xmm7,%xmm5
2344
2345# qhasm: xmm6 ^= xmm1
2346# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
2347# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
2348pxor %xmm4,%xmm3
2349
2350# qhasm: xmm3 = xmm7
2351# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
2352# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
2353movdqa %xmm5,%xmm2
2354
2355# qhasm: xmm3 ^= xmm6
2356# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
2357# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
2358pxor %xmm3,%xmm2
2359
2360# qhasm: xmm3 &= xmm4
2361# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
2362# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
2363pand %xmm0,%xmm2
2364
2365# qhasm: xmm4 ^= xmm0
2366# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
2367# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
2368pxor %xmm1,%xmm0
2369
2370# qhasm: xmm4 &= xmm6
2371# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
2372# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
2373pand %xmm3,%xmm0
2374
2375# qhasm: xmm0 &= xmm7
2376# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
2377# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
2378pand %xmm5,%xmm1
2379
2380# qhasm: xmm0 ^= xmm4
2381# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
2382# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
2383pxor %xmm0,%xmm1
2384
2385# qhasm: xmm4 ^= xmm3
2386# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
2387# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
2388pxor %xmm2,%xmm0
2389
2390# qhasm: xmm2 = xmm5
2391# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
2392# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
2393movdqa %xmm7,%xmm2
2394
2395# qhasm: xmm2 ^= xmm1
2396# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
2397# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
2398pxor %xmm4,%xmm2
2399
2400# qhasm: xmm2 &= xmm8
2401# asm 1: pand <xmm8=int6464#9,<xmm2=int6464#3
2402# asm 2: pand <xmm8=%xmm8,<xmm2=%xmm2
2403pand %xmm8,%xmm2
2404
2405# qhasm: xmm8 ^= xmm11
2406# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
2407# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
2408pxor %xmm11,%xmm8
2409
2410# qhasm: xmm8 &= xmm1
2411# asm 1: pand <xmm1=int6464#5,<xmm8=int6464#9
2412# asm 2: pand <xmm1=%xmm4,<xmm8=%xmm8
2413pand %xmm4,%xmm8
2414
2415# qhasm: xmm11 &= xmm5
2416# asm 1: pand <xmm5=int6464#8,<xmm11=int6464#12
2417# asm 2: pand <xmm5=%xmm7,<xmm11=%xmm11
2418pand %xmm7,%xmm11
2419
2420# qhasm: xmm8 ^= xmm11
2421# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
2422# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
2423pxor %xmm11,%xmm8
2424
2425# qhasm: xmm11 ^= xmm2
2426# asm 1: pxor <xmm2=int6464#3,<xmm11=int6464#12
2427# asm 2: pxor <xmm2=%xmm2,<xmm11=%xmm11
2428pxor %xmm2,%xmm11
2429
2430# qhasm: xmm14 ^= xmm4
2431# asm 1: pxor <xmm4=int6464#1,<xmm14=int6464#15
2432# asm 2: pxor <xmm4=%xmm0,<xmm14=%xmm14
2433pxor %xmm0,%xmm14
2434
2435# qhasm: xmm8 ^= xmm4
2436# asm 1: pxor <xmm4=int6464#1,<xmm8=int6464#9
2437# asm 2: pxor <xmm4=%xmm0,<xmm8=%xmm8
2438pxor %xmm0,%xmm8
2439
2440# qhasm: xmm13 ^= xmm0
2441# asm 1: pxor <xmm0=int6464#2,<xmm13=int6464#14
2442# asm 2: pxor <xmm0=%xmm1,<xmm13=%xmm13
2443pxor %xmm1,%xmm13
2444
2445# qhasm: xmm11 ^= xmm0
2446# asm 1: pxor <xmm0=int6464#2,<xmm11=int6464#12
2447# asm 2: pxor <xmm0=%xmm1,<xmm11=%xmm11
2448pxor %xmm1,%xmm11
2449
2450# qhasm: xmm4 = xmm15
2451# asm 1: movdqa <xmm15=int6464#16,>xmm4=int6464#1
2452# asm 2: movdqa <xmm15=%xmm15,>xmm4=%xmm0
2453movdqa %xmm15,%xmm0
2454
2455# qhasm: xmm0 = xmm9
2456# asm 1: movdqa <xmm9=int6464#10,>xmm0=int6464#2
2457# asm 2: movdqa <xmm9=%xmm9,>xmm0=%xmm1
2458movdqa %xmm9,%xmm1
2459
2460# qhasm: xmm4 ^= xmm12
2461# asm 1: pxor <xmm12=int6464#13,<xmm4=int6464#1
2462# asm 2: pxor <xmm12=%xmm12,<xmm4=%xmm0
2463pxor %xmm12,%xmm0
2464
2465# qhasm: xmm0 ^= xmm10
2466# asm 1: pxor <xmm10=int6464#11,<xmm0=int6464#2
2467# asm 2: pxor <xmm10=%xmm10,<xmm0=%xmm1
2468pxor %xmm10,%xmm1
2469
2470# qhasm: xmm3 = xmm7
2471# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
2472# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
2473movdqa %xmm5,%xmm2
2474
2475# qhasm: xmm3 ^= xmm6
2476# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
2477# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
2478pxor %xmm3,%xmm2
2479
2480# qhasm: xmm3 &= xmm4
2481# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
2482# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
2483pand %xmm0,%xmm2
2484
2485# qhasm: xmm4 ^= xmm0
2486# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
2487# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
2488pxor %xmm1,%xmm0
2489
2490# qhasm: xmm4 &= xmm6
2491# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
2492# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
2493pand %xmm3,%xmm0
2494
2495# qhasm: xmm0 &= xmm7
2496# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
2497# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
2498pand %xmm5,%xmm1
2499
2500# qhasm: xmm0 ^= xmm4
2501# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
2502# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
2503pxor %xmm0,%xmm1
2504
2505# qhasm: xmm4 ^= xmm3
2506# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
2507# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
2508pxor %xmm2,%xmm0
2509
2510# qhasm: xmm2 = xmm5
2511# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
2512# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
2513movdqa %xmm7,%xmm2
2514
2515# qhasm: xmm2 ^= xmm1
2516# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
2517# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
2518pxor %xmm4,%xmm2
2519
2520# qhasm: xmm2 &= xmm12
2521# asm 1: pand <xmm12=int6464#13,<xmm2=int6464#3
2522# asm 2: pand <xmm12=%xmm12,<xmm2=%xmm2
2523pand %xmm12,%xmm2
2524
2525# qhasm: xmm12 ^= xmm10
2526# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
2527# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
2528pxor %xmm10,%xmm12
2529
2530# qhasm: xmm12 &= xmm1
2531# asm 1: pand <xmm1=int6464#5,<xmm12=int6464#13
2532# asm 2: pand <xmm1=%xmm4,<xmm12=%xmm12
2533pand %xmm4,%xmm12
2534
2535# qhasm: xmm10 &= xmm5
2536# asm 1: pand <xmm5=int6464#8,<xmm10=int6464#11
2537# asm 2: pand <xmm5=%xmm7,<xmm10=%xmm10
2538pand %xmm7,%xmm10
2539
2540# qhasm: xmm12 ^= xmm10
2541# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
2542# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
2543pxor %xmm10,%xmm12
2544
2545# qhasm: xmm10 ^= xmm2
2546# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#11
2547# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm10
2548pxor %xmm2,%xmm10
2549
2550# qhasm: xmm7 ^= xmm5
2551# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
2552# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
2553pxor %xmm7,%xmm5
2554
2555# qhasm: xmm6 ^= xmm1
2556# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
2557# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
2558pxor %xmm4,%xmm3
2559
2560# qhasm: xmm3 = xmm7
2561# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
2562# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
2563movdqa %xmm5,%xmm2
2564
2565# qhasm: xmm3 ^= xmm6
2566# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
2567# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
2568pxor %xmm3,%xmm2
2569
2570# qhasm: xmm3 &= xmm15
2571# asm 1: pand <xmm15=int6464#16,<xmm3=int6464#3
2572# asm 2: pand <xmm15=%xmm15,<xmm3=%xmm2
2573pand %xmm15,%xmm2
2574
2575# qhasm: xmm15 ^= xmm9
2576# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
2577# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
2578pxor %xmm9,%xmm15
2579
2580# qhasm: xmm15 &= xmm6
2581# asm 1: pand <xmm6=int6464#4,<xmm15=int6464#16
2582# asm 2: pand <xmm6=%xmm3,<xmm15=%xmm15
2583pand %xmm3,%xmm15
2584
2585# qhasm: xmm9 &= xmm7
2586# asm 1: pand <xmm7=int6464#6,<xmm9=int6464#10
2587# asm 2: pand <xmm7=%xmm5,<xmm9=%xmm9
2588pand %xmm5,%xmm9
2589
2590# qhasm: xmm15 ^= xmm9
2591# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
2592# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
2593pxor %xmm9,%xmm15
2594
2595# qhasm: xmm9 ^= xmm3
2596# asm 1: pxor <xmm3=int6464#3,<xmm9=int6464#10
2597# asm 2: pxor <xmm3=%xmm2,<xmm9=%xmm9
2598pxor %xmm2,%xmm9
2599
2600# qhasm: xmm15 ^= xmm4
2601# asm 1: pxor <xmm4=int6464#1,<xmm15=int6464#16
2602# asm 2: pxor <xmm4=%xmm0,<xmm15=%xmm15
2603pxor %xmm0,%xmm15
2604
2605# qhasm: xmm12 ^= xmm4
2606# asm 1: pxor <xmm4=int6464#1,<xmm12=int6464#13
2607# asm 2: pxor <xmm4=%xmm0,<xmm12=%xmm12
2608pxor %xmm0,%xmm12
2609
2610# qhasm: xmm9 ^= xmm0
2611# asm 1: pxor <xmm0=int6464#2,<xmm9=int6464#10
2612# asm 2: pxor <xmm0=%xmm1,<xmm9=%xmm9
2613pxor %xmm1,%xmm9
2614
2615# qhasm: xmm10 ^= xmm0
2616# asm 1: pxor <xmm0=int6464#2,<xmm10=int6464#11
2617# asm 2: pxor <xmm0=%xmm1,<xmm10=%xmm10
2618pxor %xmm1,%xmm10
2619
2620# qhasm: xmm15 ^= xmm8
2621# asm 1: pxor <xmm8=int6464#9,<xmm15=int6464#16
2622# asm 2: pxor <xmm8=%xmm8,<xmm15=%xmm15
2623pxor %xmm8,%xmm15
2624
2625# qhasm: xmm9 ^= xmm14
2626# asm 1: pxor <xmm14=int6464#15,<xmm9=int6464#10
2627# asm 2: pxor <xmm14=%xmm14,<xmm9=%xmm9
2628pxor %xmm14,%xmm9
2629
2630# qhasm: xmm12 ^= xmm15
2631# asm 1: pxor <xmm15=int6464#16,<xmm12=int6464#13
2632# asm 2: pxor <xmm15=%xmm15,<xmm12=%xmm12
2633pxor %xmm15,%xmm12
2634
2635# qhasm: xmm14 ^= xmm8
2636# asm 1: pxor <xmm8=int6464#9,<xmm14=int6464#15
2637# asm 2: pxor <xmm8=%xmm8,<xmm14=%xmm14
2638pxor %xmm8,%xmm14
2639
2640# qhasm: xmm8 ^= xmm9
2641# asm 1: pxor <xmm9=int6464#10,<xmm8=int6464#9
2642# asm 2: pxor <xmm9=%xmm9,<xmm8=%xmm8
2643pxor %xmm9,%xmm8
2644
2645# qhasm: xmm9 ^= xmm13
2646# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
2647# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
2648pxor %xmm13,%xmm9
2649
2650# qhasm: xmm13 ^= xmm10
2651# asm 1: pxor <xmm10=int6464#11,<xmm13=int6464#14
2652# asm 2: pxor <xmm10=%xmm10,<xmm13=%xmm13
2653pxor %xmm10,%xmm13
2654
2655# qhasm: xmm12 ^= xmm13
2656# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
2657# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
2658pxor %xmm13,%xmm12
2659
2660# qhasm: xmm10 ^= xmm11
2661# asm 1: pxor <xmm11=int6464#12,<xmm10=int6464#11
2662# asm 2: pxor <xmm11=%xmm11,<xmm10=%xmm10
2663pxor %xmm11,%xmm10
2664
2665# qhasm: xmm11 ^= xmm13
2666# asm 1: pxor <xmm13=int6464#14,<xmm11=int6464#12
2667# asm 2: pxor <xmm13=%xmm13,<xmm11=%xmm11
2668pxor %xmm13,%xmm11
2669
2670# qhasm: xmm14 ^= xmm11
2671# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
2672# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
2673pxor %xmm11,%xmm14
2674
2675# qhasm: xmm0 = shuffle dwords of xmm8 by 0x93
2676# asm 1: pshufd $0x93,<xmm8=int6464#9,>xmm0=int6464#1
2677# asm 2: pshufd $0x93,<xmm8=%xmm8,>xmm0=%xmm0
2678pshufd $0x93,%xmm8,%xmm0
2679
2680# qhasm: xmm1 = shuffle dwords of xmm9 by 0x93
2681# asm 1: pshufd $0x93,<xmm9=int6464#10,>xmm1=int6464#2
2682# asm 2: pshufd $0x93,<xmm9=%xmm9,>xmm1=%xmm1
2683pshufd $0x93,%xmm9,%xmm1
2684
2685# qhasm: xmm2 = shuffle dwords of xmm12 by 0x93
2686# asm 1: pshufd $0x93,<xmm12=int6464#13,>xmm2=int6464#3
2687# asm 2: pshufd $0x93,<xmm12=%xmm12,>xmm2=%xmm2
2688pshufd $0x93,%xmm12,%xmm2
2689
2690# qhasm: xmm3 = shuffle dwords of xmm14 by 0x93
2691# asm 1: pshufd $0x93,<xmm14=int6464#15,>xmm3=int6464#4
2692# asm 2: pshufd $0x93,<xmm14=%xmm14,>xmm3=%xmm3
2693pshufd $0x93,%xmm14,%xmm3
2694
2695# qhasm: xmm4 = shuffle dwords of xmm11 by 0x93
2696# asm 1: pshufd $0x93,<xmm11=int6464#12,>xmm4=int6464#5
2697# asm 2: pshufd $0x93,<xmm11=%xmm11,>xmm4=%xmm4
2698pshufd $0x93,%xmm11,%xmm4
2699
2700# qhasm: xmm5 = shuffle dwords of xmm15 by 0x93
2701# asm 1: pshufd $0x93,<xmm15=int6464#16,>xmm5=int6464#6
2702# asm 2: pshufd $0x93,<xmm15=%xmm15,>xmm5=%xmm5
2703pshufd $0x93,%xmm15,%xmm5
2704
2705# qhasm: xmm6 = shuffle dwords of xmm10 by 0x93
2706# asm 1: pshufd $0x93,<xmm10=int6464#11,>xmm6=int6464#7
2707# asm 2: pshufd $0x93,<xmm10=%xmm10,>xmm6=%xmm6
2708pshufd $0x93,%xmm10,%xmm6
2709
2710# qhasm: xmm7 = shuffle dwords of xmm13 by 0x93
2711# asm 1: pshufd $0x93,<xmm13=int6464#14,>xmm7=int6464#8
2712# asm 2: pshufd $0x93,<xmm13=%xmm13,>xmm7=%xmm7
2713pshufd $0x93,%xmm13,%xmm7
2714
2715# qhasm: xmm8 ^= xmm0
2716# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
2717# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
2718pxor %xmm0,%xmm8
2719
2720# qhasm: xmm9 ^= xmm1
2721# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
2722# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
2723pxor %xmm1,%xmm9
2724
2725# qhasm: xmm12 ^= xmm2
2726# asm 1: pxor <xmm2=int6464#3,<xmm12=int6464#13
2727# asm 2: pxor <xmm2=%xmm2,<xmm12=%xmm12
2728pxor %xmm2,%xmm12
2729
2730# qhasm: xmm14 ^= xmm3
2731# asm 1: pxor <xmm3=int6464#4,<xmm14=int6464#15
2732# asm 2: pxor <xmm3=%xmm3,<xmm14=%xmm14
2733pxor %xmm3,%xmm14
2734
2735# qhasm: xmm11 ^= xmm4
2736# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
2737# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
2738pxor %xmm4,%xmm11
2739
2740# qhasm: xmm15 ^= xmm5
2741# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
2742# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
2743pxor %xmm5,%xmm15
2744
2745# qhasm: xmm10 ^= xmm6
2746# asm 1: pxor <xmm6=int6464#7,<xmm10=int6464#11
2747# asm 2: pxor <xmm6=%xmm6,<xmm10=%xmm10
2748pxor %xmm6,%xmm10
2749
2750# qhasm: xmm13 ^= xmm7
2751# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
2752# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
2753pxor %xmm7,%xmm13
2754
2755# qhasm: xmm0 ^= xmm13
2756# asm 1: pxor <xmm13=int6464#14,<xmm0=int6464#1
2757# asm 2: pxor <xmm13=%xmm13,<xmm0=%xmm0
2758pxor %xmm13,%xmm0
2759
2760# qhasm: xmm1 ^= xmm8
2761# asm 1: pxor <xmm8=int6464#9,<xmm1=int6464#2
2762# asm 2: pxor <xmm8=%xmm8,<xmm1=%xmm1
2763pxor %xmm8,%xmm1
2764
2765# qhasm: xmm2 ^= xmm9
2766# asm 1: pxor <xmm9=int6464#10,<xmm2=int6464#3
2767# asm 2: pxor <xmm9=%xmm9,<xmm2=%xmm2
2768pxor %xmm9,%xmm2
2769
2770# qhasm: xmm1 ^= xmm13
2771# asm 1: pxor <xmm13=int6464#14,<xmm1=int6464#2
2772# asm 2: pxor <xmm13=%xmm13,<xmm1=%xmm1
2773pxor %xmm13,%xmm1
2774
2775# qhasm: xmm3 ^= xmm12
2776# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
2777# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
2778pxor %xmm12,%xmm3
2779
2780# qhasm: xmm4 ^= xmm14
2781# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#5
2782# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm4
2783pxor %xmm14,%xmm4
2784
2785# qhasm: xmm5 ^= xmm11
2786# asm 1: pxor <xmm11=int6464#12,<xmm5=int6464#6
2787# asm 2: pxor <xmm11=%xmm11,<xmm5=%xmm5
2788pxor %xmm11,%xmm5
2789
2790# qhasm: xmm3 ^= xmm13
2791# asm 1: pxor <xmm13=int6464#14,<xmm3=int6464#4
2792# asm 2: pxor <xmm13=%xmm13,<xmm3=%xmm3
2793pxor %xmm13,%xmm3
2794
2795# qhasm: xmm6 ^= xmm15
2796# asm 1: pxor <xmm15=int6464#16,<xmm6=int6464#7
2797# asm 2: pxor <xmm15=%xmm15,<xmm6=%xmm6
2798pxor %xmm15,%xmm6
2799
2800# qhasm: xmm7 ^= xmm10
2801# asm 1: pxor <xmm10=int6464#11,<xmm7=int6464#8
2802# asm 2: pxor <xmm10=%xmm10,<xmm7=%xmm7
2803pxor %xmm10,%xmm7
2804
2805# qhasm: xmm4 ^= xmm13
2806# asm 1: pxor <xmm13=int6464#14,<xmm4=int6464#5
2807# asm 2: pxor <xmm13=%xmm13,<xmm4=%xmm4
2808pxor %xmm13,%xmm4
2809
2810# qhasm: xmm8 = shuffle dwords of xmm8 by 0x4E
2811# asm 1: pshufd $0x4E,<xmm8=int6464#9,>xmm8=int6464#9
2812# asm 2: pshufd $0x4E,<xmm8=%xmm8,>xmm8=%xmm8
2813pshufd $0x4E,%xmm8,%xmm8
2814
2815# qhasm: xmm9 = shuffle dwords of xmm9 by 0x4E
2816# asm 1: pshufd $0x4E,<xmm9=int6464#10,>xmm9=int6464#10
2817# asm 2: pshufd $0x4E,<xmm9=%xmm9,>xmm9=%xmm9
2818pshufd $0x4E,%xmm9,%xmm9
2819
2820# qhasm: xmm12 = shuffle dwords of xmm12 by 0x4E
2821# asm 1: pshufd $0x4E,<xmm12=int6464#13,>xmm12=int6464#13
2822# asm 2: pshufd $0x4E,<xmm12=%xmm12,>xmm12=%xmm12
2823pshufd $0x4E,%xmm12,%xmm12
2824
2825# qhasm: xmm14 = shuffle dwords of xmm14 by 0x4E
2826# asm 1: pshufd $0x4E,<xmm14=int6464#15,>xmm14=int6464#15
2827# asm 2: pshufd $0x4E,<xmm14=%xmm14,>xmm14=%xmm14
2828pshufd $0x4E,%xmm14,%xmm14
2829
2830# qhasm: xmm11 = shuffle dwords of xmm11 by 0x4E
2831# asm 1: pshufd $0x4E,<xmm11=int6464#12,>xmm11=int6464#12
2832# asm 2: pshufd $0x4E,<xmm11=%xmm11,>xmm11=%xmm11
2833pshufd $0x4E,%xmm11,%xmm11
2834
2835# qhasm: xmm15 = shuffle dwords of xmm15 by 0x4E
2836# asm 1: pshufd $0x4E,<xmm15=int6464#16,>xmm15=int6464#16
2837# asm 2: pshufd $0x4E,<xmm15=%xmm15,>xmm15=%xmm15
2838pshufd $0x4E,%xmm15,%xmm15
2839
2840# qhasm: xmm10 = shuffle dwords of xmm10 by 0x4E
2841# asm 1: pshufd $0x4E,<xmm10=int6464#11,>xmm10=int6464#11
2842# asm 2: pshufd $0x4E,<xmm10=%xmm10,>xmm10=%xmm10
2843pshufd $0x4E,%xmm10,%xmm10
2844
2845# qhasm: xmm13 = shuffle dwords of xmm13 by 0x4E
2846# asm 1: pshufd $0x4E,<xmm13=int6464#14,>xmm13=int6464#14
2847# asm 2: pshufd $0x4E,<xmm13=%xmm13,>xmm13=%xmm13
2848pshufd $0x4E,%xmm13,%xmm13
2849
2850# qhasm: xmm0 ^= xmm8
2851# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
2852# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
2853pxor %xmm8,%xmm0
2854
2855# qhasm: xmm1 ^= xmm9
2856# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
2857# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
2858pxor %xmm9,%xmm1
2859
2860# qhasm: xmm2 ^= xmm12
2861# asm 1: pxor <xmm12=int6464#13,<xmm2=int6464#3
2862# asm 2: pxor <xmm12=%xmm12,<xmm2=%xmm2
2863pxor %xmm12,%xmm2
2864
2865# qhasm: xmm3 ^= xmm14
2866# asm 1: pxor <xmm14=int6464#15,<xmm3=int6464#4
2867# asm 2: pxor <xmm14=%xmm14,<xmm3=%xmm3
2868pxor %xmm14,%xmm3
2869
2870# qhasm: xmm4 ^= xmm11
2871# asm 1: pxor <xmm11=int6464#12,<xmm4=int6464#5
2872# asm 2: pxor <xmm11=%xmm11,<xmm4=%xmm4
2873pxor %xmm11,%xmm4
2874
2875# qhasm: xmm5 ^= xmm15
2876# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
2877# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
2878pxor %xmm15,%xmm5
2879
2880# qhasm: xmm6 ^= xmm10
2881# asm 1: pxor <xmm10=int6464#11,<xmm6=int6464#7
2882# asm 2: pxor <xmm10=%xmm10,<xmm6=%xmm6
2883pxor %xmm10,%xmm6
2884
2885# qhasm: xmm7 ^= xmm13
2886# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
2887# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
2888pxor %xmm13,%xmm7
2889
2890# qhasm: xmm0 ^= *(int128 *)(c + 256)
2891# asm 1: pxor 256(<c=int64#4),<xmm0=int6464#1
2892# asm 2: pxor 256(<c=%rcx),<xmm0=%xmm0
2893pxor 256(%rcx),%xmm0
2894
2895# qhasm: shuffle bytes of xmm0 by SR
2896# asm 1: pshufb SR,<xmm0=int6464#1
2897# asm 2: pshufb SR,<xmm0=%xmm0
2898pshufb SR,%xmm0
2899
2900# qhasm: xmm1 ^= *(int128 *)(c + 272)
2901# asm 1: pxor 272(<c=int64#4),<xmm1=int6464#2
2902# asm 2: pxor 272(<c=%rcx),<xmm1=%xmm1
2903pxor 272(%rcx),%xmm1
2904
2905# qhasm: shuffle bytes of xmm1 by SR
2906# asm 1: pshufb SR,<xmm1=int6464#2
2907# asm 2: pshufb SR,<xmm1=%xmm1
2908pshufb SR,%xmm1
2909
2910# qhasm: xmm2 ^= *(int128 *)(c + 288)
2911# asm 1: pxor 288(<c=int64#4),<xmm2=int6464#3
2912# asm 2: pxor 288(<c=%rcx),<xmm2=%xmm2
2913pxor 288(%rcx),%xmm2
2914
2915# qhasm: shuffle bytes of xmm2 by SR
2916# asm 1: pshufb SR,<xmm2=int6464#3
2917# asm 2: pshufb SR,<xmm2=%xmm2
2918pshufb SR,%xmm2
2919
2920# qhasm: xmm3 ^= *(int128 *)(c + 304)
2921# asm 1: pxor 304(<c=int64#4),<xmm3=int6464#4
2922# asm 2: pxor 304(<c=%rcx),<xmm3=%xmm3
2923pxor 304(%rcx),%xmm3
2924
2925# qhasm: shuffle bytes of xmm3 by SR
2926# asm 1: pshufb SR,<xmm3=int6464#4
2927# asm 2: pshufb SR,<xmm3=%xmm3
2928pshufb SR,%xmm3
2929
2930# qhasm: xmm4 ^= *(int128 *)(c + 320)
2931# asm 1: pxor 320(<c=int64#4),<xmm4=int6464#5
2932# asm 2: pxor 320(<c=%rcx),<xmm4=%xmm4
2933pxor 320(%rcx),%xmm4
2934
2935# qhasm: shuffle bytes of xmm4 by SR
2936# asm 1: pshufb SR,<xmm4=int6464#5
2937# asm 2: pshufb SR,<xmm4=%xmm4
2938pshufb SR,%xmm4
2939
2940# qhasm: xmm5 ^= *(int128 *)(c + 336)
2941# asm 1: pxor 336(<c=int64#4),<xmm5=int6464#6
2942# asm 2: pxor 336(<c=%rcx),<xmm5=%xmm5
2943pxor 336(%rcx),%xmm5
2944
2945# qhasm: shuffle bytes of xmm5 by SR
2946# asm 1: pshufb SR,<xmm5=int6464#6
2947# asm 2: pshufb SR,<xmm5=%xmm5
2948pshufb SR,%xmm5
2949
2950# qhasm: xmm6 ^= *(int128 *)(c + 352)
2951# asm 1: pxor 352(<c=int64#4),<xmm6=int6464#7
2952# asm 2: pxor 352(<c=%rcx),<xmm6=%xmm6
2953pxor 352(%rcx),%xmm6
2954
2955# qhasm: shuffle bytes of xmm6 by SR
2956# asm 1: pshufb SR,<xmm6=int6464#7
2957# asm 2: pshufb SR,<xmm6=%xmm6
2958pshufb SR,%xmm6
2959
2960# qhasm: xmm7 ^= *(int128 *)(c + 368)
2961# asm 1: pxor 368(<c=int64#4),<xmm7=int6464#8
2962# asm 2: pxor 368(<c=%rcx),<xmm7=%xmm7
2963pxor 368(%rcx),%xmm7
2964
2965# qhasm: shuffle bytes of xmm7 by SR
2966# asm 1: pshufb SR,<xmm7=int6464#8
2967# asm 2: pshufb SR,<xmm7=%xmm7
2968pshufb SR,%xmm7
2969
2970# qhasm: xmm5 ^= xmm6
2971# asm 1: pxor <xmm6=int6464#7,<xmm5=int6464#6
2972# asm 2: pxor <xmm6=%xmm6,<xmm5=%xmm5
2973pxor %xmm6,%xmm5
2974
2975# qhasm: xmm2 ^= xmm1
2976# asm 1: pxor <xmm1=int6464#2,<xmm2=int6464#3
2977# asm 2: pxor <xmm1=%xmm1,<xmm2=%xmm2
2978pxor %xmm1,%xmm2
2979
2980# qhasm: xmm5 ^= xmm0
2981# asm 1: pxor <xmm0=int6464#1,<xmm5=int6464#6
2982# asm 2: pxor <xmm0=%xmm0,<xmm5=%xmm5
2983pxor %xmm0,%xmm5
2984
2985# qhasm: xmm6 ^= xmm2
2986# asm 1: pxor <xmm2=int6464#3,<xmm6=int6464#7
2987# asm 2: pxor <xmm2=%xmm2,<xmm6=%xmm6
2988pxor %xmm2,%xmm6
2989
2990# qhasm: xmm3 ^= xmm0
2991# asm 1: pxor <xmm0=int6464#1,<xmm3=int6464#4
2992# asm 2: pxor <xmm0=%xmm0,<xmm3=%xmm3
2993pxor %xmm0,%xmm3
2994
2995# qhasm: xmm6 ^= xmm3
2996# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
2997# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
2998pxor %xmm3,%xmm6
2999
3000# qhasm: xmm3 ^= xmm7
3001# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#4
3002# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm3
3003pxor %xmm7,%xmm3
3004
3005# qhasm: xmm3 ^= xmm4
3006# asm 1: pxor <xmm4=int6464#5,<xmm3=int6464#4
3007# asm 2: pxor <xmm4=%xmm4,<xmm3=%xmm3
3008pxor %xmm4,%xmm3
3009
3010# qhasm: xmm7 ^= xmm5
3011# asm 1: pxor <xmm5=int6464#6,<xmm7=int6464#8
3012# asm 2: pxor <xmm5=%xmm5,<xmm7=%xmm7
3013pxor %xmm5,%xmm7
3014
3015# qhasm: xmm3 ^= xmm1
3016# asm 1: pxor <xmm1=int6464#2,<xmm3=int6464#4
3017# asm 2: pxor <xmm1=%xmm1,<xmm3=%xmm3
3018pxor %xmm1,%xmm3
3019
3020# qhasm: xmm4 ^= xmm5
3021# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
3022# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
3023pxor %xmm5,%xmm4
3024
3025# qhasm: xmm2 ^= xmm7
3026# asm 1: pxor <xmm7=int6464#8,<xmm2=int6464#3
3027# asm 2: pxor <xmm7=%xmm7,<xmm2=%xmm2
3028pxor %xmm7,%xmm2
3029
3030# qhasm: xmm1 ^= xmm5
3031# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
3032# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
3033pxor %xmm5,%xmm1
3034
3035# qhasm: xmm11 = xmm7
3036# asm 1: movdqa <xmm7=int6464#8,>xmm11=int6464#9
3037# asm 2: movdqa <xmm7=%xmm7,>xmm11=%xmm8
3038movdqa %xmm7,%xmm8
3039
3040# qhasm: xmm10 = xmm1
3041# asm 1: movdqa <xmm1=int6464#2,>xmm10=int6464#10
3042# asm 2: movdqa <xmm1=%xmm1,>xmm10=%xmm9
3043movdqa %xmm1,%xmm9
3044
3045# qhasm: xmm9 = xmm5
3046# asm 1: movdqa <xmm5=int6464#6,>xmm9=int6464#11
3047# asm 2: movdqa <xmm5=%xmm5,>xmm9=%xmm10
3048movdqa %xmm5,%xmm10
3049
3050# qhasm: xmm13 = xmm2
3051# asm 1: movdqa <xmm2=int6464#3,>xmm13=int6464#12
3052# asm 2: movdqa <xmm2=%xmm2,>xmm13=%xmm11
3053movdqa %xmm2,%xmm11
3054
3055# qhasm: xmm12 = xmm6
3056# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#13
3057# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm12
3058movdqa %xmm6,%xmm12
3059
3060# qhasm: xmm11 ^= xmm4
3061# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#9
3062# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm8
3063pxor %xmm4,%xmm8
3064
3065# qhasm: xmm10 ^= xmm2
3066# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#10
3067# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm9
3068pxor %xmm2,%xmm9
3069
3070# qhasm: xmm9 ^= xmm3
3071# asm 1: pxor <xmm3=int6464#4,<xmm9=int6464#11
3072# asm 2: pxor <xmm3=%xmm3,<xmm9=%xmm10
3073pxor %xmm3,%xmm10
3074
3075# qhasm: xmm13 ^= xmm4
3076# asm 1: pxor <xmm4=int6464#5,<xmm13=int6464#12
3077# asm 2: pxor <xmm4=%xmm4,<xmm13=%xmm11
3078pxor %xmm4,%xmm11
3079
3080# qhasm: xmm12 ^= xmm0
3081# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
3082# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
3083pxor %xmm0,%xmm12
3084
3085# qhasm: xmm14 = xmm11
3086# asm 1: movdqa <xmm11=int6464#9,>xmm14=int6464#14
3087# asm 2: movdqa <xmm11=%xmm8,>xmm14=%xmm13
3088movdqa %xmm8,%xmm13
3089
3090# qhasm: xmm8 = xmm10
3091# asm 1: movdqa <xmm10=int6464#10,>xmm8=int6464#15
3092# asm 2: movdqa <xmm10=%xmm9,>xmm8=%xmm14
3093movdqa %xmm9,%xmm14
3094
3095# qhasm: xmm15 = xmm11
3096# asm 1: movdqa <xmm11=int6464#9,>xmm15=int6464#16
3097# asm 2: movdqa <xmm11=%xmm8,>xmm15=%xmm15
3098movdqa %xmm8,%xmm15
3099
3100# qhasm: xmm10 |= xmm9
3101# asm 1: por <xmm9=int6464#11,<xmm10=int6464#10
3102# asm 2: por <xmm9=%xmm10,<xmm10=%xmm9
3103por %xmm10,%xmm9
3104
3105# qhasm: xmm11 |= xmm12
3106# asm 1: por <xmm12=int6464#13,<xmm11=int6464#9
3107# asm 2: por <xmm12=%xmm12,<xmm11=%xmm8
3108por %xmm12,%xmm8
3109
3110# qhasm: xmm15 ^= xmm8
3111# asm 1: pxor <xmm8=int6464#15,<xmm15=int6464#16
3112# asm 2: pxor <xmm8=%xmm14,<xmm15=%xmm15
3113pxor %xmm14,%xmm15
3114
3115# qhasm: xmm14 &= xmm12
3116# asm 1: pand <xmm12=int6464#13,<xmm14=int6464#14
3117# asm 2: pand <xmm12=%xmm12,<xmm14=%xmm13
3118pand %xmm12,%xmm13
3119
3120# qhasm: xmm8 &= xmm9
3121# asm 1: pand <xmm9=int6464#11,<xmm8=int6464#15
3122# asm 2: pand <xmm9=%xmm10,<xmm8=%xmm14
3123pand %xmm10,%xmm14
3124
3125# qhasm: xmm12 ^= xmm9
3126# asm 1: pxor <xmm9=int6464#11,<xmm12=int6464#13
3127# asm 2: pxor <xmm9=%xmm10,<xmm12=%xmm12
3128pxor %xmm10,%xmm12
3129
3130# qhasm: xmm15 &= xmm12
3131# asm 1: pand <xmm12=int6464#13,<xmm15=int6464#16
3132# asm 2: pand <xmm12=%xmm12,<xmm15=%xmm15
3133pand %xmm12,%xmm15
3134
3135# qhasm: xmm12 = xmm3
3136# asm 1: movdqa <xmm3=int6464#4,>xmm12=int6464#11
3137# asm 2: movdqa <xmm3=%xmm3,>xmm12=%xmm10
3138movdqa %xmm3,%xmm10
3139
3140# qhasm: xmm12 ^= xmm0
3141# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#11
3142# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm10
3143pxor %xmm0,%xmm10
3144
3145# qhasm: xmm13 &= xmm12
3146# asm 1: pand <xmm12=int6464#11,<xmm13=int6464#12
3147# asm 2: pand <xmm12=%xmm10,<xmm13=%xmm11
3148pand %xmm10,%xmm11
3149
3150# qhasm: xmm11 ^= xmm13
3151# asm 1: pxor <xmm13=int6464#12,<xmm11=int6464#9
3152# asm 2: pxor <xmm13=%xmm11,<xmm11=%xmm8
3153pxor %xmm11,%xmm8
3154
3155# qhasm: xmm10 ^= xmm13
3156# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
3157# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
3158pxor %xmm11,%xmm9
3159
3160# qhasm: xmm13 = xmm7
3161# asm 1: movdqa <xmm7=int6464#8,>xmm13=int6464#11
3162# asm 2: movdqa <xmm7=%xmm7,>xmm13=%xmm10
3163movdqa %xmm7,%xmm10
3164
3165# qhasm: xmm13 ^= xmm1
3166# asm 1: pxor <xmm1=int6464#2,<xmm13=int6464#11
3167# asm 2: pxor <xmm1=%xmm1,<xmm13=%xmm10
3168pxor %xmm1,%xmm10
3169
3170# qhasm: xmm12 = xmm5
3171# asm 1: movdqa <xmm5=int6464#6,>xmm12=int6464#12
3172# asm 2: movdqa <xmm5=%xmm5,>xmm12=%xmm11
3173movdqa %xmm5,%xmm11
3174
3175# qhasm: xmm9 = xmm13
3176# asm 1: movdqa <xmm13=int6464#11,>xmm9=int6464#13
3177# asm 2: movdqa <xmm13=%xmm10,>xmm9=%xmm12
3178movdqa %xmm10,%xmm12
3179
3180# qhasm: xmm12 ^= xmm6
3181# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#12
3182# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm11
3183pxor %xmm6,%xmm11
3184
3185# qhasm: xmm9 |= xmm12
3186# asm 1: por <xmm12=int6464#12,<xmm9=int6464#13
3187# asm 2: por <xmm12=%xmm11,<xmm9=%xmm12
3188por %xmm11,%xmm12
3189
3190# qhasm: xmm13 &= xmm12
3191# asm 1: pand <xmm12=int6464#12,<xmm13=int6464#11
3192# asm 2: pand <xmm12=%xmm11,<xmm13=%xmm10
3193pand %xmm11,%xmm10
3194
3195# qhasm: xmm8 ^= xmm13
3196# asm 1: pxor <xmm13=int6464#11,<xmm8=int6464#15
3197# asm 2: pxor <xmm13=%xmm10,<xmm8=%xmm14
3198pxor %xmm10,%xmm14
3199
3200# qhasm: xmm11 ^= xmm15
3201# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#9
3202# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm8
3203pxor %xmm15,%xmm8
3204
3205# qhasm: xmm10 ^= xmm14
3206# asm 1: pxor <xmm14=int6464#14,<xmm10=int6464#10
3207# asm 2: pxor <xmm14=%xmm13,<xmm10=%xmm9
3208pxor %xmm13,%xmm9
3209
3210# qhasm: xmm9 ^= xmm15
3211# asm 1: pxor <xmm15=int6464#16,<xmm9=int6464#13
3212# asm 2: pxor <xmm15=%xmm15,<xmm9=%xmm12
3213pxor %xmm15,%xmm12
3214
3215# qhasm: xmm8 ^= xmm14
3216# asm 1: pxor <xmm14=int6464#14,<xmm8=int6464#15
3217# asm 2: pxor <xmm14=%xmm13,<xmm8=%xmm14
3218pxor %xmm13,%xmm14
3219
3220# qhasm: xmm9 ^= xmm14
3221# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
3222# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
3223pxor %xmm13,%xmm12
3224
3225# qhasm: xmm12 = xmm2
3226# asm 1: movdqa <xmm2=int6464#3,>xmm12=int6464#11
3227# asm 2: movdqa <xmm2=%xmm2,>xmm12=%xmm10
3228movdqa %xmm2,%xmm10
3229
3230# qhasm: xmm13 = xmm4
3231# asm 1: movdqa <xmm4=int6464#5,>xmm13=int6464#12
3232# asm 2: movdqa <xmm4=%xmm4,>xmm13=%xmm11
3233movdqa %xmm4,%xmm11
3234
3235# qhasm: xmm14 = xmm1
3236# asm 1: movdqa <xmm1=int6464#2,>xmm14=int6464#14
3237# asm 2: movdqa <xmm1=%xmm1,>xmm14=%xmm13
3238movdqa %xmm1,%xmm13
3239
3240# qhasm: xmm15 = xmm7
3241# asm 1: movdqa <xmm7=int6464#8,>xmm15=int6464#16
3242# asm 2: movdqa <xmm7=%xmm7,>xmm15=%xmm15
3243movdqa %xmm7,%xmm15
3244
3245# qhasm: xmm12 &= xmm3
3246# asm 1: pand <xmm3=int6464#4,<xmm12=int6464#11
3247# asm 2: pand <xmm3=%xmm3,<xmm12=%xmm10
3248pand %xmm3,%xmm10
3249
3250# qhasm: xmm13 &= xmm0
3251# asm 1: pand <xmm0=int6464#1,<xmm13=int6464#12
3252# asm 2: pand <xmm0=%xmm0,<xmm13=%xmm11
3253pand %xmm0,%xmm11
3254
3255# qhasm: xmm14 &= xmm5
3256# asm 1: pand <xmm5=int6464#6,<xmm14=int6464#14
3257# asm 2: pand <xmm5=%xmm5,<xmm14=%xmm13
3258pand %xmm5,%xmm13
3259
3260# qhasm: xmm15 |= xmm6
3261# asm 1: por <xmm6=int6464#7,<xmm15=int6464#16
3262# asm 2: por <xmm6=%xmm6,<xmm15=%xmm15
3263por %xmm6,%xmm15
3264
3265# qhasm: xmm11 ^= xmm12
3266# asm 1: pxor <xmm12=int6464#11,<xmm11=int6464#9
3267# asm 2: pxor <xmm12=%xmm10,<xmm11=%xmm8
3268pxor %xmm10,%xmm8
3269
3270# qhasm: xmm10 ^= xmm13
3271# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
3272# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
3273pxor %xmm11,%xmm9
3274
3275# qhasm: xmm9 ^= xmm14
3276# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
3277# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
3278pxor %xmm13,%xmm12
3279
3280# qhasm: xmm8 ^= xmm15
3281# asm 1: pxor <xmm15=int6464#16,<xmm8=int6464#15
3282# asm 2: pxor <xmm15=%xmm15,<xmm8=%xmm14
3283pxor %xmm15,%xmm14
3284
3285# qhasm: xmm12 = xmm11
3286# asm 1: movdqa <xmm11=int6464#9,>xmm12=int6464#11
3287# asm 2: movdqa <xmm11=%xmm8,>xmm12=%xmm10
3288movdqa %xmm8,%xmm10
3289
3290# qhasm: xmm12 ^= xmm10
3291# asm 1: pxor <xmm10=int6464#10,<xmm12=int6464#11
3292# asm 2: pxor <xmm10=%xmm9,<xmm12=%xmm10
3293pxor %xmm9,%xmm10
3294
3295# qhasm: xmm11 &= xmm9
3296# asm 1: pand <xmm9=int6464#13,<xmm11=int6464#9
3297# asm 2: pand <xmm9=%xmm12,<xmm11=%xmm8
3298pand %xmm12,%xmm8
3299
3300# qhasm: xmm14 = xmm8
3301# asm 1: movdqa <xmm8=int6464#15,>xmm14=int6464#12
3302# asm 2: movdqa <xmm8=%xmm14,>xmm14=%xmm11
3303movdqa %xmm14,%xmm11
3304
3305# qhasm: xmm14 ^= xmm11
3306# asm 1: pxor <xmm11=int6464#9,<xmm14=int6464#12
3307# asm 2: pxor <xmm11=%xmm8,<xmm14=%xmm11
3308pxor %xmm8,%xmm11
3309
3310# qhasm: xmm15 = xmm12
3311# asm 1: movdqa <xmm12=int6464#11,>xmm15=int6464#14
3312# asm 2: movdqa <xmm12=%xmm10,>xmm15=%xmm13
3313movdqa %xmm10,%xmm13
3314
3315# qhasm: xmm15 &= xmm14
3316# asm 1: pand <xmm14=int6464#12,<xmm15=int6464#14
3317# asm 2: pand <xmm14=%xmm11,<xmm15=%xmm13
3318pand %xmm11,%xmm13
3319
3320# qhasm: xmm15 ^= xmm10
3321# asm 1: pxor <xmm10=int6464#10,<xmm15=int6464#14
3322# asm 2: pxor <xmm10=%xmm9,<xmm15=%xmm13
3323pxor %xmm9,%xmm13
3324
3325# qhasm: xmm13 = xmm9
3326# asm 1: movdqa <xmm9=int6464#13,>xmm13=int6464#16
3327# asm 2: movdqa <xmm9=%xmm12,>xmm13=%xmm15
3328movdqa %xmm12,%xmm15
3329
3330# qhasm: xmm13 ^= xmm8
3331# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
3332# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
3333pxor %xmm14,%xmm15
3334
3335# qhasm: xmm11 ^= xmm10
3336# asm 1: pxor <xmm10=int6464#10,<xmm11=int6464#9
3337# asm 2: pxor <xmm10=%xmm9,<xmm11=%xmm8
3338pxor %xmm9,%xmm8
3339
3340# qhasm: xmm13 &= xmm11
3341# asm 1: pand <xmm11=int6464#9,<xmm13=int6464#16
3342# asm 2: pand <xmm11=%xmm8,<xmm13=%xmm15
3343pand %xmm8,%xmm15
3344
3345# qhasm: xmm13 ^= xmm8
3346# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
3347# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
3348pxor %xmm14,%xmm15
3349
3350# qhasm: xmm9 ^= xmm13
3351# asm 1: pxor <xmm13=int6464#16,<xmm9=int6464#13
3352# asm 2: pxor <xmm13=%xmm15,<xmm9=%xmm12
3353pxor %xmm15,%xmm12
3354
3355# qhasm: xmm10 = xmm14
3356# asm 1: movdqa <xmm14=int6464#12,>xmm10=int6464#9
3357# asm 2: movdqa <xmm14=%xmm11,>xmm10=%xmm8
3358movdqa %xmm11,%xmm8
3359
3360# qhasm: xmm10 ^= xmm13
3361# asm 1: pxor <xmm13=int6464#16,<xmm10=int6464#9
3362# asm 2: pxor <xmm13=%xmm15,<xmm10=%xmm8
3363pxor %xmm15,%xmm8
3364
3365# qhasm: xmm10 &= xmm8
3366# asm 1: pand <xmm8=int6464#15,<xmm10=int6464#9
3367# asm 2: pand <xmm8=%xmm14,<xmm10=%xmm8
3368pand %xmm14,%xmm8
3369
3370# qhasm: xmm9 ^= xmm10
3371# asm 1: pxor <xmm10=int6464#9,<xmm9=int6464#13
3372# asm 2: pxor <xmm10=%xmm8,<xmm9=%xmm12
3373pxor %xmm8,%xmm12
3374
3375# qhasm: xmm14 ^= xmm10
3376# asm 1: pxor <xmm10=int6464#9,<xmm14=int6464#12
3377# asm 2: pxor <xmm10=%xmm8,<xmm14=%xmm11
3378pxor %xmm8,%xmm11
3379
3380# qhasm: xmm14 &= xmm15
3381# asm 1: pand <xmm15=int6464#14,<xmm14=int6464#12
3382# asm 2: pand <xmm15=%xmm13,<xmm14=%xmm11
3383pand %xmm13,%xmm11
3384
3385# qhasm: xmm14 ^= xmm12
3386# asm 1: pxor <xmm12=int6464#11,<xmm14=int6464#12
3387# asm 2: pxor <xmm12=%xmm10,<xmm14=%xmm11
3388pxor %xmm10,%xmm11
3389
3390# qhasm: xmm12 = xmm6
3391# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#9
3392# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm8
3393movdqa %xmm6,%xmm8
3394
3395# qhasm: xmm8 = xmm5
3396# asm 1: movdqa <xmm5=int6464#6,>xmm8=int6464#10
3397# asm 2: movdqa <xmm5=%xmm5,>xmm8=%xmm9
3398movdqa %xmm5,%xmm9
3399
3400# qhasm: xmm10 = xmm15
3401# asm 1: movdqa <xmm15=int6464#14,>xmm10=int6464#11
3402# asm 2: movdqa <xmm15=%xmm13,>xmm10=%xmm10
3403movdqa %xmm13,%xmm10
3404
3405# qhasm: xmm10 ^= xmm14
3406# asm 1: pxor <xmm14=int6464#12,<xmm10=int6464#11
3407# asm 2: pxor <xmm14=%xmm11,<xmm10=%xmm10
3408pxor %xmm11,%xmm10
3409
3410# qhasm: xmm10 &= xmm6
3411# asm 1: pand <xmm6=int6464#7,<xmm10=int6464#11
3412# asm 2: pand <xmm6=%xmm6,<xmm10=%xmm10
3413pand %xmm6,%xmm10
3414
3415# qhasm: xmm6 ^= xmm5
3416# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
3417# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
3418pxor %xmm5,%xmm6
3419
3420# qhasm: xmm6 &= xmm14
3421# asm 1: pand <xmm14=int6464#12,<xmm6=int6464#7
3422# asm 2: pand <xmm14=%xmm11,<xmm6=%xmm6
3423pand %xmm11,%xmm6
3424
3425# qhasm: xmm5 &= xmm15
3426# asm 1: pand <xmm15=int6464#14,<xmm5=int6464#6
3427# asm 2: pand <xmm15=%xmm13,<xmm5=%xmm5
3428pand %xmm13,%xmm5
3429
3430# qhasm: xmm6 ^= xmm5
3431# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
3432# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
3433pxor %xmm5,%xmm6
3434
3435# qhasm: xmm5 ^= xmm10
3436# asm 1: pxor <xmm10=int6464#11,<xmm5=int6464#6
3437# asm 2: pxor <xmm10=%xmm10,<xmm5=%xmm5
3438pxor %xmm10,%xmm5
3439
3440# qhasm: xmm12 ^= xmm0
3441# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#9
3442# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm8
3443pxor %xmm0,%xmm8
3444
3445# qhasm: xmm8 ^= xmm3
3446# asm 1: pxor <xmm3=int6464#4,<xmm8=int6464#10
3447# asm 2: pxor <xmm3=%xmm3,<xmm8=%xmm9
3448pxor %xmm3,%xmm9
3449
3450# qhasm: xmm15 ^= xmm13
3451# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
3452# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
3453pxor %xmm15,%xmm13
3454
3455# qhasm: xmm14 ^= xmm9
3456# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
3457# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
3458pxor %xmm12,%xmm11
3459
3460# qhasm: xmm11 = xmm15
3461# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
3462# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
3463movdqa %xmm13,%xmm10
3464
3465# qhasm: xmm11 ^= xmm14
3466# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
3467# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
3468pxor %xmm11,%xmm10
3469
3470# qhasm: xmm11 &= xmm12
3471# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
3472# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
3473pand %xmm8,%xmm10
3474
3475# qhasm: xmm12 ^= xmm8
3476# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
3477# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
3478pxor %xmm9,%xmm8
3479
3480# qhasm: xmm12 &= xmm14
3481# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
3482# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
3483pand %xmm11,%xmm8
3484
3485# qhasm: xmm8 &= xmm15
3486# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
3487# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
3488pand %xmm13,%xmm9
3489
3490# qhasm: xmm8 ^= xmm12
3491# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
3492# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
3493pxor %xmm8,%xmm9
3494
3495# qhasm: xmm12 ^= xmm11
3496# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
3497# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
3498pxor %xmm10,%xmm8
3499
3500# qhasm: xmm10 = xmm13
3501# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
3502# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
3503movdqa %xmm15,%xmm10
3504
3505# qhasm: xmm10 ^= xmm9
3506# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
3507# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
3508pxor %xmm12,%xmm10
3509
3510# qhasm: xmm10 &= xmm0
3511# asm 1: pand <xmm0=int6464#1,<xmm10=int6464#11
3512# asm 2: pand <xmm0=%xmm0,<xmm10=%xmm10
3513pand %xmm0,%xmm10
3514
3515# qhasm: xmm0 ^= xmm3
3516# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
3517# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
3518pxor %xmm3,%xmm0
3519
3520# qhasm: xmm0 &= xmm9
3521# asm 1: pand <xmm9=int6464#13,<xmm0=int6464#1
3522# asm 2: pand <xmm9=%xmm12,<xmm0=%xmm0
3523pand %xmm12,%xmm0
3524
3525# qhasm: xmm3 &= xmm13
3526# asm 1: pand <xmm13=int6464#16,<xmm3=int6464#4
3527# asm 2: pand <xmm13=%xmm15,<xmm3=%xmm3
3528pand %xmm15,%xmm3
3529
3530# qhasm: xmm0 ^= xmm3
3531# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
3532# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
3533pxor %xmm3,%xmm0
3534
3535# qhasm: xmm3 ^= xmm10
3536# asm 1: pxor <xmm10=int6464#11,<xmm3=int6464#4
3537# asm 2: pxor <xmm10=%xmm10,<xmm3=%xmm3
3538pxor %xmm10,%xmm3
3539
3540# qhasm: xmm6 ^= xmm12
3541# asm 1: pxor <xmm12=int6464#9,<xmm6=int6464#7
3542# asm 2: pxor <xmm12=%xmm8,<xmm6=%xmm6
3543pxor %xmm8,%xmm6
3544
3545# qhasm: xmm0 ^= xmm12
3546# asm 1: pxor <xmm12=int6464#9,<xmm0=int6464#1
3547# asm 2: pxor <xmm12=%xmm8,<xmm0=%xmm0
3548pxor %xmm8,%xmm0
3549
3550# qhasm: xmm5 ^= xmm8
3551# asm 1: pxor <xmm8=int6464#10,<xmm5=int6464#6
3552# asm 2: pxor <xmm8=%xmm9,<xmm5=%xmm5
3553pxor %xmm9,%xmm5
3554
3555# qhasm: xmm3 ^= xmm8
3556# asm 1: pxor <xmm8=int6464#10,<xmm3=int6464#4
3557# asm 2: pxor <xmm8=%xmm9,<xmm3=%xmm3
3558pxor %xmm9,%xmm3
3559
3560# qhasm: xmm12 = xmm7
3561# asm 1: movdqa <xmm7=int6464#8,>xmm12=int6464#9
3562# asm 2: movdqa <xmm7=%xmm7,>xmm12=%xmm8
3563movdqa %xmm7,%xmm8
3564
3565# qhasm: xmm8 = xmm1
3566# asm 1: movdqa <xmm1=int6464#2,>xmm8=int6464#10
3567# asm 2: movdqa <xmm1=%xmm1,>xmm8=%xmm9
3568movdqa %xmm1,%xmm9
3569
3570# qhasm: xmm12 ^= xmm4
3571# asm 1: pxor <xmm4=int6464#5,<xmm12=int6464#9
3572# asm 2: pxor <xmm4=%xmm4,<xmm12=%xmm8
3573pxor %xmm4,%xmm8
3574
3575# qhasm: xmm8 ^= xmm2
3576# asm 1: pxor <xmm2=int6464#3,<xmm8=int6464#10
3577# asm 2: pxor <xmm2=%xmm2,<xmm8=%xmm9
3578pxor %xmm2,%xmm9
3579
3580# qhasm: xmm11 = xmm15
3581# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
3582# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
3583movdqa %xmm13,%xmm10
3584
3585# qhasm: xmm11 ^= xmm14
3586# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
3587# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
3588pxor %xmm11,%xmm10
3589
3590# qhasm: xmm11 &= xmm12
3591# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
3592# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
3593pand %xmm8,%xmm10
3594
3595# qhasm: xmm12 ^= xmm8
3596# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
3597# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
3598pxor %xmm9,%xmm8
3599
3600# qhasm: xmm12 &= xmm14
3601# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
3602# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
3603pand %xmm11,%xmm8
3604
3605# qhasm: xmm8 &= xmm15
3606# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
3607# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
3608pand %xmm13,%xmm9
3609
3610# qhasm: xmm8 ^= xmm12
3611# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
3612# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
3613pxor %xmm8,%xmm9
3614
3615# qhasm: xmm12 ^= xmm11
3616# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
3617# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
3618pxor %xmm10,%xmm8
3619
3620# qhasm: xmm10 = xmm13
3621# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
3622# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
3623movdqa %xmm15,%xmm10
3624
3625# qhasm: xmm10 ^= xmm9
3626# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
3627# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
3628pxor %xmm12,%xmm10
3629
3630# qhasm: xmm10 &= xmm4
3631# asm 1: pand <xmm4=int6464#5,<xmm10=int6464#11
3632# asm 2: pand <xmm4=%xmm4,<xmm10=%xmm10
3633pand %xmm4,%xmm10
3634
3635# qhasm: xmm4 ^= xmm2
3636# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
3637# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
3638pxor %xmm2,%xmm4
3639
3640# qhasm: xmm4 &= xmm9
3641# asm 1: pand <xmm9=int6464#13,<xmm4=int6464#5
3642# asm 2: pand <xmm9=%xmm12,<xmm4=%xmm4
3643pand %xmm12,%xmm4
3644
3645# qhasm: xmm2 &= xmm13
3646# asm 1: pand <xmm13=int6464#16,<xmm2=int6464#3
3647# asm 2: pand <xmm13=%xmm15,<xmm2=%xmm2
3648pand %xmm15,%xmm2
3649
3650# qhasm: xmm4 ^= xmm2
3651# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
3652# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
3653pxor %xmm2,%xmm4
3654
3655# qhasm: xmm2 ^= xmm10
3656# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#3
3657# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm2
3658pxor %xmm10,%xmm2
3659
3660# qhasm: xmm15 ^= xmm13
3661# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
3662# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
3663pxor %xmm15,%xmm13
3664
3665# qhasm: xmm14 ^= xmm9
3666# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
3667# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
3668pxor %xmm12,%xmm11
3669
3670# qhasm: xmm11 = xmm15
3671# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
3672# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
3673movdqa %xmm13,%xmm10
3674
3675# qhasm: xmm11 ^= xmm14
3676# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
3677# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
3678pxor %xmm11,%xmm10
3679
3680# qhasm: xmm11 &= xmm7
3681# asm 1: pand <xmm7=int6464#8,<xmm11=int6464#11
3682# asm 2: pand <xmm7=%xmm7,<xmm11=%xmm10
3683pand %xmm7,%xmm10
3684
3685# qhasm: xmm7 ^= xmm1
3686# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
3687# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
3688pxor %xmm1,%xmm7
3689
3690# qhasm: xmm7 &= xmm14
3691# asm 1: pand <xmm14=int6464#12,<xmm7=int6464#8
3692# asm 2: pand <xmm14=%xmm11,<xmm7=%xmm7
3693pand %xmm11,%xmm7
3694
3695# qhasm: xmm1 &= xmm15
3696# asm 1: pand <xmm15=int6464#14,<xmm1=int6464#2
3697# asm 2: pand <xmm15=%xmm13,<xmm1=%xmm1
3698pand %xmm13,%xmm1
3699
3700# qhasm: xmm7 ^= xmm1
3701# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
3702# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
3703pxor %xmm1,%xmm7
3704
3705# qhasm: xmm1 ^= xmm11
3706# asm 1: pxor <xmm11=int6464#11,<xmm1=int6464#2
3707# asm 2: pxor <xmm11=%xmm10,<xmm1=%xmm1
3708pxor %xmm10,%xmm1
3709
3710# qhasm: xmm7 ^= xmm12
3711# asm 1: pxor <xmm12=int6464#9,<xmm7=int6464#8
3712# asm 2: pxor <xmm12=%xmm8,<xmm7=%xmm7
3713pxor %xmm8,%xmm7
3714
3715# qhasm: xmm4 ^= xmm12
3716# asm 1: pxor <xmm12=int6464#9,<xmm4=int6464#5
3717# asm 2: pxor <xmm12=%xmm8,<xmm4=%xmm4
3718pxor %xmm8,%xmm4
3719
3720# qhasm: xmm1 ^= xmm8
3721# asm 1: pxor <xmm8=int6464#10,<xmm1=int6464#2
3722# asm 2: pxor <xmm8=%xmm9,<xmm1=%xmm1
3723pxor %xmm9,%xmm1
3724
3725# qhasm: xmm2 ^= xmm8
3726# asm 1: pxor <xmm8=int6464#10,<xmm2=int6464#3
3727# asm 2: pxor <xmm8=%xmm9,<xmm2=%xmm2
3728pxor %xmm9,%xmm2
3729
3730# qhasm: xmm7 ^= xmm0
3731# asm 1: pxor <xmm0=int6464#1,<xmm7=int6464#8
3732# asm 2: pxor <xmm0=%xmm0,<xmm7=%xmm7
3733pxor %xmm0,%xmm7
3734
3735# qhasm: xmm1 ^= xmm6
3736# asm 1: pxor <xmm6=int6464#7,<xmm1=int6464#2
3737# asm 2: pxor <xmm6=%xmm6,<xmm1=%xmm1
3738pxor %xmm6,%xmm1
3739
3740# qhasm: xmm4 ^= xmm7
3741# asm 1: pxor <xmm7=int6464#8,<xmm4=int6464#5
3742# asm 2: pxor <xmm7=%xmm7,<xmm4=%xmm4
3743pxor %xmm7,%xmm4
3744
3745# qhasm: xmm6 ^= xmm0
3746# asm 1: pxor <xmm0=int6464#1,<xmm6=int6464#7
3747# asm 2: pxor <xmm0=%xmm0,<xmm6=%xmm6
3748pxor %xmm0,%xmm6
3749
3750# qhasm: xmm0 ^= xmm1
3751# asm 1: pxor <xmm1=int6464#2,<xmm0=int6464#1
3752# asm 2: pxor <xmm1=%xmm1,<xmm0=%xmm0
3753pxor %xmm1,%xmm0
3754
3755# qhasm: xmm1 ^= xmm5
3756# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
3757# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
3758pxor %xmm5,%xmm1
3759
3760# qhasm: xmm5 ^= xmm2
3761# asm 1: pxor <xmm2=int6464#3,<xmm5=int6464#6
3762# asm 2: pxor <xmm2=%xmm2,<xmm5=%xmm5
3763pxor %xmm2,%xmm5
3764
3765# qhasm: xmm4 ^= xmm5
3766# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
3767# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
3768pxor %xmm5,%xmm4
3769
3770# qhasm: xmm2 ^= xmm3
3771# asm 1: pxor <xmm3=int6464#4,<xmm2=int6464#3
3772# asm 2: pxor <xmm3=%xmm3,<xmm2=%xmm2
3773pxor %xmm3,%xmm2
3774
3775# qhasm: xmm3 ^= xmm5
3776# asm 1: pxor <xmm5=int6464#6,<xmm3=int6464#4
3777# asm 2: pxor <xmm5=%xmm5,<xmm3=%xmm3
3778pxor %xmm5,%xmm3
3779
3780# qhasm: xmm6 ^= xmm3
3781# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
3782# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
3783pxor %xmm3,%xmm6
3784
3785# qhasm: xmm8 = shuffle dwords of xmm0 by 0x93
3786# asm 1: pshufd $0x93,<xmm0=int6464#1,>xmm8=int6464#9
3787# asm 2: pshufd $0x93,<xmm0=%xmm0,>xmm8=%xmm8
3788pshufd $0x93,%xmm0,%xmm8
3789
3790# qhasm: xmm9 = shuffle dwords of xmm1 by 0x93
3791# asm 1: pshufd $0x93,<xmm1=int6464#2,>xmm9=int6464#10
3792# asm 2: pshufd $0x93,<xmm1=%xmm1,>xmm9=%xmm9
3793pshufd $0x93,%xmm1,%xmm9
3794
3795# qhasm: xmm10 = shuffle dwords of xmm4 by 0x93
3796# asm 1: pshufd $0x93,<xmm4=int6464#5,>xmm10=int6464#11
3797# asm 2: pshufd $0x93,<xmm4=%xmm4,>xmm10=%xmm10
3798pshufd $0x93,%xmm4,%xmm10
3799
3800# qhasm: xmm11 = shuffle dwords of xmm6 by 0x93
3801# asm 1: pshufd $0x93,<xmm6=int6464#7,>xmm11=int6464#12
3802# asm 2: pshufd $0x93,<xmm6=%xmm6,>xmm11=%xmm11
3803pshufd $0x93,%xmm6,%xmm11
3804
3805# qhasm: xmm12 = shuffle dwords of xmm3 by 0x93
3806# asm 1: pshufd $0x93,<xmm3=int6464#4,>xmm12=int6464#13
3807# asm 2: pshufd $0x93,<xmm3=%xmm3,>xmm12=%xmm12
3808pshufd $0x93,%xmm3,%xmm12
3809
3810# qhasm: xmm13 = shuffle dwords of xmm7 by 0x93
3811# asm 1: pshufd $0x93,<xmm7=int6464#8,>xmm13=int6464#14
3812# asm 2: pshufd $0x93,<xmm7=%xmm7,>xmm13=%xmm13
3813pshufd $0x93,%xmm7,%xmm13
3814
3815# qhasm: xmm14 = shuffle dwords of xmm2 by 0x93
3816# asm 1: pshufd $0x93,<xmm2=int6464#3,>xmm14=int6464#15
3817# asm 2: pshufd $0x93,<xmm2=%xmm2,>xmm14=%xmm14
3818pshufd $0x93,%xmm2,%xmm14
3819
3820# qhasm: xmm15 = shuffle dwords of xmm5 by 0x93
3821# asm 1: pshufd $0x93,<xmm5=int6464#6,>xmm15=int6464#16
3822# asm 2: pshufd $0x93,<xmm5=%xmm5,>xmm15=%xmm15
3823pshufd $0x93,%xmm5,%xmm15
3824
3825# qhasm: xmm0 ^= xmm8
3826# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
3827# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
3828pxor %xmm8,%xmm0
3829
3830# qhasm: xmm1 ^= xmm9
3831# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
3832# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
3833pxor %xmm9,%xmm1
3834
3835# qhasm: xmm4 ^= xmm10
3836# asm 1: pxor <xmm10=int6464#11,<xmm4=int6464#5
3837# asm 2: pxor <xmm10=%xmm10,<xmm4=%xmm4
3838pxor %xmm10,%xmm4
3839
3840# qhasm: xmm6 ^= xmm11
3841# asm 1: pxor <xmm11=int6464#12,<xmm6=int6464#7
3842# asm 2: pxor <xmm11=%xmm11,<xmm6=%xmm6
3843pxor %xmm11,%xmm6
3844
3845# qhasm: xmm3 ^= xmm12
3846# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
3847# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
3848pxor %xmm12,%xmm3
3849
3850# qhasm: xmm7 ^= xmm13
3851# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
3852# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
3853pxor %xmm13,%xmm7
3854
3855# qhasm: xmm2 ^= xmm14
3856# asm 1: pxor <xmm14=int6464#15,<xmm2=int6464#3
3857# asm 2: pxor <xmm14=%xmm14,<xmm2=%xmm2
3858pxor %xmm14,%xmm2
3859
3860# qhasm: xmm5 ^= xmm15
3861# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
3862# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
3863pxor %xmm15,%xmm5
3864
3865# qhasm: xmm8 ^= xmm5
3866# asm 1: pxor <xmm5=int6464#6,<xmm8=int6464#9
3867# asm 2: pxor <xmm5=%xmm5,<xmm8=%xmm8
3868pxor %xmm5,%xmm8
3869
3870# qhasm: xmm9 ^= xmm0
3871# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
3872# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
3873pxor %xmm0,%xmm9
3874
3875# qhasm: xmm10 ^= xmm1
3876# asm 1: pxor <xmm1=int6464#2,<xmm10=int6464#11
3877# asm 2: pxor <xmm1=%xmm1,<xmm10=%xmm10
3878pxor %xmm1,%xmm10
3879
3880# qhasm: xmm9 ^= xmm5
3881# asm 1: pxor <xmm5=int6464#6,<xmm9=int6464#10
3882# asm 2: pxor <xmm5=%xmm5,<xmm9=%xmm9
3883pxor %xmm5,%xmm9
3884
3885# qhasm: xmm11 ^= xmm4
3886# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
3887# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
3888pxor %xmm4,%xmm11
3889
3890# qhasm: xmm12 ^= xmm6
3891# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#13
3892# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm12
3893pxor %xmm6,%xmm12
3894
3895# qhasm: xmm13 ^= xmm3
3896# asm 1: pxor <xmm3=int6464#4,<xmm13=int6464#14
3897# asm 2: pxor <xmm3=%xmm3,<xmm13=%xmm13
3898pxor %xmm3,%xmm13
3899
3900# qhasm: xmm11 ^= xmm5
3901# asm 1: pxor <xmm5=int6464#6,<xmm11=int6464#12
3902# asm 2: pxor <xmm5=%xmm5,<xmm11=%xmm11
3903pxor %xmm5,%xmm11
3904
3905# qhasm: xmm14 ^= xmm7
3906# asm 1: pxor <xmm7=int6464#8,<xmm14=int6464#15
3907# asm 2: pxor <xmm7=%xmm7,<xmm14=%xmm14
3908pxor %xmm7,%xmm14
3909
3910# qhasm: xmm15 ^= xmm2
3911# asm 1: pxor <xmm2=int6464#3,<xmm15=int6464#16
3912# asm 2: pxor <xmm2=%xmm2,<xmm15=%xmm15
3913pxor %xmm2,%xmm15
3914
3915# qhasm: xmm12 ^= xmm5
3916# asm 1: pxor <xmm5=int6464#6,<xmm12=int6464#13
3917# asm 2: pxor <xmm5=%xmm5,<xmm12=%xmm12
3918pxor %xmm5,%xmm12
3919
3920# qhasm: xmm0 = shuffle dwords of xmm0 by 0x4E
3921# asm 1: pshufd $0x4E,<xmm0=int6464#1,>xmm0=int6464#1
3922# asm 2: pshufd $0x4E,<xmm0=%xmm0,>xmm0=%xmm0
3923pshufd $0x4E,%xmm0,%xmm0
3924
3925# qhasm: xmm1 = shuffle dwords of xmm1 by 0x4E
3926# asm 1: pshufd $0x4E,<xmm1=int6464#2,>xmm1=int6464#2
3927# asm 2: pshufd $0x4E,<xmm1=%xmm1,>xmm1=%xmm1
3928pshufd $0x4E,%xmm1,%xmm1
3929
3930# qhasm: xmm4 = shuffle dwords of xmm4 by 0x4E
3931# asm 1: pshufd $0x4E,<xmm4=int6464#5,>xmm4=int6464#5
3932# asm 2: pshufd $0x4E,<xmm4=%xmm4,>xmm4=%xmm4
3933pshufd $0x4E,%xmm4,%xmm4
3934
3935# qhasm: xmm6 = shuffle dwords of xmm6 by 0x4E
3936# asm 1: pshufd $0x4E,<xmm6=int6464#7,>xmm6=int6464#7
3937# asm 2: pshufd $0x4E,<xmm6=%xmm6,>xmm6=%xmm6
3938pshufd $0x4E,%xmm6,%xmm6
3939
3940# qhasm: xmm3 = shuffle dwords of xmm3 by 0x4E
3941# asm 1: pshufd $0x4E,<xmm3=int6464#4,>xmm3=int6464#4
3942# asm 2: pshufd $0x4E,<xmm3=%xmm3,>xmm3=%xmm3
3943pshufd $0x4E,%xmm3,%xmm3
3944
3945# qhasm: xmm7 = shuffle dwords of xmm7 by 0x4E
3946# asm 1: pshufd $0x4E,<xmm7=int6464#8,>xmm7=int6464#8
3947# asm 2: pshufd $0x4E,<xmm7=%xmm7,>xmm7=%xmm7
3948pshufd $0x4E,%xmm7,%xmm7
3949
3950# qhasm: xmm2 = shuffle dwords of xmm2 by 0x4E
3951# asm 1: pshufd $0x4E,<xmm2=int6464#3,>xmm2=int6464#3
3952# asm 2: pshufd $0x4E,<xmm2=%xmm2,>xmm2=%xmm2
3953pshufd $0x4E,%xmm2,%xmm2
3954
3955# qhasm: xmm5 = shuffle dwords of xmm5 by 0x4E
3956# asm 1: pshufd $0x4E,<xmm5=int6464#6,>xmm5=int6464#6
3957# asm 2: pshufd $0x4E,<xmm5=%xmm5,>xmm5=%xmm5
3958pshufd $0x4E,%xmm5,%xmm5
3959
3960# qhasm: xmm8 ^= xmm0
3961# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
3962# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
3963pxor %xmm0,%xmm8
3964
3965# qhasm: xmm9 ^= xmm1
3966# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
3967# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
3968pxor %xmm1,%xmm9
3969
3970# qhasm: xmm10 ^= xmm4
3971# asm 1: pxor <xmm4=int6464#5,<xmm10=int6464#11
3972# asm 2: pxor <xmm4=%xmm4,<xmm10=%xmm10
3973pxor %xmm4,%xmm10
3974
3975# qhasm: xmm11 ^= xmm6
3976# asm 1: pxor <xmm6=int6464#7,<xmm11=int6464#12
3977# asm 2: pxor <xmm6=%xmm6,<xmm11=%xmm11
3978pxor %xmm6,%xmm11
3979
3980# qhasm: xmm12 ^= xmm3
3981# asm 1: pxor <xmm3=int6464#4,<xmm12=int6464#13
3982# asm 2: pxor <xmm3=%xmm3,<xmm12=%xmm12
3983pxor %xmm3,%xmm12
3984
3985# qhasm: xmm13 ^= xmm7
3986# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
3987# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
3988pxor %xmm7,%xmm13
3989
3990# qhasm: xmm14 ^= xmm2
3991# asm 1: pxor <xmm2=int6464#3,<xmm14=int6464#15
3992# asm 2: pxor <xmm2=%xmm2,<xmm14=%xmm14
3993pxor %xmm2,%xmm14
3994
3995# qhasm: xmm15 ^= xmm5
3996# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
3997# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
3998pxor %xmm5,%xmm15
3999
4000# qhasm: xmm8 ^= *(int128 *)(c + 384)
4001# asm 1: pxor 384(<c=int64#4),<xmm8=int6464#9
4002# asm 2: pxor 384(<c=%rcx),<xmm8=%xmm8
4003pxor 384(%rcx),%xmm8
4004
4005# qhasm: shuffle bytes of xmm8 by SR
4006# asm 1: pshufb SR,<xmm8=int6464#9
4007# asm 2: pshufb SR,<xmm8=%xmm8
4008pshufb SR,%xmm8
4009
4010# qhasm: xmm9 ^= *(int128 *)(c + 400)
4011# asm 1: pxor 400(<c=int64#4),<xmm9=int6464#10
4012# asm 2: pxor 400(<c=%rcx),<xmm9=%xmm9
4013pxor 400(%rcx),%xmm9
4014
4015# qhasm: shuffle bytes of xmm9 by SR
4016# asm 1: pshufb SR,<xmm9=int6464#10
4017# asm 2: pshufb SR,<xmm9=%xmm9
4018pshufb SR,%xmm9
4019
4020# qhasm: xmm10 ^= *(int128 *)(c + 416)
4021# asm 1: pxor 416(<c=int64#4),<xmm10=int6464#11
4022# asm 2: pxor 416(<c=%rcx),<xmm10=%xmm10
4023pxor 416(%rcx),%xmm10
4024
4025# qhasm: shuffle bytes of xmm10 by SR
4026# asm 1: pshufb SR,<xmm10=int6464#11
4027# asm 2: pshufb SR,<xmm10=%xmm10
4028pshufb SR,%xmm10
4029
4030# qhasm: xmm11 ^= *(int128 *)(c + 432)
4031# asm 1: pxor 432(<c=int64#4),<xmm11=int6464#12
4032# asm 2: pxor 432(<c=%rcx),<xmm11=%xmm11
4033pxor 432(%rcx),%xmm11
4034
4035# qhasm: shuffle bytes of xmm11 by SR
4036# asm 1: pshufb SR,<xmm11=int6464#12
4037# asm 2: pshufb SR,<xmm11=%xmm11
4038pshufb SR,%xmm11
4039
4040# qhasm: xmm12 ^= *(int128 *)(c + 448)
4041# asm 1: pxor 448(<c=int64#4),<xmm12=int6464#13
4042# asm 2: pxor 448(<c=%rcx),<xmm12=%xmm12
4043pxor 448(%rcx),%xmm12
4044
4045# qhasm: shuffle bytes of xmm12 by SR
4046# asm 1: pshufb SR,<xmm12=int6464#13
4047# asm 2: pshufb SR,<xmm12=%xmm12
4048pshufb SR,%xmm12
4049
4050# qhasm: xmm13 ^= *(int128 *)(c + 464)
4051# asm 1: pxor 464(<c=int64#4),<xmm13=int6464#14
4052# asm 2: pxor 464(<c=%rcx),<xmm13=%xmm13
4053pxor 464(%rcx),%xmm13
4054
4055# qhasm: shuffle bytes of xmm13 by SR
4056# asm 1: pshufb SR,<xmm13=int6464#14
4057# asm 2: pshufb SR,<xmm13=%xmm13
4058pshufb SR,%xmm13
4059
4060# qhasm: xmm14 ^= *(int128 *)(c + 480)
4061# asm 1: pxor 480(<c=int64#4),<xmm14=int6464#15
4062# asm 2: pxor 480(<c=%rcx),<xmm14=%xmm14
4063pxor 480(%rcx),%xmm14
4064
4065# qhasm: shuffle bytes of xmm14 by SR
4066# asm 1: pshufb SR,<xmm14=int6464#15
4067# asm 2: pshufb SR,<xmm14=%xmm14
4068pshufb SR,%xmm14
4069
4070# qhasm: xmm15 ^= *(int128 *)(c + 496)
4071# asm 1: pxor 496(<c=int64#4),<xmm15=int6464#16
4072# asm 2: pxor 496(<c=%rcx),<xmm15=%xmm15
4073pxor 496(%rcx),%xmm15
4074
4075# qhasm: shuffle bytes of xmm15 by SR
4076# asm 1: pshufb SR,<xmm15=int6464#16
4077# asm 2: pshufb SR,<xmm15=%xmm15
4078pshufb SR,%xmm15
4079
4080# qhasm: xmm13 ^= xmm14
4081# asm 1: pxor <xmm14=int6464#15,<xmm13=int6464#14
4082# asm 2: pxor <xmm14=%xmm14,<xmm13=%xmm13
4083pxor %xmm14,%xmm13
4084
4085# qhasm: xmm10 ^= xmm9
4086# asm 1: pxor <xmm9=int6464#10,<xmm10=int6464#11
4087# asm 2: pxor <xmm9=%xmm9,<xmm10=%xmm10
4088pxor %xmm9,%xmm10
4089
4090# qhasm: xmm13 ^= xmm8
4091# asm 1: pxor <xmm8=int6464#9,<xmm13=int6464#14
4092# asm 2: pxor <xmm8=%xmm8,<xmm13=%xmm13
4093pxor %xmm8,%xmm13
4094
4095# qhasm: xmm14 ^= xmm10
4096# asm 1: pxor <xmm10=int6464#11,<xmm14=int6464#15
4097# asm 2: pxor <xmm10=%xmm10,<xmm14=%xmm14
4098pxor %xmm10,%xmm14
4099
4100# qhasm: xmm11 ^= xmm8
4101# asm 1: pxor <xmm8=int6464#9,<xmm11=int6464#12
4102# asm 2: pxor <xmm8=%xmm8,<xmm11=%xmm11
4103pxor %xmm8,%xmm11
4104
4105# qhasm: xmm14 ^= xmm11
4106# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
4107# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
4108pxor %xmm11,%xmm14
4109
4110# qhasm: xmm11 ^= xmm15
4111# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#12
4112# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm11
4113pxor %xmm15,%xmm11
4114
4115# qhasm: xmm11 ^= xmm12
4116# asm 1: pxor <xmm12=int6464#13,<xmm11=int6464#12
4117# asm 2: pxor <xmm12=%xmm12,<xmm11=%xmm11
4118pxor %xmm12,%xmm11
4119
4120# qhasm: xmm15 ^= xmm13
4121# asm 1: pxor <xmm13=int6464#14,<xmm15=int6464#16
4122# asm 2: pxor <xmm13=%xmm13,<xmm15=%xmm15
4123pxor %xmm13,%xmm15
4124
4125# qhasm: xmm11 ^= xmm9
4126# asm 1: pxor <xmm9=int6464#10,<xmm11=int6464#12
4127# asm 2: pxor <xmm9=%xmm9,<xmm11=%xmm11
4128pxor %xmm9,%xmm11
4129
4130# qhasm: xmm12 ^= xmm13
4131# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
4132# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
4133pxor %xmm13,%xmm12
4134
4135# qhasm: xmm10 ^= xmm15
4136# asm 1: pxor <xmm15=int6464#16,<xmm10=int6464#11
4137# asm 2: pxor <xmm15=%xmm15,<xmm10=%xmm10
4138pxor %xmm15,%xmm10
4139
4140# qhasm: xmm9 ^= xmm13
4141# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
4142# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
4143pxor %xmm13,%xmm9
4144
4145# qhasm: xmm3 = xmm15
4146# asm 1: movdqa <xmm15=int6464#16,>xmm3=int6464#1
4147# asm 2: movdqa <xmm15=%xmm15,>xmm3=%xmm0
4148movdqa %xmm15,%xmm0
4149
4150# qhasm: xmm2 = xmm9
4151# asm 1: movdqa <xmm9=int6464#10,>xmm2=int6464#2
4152# asm 2: movdqa <xmm9=%xmm9,>xmm2=%xmm1
4153movdqa %xmm9,%xmm1
4154
4155# qhasm: xmm1 = xmm13
4156# asm 1: movdqa <xmm13=int6464#14,>xmm1=int6464#3
4157# asm 2: movdqa <xmm13=%xmm13,>xmm1=%xmm2
4158movdqa %xmm13,%xmm2
4159
4160# qhasm: xmm5 = xmm10
4161# asm 1: movdqa <xmm10=int6464#11,>xmm5=int6464#4
4162# asm 2: movdqa <xmm10=%xmm10,>xmm5=%xmm3
4163movdqa %xmm10,%xmm3
4164
4165# qhasm: xmm4 = xmm14
4166# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#5
4167# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm4
4168movdqa %xmm14,%xmm4
4169
4170# qhasm: xmm3 ^= xmm12
4171# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#1
4172# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm0
4173pxor %xmm12,%xmm0
4174
4175# qhasm: xmm2 ^= xmm10
4176# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#2
4177# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm1
4178pxor %xmm10,%xmm1
4179
4180# qhasm: xmm1 ^= xmm11
4181# asm 1: pxor <xmm11=int6464#12,<xmm1=int6464#3
4182# asm 2: pxor <xmm11=%xmm11,<xmm1=%xmm2
4183pxor %xmm11,%xmm2
4184
4185# qhasm: xmm5 ^= xmm12
4186# asm 1: pxor <xmm12=int6464#13,<xmm5=int6464#4
4187# asm 2: pxor <xmm12=%xmm12,<xmm5=%xmm3
4188pxor %xmm12,%xmm3
4189
4190# qhasm: xmm4 ^= xmm8
4191# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
4192# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
4193pxor %xmm8,%xmm4
4194
4195# qhasm: xmm6 = xmm3
4196# asm 1: movdqa <xmm3=int6464#1,>xmm6=int6464#6
4197# asm 2: movdqa <xmm3=%xmm0,>xmm6=%xmm5
4198movdqa %xmm0,%xmm5
4199
4200# qhasm: xmm0 = xmm2
4201# asm 1: movdqa <xmm2=int6464#2,>xmm0=int6464#7
4202# asm 2: movdqa <xmm2=%xmm1,>xmm0=%xmm6
4203movdqa %xmm1,%xmm6
4204
4205# qhasm: xmm7 = xmm3
4206# asm 1: movdqa <xmm3=int6464#1,>xmm7=int6464#8
4207# asm 2: movdqa <xmm3=%xmm0,>xmm7=%xmm7
4208movdqa %xmm0,%xmm7
4209
4210# qhasm: xmm2 |= xmm1
4211# asm 1: por <xmm1=int6464#3,<xmm2=int6464#2
4212# asm 2: por <xmm1=%xmm2,<xmm2=%xmm1
4213por %xmm2,%xmm1
4214
4215# qhasm: xmm3 |= xmm4
4216# asm 1: por <xmm4=int6464#5,<xmm3=int6464#1
4217# asm 2: por <xmm4=%xmm4,<xmm3=%xmm0
4218por %xmm4,%xmm0
4219
4220# qhasm: xmm7 ^= xmm0
4221# asm 1: pxor <xmm0=int6464#7,<xmm7=int6464#8
4222# asm 2: pxor <xmm0=%xmm6,<xmm7=%xmm7
4223pxor %xmm6,%xmm7
4224
4225# qhasm: xmm6 &= xmm4
4226# asm 1: pand <xmm4=int6464#5,<xmm6=int6464#6
4227# asm 2: pand <xmm4=%xmm4,<xmm6=%xmm5
4228pand %xmm4,%xmm5
4229
4230# qhasm: xmm0 &= xmm1
4231# asm 1: pand <xmm1=int6464#3,<xmm0=int6464#7
4232# asm 2: pand <xmm1=%xmm2,<xmm0=%xmm6
4233pand %xmm2,%xmm6
4234
4235# qhasm: xmm4 ^= xmm1
4236# asm 1: pxor <xmm1=int6464#3,<xmm4=int6464#5
4237# asm 2: pxor <xmm1=%xmm2,<xmm4=%xmm4
4238pxor %xmm2,%xmm4
4239
4240# qhasm: xmm7 &= xmm4
4241# asm 1: pand <xmm4=int6464#5,<xmm7=int6464#8
4242# asm 2: pand <xmm4=%xmm4,<xmm7=%xmm7
4243pand %xmm4,%xmm7
4244
4245# qhasm: xmm4 = xmm11
4246# asm 1: movdqa <xmm11=int6464#12,>xmm4=int6464#3
4247# asm 2: movdqa <xmm11=%xmm11,>xmm4=%xmm2
4248movdqa %xmm11,%xmm2
4249
4250# qhasm: xmm4 ^= xmm8
4251# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#3
4252# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm2
4253pxor %xmm8,%xmm2
4254
4255# qhasm: xmm5 &= xmm4
4256# asm 1: pand <xmm4=int6464#3,<xmm5=int6464#4
4257# asm 2: pand <xmm4=%xmm2,<xmm5=%xmm3
4258pand %xmm2,%xmm3
4259
4260# qhasm: xmm3 ^= xmm5
4261# asm 1: pxor <xmm5=int6464#4,<xmm3=int6464#1
4262# asm 2: pxor <xmm5=%xmm3,<xmm3=%xmm0
4263pxor %xmm3,%xmm0
4264
4265# qhasm: xmm2 ^= xmm5
4266# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
4267# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
4268pxor %xmm3,%xmm1
4269
4270# qhasm: xmm5 = xmm15
4271# asm 1: movdqa <xmm15=int6464#16,>xmm5=int6464#3
4272# asm 2: movdqa <xmm15=%xmm15,>xmm5=%xmm2
4273movdqa %xmm15,%xmm2
4274
4275# qhasm: xmm5 ^= xmm9
4276# asm 1: pxor <xmm9=int6464#10,<xmm5=int6464#3
4277# asm 2: pxor <xmm9=%xmm9,<xmm5=%xmm2
4278pxor %xmm9,%xmm2
4279
4280# qhasm: xmm4 = xmm13
4281# asm 1: movdqa <xmm13=int6464#14,>xmm4=int6464#4
4282# asm 2: movdqa <xmm13=%xmm13,>xmm4=%xmm3
4283movdqa %xmm13,%xmm3
4284
4285# qhasm: xmm1 = xmm5
4286# asm 1: movdqa <xmm5=int6464#3,>xmm1=int6464#5
4287# asm 2: movdqa <xmm5=%xmm2,>xmm1=%xmm4
4288movdqa %xmm2,%xmm4
4289
4290# qhasm: xmm4 ^= xmm14
4291# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#4
4292# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm3
4293pxor %xmm14,%xmm3
4294
4295# qhasm: xmm1 |= xmm4
4296# asm 1: por <xmm4=int6464#4,<xmm1=int6464#5
4297# asm 2: por <xmm4=%xmm3,<xmm1=%xmm4
4298por %xmm3,%xmm4
4299
4300# qhasm: xmm5 &= xmm4
4301# asm 1: pand <xmm4=int6464#4,<xmm5=int6464#3
4302# asm 2: pand <xmm4=%xmm3,<xmm5=%xmm2
4303pand %xmm3,%xmm2
4304
4305# qhasm: xmm0 ^= xmm5
4306# asm 1: pxor <xmm5=int6464#3,<xmm0=int6464#7
4307# asm 2: pxor <xmm5=%xmm2,<xmm0=%xmm6
4308pxor %xmm2,%xmm6
4309
4310# qhasm: xmm3 ^= xmm7
4311# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#1
4312# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm0
4313pxor %xmm7,%xmm0
4314
4315# qhasm: xmm2 ^= xmm6
4316# asm 1: pxor <xmm6=int6464#6,<xmm2=int6464#2
4317# asm 2: pxor <xmm6=%xmm5,<xmm2=%xmm1
4318pxor %xmm5,%xmm1
4319
4320# qhasm: xmm1 ^= xmm7
4321# asm 1: pxor <xmm7=int6464#8,<xmm1=int6464#5
4322# asm 2: pxor <xmm7=%xmm7,<xmm1=%xmm4
4323pxor %xmm7,%xmm4
4324
4325# qhasm: xmm0 ^= xmm6
4326# asm 1: pxor <xmm6=int6464#6,<xmm0=int6464#7
4327# asm 2: pxor <xmm6=%xmm5,<xmm0=%xmm6
4328pxor %xmm5,%xmm6
4329
4330# qhasm: xmm1 ^= xmm6
4331# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
4332# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
4333pxor %xmm5,%xmm4
4334
4335# qhasm: xmm4 = xmm10
4336# asm 1: movdqa <xmm10=int6464#11,>xmm4=int6464#3
4337# asm 2: movdqa <xmm10=%xmm10,>xmm4=%xmm2
4338movdqa %xmm10,%xmm2
4339
4340# qhasm: xmm5 = xmm12
4341# asm 1: movdqa <xmm12=int6464#13,>xmm5=int6464#4
4342# asm 2: movdqa <xmm12=%xmm12,>xmm5=%xmm3
4343movdqa %xmm12,%xmm3
4344
4345# qhasm: xmm6 = xmm9
4346# asm 1: movdqa <xmm9=int6464#10,>xmm6=int6464#6
4347# asm 2: movdqa <xmm9=%xmm9,>xmm6=%xmm5
4348movdqa %xmm9,%xmm5
4349
4350# qhasm: xmm7 = xmm15
4351# asm 1: movdqa <xmm15=int6464#16,>xmm7=int6464#8
4352# asm 2: movdqa <xmm15=%xmm15,>xmm7=%xmm7
4353movdqa %xmm15,%xmm7
4354
4355# qhasm: xmm4 &= xmm11
4356# asm 1: pand <xmm11=int6464#12,<xmm4=int6464#3
4357# asm 2: pand <xmm11=%xmm11,<xmm4=%xmm2
4358pand %xmm11,%xmm2
4359
4360# qhasm: xmm5 &= xmm8
4361# asm 1: pand <xmm8=int6464#9,<xmm5=int6464#4
4362# asm 2: pand <xmm8=%xmm8,<xmm5=%xmm3
4363pand %xmm8,%xmm3
4364
4365# qhasm: xmm6 &= xmm13
4366# asm 1: pand <xmm13=int6464#14,<xmm6=int6464#6
4367# asm 2: pand <xmm13=%xmm13,<xmm6=%xmm5
4368pand %xmm13,%xmm5
4369
4370# qhasm: xmm7 |= xmm14
4371# asm 1: por <xmm14=int6464#15,<xmm7=int6464#8
4372# asm 2: por <xmm14=%xmm14,<xmm7=%xmm7
4373por %xmm14,%xmm7
4374
4375# qhasm: xmm3 ^= xmm4
4376# asm 1: pxor <xmm4=int6464#3,<xmm3=int6464#1
4377# asm 2: pxor <xmm4=%xmm2,<xmm3=%xmm0
4378pxor %xmm2,%xmm0
4379
4380# qhasm: xmm2 ^= xmm5
4381# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
4382# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
4383pxor %xmm3,%xmm1
4384
4385# qhasm: xmm1 ^= xmm6
4386# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
4387# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
4388pxor %xmm5,%xmm4
4389
4390# qhasm: xmm0 ^= xmm7
4391# asm 1: pxor <xmm7=int6464#8,<xmm0=int6464#7
4392# asm 2: pxor <xmm7=%xmm7,<xmm0=%xmm6
4393pxor %xmm7,%xmm6
4394
4395# qhasm: xmm4 = xmm3
4396# asm 1: movdqa <xmm3=int6464#1,>xmm4=int6464#3
4397# asm 2: movdqa <xmm3=%xmm0,>xmm4=%xmm2
4398movdqa %xmm0,%xmm2
4399
4400# qhasm: xmm4 ^= xmm2
4401# asm 1: pxor <xmm2=int6464#2,<xmm4=int6464#3
4402# asm 2: pxor <xmm2=%xmm1,<xmm4=%xmm2
4403pxor %xmm1,%xmm2
4404
4405# qhasm: xmm3 &= xmm1
4406# asm 1: pand <xmm1=int6464#5,<xmm3=int6464#1
4407# asm 2: pand <xmm1=%xmm4,<xmm3=%xmm0
4408pand %xmm4,%xmm0
4409
4410# qhasm: xmm6 = xmm0
4411# asm 1: movdqa <xmm0=int6464#7,>xmm6=int6464#4
4412# asm 2: movdqa <xmm0=%xmm6,>xmm6=%xmm3
4413movdqa %xmm6,%xmm3
4414
4415# qhasm: xmm6 ^= xmm3
4416# asm 1: pxor <xmm3=int6464#1,<xmm6=int6464#4
4417# asm 2: pxor <xmm3=%xmm0,<xmm6=%xmm3
4418pxor %xmm0,%xmm3
4419
4420# qhasm: xmm7 = xmm4
4421# asm 1: movdqa <xmm4=int6464#3,>xmm7=int6464#6
4422# asm 2: movdqa <xmm4=%xmm2,>xmm7=%xmm5
4423movdqa %xmm2,%xmm5
4424
4425# qhasm: xmm7 &= xmm6
4426# asm 1: pand <xmm6=int6464#4,<xmm7=int6464#6
4427# asm 2: pand <xmm6=%xmm3,<xmm7=%xmm5
4428pand %xmm3,%xmm5
4429
4430# qhasm: xmm7 ^= xmm2
4431# asm 1: pxor <xmm2=int6464#2,<xmm7=int6464#6
4432# asm 2: pxor <xmm2=%xmm1,<xmm7=%xmm5
4433pxor %xmm1,%xmm5
4434
4435# qhasm: xmm5 = xmm1
4436# asm 1: movdqa <xmm1=int6464#5,>xmm5=int6464#8
4437# asm 2: movdqa <xmm1=%xmm4,>xmm5=%xmm7
4438movdqa %xmm4,%xmm7
4439
4440# qhasm: xmm5 ^= xmm0
4441# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
4442# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
4443pxor %xmm6,%xmm7
4444
4445# qhasm: xmm3 ^= xmm2
4446# asm 1: pxor <xmm2=int6464#2,<xmm3=int6464#1
4447# asm 2: pxor <xmm2=%xmm1,<xmm3=%xmm0
4448pxor %xmm1,%xmm0
4449
4450# qhasm: xmm5 &= xmm3
4451# asm 1: pand <xmm3=int6464#1,<xmm5=int6464#8
4452# asm 2: pand <xmm3=%xmm0,<xmm5=%xmm7
4453pand %xmm0,%xmm7
4454
4455# qhasm: xmm5 ^= xmm0
4456# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
4457# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
4458pxor %xmm6,%xmm7
4459
4460# qhasm: xmm1 ^= xmm5
4461# asm 1: pxor <xmm5=int6464#8,<xmm1=int6464#5
4462# asm 2: pxor <xmm5=%xmm7,<xmm1=%xmm4
4463pxor %xmm7,%xmm4
4464
4465# qhasm: xmm2 = xmm6
4466# asm 1: movdqa <xmm6=int6464#4,>xmm2=int6464#1
4467# asm 2: movdqa <xmm6=%xmm3,>xmm2=%xmm0
4468movdqa %xmm3,%xmm0
4469
4470# qhasm: xmm2 ^= xmm5
4471# asm 1: pxor <xmm5=int6464#8,<xmm2=int6464#1
4472# asm 2: pxor <xmm5=%xmm7,<xmm2=%xmm0
4473pxor %xmm7,%xmm0
4474
4475# qhasm: xmm2 &= xmm0
4476# asm 1: pand <xmm0=int6464#7,<xmm2=int6464#1
4477# asm 2: pand <xmm0=%xmm6,<xmm2=%xmm0
4478pand %xmm6,%xmm0
4479
4480# qhasm: xmm1 ^= xmm2
4481# asm 1: pxor <xmm2=int6464#1,<xmm1=int6464#5
4482# asm 2: pxor <xmm2=%xmm0,<xmm1=%xmm4
4483pxor %xmm0,%xmm4
4484
4485# qhasm: xmm6 ^= xmm2
4486# asm 1: pxor <xmm2=int6464#1,<xmm6=int6464#4
4487# asm 2: pxor <xmm2=%xmm0,<xmm6=%xmm3
4488pxor %xmm0,%xmm3
4489
4490# qhasm: xmm6 &= xmm7
4491# asm 1: pand <xmm7=int6464#6,<xmm6=int6464#4
4492# asm 2: pand <xmm7=%xmm5,<xmm6=%xmm3
4493pand %xmm5,%xmm3
4494
4495# qhasm: xmm6 ^= xmm4
4496# asm 1: pxor <xmm4=int6464#3,<xmm6=int6464#4
4497# asm 2: pxor <xmm4=%xmm2,<xmm6=%xmm3
4498pxor %xmm2,%xmm3
4499
4500# qhasm: xmm4 = xmm14
4501# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#1
4502# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm0
4503movdqa %xmm14,%xmm0
4504
4505# qhasm: xmm0 = xmm13
4506# asm 1: movdqa <xmm13=int6464#14,>xmm0=int6464#2
4507# asm 2: movdqa <xmm13=%xmm13,>xmm0=%xmm1
4508movdqa %xmm13,%xmm1
4509
4510# qhasm: xmm2 = xmm7
4511# asm 1: movdqa <xmm7=int6464#6,>xmm2=int6464#3
4512# asm 2: movdqa <xmm7=%xmm5,>xmm2=%xmm2
4513movdqa %xmm5,%xmm2
4514
4515# qhasm: xmm2 ^= xmm6
4516# asm 1: pxor <xmm6=int6464#4,<xmm2=int6464#3
4517# asm 2: pxor <xmm6=%xmm3,<xmm2=%xmm2
4518pxor %xmm3,%xmm2
4519
4520# qhasm: xmm2 &= xmm14
4521# asm 1: pand <xmm14=int6464#15,<xmm2=int6464#3
4522# asm 2: pand <xmm14=%xmm14,<xmm2=%xmm2
4523pand %xmm14,%xmm2
4524
4525# qhasm: xmm14 ^= xmm13
4526# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
4527# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
4528pxor %xmm13,%xmm14
4529
4530# qhasm: xmm14 &= xmm6
4531# asm 1: pand <xmm6=int6464#4,<xmm14=int6464#15
4532# asm 2: pand <xmm6=%xmm3,<xmm14=%xmm14
4533pand %xmm3,%xmm14
4534
4535# qhasm: xmm13 &= xmm7
4536# asm 1: pand <xmm7=int6464#6,<xmm13=int6464#14
4537# asm 2: pand <xmm7=%xmm5,<xmm13=%xmm13
4538pand %xmm5,%xmm13
4539
4540# qhasm: xmm14 ^= xmm13
4541# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
4542# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
4543pxor %xmm13,%xmm14
4544
4545# qhasm: xmm13 ^= xmm2
4546# asm 1: pxor <xmm2=int6464#3,<xmm13=int6464#14
4547# asm 2: pxor <xmm2=%xmm2,<xmm13=%xmm13
4548pxor %xmm2,%xmm13
4549
4550# qhasm: xmm4 ^= xmm8
4551# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#1
4552# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm0
4553pxor %xmm8,%xmm0
4554
4555# qhasm: xmm0 ^= xmm11
4556# asm 1: pxor <xmm11=int6464#12,<xmm0=int6464#2
4557# asm 2: pxor <xmm11=%xmm11,<xmm0=%xmm1
4558pxor %xmm11,%xmm1
4559
4560# qhasm: xmm7 ^= xmm5
4561# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
4562# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
4563pxor %xmm7,%xmm5
4564
4565# qhasm: xmm6 ^= xmm1
4566# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
4567# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
4568pxor %xmm4,%xmm3
4569
4570# qhasm: xmm3 = xmm7
4571# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
4572# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
4573movdqa %xmm5,%xmm2
4574
4575# qhasm: xmm3 ^= xmm6
4576# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
4577# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
4578pxor %xmm3,%xmm2
4579
4580# qhasm: xmm3 &= xmm4
4581# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
4582# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
4583pand %xmm0,%xmm2
4584
4585# qhasm: xmm4 ^= xmm0
4586# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
4587# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
4588pxor %xmm1,%xmm0
4589
4590# qhasm: xmm4 &= xmm6
4591# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
4592# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
4593pand %xmm3,%xmm0
4594
4595# qhasm: xmm0 &= xmm7
4596# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
4597# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
4598pand %xmm5,%xmm1
4599
4600# qhasm: xmm0 ^= xmm4
4601# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
4602# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
4603pxor %xmm0,%xmm1
4604
4605# qhasm: xmm4 ^= xmm3
4606# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
4607# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
4608pxor %xmm2,%xmm0
4609
4610# qhasm: xmm2 = xmm5
4611# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
4612# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
4613movdqa %xmm7,%xmm2
4614
4615# qhasm: xmm2 ^= xmm1
4616# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
4617# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
4618pxor %xmm4,%xmm2
4619
4620# qhasm: xmm2 &= xmm8
4621# asm 1: pand <xmm8=int6464#9,<xmm2=int6464#3
4622# asm 2: pand <xmm8=%xmm8,<xmm2=%xmm2
4623pand %xmm8,%xmm2
4624
4625# qhasm: xmm8 ^= xmm11
4626# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
4627# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
4628pxor %xmm11,%xmm8
4629
4630# qhasm: xmm8 &= xmm1
4631# asm 1: pand <xmm1=int6464#5,<xmm8=int6464#9
4632# asm 2: pand <xmm1=%xmm4,<xmm8=%xmm8
4633pand %xmm4,%xmm8
4634
4635# qhasm: xmm11 &= xmm5
4636# asm 1: pand <xmm5=int6464#8,<xmm11=int6464#12
4637# asm 2: pand <xmm5=%xmm7,<xmm11=%xmm11
4638pand %xmm7,%xmm11
4639
4640# qhasm: xmm8 ^= xmm11
4641# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
4642# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
4643pxor %xmm11,%xmm8
4644
4645# qhasm: xmm11 ^= xmm2
4646# asm 1: pxor <xmm2=int6464#3,<xmm11=int6464#12
4647# asm 2: pxor <xmm2=%xmm2,<xmm11=%xmm11
4648pxor %xmm2,%xmm11
4649
4650# qhasm: xmm14 ^= xmm4
4651# asm 1: pxor <xmm4=int6464#1,<xmm14=int6464#15
4652# asm 2: pxor <xmm4=%xmm0,<xmm14=%xmm14
4653pxor %xmm0,%xmm14
4654
4655# qhasm: xmm8 ^= xmm4
4656# asm 1: pxor <xmm4=int6464#1,<xmm8=int6464#9
4657# asm 2: pxor <xmm4=%xmm0,<xmm8=%xmm8
4658pxor %xmm0,%xmm8
4659
4660# qhasm: xmm13 ^= xmm0
4661# asm 1: pxor <xmm0=int6464#2,<xmm13=int6464#14
4662# asm 2: pxor <xmm0=%xmm1,<xmm13=%xmm13
4663pxor %xmm1,%xmm13
4664
4665# qhasm: xmm11 ^= xmm0
4666# asm 1: pxor <xmm0=int6464#2,<xmm11=int6464#12
4667# asm 2: pxor <xmm0=%xmm1,<xmm11=%xmm11
4668pxor %xmm1,%xmm11
4669
4670# qhasm: xmm4 = xmm15
4671# asm 1: movdqa <xmm15=int6464#16,>xmm4=int6464#1
4672# asm 2: movdqa <xmm15=%xmm15,>xmm4=%xmm0
4673movdqa %xmm15,%xmm0
4674
4675# qhasm: xmm0 = xmm9
4676# asm 1: movdqa <xmm9=int6464#10,>xmm0=int6464#2
4677# asm 2: movdqa <xmm9=%xmm9,>xmm0=%xmm1
4678movdqa %xmm9,%xmm1
4679
4680# qhasm: xmm4 ^= xmm12
4681# asm 1: pxor <xmm12=int6464#13,<xmm4=int6464#1
4682# asm 2: pxor <xmm12=%xmm12,<xmm4=%xmm0
4683pxor %xmm12,%xmm0
4684
4685# qhasm: xmm0 ^= xmm10
4686# asm 1: pxor <xmm10=int6464#11,<xmm0=int6464#2
4687# asm 2: pxor <xmm10=%xmm10,<xmm0=%xmm1
4688pxor %xmm10,%xmm1
4689
4690# qhasm: xmm3 = xmm7
4691# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
4692# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
4693movdqa %xmm5,%xmm2
4694
4695# qhasm: xmm3 ^= xmm6
4696# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
4697# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
4698pxor %xmm3,%xmm2
4699
4700# qhasm: xmm3 &= xmm4
4701# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
4702# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
4703pand %xmm0,%xmm2
4704
4705# qhasm: xmm4 ^= xmm0
4706# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
4707# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
4708pxor %xmm1,%xmm0
4709
4710# qhasm: xmm4 &= xmm6
4711# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
4712# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
4713pand %xmm3,%xmm0
4714
4715# qhasm: xmm0 &= xmm7
4716# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
4717# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
4718pand %xmm5,%xmm1
4719
4720# qhasm: xmm0 ^= xmm4
4721# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
4722# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
4723pxor %xmm0,%xmm1
4724
4725# qhasm: xmm4 ^= xmm3
4726# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
4727# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
4728pxor %xmm2,%xmm0
4729
4730# qhasm: xmm2 = xmm5
4731# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
4732# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
4733movdqa %xmm7,%xmm2
4734
4735# qhasm: xmm2 ^= xmm1
4736# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
4737# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
4738pxor %xmm4,%xmm2
4739
4740# qhasm: xmm2 &= xmm12
4741# asm 1: pand <xmm12=int6464#13,<xmm2=int6464#3
4742# asm 2: pand <xmm12=%xmm12,<xmm2=%xmm2
4743pand %xmm12,%xmm2
4744
4745# qhasm: xmm12 ^= xmm10
4746# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
4747# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
4748pxor %xmm10,%xmm12
4749
4750# qhasm: xmm12 &= xmm1
4751# asm 1: pand <xmm1=int6464#5,<xmm12=int6464#13
4752# asm 2: pand <xmm1=%xmm4,<xmm12=%xmm12
4753pand %xmm4,%xmm12
4754
4755# qhasm: xmm10 &= xmm5
4756# asm 1: pand <xmm5=int6464#8,<xmm10=int6464#11
4757# asm 2: pand <xmm5=%xmm7,<xmm10=%xmm10
4758pand %xmm7,%xmm10
4759
4760# qhasm: xmm12 ^= xmm10
4761# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
4762# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
4763pxor %xmm10,%xmm12
4764
4765# qhasm: xmm10 ^= xmm2
4766# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#11
4767# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm10
4768pxor %xmm2,%xmm10
4769
4770# qhasm: xmm7 ^= xmm5
4771# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
4772# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
4773pxor %xmm7,%xmm5
4774
4775# qhasm: xmm6 ^= xmm1
4776# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
4777# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
4778pxor %xmm4,%xmm3
4779
4780# qhasm: xmm3 = xmm7
4781# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
4782# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
4783movdqa %xmm5,%xmm2
4784
4785# qhasm: xmm3 ^= xmm6
4786# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
4787# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
4788pxor %xmm3,%xmm2
4789
4790# qhasm: xmm3 &= xmm15
4791# asm 1: pand <xmm15=int6464#16,<xmm3=int6464#3
4792# asm 2: pand <xmm15=%xmm15,<xmm3=%xmm2
4793pand %xmm15,%xmm2
4794
4795# qhasm: xmm15 ^= xmm9
4796# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
4797# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
4798pxor %xmm9,%xmm15
4799
4800# qhasm: xmm15 &= xmm6
4801# asm 1: pand <xmm6=int6464#4,<xmm15=int6464#16
4802# asm 2: pand <xmm6=%xmm3,<xmm15=%xmm15
4803pand %xmm3,%xmm15
4804
4805# qhasm: xmm9 &= xmm7
4806# asm 1: pand <xmm7=int6464#6,<xmm9=int6464#10
4807# asm 2: pand <xmm7=%xmm5,<xmm9=%xmm9
4808pand %xmm5,%xmm9
4809
4810# qhasm: xmm15 ^= xmm9
4811# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
4812# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
4813pxor %xmm9,%xmm15
4814
4815# qhasm: xmm9 ^= xmm3
4816# asm 1: pxor <xmm3=int6464#3,<xmm9=int6464#10
4817# asm 2: pxor <xmm3=%xmm2,<xmm9=%xmm9
4818pxor %xmm2,%xmm9
4819
4820# qhasm: xmm15 ^= xmm4
4821# asm 1: pxor <xmm4=int6464#1,<xmm15=int6464#16
4822# asm 2: pxor <xmm4=%xmm0,<xmm15=%xmm15
4823pxor %xmm0,%xmm15
4824
4825# qhasm: xmm12 ^= xmm4
4826# asm 1: pxor <xmm4=int6464#1,<xmm12=int6464#13
4827# asm 2: pxor <xmm4=%xmm0,<xmm12=%xmm12
4828pxor %xmm0,%xmm12
4829
4830# qhasm: xmm9 ^= xmm0
4831# asm 1: pxor <xmm0=int6464#2,<xmm9=int6464#10
4832# asm 2: pxor <xmm0=%xmm1,<xmm9=%xmm9
4833pxor %xmm1,%xmm9
4834
4835# qhasm: xmm10 ^= xmm0
4836# asm 1: pxor <xmm0=int6464#2,<xmm10=int6464#11
4837# asm 2: pxor <xmm0=%xmm1,<xmm10=%xmm10
4838pxor %xmm1,%xmm10
4839
4840# qhasm: xmm15 ^= xmm8
4841# asm 1: pxor <xmm8=int6464#9,<xmm15=int6464#16
4842# asm 2: pxor <xmm8=%xmm8,<xmm15=%xmm15
4843pxor %xmm8,%xmm15
4844
4845# qhasm: xmm9 ^= xmm14
4846# asm 1: pxor <xmm14=int6464#15,<xmm9=int6464#10
4847# asm 2: pxor <xmm14=%xmm14,<xmm9=%xmm9
4848pxor %xmm14,%xmm9
4849
4850# qhasm: xmm12 ^= xmm15
4851# asm 1: pxor <xmm15=int6464#16,<xmm12=int6464#13
4852# asm 2: pxor <xmm15=%xmm15,<xmm12=%xmm12
4853pxor %xmm15,%xmm12
4854
4855# qhasm: xmm14 ^= xmm8
4856# asm 1: pxor <xmm8=int6464#9,<xmm14=int6464#15
4857# asm 2: pxor <xmm8=%xmm8,<xmm14=%xmm14
4858pxor %xmm8,%xmm14
4859
4860# qhasm: xmm8 ^= xmm9
4861# asm 1: pxor <xmm9=int6464#10,<xmm8=int6464#9
4862# asm 2: pxor <xmm9=%xmm9,<xmm8=%xmm8
4863pxor %xmm9,%xmm8
4864
4865# qhasm: xmm9 ^= xmm13
4866# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
4867# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
4868pxor %xmm13,%xmm9
4869
4870# qhasm: xmm13 ^= xmm10
4871# asm 1: pxor <xmm10=int6464#11,<xmm13=int6464#14
4872# asm 2: pxor <xmm10=%xmm10,<xmm13=%xmm13
4873pxor %xmm10,%xmm13
4874
4875# qhasm: xmm12 ^= xmm13
4876# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
4877# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
4878pxor %xmm13,%xmm12
4879
4880# qhasm: xmm10 ^= xmm11
4881# asm 1: pxor <xmm11=int6464#12,<xmm10=int6464#11
4882# asm 2: pxor <xmm11=%xmm11,<xmm10=%xmm10
4883pxor %xmm11,%xmm10
4884
4885# qhasm: xmm11 ^= xmm13
4886# asm 1: pxor <xmm13=int6464#14,<xmm11=int6464#12
4887# asm 2: pxor <xmm13=%xmm13,<xmm11=%xmm11
4888pxor %xmm13,%xmm11
4889
4890# qhasm: xmm14 ^= xmm11
4891# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
4892# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
4893pxor %xmm11,%xmm14
4894
4895# qhasm: xmm0 = shuffle dwords of xmm8 by 0x93
4896# asm 1: pshufd $0x93,<xmm8=int6464#9,>xmm0=int6464#1
4897# asm 2: pshufd $0x93,<xmm8=%xmm8,>xmm0=%xmm0
4898pshufd $0x93,%xmm8,%xmm0
4899
4900# qhasm: xmm1 = shuffle dwords of xmm9 by 0x93
4901# asm 1: pshufd $0x93,<xmm9=int6464#10,>xmm1=int6464#2
4902# asm 2: pshufd $0x93,<xmm9=%xmm9,>xmm1=%xmm1
4903pshufd $0x93,%xmm9,%xmm1
4904
4905# qhasm: xmm2 = shuffle dwords of xmm12 by 0x93
4906# asm 1: pshufd $0x93,<xmm12=int6464#13,>xmm2=int6464#3
4907# asm 2: pshufd $0x93,<xmm12=%xmm12,>xmm2=%xmm2
4908pshufd $0x93,%xmm12,%xmm2
4909
4910# qhasm: xmm3 = shuffle dwords of xmm14 by 0x93
4911# asm 1: pshufd $0x93,<xmm14=int6464#15,>xmm3=int6464#4
4912# asm 2: pshufd $0x93,<xmm14=%xmm14,>xmm3=%xmm3
4913pshufd $0x93,%xmm14,%xmm3
4914
4915# qhasm: xmm4 = shuffle dwords of xmm11 by 0x93
4916# asm 1: pshufd $0x93,<xmm11=int6464#12,>xmm4=int6464#5
4917# asm 2: pshufd $0x93,<xmm11=%xmm11,>xmm4=%xmm4
4918pshufd $0x93,%xmm11,%xmm4
4919
4920# qhasm: xmm5 = shuffle dwords of xmm15 by 0x93
4921# asm 1: pshufd $0x93,<xmm15=int6464#16,>xmm5=int6464#6
4922# asm 2: pshufd $0x93,<xmm15=%xmm15,>xmm5=%xmm5
4923pshufd $0x93,%xmm15,%xmm5
4924
4925# qhasm: xmm6 = shuffle dwords of xmm10 by 0x93
4926# asm 1: pshufd $0x93,<xmm10=int6464#11,>xmm6=int6464#7
4927# asm 2: pshufd $0x93,<xmm10=%xmm10,>xmm6=%xmm6
4928pshufd $0x93,%xmm10,%xmm6
4929
4930# qhasm: xmm7 = shuffle dwords of xmm13 by 0x93
4931# asm 1: pshufd $0x93,<xmm13=int6464#14,>xmm7=int6464#8
4932# asm 2: pshufd $0x93,<xmm13=%xmm13,>xmm7=%xmm7
4933pshufd $0x93,%xmm13,%xmm7
4934
4935# qhasm: xmm8 ^= xmm0
4936# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
4937# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
4938pxor %xmm0,%xmm8
4939
4940# qhasm: xmm9 ^= xmm1
4941# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
4942# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
4943pxor %xmm1,%xmm9
4944
4945# qhasm: xmm12 ^= xmm2
4946# asm 1: pxor <xmm2=int6464#3,<xmm12=int6464#13
4947# asm 2: pxor <xmm2=%xmm2,<xmm12=%xmm12
4948pxor %xmm2,%xmm12
4949
4950# qhasm: xmm14 ^= xmm3
4951# asm 1: pxor <xmm3=int6464#4,<xmm14=int6464#15
4952# asm 2: pxor <xmm3=%xmm3,<xmm14=%xmm14
4953pxor %xmm3,%xmm14
4954
4955# qhasm: xmm11 ^= xmm4
4956# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
4957# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
4958pxor %xmm4,%xmm11
4959
4960# qhasm: xmm15 ^= xmm5
4961# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
4962# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
4963pxor %xmm5,%xmm15
4964
4965# qhasm: xmm10 ^= xmm6
4966# asm 1: pxor <xmm6=int6464#7,<xmm10=int6464#11
4967# asm 2: pxor <xmm6=%xmm6,<xmm10=%xmm10
4968pxor %xmm6,%xmm10
4969
4970# qhasm: xmm13 ^= xmm7
4971# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
4972# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
4973pxor %xmm7,%xmm13
4974
4975# qhasm: xmm0 ^= xmm13
4976# asm 1: pxor <xmm13=int6464#14,<xmm0=int6464#1
4977# asm 2: pxor <xmm13=%xmm13,<xmm0=%xmm0
4978pxor %xmm13,%xmm0
4979
4980# qhasm: xmm1 ^= xmm8
4981# asm 1: pxor <xmm8=int6464#9,<xmm1=int6464#2
4982# asm 2: pxor <xmm8=%xmm8,<xmm1=%xmm1
4983pxor %xmm8,%xmm1
4984
4985# qhasm: xmm2 ^= xmm9
4986# asm 1: pxor <xmm9=int6464#10,<xmm2=int6464#3
4987# asm 2: pxor <xmm9=%xmm9,<xmm2=%xmm2
4988pxor %xmm9,%xmm2
4989
4990# qhasm: xmm1 ^= xmm13
4991# asm 1: pxor <xmm13=int6464#14,<xmm1=int6464#2
4992# asm 2: pxor <xmm13=%xmm13,<xmm1=%xmm1
4993pxor %xmm13,%xmm1
4994
4995# qhasm: xmm3 ^= xmm12
4996# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
4997# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
4998pxor %xmm12,%xmm3
4999
5000# qhasm: xmm4 ^= xmm14
5001# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#5
5002# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm4
5003pxor %xmm14,%xmm4
5004
5005# qhasm: xmm5 ^= xmm11
5006# asm 1: pxor <xmm11=int6464#12,<xmm5=int6464#6
5007# asm 2: pxor <xmm11=%xmm11,<xmm5=%xmm5
5008pxor %xmm11,%xmm5
5009
5010# qhasm: xmm3 ^= xmm13
5011# asm 1: pxor <xmm13=int6464#14,<xmm3=int6464#4
5012# asm 2: pxor <xmm13=%xmm13,<xmm3=%xmm3
5013pxor %xmm13,%xmm3
5014
5015# qhasm: xmm6 ^= xmm15
5016# asm 1: pxor <xmm15=int6464#16,<xmm6=int6464#7
5017# asm 2: pxor <xmm15=%xmm15,<xmm6=%xmm6
5018pxor %xmm15,%xmm6
5019
5020# qhasm: xmm7 ^= xmm10
5021# asm 1: pxor <xmm10=int6464#11,<xmm7=int6464#8
5022# asm 2: pxor <xmm10=%xmm10,<xmm7=%xmm7
5023pxor %xmm10,%xmm7
5024
5025# qhasm: xmm4 ^= xmm13
5026# asm 1: pxor <xmm13=int6464#14,<xmm4=int6464#5
5027# asm 2: pxor <xmm13=%xmm13,<xmm4=%xmm4
5028pxor %xmm13,%xmm4
5029
5030# qhasm: xmm8 = shuffle dwords of xmm8 by 0x4E
5031# asm 1: pshufd $0x4E,<xmm8=int6464#9,>xmm8=int6464#9
5032# asm 2: pshufd $0x4E,<xmm8=%xmm8,>xmm8=%xmm8
5033pshufd $0x4E,%xmm8,%xmm8
5034
5035# qhasm: xmm9 = shuffle dwords of xmm9 by 0x4E
5036# asm 1: pshufd $0x4E,<xmm9=int6464#10,>xmm9=int6464#10
5037# asm 2: pshufd $0x4E,<xmm9=%xmm9,>xmm9=%xmm9
5038pshufd $0x4E,%xmm9,%xmm9
5039
5040# qhasm: xmm12 = shuffle dwords of xmm12 by 0x4E
5041# asm 1: pshufd $0x4E,<xmm12=int6464#13,>xmm12=int6464#13
5042# asm 2: pshufd $0x4E,<xmm12=%xmm12,>xmm12=%xmm12
5043pshufd $0x4E,%xmm12,%xmm12
5044
5045# qhasm: xmm14 = shuffle dwords of xmm14 by 0x4E
5046# asm 1: pshufd $0x4E,<xmm14=int6464#15,>xmm14=int6464#15
5047# asm 2: pshufd $0x4E,<xmm14=%xmm14,>xmm14=%xmm14
5048pshufd $0x4E,%xmm14,%xmm14
5049
5050# qhasm: xmm11 = shuffle dwords of xmm11 by 0x4E
5051# asm 1: pshufd $0x4E,<xmm11=int6464#12,>xmm11=int6464#12
5052# asm 2: pshufd $0x4E,<xmm11=%xmm11,>xmm11=%xmm11
5053pshufd $0x4E,%xmm11,%xmm11
5054
5055# qhasm: xmm15 = shuffle dwords of xmm15 by 0x4E
5056# asm 1: pshufd $0x4E,<xmm15=int6464#16,>xmm15=int6464#16
5057# asm 2: pshufd $0x4E,<xmm15=%xmm15,>xmm15=%xmm15
5058pshufd $0x4E,%xmm15,%xmm15
5059
5060# qhasm: xmm10 = shuffle dwords of xmm10 by 0x4E
5061# asm 1: pshufd $0x4E,<xmm10=int6464#11,>xmm10=int6464#11
5062# asm 2: pshufd $0x4E,<xmm10=%xmm10,>xmm10=%xmm10
5063pshufd $0x4E,%xmm10,%xmm10
5064
5065# qhasm: xmm13 = shuffle dwords of xmm13 by 0x4E
5066# asm 1: pshufd $0x4E,<xmm13=int6464#14,>xmm13=int6464#14
5067# asm 2: pshufd $0x4E,<xmm13=%xmm13,>xmm13=%xmm13
5068pshufd $0x4E,%xmm13,%xmm13
5069
5070# qhasm: xmm0 ^= xmm8
5071# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
5072# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
5073pxor %xmm8,%xmm0
5074
5075# qhasm: xmm1 ^= xmm9
5076# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
5077# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
5078pxor %xmm9,%xmm1
5079
5080# qhasm: xmm2 ^= xmm12
5081# asm 1: pxor <xmm12=int6464#13,<xmm2=int6464#3
5082# asm 2: pxor <xmm12=%xmm12,<xmm2=%xmm2
5083pxor %xmm12,%xmm2
5084
5085# qhasm: xmm3 ^= xmm14
5086# asm 1: pxor <xmm14=int6464#15,<xmm3=int6464#4
5087# asm 2: pxor <xmm14=%xmm14,<xmm3=%xmm3
5088pxor %xmm14,%xmm3
5089
5090# qhasm: xmm4 ^= xmm11
5091# asm 1: pxor <xmm11=int6464#12,<xmm4=int6464#5
5092# asm 2: pxor <xmm11=%xmm11,<xmm4=%xmm4
5093pxor %xmm11,%xmm4
5094
5095# qhasm: xmm5 ^= xmm15
5096# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
5097# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
5098pxor %xmm15,%xmm5
5099
5100# qhasm: xmm6 ^= xmm10
5101# asm 1: pxor <xmm10=int6464#11,<xmm6=int6464#7
5102# asm 2: pxor <xmm10=%xmm10,<xmm6=%xmm6
5103pxor %xmm10,%xmm6
5104
5105# qhasm: xmm7 ^= xmm13
5106# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
5107# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
5108pxor %xmm13,%xmm7
5109
5110# qhasm: xmm0 ^= *(int128 *)(c + 512)
5111# asm 1: pxor 512(<c=int64#4),<xmm0=int6464#1
5112# asm 2: pxor 512(<c=%rcx),<xmm0=%xmm0
5113pxor 512(%rcx),%xmm0
5114
5115# qhasm: shuffle bytes of xmm0 by SR
5116# asm 1: pshufb SR,<xmm0=int6464#1
5117# asm 2: pshufb SR,<xmm0=%xmm0
5118pshufb SR,%xmm0
5119
5120# qhasm: xmm1 ^= *(int128 *)(c + 528)
5121# asm 1: pxor 528(<c=int64#4),<xmm1=int6464#2
5122# asm 2: pxor 528(<c=%rcx),<xmm1=%xmm1
5123pxor 528(%rcx),%xmm1
5124
5125# qhasm: shuffle bytes of xmm1 by SR
5126# asm 1: pshufb SR,<xmm1=int6464#2
5127# asm 2: pshufb SR,<xmm1=%xmm1
5128pshufb SR,%xmm1
5129
5130# qhasm: xmm2 ^= *(int128 *)(c + 544)
5131# asm 1: pxor 544(<c=int64#4),<xmm2=int6464#3
5132# asm 2: pxor 544(<c=%rcx),<xmm2=%xmm2
5133pxor 544(%rcx),%xmm2
5134
5135# qhasm: shuffle bytes of xmm2 by SR
5136# asm 1: pshufb SR,<xmm2=int6464#3
5137# asm 2: pshufb SR,<xmm2=%xmm2
5138pshufb SR,%xmm2
5139
5140# qhasm: xmm3 ^= *(int128 *)(c + 560)
5141# asm 1: pxor 560(<c=int64#4),<xmm3=int6464#4
5142# asm 2: pxor 560(<c=%rcx),<xmm3=%xmm3
5143pxor 560(%rcx),%xmm3
5144
5145# qhasm: shuffle bytes of xmm3 by SR
5146# asm 1: pshufb SR,<xmm3=int6464#4
5147# asm 2: pshufb SR,<xmm3=%xmm3
5148pshufb SR,%xmm3
5149
5150# qhasm: xmm4 ^= *(int128 *)(c + 576)
5151# asm 1: pxor 576(<c=int64#4),<xmm4=int6464#5
5152# asm 2: pxor 576(<c=%rcx),<xmm4=%xmm4
5153pxor 576(%rcx),%xmm4
5154
5155# qhasm: shuffle bytes of xmm4 by SR
5156# asm 1: pshufb SR,<xmm4=int6464#5
5157# asm 2: pshufb SR,<xmm4=%xmm4
5158pshufb SR,%xmm4
5159
5160# qhasm: xmm5 ^= *(int128 *)(c + 592)
5161# asm 1: pxor 592(<c=int64#4),<xmm5=int6464#6
5162# asm 2: pxor 592(<c=%rcx),<xmm5=%xmm5
5163pxor 592(%rcx),%xmm5
5164
5165# qhasm: shuffle bytes of xmm5 by SR
5166# asm 1: pshufb SR,<xmm5=int6464#6
5167# asm 2: pshufb SR,<xmm5=%xmm5
5168pshufb SR,%xmm5
5169
5170# qhasm: xmm6 ^= *(int128 *)(c + 608)
5171# asm 1: pxor 608(<c=int64#4),<xmm6=int6464#7
5172# asm 2: pxor 608(<c=%rcx),<xmm6=%xmm6
5173pxor 608(%rcx),%xmm6
5174
5175# qhasm: shuffle bytes of xmm6 by SR
5176# asm 1: pshufb SR,<xmm6=int6464#7
5177# asm 2: pshufb SR,<xmm6=%xmm6
5178pshufb SR,%xmm6
5179
5180# qhasm: xmm7 ^= *(int128 *)(c + 624)
5181# asm 1: pxor 624(<c=int64#4),<xmm7=int6464#8
5182# asm 2: pxor 624(<c=%rcx),<xmm7=%xmm7
5183pxor 624(%rcx),%xmm7
5184
5185# qhasm: shuffle bytes of xmm7 by SR
5186# asm 1: pshufb SR,<xmm7=int6464#8
5187# asm 2: pshufb SR,<xmm7=%xmm7
5188pshufb SR,%xmm7
5189
5190# qhasm: xmm5 ^= xmm6
5191# asm 1: pxor <xmm6=int6464#7,<xmm5=int6464#6
5192# asm 2: pxor <xmm6=%xmm6,<xmm5=%xmm5
5193pxor %xmm6,%xmm5
5194
5195# qhasm: xmm2 ^= xmm1
5196# asm 1: pxor <xmm1=int6464#2,<xmm2=int6464#3
5197# asm 2: pxor <xmm1=%xmm1,<xmm2=%xmm2
5198pxor %xmm1,%xmm2
5199
5200# qhasm: xmm5 ^= xmm0
5201# asm 1: pxor <xmm0=int6464#1,<xmm5=int6464#6
5202# asm 2: pxor <xmm0=%xmm0,<xmm5=%xmm5
5203pxor %xmm0,%xmm5
5204
5205# qhasm: xmm6 ^= xmm2
5206# asm 1: pxor <xmm2=int6464#3,<xmm6=int6464#7
5207# asm 2: pxor <xmm2=%xmm2,<xmm6=%xmm6
5208pxor %xmm2,%xmm6
5209
5210# qhasm: xmm3 ^= xmm0
5211# asm 1: pxor <xmm0=int6464#1,<xmm3=int6464#4
5212# asm 2: pxor <xmm0=%xmm0,<xmm3=%xmm3
5213pxor %xmm0,%xmm3
5214
5215# qhasm: xmm6 ^= xmm3
5216# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
5217# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
5218pxor %xmm3,%xmm6
5219
5220# qhasm: xmm3 ^= xmm7
5221# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#4
5222# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm3
5223pxor %xmm7,%xmm3
5224
5225# qhasm: xmm3 ^= xmm4
5226# asm 1: pxor <xmm4=int6464#5,<xmm3=int6464#4
5227# asm 2: pxor <xmm4=%xmm4,<xmm3=%xmm3
5228pxor %xmm4,%xmm3
5229
5230# qhasm: xmm7 ^= xmm5
5231# asm 1: pxor <xmm5=int6464#6,<xmm7=int6464#8
5232# asm 2: pxor <xmm5=%xmm5,<xmm7=%xmm7
5233pxor %xmm5,%xmm7
5234
5235# qhasm: xmm3 ^= xmm1
5236# asm 1: pxor <xmm1=int6464#2,<xmm3=int6464#4
5237# asm 2: pxor <xmm1=%xmm1,<xmm3=%xmm3
5238pxor %xmm1,%xmm3
5239
5240# qhasm: xmm4 ^= xmm5
5241# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
5242# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
5243pxor %xmm5,%xmm4
5244
5245# qhasm: xmm2 ^= xmm7
5246# asm 1: pxor <xmm7=int6464#8,<xmm2=int6464#3
5247# asm 2: pxor <xmm7=%xmm7,<xmm2=%xmm2
5248pxor %xmm7,%xmm2
5249
5250# qhasm: xmm1 ^= xmm5
5251# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
5252# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
5253pxor %xmm5,%xmm1
5254
5255# qhasm: xmm11 = xmm7
5256# asm 1: movdqa <xmm7=int6464#8,>xmm11=int6464#9
5257# asm 2: movdqa <xmm7=%xmm7,>xmm11=%xmm8
5258movdqa %xmm7,%xmm8
5259
5260# qhasm: xmm10 = xmm1
5261# asm 1: movdqa <xmm1=int6464#2,>xmm10=int6464#10
5262# asm 2: movdqa <xmm1=%xmm1,>xmm10=%xmm9
5263movdqa %xmm1,%xmm9
5264
5265# qhasm: xmm9 = xmm5
5266# asm 1: movdqa <xmm5=int6464#6,>xmm9=int6464#11
5267# asm 2: movdqa <xmm5=%xmm5,>xmm9=%xmm10
5268movdqa %xmm5,%xmm10
5269
5270# qhasm: xmm13 = xmm2
5271# asm 1: movdqa <xmm2=int6464#3,>xmm13=int6464#12
5272# asm 2: movdqa <xmm2=%xmm2,>xmm13=%xmm11
5273movdqa %xmm2,%xmm11
5274
5275# qhasm: xmm12 = xmm6
5276# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#13
5277# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm12
5278movdqa %xmm6,%xmm12
5279
5280# qhasm: xmm11 ^= xmm4
5281# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#9
5282# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm8
5283pxor %xmm4,%xmm8
5284
5285# qhasm: xmm10 ^= xmm2
5286# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#10
5287# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm9
5288pxor %xmm2,%xmm9
5289
5290# qhasm: xmm9 ^= xmm3
5291# asm 1: pxor <xmm3=int6464#4,<xmm9=int6464#11
5292# asm 2: pxor <xmm3=%xmm3,<xmm9=%xmm10
5293pxor %xmm3,%xmm10
5294
5295# qhasm: xmm13 ^= xmm4
5296# asm 1: pxor <xmm4=int6464#5,<xmm13=int6464#12
5297# asm 2: pxor <xmm4=%xmm4,<xmm13=%xmm11
5298pxor %xmm4,%xmm11
5299
5300# qhasm: xmm12 ^= xmm0
5301# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
5302# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
5303pxor %xmm0,%xmm12
5304
5305# qhasm: xmm14 = xmm11
5306# asm 1: movdqa <xmm11=int6464#9,>xmm14=int6464#14
5307# asm 2: movdqa <xmm11=%xmm8,>xmm14=%xmm13
5308movdqa %xmm8,%xmm13
5309
5310# qhasm: xmm8 = xmm10
5311# asm 1: movdqa <xmm10=int6464#10,>xmm8=int6464#15
5312# asm 2: movdqa <xmm10=%xmm9,>xmm8=%xmm14
5313movdqa %xmm9,%xmm14
5314
5315# qhasm: xmm15 = xmm11
5316# asm 1: movdqa <xmm11=int6464#9,>xmm15=int6464#16
5317# asm 2: movdqa <xmm11=%xmm8,>xmm15=%xmm15
5318movdqa %xmm8,%xmm15
5319
5320# qhasm: xmm10 |= xmm9
5321# asm 1: por <xmm9=int6464#11,<xmm10=int6464#10
5322# asm 2: por <xmm9=%xmm10,<xmm10=%xmm9
5323por %xmm10,%xmm9
5324
5325# qhasm: xmm11 |= xmm12
5326# asm 1: por <xmm12=int6464#13,<xmm11=int6464#9
5327# asm 2: por <xmm12=%xmm12,<xmm11=%xmm8
5328por %xmm12,%xmm8
5329
5330# qhasm: xmm15 ^= xmm8
5331# asm 1: pxor <xmm8=int6464#15,<xmm15=int6464#16
5332# asm 2: pxor <xmm8=%xmm14,<xmm15=%xmm15
5333pxor %xmm14,%xmm15
5334
5335# qhasm: xmm14 &= xmm12
5336# asm 1: pand <xmm12=int6464#13,<xmm14=int6464#14
5337# asm 2: pand <xmm12=%xmm12,<xmm14=%xmm13
5338pand %xmm12,%xmm13
5339
5340# qhasm: xmm8 &= xmm9
5341# asm 1: pand <xmm9=int6464#11,<xmm8=int6464#15
5342# asm 2: pand <xmm9=%xmm10,<xmm8=%xmm14
5343pand %xmm10,%xmm14
5344
5345# qhasm: xmm12 ^= xmm9
5346# asm 1: pxor <xmm9=int6464#11,<xmm12=int6464#13
5347# asm 2: pxor <xmm9=%xmm10,<xmm12=%xmm12
5348pxor %xmm10,%xmm12
5349
5350# qhasm: xmm15 &= xmm12
5351# asm 1: pand <xmm12=int6464#13,<xmm15=int6464#16
5352# asm 2: pand <xmm12=%xmm12,<xmm15=%xmm15
5353pand %xmm12,%xmm15
5354
5355# qhasm: xmm12 = xmm3
5356# asm 1: movdqa <xmm3=int6464#4,>xmm12=int6464#11
5357# asm 2: movdqa <xmm3=%xmm3,>xmm12=%xmm10
5358movdqa %xmm3,%xmm10
5359
5360# qhasm: xmm12 ^= xmm0
5361# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#11
5362# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm10
5363pxor %xmm0,%xmm10
5364
5365# qhasm: xmm13 &= xmm12
5366# asm 1: pand <xmm12=int6464#11,<xmm13=int6464#12
5367# asm 2: pand <xmm12=%xmm10,<xmm13=%xmm11
5368pand %xmm10,%xmm11
5369
5370# qhasm: xmm11 ^= xmm13
5371# asm 1: pxor <xmm13=int6464#12,<xmm11=int6464#9
5372# asm 2: pxor <xmm13=%xmm11,<xmm11=%xmm8
5373pxor %xmm11,%xmm8
5374
5375# qhasm: xmm10 ^= xmm13
5376# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
5377# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
5378pxor %xmm11,%xmm9
5379
5380# qhasm: xmm13 = xmm7
5381# asm 1: movdqa <xmm7=int6464#8,>xmm13=int6464#11
5382# asm 2: movdqa <xmm7=%xmm7,>xmm13=%xmm10
5383movdqa %xmm7,%xmm10
5384
5385# qhasm: xmm13 ^= xmm1
5386# asm 1: pxor <xmm1=int6464#2,<xmm13=int6464#11
5387# asm 2: pxor <xmm1=%xmm1,<xmm13=%xmm10
5388pxor %xmm1,%xmm10
5389
5390# qhasm: xmm12 = xmm5
5391# asm 1: movdqa <xmm5=int6464#6,>xmm12=int6464#12
5392# asm 2: movdqa <xmm5=%xmm5,>xmm12=%xmm11
5393movdqa %xmm5,%xmm11
5394
5395# qhasm: xmm9 = xmm13
5396# asm 1: movdqa <xmm13=int6464#11,>xmm9=int6464#13
5397# asm 2: movdqa <xmm13=%xmm10,>xmm9=%xmm12
5398movdqa %xmm10,%xmm12
5399
5400# qhasm: xmm12 ^= xmm6
5401# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#12
5402# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm11
5403pxor %xmm6,%xmm11
5404
5405# qhasm: xmm9 |= xmm12
5406# asm 1: por <xmm12=int6464#12,<xmm9=int6464#13
5407# asm 2: por <xmm12=%xmm11,<xmm9=%xmm12
5408por %xmm11,%xmm12
5409
5410# qhasm: xmm13 &= xmm12
5411# asm 1: pand <xmm12=int6464#12,<xmm13=int6464#11
5412# asm 2: pand <xmm12=%xmm11,<xmm13=%xmm10
5413pand %xmm11,%xmm10
5414
5415# qhasm: xmm8 ^= xmm13
5416# asm 1: pxor <xmm13=int6464#11,<xmm8=int6464#15
5417# asm 2: pxor <xmm13=%xmm10,<xmm8=%xmm14
5418pxor %xmm10,%xmm14
5419
5420# qhasm: xmm11 ^= xmm15
5421# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#9
5422# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm8
5423pxor %xmm15,%xmm8
5424
5425# qhasm: xmm10 ^= xmm14
5426# asm 1: pxor <xmm14=int6464#14,<xmm10=int6464#10
5427# asm 2: pxor <xmm14=%xmm13,<xmm10=%xmm9
5428pxor %xmm13,%xmm9
5429
5430# qhasm: xmm9 ^= xmm15
5431# asm 1: pxor <xmm15=int6464#16,<xmm9=int6464#13
5432# asm 2: pxor <xmm15=%xmm15,<xmm9=%xmm12
5433pxor %xmm15,%xmm12
5434
5435# qhasm: xmm8 ^= xmm14
5436# asm 1: pxor <xmm14=int6464#14,<xmm8=int6464#15
5437# asm 2: pxor <xmm14=%xmm13,<xmm8=%xmm14
5438pxor %xmm13,%xmm14
5439
5440# qhasm: xmm9 ^= xmm14
5441# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
5442# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
5443pxor %xmm13,%xmm12
5444
5445# qhasm: xmm12 = xmm2
5446# asm 1: movdqa <xmm2=int6464#3,>xmm12=int6464#11
5447# asm 2: movdqa <xmm2=%xmm2,>xmm12=%xmm10
5448movdqa %xmm2,%xmm10
5449
5450# qhasm: xmm13 = xmm4
5451# asm 1: movdqa <xmm4=int6464#5,>xmm13=int6464#12
5452# asm 2: movdqa <xmm4=%xmm4,>xmm13=%xmm11
5453movdqa %xmm4,%xmm11
5454
5455# qhasm: xmm14 = xmm1
5456# asm 1: movdqa <xmm1=int6464#2,>xmm14=int6464#14
5457# asm 2: movdqa <xmm1=%xmm1,>xmm14=%xmm13
5458movdqa %xmm1,%xmm13
5459
5460# qhasm: xmm15 = xmm7
5461# asm 1: movdqa <xmm7=int6464#8,>xmm15=int6464#16
5462# asm 2: movdqa <xmm7=%xmm7,>xmm15=%xmm15
5463movdqa %xmm7,%xmm15
5464
5465# qhasm: xmm12 &= xmm3
5466# asm 1: pand <xmm3=int6464#4,<xmm12=int6464#11
5467# asm 2: pand <xmm3=%xmm3,<xmm12=%xmm10
5468pand %xmm3,%xmm10
5469
5470# qhasm: xmm13 &= xmm0
5471# asm 1: pand <xmm0=int6464#1,<xmm13=int6464#12
5472# asm 2: pand <xmm0=%xmm0,<xmm13=%xmm11
5473pand %xmm0,%xmm11
5474
5475# qhasm: xmm14 &= xmm5
5476# asm 1: pand <xmm5=int6464#6,<xmm14=int6464#14
5477# asm 2: pand <xmm5=%xmm5,<xmm14=%xmm13
5478pand %xmm5,%xmm13
5479
5480# qhasm: xmm15 |= xmm6
5481# asm 1: por <xmm6=int6464#7,<xmm15=int6464#16
5482# asm 2: por <xmm6=%xmm6,<xmm15=%xmm15
5483por %xmm6,%xmm15
5484
5485# qhasm: xmm11 ^= xmm12
5486# asm 1: pxor <xmm12=int6464#11,<xmm11=int6464#9
5487# asm 2: pxor <xmm12=%xmm10,<xmm11=%xmm8
5488pxor %xmm10,%xmm8
5489
5490# qhasm: xmm10 ^= xmm13
5491# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
5492# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
5493pxor %xmm11,%xmm9
5494
5495# qhasm: xmm9 ^= xmm14
5496# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
5497# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
5498pxor %xmm13,%xmm12
5499
5500# qhasm: xmm8 ^= xmm15
5501# asm 1: pxor <xmm15=int6464#16,<xmm8=int6464#15
5502# asm 2: pxor <xmm15=%xmm15,<xmm8=%xmm14
5503pxor %xmm15,%xmm14
5504
5505# qhasm: xmm12 = xmm11
5506# asm 1: movdqa <xmm11=int6464#9,>xmm12=int6464#11
5507# asm 2: movdqa <xmm11=%xmm8,>xmm12=%xmm10
5508movdqa %xmm8,%xmm10
5509
5510# qhasm: xmm12 ^= xmm10
5511# asm 1: pxor <xmm10=int6464#10,<xmm12=int6464#11
5512# asm 2: pxor <xmm10=%xmm9,<xmm12=%xmm10
5513pxor %xmm9,%xmm10
5514
5515# qhasm: xmm11 &= xmm9
5516# asm 1: pand <xmm9=int6464#13,<xmm11=int6464#9
5517# asm 2: pand <xmm9=%xmm12,<xmm11=%xmm8
5518pand %xmm12,%xmm8
5519
5520# qhasm: xmm14 = xmm8
5521# asm 1: movdqa <xmm8=int6464#15,>xmm14=int6464#12
5522# asm 2: movdqa <xmm8=%xmm14,>xmm14=%xmm11
5523movdqa %xmm14,%xmm11
5524
5525# qhasm: xmm14 ^= xmm11
5526# asm 1: pxor <xmm11=int6464#9,<xmm14=int6464#12
5527# asm 2: pxor <xmm11=%xmm8,<xmm14=%xmm11
5528pxor %xmm8,%xmm11
5529
5530# qhasm: xmm15 = xmm12
5531# asm 1: movdqa <xmm12=int6464#11,>xmm15=int6464#14
5532# asm 2: movdqa <xmm12=%xmm10,>xmm15=%xmm13
5533movdqa %xmm10,%xmm13
5534
5535# qhasm: xmm15 &= xmm14
5536# asm 1: pand <xmm14=int6464#12,<xmm15=int6464#14
5537# asm 2: pand <xmm14=%xmm11,<xmm15=%xmm13
5538pand %xmm11,%xmm13
5539
5540# qhasm: xmm15 ^= xmm10
5541# asm 1: pxor <xmm10=int6464#10,<xmm15=int6464#14
5542# asm 2: pxor <xmm10=%xmm9,<xmm15=%xmm13
5543pxor %xmm9,%xmm13
5544
5545# qhasm: xmm13 = xmm9
5546# asm 1: movdqa <xmm9=int6464#13,>xmm13=int6464#16
5547# asm 2: movdqa <xmm9=%xmm12,>xmm13=%xmm15
5548movdqa %xmm12,%xmm15
5549
5550# qhasm: xmm13 ^= xmm8
5551# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
5552# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
5553pxor %xmm14,%xmm15
5554
5555# qhasm: xmm11 ^= xmm10
5556# asm 1: pxor <xmm10=int6464#10,<xmm11=int6464#9
5557# asm 2: pxor <xmm10=%xmm9,<xmm11=%xmm8
5558pxor %xmm9,%xmm8
5559
5560# qhasm: xmm13 &= xmm11
5561# asm 1: pand <xmm11=int6464#9,<xmm13=int6464#16
5562# asm 2: pand <xmm11=%xmm8,<xmm13=%xmm15
5563pand %xmm8,%xmm15
5564
5565# qhasm: xmm13 ^= xmm8
5566# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
5567# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
5568pxor %xmm14,%xmm15
5569
5570# qhasm: xmm9 ^= xmm13
5571# asm 1: pxor <xmm13=int6464#16,<xmm9=int6464#13
5572# asm 2: pxor <xmm13=%xmm15,<xmm9=%xmm12
5573pxor %xmm15,%xmm12
5574
5575# qhasm: xmm10 = xmm14
5576# asm 1: movdqa <xmm14=int6464#12,>xmm10=int6464#9
5577# asm 2: movdqa <xmm14=%xmm11,>xmm10=%xmm8
5578movdqa %xmm11,%xmm8
5579
5580# qhasm: xmm10 ^= xmm13
5581# asm 1: pxor <xmm13=int6464#16,<xmm10=int6464#9
5582# asm 2: pxor <xmm13=%xmm15,<xmm10=%xmm8
5583pxor %xmm15,%xmm8
5584
5585# qhasm: xmm10 &= xmm8
5586# asm 1: pand <xmm8=int6464#15,<xmm10=int6464#9
5587# asm 2: pand <xmm8=%xmm14,<xmm10=%xmm8
5588pand %xmm14,%xmm8
5589
5590# qhasm: xmm9 ^= xmm10
5591# asm 1: pxor <xmm10=int6464#9,<xmm9=int6464#13
5592# asm 2: pxor <xmm10=%xmm8,<xmm9=%xmm12
5593pxor %xmm8,%xmm12
5594
5595# qhasm: xmm14 ^= xmm10
5596# asm 1: pxor <xmm10=int6464#9,<xmm14=int6464#12
5597# asm 2: pxor <xmm10=%xmm8,<xmm14=%xmm11
5598pxor %xmm8,%xmm11
5599
5600# qhasm: xmm14 &= xmm15
5601# asm 1: pand <xmm15=int6464#14,<xmm14=int6464#12
5602# asm 2: pand <xmm15=%xmm13,<xmm14=%xmm11
5603pand %xmm13,%xmm11
5604
5605# qhasm: xmm14 ^= xmm12
5606# asm 1: pxor <xmm12=int6464#11,<xmm14=int6464#12
5607# asm 2: pxor <xmm12=%xmm10,<xmm14=%xmm11
5608pxor %xmm10,%xmm11
5609
5610# qhasm: xmm12 = xmm6
5611# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#9
5612# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm8
5613movdqa %xmm6,%xmm8
5614
5615# qhasm: xmm8 = xmm5
5616# asm 1: movdqa <xmm5=int6464#6,>xmm8=int6464#10
5617# asm 2: movdqa <xmm5=%xmm5,>xmm8=%xmm9
5618movdqa %xmm5,%xmm9
5619
5620# qhasm: xmm10 = xmm15
5621# asm 1: movdqa <xmm15=int6464#14,>xmm10=int6464#11
5622# asm 2: movdqa <xmm15=%xmm13,>xmm10=%xmm10
5623movdqa %xmm13,%xmm10
5624
5625# qhasm: xmm10 ^= xmm14
5626# asm 1: pxor <xmm14=int6464#12,<xmm10=int6464#11
5627# asm 2: pxor <xmm14=%xmm11,<xmm10=%xmm10
5628pxor %xmm11,%xmm10
5629
5630# qhasm: xmm10 &= xmm6
5631# asm 1: pand <xmm6=int6464#7,<xmm10=int6464#11
5632# asm 2: pand <xmm6=%xmm6,<xmm10=%xmm10
5633pand %xmm6,%xmm10
5634
5635# qhasm: xmm6 ^= xmm5
5636# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
5637# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
5638pxor %xmm5,%xmm6
5639
5640# qhasm: xmm6 &= xmm14
5641# asm 1: pand <xmm14=int6464#12,<xmm6=int6464#7
5642# asm 2: pand <xmm14=%xmm11,<xmm6=%xmm6
5643pand %xmm11,%xmm6
5644
5645# qhasm: xmm5 &= xmm15
5646# asm 1: pand <xmm15=int6464#14,<xmm5=int6464#6
5647# asm 2: pand <xmm15=%xmm13,<xmm5=%xmm5
5648pand %xmm13,%xmm5
5649
5650# qhasm: xmm6 ^= xmm5
5651# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
5652# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
5653pxor %xmm5,%xmm6
5654
5655# qhasm: xmm5 ^= xmm10
5656# asm 1: pxor <xmm10=int6464#11,<xmm5=int6464#6
5657# asm 2: pxor <xmm10=%xmm10,<xmm5=%xmm5
5658pxor %xmm10,%xmm5
5659
5660# qhasm: xmm12 ^= xmm0
5661# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#9
5662# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm8
5663pxor %xmm0,%xmm8
5664
5665# qhasm: xmm8 ^= xmm3
5666# asm 1: pxor <xmm3=int6464#4,<xmm8=int6464#10
5667# asm 2: pxor <xmm3=%xmm3,<xmm8=%xmm9
5668pxor %xmm3,%xmm9
5669
5670# qhasm: xmm15 ^= xmm13
5671# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
5672# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
5673pxor %xmm15,%xmm13
5674
5675# qhasm: xmm14 ^= xmm9
5676# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
5677# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
5678pxor %xmm12,%xmm11
5679
5680# qhasm: xmm11 = xmm15
5681# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
5682# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
5683movdqa %xmm13,%xmm10
5684
5685# qhasm: xmm11 ^= xmm14
5686# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
5687# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
5688pxor %xmm11,%xmm10
5689
5690# qhasm: xmm11 &= xmm12
5691# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
5692# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
5693pand %xmm8,%xmm10
5694
5695# qhasm: xmm12 ^= xmm8
5696# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
5697# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
5698pxor %xmm9,%xmm8
5699
5700# qhasm: xmm12 &= xmm14
5701# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
5702# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
5703pand %xmm11,%xmm8
5704
5705# qhasm: xmm8 &= xmm15
5706# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
5707# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
5708pand %xmm13,%xmm9
5709
5710# qhasm: xmm8 ^= xmm12
5711# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
5712# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
5713pxor %xmm8,%xmm9
5714
5715# qhasm: xmm12 ^= xmm11
5716# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
5717# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
5718pxor %xmm10,%xmm8
5719
5720# qhasm: xmm10 = xmm13
5721# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
5722# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
5723movdqa %xmm15,%xmm10
5724
5725# qhasm: xmm10 ^= xmm9
5726# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
5727# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
5728pxor %xmm12,%xmm10
5729
5730# qhasm: xmm10 &= xmm0
5731# asm 1: pand <xmm0=int6464#1,<xmm10=int6464#11
5732# asm 2: pand <xmm0=%xmm0,<xmm10=%xmm10
5733pand %xmm0,%xmm10
5734
5735# qhasm: xmm0 ^= xmm3
5736# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
5737# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
5738pxor %xmm3,%xmm0
5739
5740# qhasm: xmm0 &= xmm9
5741# asm 1: pand <xmm9=int6464#13,<xmm0=int6464#1
5742# asm 2: pand <xmm9=%xmm12,<xmm0=%xmm0
5743pand %xmm12,%xmm0
5744
5745# qhasm: xmm3 &= xmm13
5746# asm 1: pand <xmm13=int6464#16,<xmm3=int6464#4
5747# asm 2: pand <xmm13=%xmm15,<xmm3=%xmm3
5748pand %xmm15,%xmm3
5749
5750# qhasm: xmm0 ^= xmm3
5751# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
5752# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
5753pxor %xmm3,%xmm0
5754
5755# qhasm: xmm3 ^= xmm10
5756# asm 1: pxor <xmm10=int6464#11,<xmm3=int6464#4
5757# asm 2: pxor <xmm10=%xmm10,<xmm3=%xmm3
5758pxor %xmm10,%xmm3
5759
5760# qhasm: xmm6 ^= xmm12
5761# asm 1: pxor <xmm12=int6464#9,<xmm6=int6464#7
5762# asm 2: pxor <xmm12=%xmm8,<xmm6=%xmm6
5763pxor %xmm8,%xmm6
5764
5765# qhasm: xmm0 ^= xmm12
5766# asm 1: pxor <xmm12=int6464#9,<xmm0=int6464#1
5767# asm 2: pxor <xmm12=%xmm8,<xmm0=%xmm0
5768pxor %xmm8,%xmm0
5769
5770# qhasm: xmm5 ^= xmm8
5771# asm 1: pxor <xmm8=int6464#10,<xmm5=int6464#6
5772# asm 2: pxor <xmm8=%xmm9,<xmm5=%xmm5
5773pxor %xmm9,%xmm5
5774
5775# qhasm: xmm3 ^= xmm8
5776# asm 1: pxor <xmm8=int6464#10,<xmm3=int6464#4
5777# asm 2: pxor <xmm8=%xmm9,<xmm3=%xmm3
5778pxor %xmm9,%xmm3
5779
5780# qhasm: xmm12 = xmm7
5781# asm 1: movdqa <xmm7=int6464#8,>xmm12=int6464#9
5782# asm 2: movdqa <xmm7=%xmm7,>xmm12=%xmm8
5783movdqa %xmm7,%xmm8
5784
5785# qhasm: xmm8 = xmm1
5786# asm 1: movdqa <xmm1=int6464#2,>xmm8=int6464#10
5787# asm 2: movdqa <xmm1=%xmm1,>xmm8=%xmm9
5788movdqa %xmm1,%xmm9
5789
5790# qhasm: xmm12 ^= xmm4
5791# asm 1: pxor <xmm4=int6464#5,<xmm12=int6464#9
5792# asm 2: pxor <xmm4=%xmm4,<xmm12=%xmm8
5793pxor %xmm4,%xmm8
5794
5795# qhasm: xmm8 ^= xmm2
5796# asm 1: pxor <xmm2=int6464#3,<xmm8=int6464#10
5797# asm 2: pxor <xmm2=%xmm2,<xmm8=%xmm9
5798pxor %xmm2,%xmm9
5799
5800# qhasm: xmm11 = xmm15
5801# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
5802# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
5803movdqa %xmm13,%xmm10
5804
5805# qhasm: xmm11 ^= xmm14
5806# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
5807# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
5808pxor %xmm11,%xmm10
5809
5810# qhasm: xmm11 &= xmm12
5811# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
5812# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
5813pand %xmm8,%xmm10
5814
5815# qhasm: xmm12 ^= xmm8
5816# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
5817# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
5818pxor %xmm9,%xmm8
5819
5820# qhasm: xmm12 &= xmm14
5821# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
5822# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
5823pand %xmm11,%xmm8
5824
5825# qhasm: xmm8 &= xmm15
5826# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
5827# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
5828pand %xmm13,%xmm9
5829
5830# qhasm: xmm8 ^= xmm12
5831# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
5832# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
5833pxor %xmm8,%xmm9
5834
5835# qhasm: xmm12 ^= xmm11
5836# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
5837# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
5838pxor %xmm10,%xmm8
5839
5840# qhasm: xmm10 = xmm13
5841# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
5842# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
5843movdqa %xmm15,%xmm10
5844
5845# qhasm: xmm10 ^= xmm9
5846# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
5847# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
5848pxor %xmm12,%xmm10
5849
5850# qhasm: xmm10 &= xmm4
5851# asm 1: pand <xmm4=int6464#5,<xmm10=int6464#11
5852# asm 2: pand <xmm4=%xmm4,<xmm10=%xmm10
5853pand %xmm4,%xmm10
5854
5855# qhasm: xmm4 ^= xmm2
5856# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
5857# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
5858pxor %xmm2,%xmm4
5859
5860# qhasm: xmm4 &= xmm9
5861# asm 1: pand <xmm9=int6464#13,<xmm4=int6464#5
5862# asm 2: pand <xmm9=%xmm12,<xmm4=%xmm4
5863pand %xmm12,%xmm4
5864
5865# qhasm: xmm2 &= xmm13
5866# asm 1: pand <xmm13=int6464#16,<xmm2=int6464#3
5867# asm 2: pand <xmm13=%xmm15,<xmm2=%xmm2
5868pand %xmm15,%xmm2
5869
5870# qhasm: xmm4 ^= xmm2
5871# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
5872# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
5873pxor %xmm2,%xmm4
5874
5875# qhasm: xmm2 ^= xmm10
5876# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#3
5877# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm2
5878pxor %xmm10,%xmm2
5879
5880# qhasm: xmm15 ^= xmm13
5881# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
5882# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
5883pxor %xmm15,%xmm13
5884
5885# qhasm: xmm14 ^= xmm9
5886# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
5887# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
5888pxor %xmm12,%xmm11
5889
5890# qhasm: xmm11 = xmm15
5891# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
5892# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
5893movdqa %xmm13,%xmm10
5894
5895# qhasm: xmm11 ^= xmm14
5896# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
5897# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
5898pxor %xmm11,%xmm10
5899
5900# qhasm: xmm11 &= xmm7
5901# asm 1: pand <xmm7=int6464#8,<xmm11=int6464#11
5902# asm 2: pand <xmm7=%xmm7,<xmm11=%xmm10
5903pand %xmm7,%xmm10
5904
5905# qhasm: xmm7 ^= xmm1
5906# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
5907# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
5908pxor %xmm1,%xmm7
5909
5910# qhasm: xmm7 &= xmm14
5911# asm 1: pand <xmm14=int6464#12,<xmm7=int6464#8
5912# asm 2: pand <xmm14=%xmm11,<xmm7=%xmm7
5913pand %xmm11,%xmm7
5914
5915# qhasm: xmm1 &= xmm15
5916# asm 1: pand <xmm15=int6464#14,<xmm1=int6464#2
5917# asm 2: pand <xmm15=%xmm13,<xmm1=%xmm1
5918pand %xmm13,%xmm1
5919
5920# qhasm: xmm7 ^= xmm1
5921# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
5922# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
5923pxor %xmm1,%xmm7
5924
5925# qhasm: xmm1 ^= xmm11
5926# asm 1: pxor <xmm11=int6464#11,<xmm1=int6464#2
5927# asm 2: pxor <xmm11=%xmm10,<xmm1=%xmm1
5928pxor %xmm10,%xmm1
5929
5930# qhasm: xmm7 ^= xmm12
5931# asm 1: pxor <xmm12=int6464#9,<xmm7=int6464#8
5932# asm 2: pxor <xmm12=%xmm8,<xmm7=%xmm7
5933pxor %xmm8,%xmm7
5934
5935# qhasm: xmm4 ^= xmm12
5936# asm 1: pxor <xmm12=int6464#9,<xmm4=int6464#5
5937# asm 2: pxor <xmm12=%xmm8,<xmm4=%xmm4
5938pxor %xmm8,%xmm4
5939
5940# qhasm: xmm1 ^= xmm8
5941# asm 1: pxor <xmm8=int6464#10,<xmm1=int6464#2
5942# asm 2: pxor <xmm8=%xmm9,<xmm1=%xmm1
5943pxor %xmm9,%xmm1
5944
5945# qhasm: xmm2 ^= xmm8
5946# asm 1: pxor <xmm8=int6464#10,<xmm2=int6464#3
5947# asm 2: pxor <xmm8=%xmm9,<xmm2=%xmm2
5948pxor %xmm9,%xmm2
5949
5950# qhasm: xmm7 ^= xmm0
5951# asm 1: pxor <xmm0=int6464#1,<xmm7=int6464#8
5952# asm 2: pxor <xmm0=%xmm0,<xmm7=%xmm7
5953pxor %xmm0,%xmm7
5954
5955# qhasm: xmm1 ^= xmm6
5956# asm 1: pxor <xmm6=int6464#7,<xmm1=int6464#2
5957# asm 2: pxor <xmm6=%xmm6,<xmm1=%xmm1
5958pxor %xmm6,%xmm1
5959
5960# qhasm: xmm4 ^= xmm7
5961# asm 1: pxor <xmm7=int6464#8,<xmm4=int6464#5
5962# asm 2: pxor <xmm7=%xmm7,<xmm4=%xmm4
5963pxor %xmm7,%xmm4
5964
5965# qhasm: xmm6 ^= xmm0
5966# asm 1: pxor <xmm0=int6464#1,<xmm6=int6464#7
5967# asm 2: pxor <xmm0=%xmm0,<xmm6=%xmm6
5968pxor %xmm0,%xmm6
5969
5970# qhasm: xmm0 ^= xmm1
5971# asm 1: pxor <xmm1=int6464#2,<xmm0=int6464#1
5972# asm 2: pxor <xmm1=%xmm1,<xmm0=%xmm0
5973pxor %xmm1,%xmm0
5974
5975# qhasm: xmm1 ^= xmm5
5976# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
5977# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
5978pxor %xmm5,%xmm1
5979
5980# qhasm: xmm5 ^= xmm2
5981# asm 1: pxor <xmm2=int6464#3,<xmm5=int6464#6
5982# asm 2: pxor <xmm2=%xmm2,<xmm5=%xmm5
5983pxor %xmm2,%xmm5
5984
5985# qhasm: xmm4 ^= xmm5
5986# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
5987# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
5988pxor %xmm5,%xmm4
5989
5990# qhasm: xmm2 ^= xmm3
5991# asm 1: pxor <xmm3=int6464#4,<xmm2=int6464#3
5992# asm 2: pxor <xmm3=%xmm3,<xmm2=%xmm2
5993pxor %xmm3,%xmm2
5994
5995# qhasm: xmm3 ^= xmm5
5996# asm 1: pxor <xmm5=int6464#6,<xmm3=int6464#4
5997# asm 2: pxor <xmm5=%xmm5,<xmm3=%xmm3
5998pxor %xmm5,%xmm3
5999
6000# qhasm: xmm6 ^= xmm3
6001# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
6002# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
6003pxor %xmm3,%xmm6
6004
6005# qhasm: xmm8 = shuffle dwords of xmm0 by 0x93
6006# asm 1: pshufd $0x93,<xmm0=int6464#1,>xmm8=int6464#9
6007# asm 2: pshufd $0x93,<xmm0=%xmm0,>xmm8=%xmm8
6008pshufd $0x93,%xmm0,%xmm8
6009
6010# qhasm: xmm9 = shuffle dwords of xmm1 by 0x93
6011# asm 1: pshufd $0x93,<xmm1=int6464#2,>xmm9=int6464#10
6012# asm 2: pshufd $0x93,<xmm1=%xmm1,>xmm9=%xmm9
6013pshufd $0x93,%xmm1,%xmm9
6014
6015# qhasm: xmm10 = shuffle dwords of xmm4 by 0x93
6016# asm 1: pshufd $0x93,<xmm4=int6464#5,>xmm10=int6464#11
6017# asm 2: pshufd $0x93,<xmm4=%xmm4,>xmm10=%xmm10
6018pshufd $0x93,%xmm4,%xmm10
6019
6020# qhasm: xmm11 = shuffle dwords of xmm6 by 0x93
6021# asm 1: pshufd $0x93,<xmm6=int6464#7,>xmm11=int6464#12
6022# asm 2: pshufd $0x93,<xmm6=%xmm6,>xmm11=%xmm11
6023pshufd $0x93,%xmm6,%xmm11
6024
6025# qhasm: xmm12 = shuffle dwords of xmm3 by 0x93
6026# asm 1: pshufd $0x93,<xmm3=int6464#4,>xmm12=int6464#13
6027# asm 2: pshufd $0x93,<xmm3=%xmm3,>xmm12=%xmm12
6028pshufd $0x93,%xmm3,%xmm12
6029
6030# qhasm: xmm13 = shuffle dwords of xmm7 by 0x93
6031# asm 1: pshufd $0x93,<xmm7=int6464#8,>xmm13=int6464#14
6032# asm 2: pshufd $0x93,<xmm7=%xmm7,>xmm13=%xmm13
6033pshufd $0x93,%xmm7,%xmm13
6034
6035# qhasm: xmm14 = shuffle dwords of xmm2 by 0x93
6036# asm 1: pshufd $0x93,<xmm2=int6464#3,>xmm14=int6464#15
6037# asm 2: pshufd $0x93,<xmm2=%xmm2,>xmm14=%xmm14
6038pshufd $0x93,%xmm2,%xmm14
6039
6040# qhasm: xmm15 = shuffle dwords of xmm5 by 0x93
6041# asm 1: pshufd $0x93,<xmm5=int6464#6,>xmm15=int6464#16
6042# asm 2: pshufd $0x93,<xmm5=%xmm5,>xmm15=%xmm15
6043pshufd $0x93,%xmm5,%xmm15
6044
6045# qhasm: xmm0 ^= xmm8
6046# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
6047# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
6048pxor %xmm8,%xmm0
6049
6050# qhasm: xmm1 ^= xmm9
6051# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
6052# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
6053pxor %xmm9,%xmm1
6054
6055# qhasm: xmm4 ^= xmm10
6056# asm 1: pxor <xmm10=int6464#11,<xmm4=int6464#5
6057# asm 2: pxor <xmm10=%xmm10,<xmm4=%xmm4
6058pxor %xmm10,%xmm4
6059
6060# qhasm: xmm6 ^= xmm11
6061# asm 1: pxor <xmm11=int6464#12,<xmm6=int6464#7
6062# asm 2: pxor <xmm11=%xmm11,<xmm6=%xmm6
6063pxor %xmm11,%xmm6
6064
6065# qhasm: xmm3 ^= xmm12
6066# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
6067# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
6068pxor %xmm12,%xmm3
6069
6070# qhasm: xmm7 ^= xmm13
6071# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
6072# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
6073pxor %xmm13,%xmm7
6074
6075# qhasm: xmm2 ^= xmm14
6076# asm 1: pxor <xmm14=int6464#15,<xmm2=int6464#3
6077# asm 2: pxor <xmm14=%xmm14,<xmm2=%xmm2
6078pxor %xmm14,%xmm2
6079
6080# qhasm: xmm5 ^= xmm15
6081# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
6082# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
6083pxor %xmm15,%xmm5
6084
6085# qhasm: xmm8 ^= xmm5
6086# asm 1: pxor <xmm5=int6464#6,<xmm8=int6464#9
6087# asm 2: pxor <xmm5=%xmm5,<xmm8=%xmm8
6088pxor %xmm5,%xmm8
6089
6090# qhasm: xmm9 ^= xmm0
6091# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
6092# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
6093pxor %xmm0,%xmm9
6094
6095# qhasm: xmm10 ^= xmm1
6096# asm 1: pxor <xmm1=int6464#2,<xmm10=int6464#11
6097# asm 2: pxor <xmm1=%xmm1,<xmm10=%xmm10
6098pxor %xmm1,%xmm10
6099
6100# qhasm: xmm9 ^= xmm5
6101# asm 1: pxor <xmm5=int6464#6,<xmm9=int6464#10
6102# asm 2: pxor <xmm5=%xmm5,<xmm9=%xmm9
6103pxor %xmm5,%xmm9
6104
6105# qhasm: xmm11 ^= xmm4
6106# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
6107# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
6108pxor %xmm4,%xmm11
6109
6110# qhasm: xmm12 ^= xmm6
6111# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#13
6112# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm12
6113pxor %xmm6,%xmm12
6114
6115# qhasm: xmm13 ^= xmm3
6116# asm 1: pxor <xmm3=int6464#4,<xmm13=int6464#14
6117# asm 2: pxor <xmm3=%xmm3,<xmm13=%xmm13
6118pxor %xmm3,%xmm13
6119
6120# qhasm: xmm11 ^= xmm5
6121# asm 1: pxor <xmm5=int6464#6,<xmm11=int6464#12
6122# asm 2: pxor <xmm5=%xmm5,<xmm11=%xmm11
6123pxor %xmm5,%xmm11
6124
6125# qhasm: xmm14 ^= xmm7
6126# asm 1: pxor <xmm7=int6464#8,<xmm14=int6464#15
6127# asm 2: pxor <xmm7=%xmm7,<xmm14=%xmm14
6128pxor %xmm7,%xmm14
6129
6130# qhasm: xmm15 ^= xmm2
6131# asm 1: pxor <xmm2=int6464#3,<xmm15=int6464#16
6132# asm 2: pxor <xmm2=%xmm2,<xmm15=%xmm15
6133pxor %xmm2,%xmm15
6134
6135# qhasm: xmm12 ^= xmm5
6136# asm 1: pxor <xmm5=int6464#6,<xmm12=int6464#13
6137# asm 2: pxor <xmm5=%xmm5,<xmm12=%xmm12
6138pxor %xmm5,%xmm12
6139
6140# qhasm: xmm0 = shuffle dwords of xmm0 by 0x4E
6141# asm 1: pshufd $0x4E,<xmm0=int6464#1,>xmm0=int6464#1
6142# asm 2: pshufd $0x4E,<xmm0=%xmm0,>xmm0=%xmm0
6143pshufd $0x4E,%xmm0,%xmm0
6144
6145# qhasm: xmm1 = shuffle dwords of xmm1 by 0x4E
6146# asm 1: pshufd $0x4E,<xmm1=int6464#2,>xmm1=int6464#2
6147# asm 2: pshufd $0x4E,<xmm1=%xmm1,>xmm1=%xmm1
6148pshufd $0x4E,%xmm1,%xmm1
6149
6150# qhasm: xmm4 = shuffle dwords of xmm4 by 0x4E
6151# asm 1: pshufd $0x4E,<xmm4=int6464#5,>xmm4=int6464#5
6152# asm 2: pshufd $0x4E,<xmm4=%xmm4,>xmm4=%xmm4
6153pshufd $0x4E,%xmm4,%xmm4
6154
6155# qhasm: xmm6 = shuffle dwords of xmm6 by 0x4E
6156# asm 1: pshufd $0x4E,<xmm6=int6464#7,>xmm6=int6464#7
6157# asm 2: pshufd $0x4E,<xmm6=%xmm6,>xmm6=%xmm6
6158pshufd $0x4E,%xmm6,%xmm6
6159
6160# qhasm: xmm3 = shuffle dwords of xmm3 by 0x4E
6161# asm 1: pshufd $0x4E,<xmm3=int6464#4,>xmm3=int6464#4
6162# asm 2: pshufd $0x4E,<xmm3=%xmm3,>xmm3=%xmm3
6163pshufd $0x4E,%xmm3,%xmm3
6164
6165# qhasm: xmm7 = shuffle dwords of xmm7 by 0x4E
6166# asm 1: pshufd $0x4E,<xmm7=int6464#8,>xmm7=int6464#8
6167# asm 2: pshufd $0x4E,<xmm7=%xmm7,>xmm7=%xmm7
6168pshufd $0x4E,%xmm7,%xmm7
6169
6170# qhasm: xmm2 = shuffle dwords of xmm2 by 0x4E
6171# asm 1: pshufd $0x4E,<xmm2=int6464#3,>xmm2=int6464#3
6172# asm 2: pshufd $0x4E,<xmm2=%xmm2,>xmm2=%xmm2
6173pshufd $0x4E,%xmm2,%xmm2
6174
6175# qhasm: xmm5 = shuffle dwords of xmm5 by 0x4E
6176# asm 1: pshufd $0x4E,<xmm5=int6464#6,>xmm5=int6464#6
6177# asm 2: pshufd $0x4E,<xmm5=%xmm5,>xmm5=%xmm5
6178pshufd $0x4E,%xmm5,%xmm5
6179
6180# qhasm: xmm8 ^= xmm0
6181# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
6182# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
6183pxor %xmm0,%xmm8
6184
6185# qhasm: xmm9 ^= xmm1
6186# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
6187# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
6188pxor %xmm1,%xmm9
6189
6190# qhasm: xmm10 ^= xmm4
6191# asm 1: pxor <xmm4=int6464#5,<xmm10=int6464#11
6192# asm 2: pxor <xmm4=%xmm4,<xmm10=%xmm10
6193pxor %xmm4,%xmm10
6194
6195# qhasm: xmm11 ^= xmm6
6196# asm 1: pxor <xmm6=int6464#7,<xmm11=int6464#12
6197# asm 2: pxor <xmm6=%xmm6,<xmm11=%xmm11
6198pxor %xmm6,%xmm11
6199
6200# qhasm: xmm12 ^= xmm3
6201# asm 1: pxor <xmm3=int6464#4,<xmm12=int6464#13
6202# asm 2: pxor <xmm3=%xmm3,<xmm12=%xmm12
6203pxor %xmm3,%xmm12
6204
6205# qhasm: xmm13 ^= xmm7
6206# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
6207# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
6208pxor %xmm7,%xmm13
6209
6210# qhasm: xmm14 ^= xmm2
6211# asm 1: pxor <xmm2=int6464#3,<xmm14=int6464#15
6212# asm 2: pxor <xmm2=%xmm2,<xmm14=%xmm14
6213pxor %xmm2,%xmm14
6214
6215# qhasm: xmm15 ^= xmm5
6216# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
6217# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
6218pxor %xmm5,%xmm15
6219
6220# qhasm: xmm8 ^= *(int128 *)(c + 640)
6221# asm 1: pxor 640(<c=int64#4),<xmm8=int6464#9
6222# asm 2: pxor 640(<c=%rcx),<xmm8=%xmm8
6223pxor 640(%rcx),%xmm8
6224
6225# qhasm: shuffle bytes of xmm8 by SR
6226# asm 1: pshufb SR,<xmm8=int6464#9
6227# asm 2: pshufb SR,<xmm8=%xmm8
6228pshufb SR,%xmm8
6229
6230# qhasm: xmm9 ^= *(int128 *)(c + 656)
6231# asm 1: pxor 656(<c=int64#4),<xmm9=int6464#10
6232# asm 2: pxor 656(<c=%rcx),<xmm9=%xmm9
6233pxor 656(%rcx),%xmm9
6234
6235# qhasm: shuffle bytes of xmm9 by SR
6236# asm 1: pshufb SR,<xmm9=int6464#10
6237# asm 2: pshufb SR,<xmm9=%xmm9
6238pshufb SR,%xmm9
6239
6240# qhasm: xmm10 ^= *(int128 *)(c + 672)
6241# asm 1: pxor 672(<c=int64#4),<xmm10=int6464#11
6242# asm 2: pxor 672(<c=%rcx),<xmm10=%xmm10
6243pxor 672(%rcx),%xmm10
6244
6245# qhasm: shuffle bytes of xmm10 by SR
6246# asm 1: pshufb SR,<xmm10=int6464#11
6247# asm 2: pshufb SR,<xmm10=%xmm10
6248pshufb SR,%xmm10
6249
6250# qhasm: xmm11 ^= *(int128 *)(c + 688)
6251# asm 1: pxor 688(<c=int64#4),<xmm11=int6464#12
6252# asm 2: pxor 688(<c=%rcx),<xmm11=%xmm11
6253pxor 688(%rcx),%xmm11
6254
6255# qhasm: shuffle bytes of xmm11 by SR
6256# asm 1: pshufb SR,<xmm11=int6464#12
6257# asm 2: pshufb SR,<xmm11=%xmm11
6258pshufb SR,%xmm11
6259
6260# qhasm: xmm12 ^= *(int128 *)(c + 704)
6261# asm 1: pxor 704(<c=int64#4),<xmm12=int6464#13
6262# asm 2: pxor 704(<c=%rcx),<xmm12=%xmm12
6263pxor 704(%rcx),%xmm12
6264
6265# qhasm: shuffle bytes of xmm12 by SR
6266# asm 1: pshufb SR,<xmm12=int6464#13
6267# asm 2: pshufb SR,<xmm12=%xmm12
6268pshufb SR,%xmm12
6269
6270# qhasm: xmm13 ^= *(int128 *)(c + 720)
6271# asm 1: pxor 720(<c=int64#4),<xmm13=int6464#14
6272# asm 2: pxor 720(<c=%rcx),<xmm13=%xmm13
6273pxor 720(%rcx),%xmm13
6274
6275# qhasm: shuffle bytes of xmm13 by SR
6276# asm 1: pshufb SR,<xmm13=int6464#14
6277# asm 2: pshufb SR,<xmm13=%xmm13
6278pshufb SR,%xmm13
6279
6280# qhasm: xmm14 ^= *(int128 *)(c + 736)
6281# asm 1: pxor 736(<c=int64#4),<xmm14=int6464#15
6282# asm 2: pxor 736(<c=%rcx),<xmm14=%xmm14
6283pxor 736(%rcx),%xmm14
6284
6285# qhasm: shuffle bytes of xmm14 by SR
6286# asm 1: pshufb SR,<xmm14=int6464#15
6287# asm 2: pshufb SR,<xmm14=%xmm14
6288pshufb SR,%xmm14
6289
6290# qhasm: xmm15 ^= *(int128 *)(c + 752)
6291# asm 1: pxor 752(<c=int64#4),<xmm15=int6464#16
6292# asm 2: pxor 752(<c=%rcx),<xmm15=%xmm15
6293pxor 752(%rcx),%xmm15
6294
6295# qhasm: shuffle bytes of xmm15 by SR
6296# asm 1: pshufb SR,<xmm15=int6464#16
6297# asm 2: pshufb SR,<xmm15=%xmm15
6298pshufb SR,%xmm15
6299
6300# qhasm: xmm13 ^= xmm14
6301# asm 1: pxor <xmm14=int6464#15,<xmm13=int6464#14
6302# asm 2: pxor <xmm14=%xmm14,<xmm13=%xmm13
6303pxor %xmm14,%xmm13
6304
6305# qhasm: xmm10 ^= xmm9
6306# asm 1: pxor <xmm9=int6464#10,<xmm10=int6464#11
6307# asm 2: pxor <xmm9=%xmm9,<xmm10=%xmm10
6308pxor %xmm9,%xmm10
6309
6310# qhasm: xmm13 ^= xmm8
6311# asm 1: pxor <xmm8=int6464#9,<xmm13=int6464#14
6312# asm 2: pxor <xmm8=%xmm8,<xmm13=%xmm13
6313pxor %xmm8,%xmm13
6314
6315# qhasm: xmm14 ^= xmm10
6316# asm 1: pxor <xmm10=int6464#11,<xmm14=int6464#15
6317# asm 2: pxor <xmm10=%xmm10,<xmm14=%xmm14
6318pxor %xmm10,%xmm14
6319
6320# qhasm: xmm11 ^= xmm8
6321# asm 1: pxor <xmm8=int6464#9,<xmm11=int6464#12
6322# asm 2: pxor <xmm8=%xmm8,<xmm11=%xmm11
6323pxor %xmm8,%xmm11
6324
6325# qhasm: xmm14 ^= xmm11
6326# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
6327# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
6328pxor %xmm11,%xmm14
6329
6330# qhasm: xmm11 ^= xmm15
6331# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#12
6332# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm11
6333pxor %xmm15,%xmm11
6334
6335# qhasm: xmm11 ^= xmm12
6336# asm 1: pxor <xmm12=int6464#13,<xmm11=int6464#12
6337# asm 2: pxor <xmm12=%xmm12,<xmm11=%xmm11
6338pxor %xmm12,%xmm11
6339
6340# qhasm: xmm15 ^= xmm13
6341# asm 1: pxor <xmm13=int6464#14,<xmm15=int6464#16
6342# asm 2: pxor <xmm13=%xmm13,<xmm15=%xmm15
6343pxor %xmm13,%xmm15
6344
6345# qhasm: xmm11 ^= xmm9
6346# asm 1: pxor <xmm9=int6464#10,<xmm11=int6464#12
6347# asm 2: pxor <xmm9=%xmm9,<xmm11=%xmm11
6348pxor %xmm9,%xmm11
6349
6350# qhasm: xmm12 ^= xmm13
6351# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
6352# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
6353pxor %xmm13,%xmm12
6354
6355# qhasm: xmm10 ^= xmm15
6356# asm 1: pxor <xmm15=int6464#16,<xmm10=int6464#11
6357# asm 2: pxor <xmm15=%xmm15,<xmm10=%xmm10
6358pxor %xmm15,%xmm10
6359
6360# qhasm: xmm9 ^= xmm13
6361# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
6362# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
6363pxor %xmm13,%xmm9
6364
6365# qhasm: xmm3 = xmm15
6366# asm 1: movdqa <xmm15=int6464#16,>xmm3=int6464#1
6367# asm 2: movdqa <xmm15=%xmm15,>xmm3=%xmm0
6368movdqa %xmm15,%xmm0
6369
6370# qhasm: xmm2 = xmm9
6371# asm 1: movdqa <xmm9=int6464#10,>xmm2=int6464#2
6372# asm 2: movdqa <xmm9=%xmm9,>xmm2=%xmm1
6373movdqa %xmm9,%xmm1
6374
6375# qhasm: xmm1 = xmm13
6376# asm 1: movdqa <xmm13=int6464#14,>xmm1=int6464#3
6377# asm 2: movdqa <xmm13=%xmm13,>xmm1=%xmm2
6378movdqa %xmm13,%xmm2
6379
6380# qhasm: xmm5 = xmm10
6381# asm 1: movdqa <xmm10=int6464#11,>xmm5=int6464#4
6382# asm 2: movdqa <xmm10=%xmm10,>xmm5=%xmm3
6383movdqa %xmm10,%xmm3
6384
6385# qhasm: xmm4 = xmm14
6386# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#5
6387# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm4
6388movdqa %xmm14,%xmm4
6389
6390# qhasm: xmm3 ^= xmm12
6391# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#1
6392# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm0
6393pxor %xmm12,%xmm0
6394
6395# qhasm: xmm2 ^= xmm10
6396# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#2
6397# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm1
6398pxor %xmm10,%xmm1
6399
6400# qhasm: xmm1 ^= xmm11
6401# asm 1: pxor <xmm11=int6464#12,<xmm1=int6464#3
6402# asm 2: pxor <xmm11=%xmm11,<xmm1=%xmm2
6403pxor %xmm11,%xmm2
6404
6405# qhasm: xmm5 ^= xmm12
6406# asm 1: pxor <xmm12=int6464#13,<xmm5=int6464#4
6407# asm 2: pxor <xmm12=%xmm12,<xmm5=%xmm3
6408pxor %xmm12,%xmm3
6409
6410# qhasm: xmm4 ^= xmm8
6411# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
6412# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
6413pxor %xmm8,%xmm4
6414
6415# qhasm: xmm6 = xmm3
6416# asm 1: movdqa <xmm3=int6464#1,>xmm6=int6464#6
6417# asm 2: movdqa <xmm3=%xmm0,>xmm6=%xmm5
6418movdqa %xmm0,%xmm5
6419
6420# qhasm: xmm0 = xmm2
6421# asm 1: movdqa <xmm2=int6464#2,>xmm0=int6464#7
6422# asm 2: movdqa <xmm2=%xmm1,>xmm0=%xmm6
6423movdqa %xmm1,%xmm6
6424
6425# qhasm: xmm7 = xmm3
6426# asm 1: movdqa <xmm3=int6464#1,>xmm7=int6464#8
6427# asm 2: movdqa <xmm3=%xmm0,>xmm7=%xmm7
6428movdqa %xmm0,%xmm7
6429
6430# qhasm: xmm2 |= xmm1
6431# asm 1: por <xmm1=int6464#3,<xmm2=int6464#2
6432# asm 2: por <xmm1=%xmm2,<xmm2=%xmm1
6433por %xmm2,%xmm1
6434
6435# qhasm: xmm3 |= xmm4
6436# asm 1: por <xmm4=int6464#5,<xmm3=int6464#1
6437# asm 2: por <xmm4=%xmm4,<xmm3=%xmm0
6438por %xmm4,%xmm0
6439
6440# qhasm: xmm7 ^= xmm0
6441# asm 1: pxor <xmm0=int6464#7,<xmm7=int6464#8
6442# asm 2: pxor <xmm0=%xmm6,<xmm7=%xmm7
6443pxor %xmm6,%xmm7
6444
6445# qhasm: xmm6 &= xmm4
6446# asm 1: pand <xmm4=int6464#5,<xmm6=int6464#6
6447# asm 2: pand <xmm4=%xmm4,<xmm6=%xmm5
6448pand %xmm4,%xmm5
6449
6450# qhasm: xmm0 &= xmm1
6451# asm 1: pand <xmm1=int6464#3,<xmm0=int6464#7
6452# asm 2: pand <xmm1=%xmm2,<xmm0=%xmm6
6453pand %xmm2,%xmm6
6454
6455# qhasm: xmm4 ^= xmm1
6456# asm 1: pxor <xmm1=int6464#3,<xmm4=int6464#5
6457# asm 2: pxor <xmm1=%xmm2,<xmm4=%xmm4
6458pxor %xmm2,%xmm4
6459
6460# qhasm: xmm7 &= xmm4
6461# asm 1: pand <xmm4=int6464#5,<xmm7=int6464#8
6462# asm 2: pand <xmm4=%xmm4,<xmm7=%xmm7
6463pand %xmm4,%xmm7
6464
6465# qhasm: xmm4 = xmm11
6466# asm 1: movdqa <xmm11=int6464#12,>xmm4=int6464#3
6467# asm 2: movdqa <xmm11=%xmm11,>xmm4=%xmm2
6468movdqa %xmm11,%xmm2
6469
6470# qhasm: xmm4 ^= xmm8
6471# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#3
6472# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm2
6473pxor %xmm8,%xmm2
6474
6475# qhasm: xmm5 &= xmm4
6476# asm 1: pand <xmm4=int6464#3,<xmm5=int6464#4
6477# asm 2: pand <xmm4=%xmm2,<xmm5=%xmm3
6478pand %xmm2,%xmm3
6479
6480# qhasm: xmm3 ^= xmm5
6481# asm 1: pxor <xmm5=int6464#4,<xmm3=int6464#1
6482# asm 2: pxor <xmm5=%xmm3,<xmm3=%xmm0
6483pxor %xmm3,%xmm0
6484
6485# qhasm: xmm2 ^= xmm5
6486# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
6487# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
6488pxor %xmm3,%xmm1
6489
6490# qhasm: xmm5 = xmm15
6491# asm 1: movdqa <xmm15=int6464#16,>xmm5=int6464#3
6492# asm 2: movdqa <xmm15=%xmm15,>xmm5=%xmm2
6493movdqa %xmm15,%xmm2
6494
6495# qhasm: xmm5 ^= xmm9
6496# asm 1: pxor <xmm9=int6464#10,<xmm5=int6464#3
6497# asm 2: pxor <xmm9=%xmm9,<xmm5=%xmm2
6498pxor %xmm9,%xmm2
6499
6500# qhasm: xmm4 = xmm13
6501# asm 1: movdqa <xmm13=int6464#14,>xmm4=int6464#4
6502# asm 2: movdqa <xmm13=%xmm13,>xmm4=%xmm3
6503movdqa %xmm13,%xmm3
6504
6505# qhasm: xmm1 = xmm5
6506# asm 1: movdqa <xmm5=int6464#3,>xmm1=int6464#5
6507# asm 2: movdqa <xmm5=%xmm2,>xmm1=%xmm4
6508movdqa %xmm2,%xmm4
6509
6510# qhasm: xmm4 ^= xmm14
6511# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#4
6512# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm3
6513pxor %xmm14,%xmm3
6514
6515# qhasm: xmm1 |= xmm4
6516# asm 1: por <xmm4=int6464#4,<xmm1=int6464#5
6517# asm 2: por <xmm4=%xmm3,<xmm1=%xmm4
6518por %xmm3,%xmm4
6519
6520# qhasm: xmm5 &= xmm4
6521# asm 1: pand <xmm4=int6464#4,<xmm5=int6464#3
6522# asm 2: pand <xmm4=%xmm3,<xmm5=%xmm2
6523pand %xmm3,%xmm2
6524
6525# qhasm: xmm0 ^= xmm5
6526# asm 1: pxor <xmm5=int6464#3,<xmm0=int6464#7
6527# asm 2: pxor <xmm5=%xmm2,<xmm0=%xmm6
6528pxor %xmm2,%xmm6
6529
6530# qhasm: xmm3 ^= xmm7
6531# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#1
6532# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm0
6533pxor %xmm7,%xmm0
6534
6535# qhasm: xmm2 ^= xmm6
6536# asm 1: pxor <xmm6=int6464#6,<xmm2=int6464#2
6537# asm 2: pxor <xmm6=%xmm5,<xmm2=%xmm1
6538pxor %xmm5,%xmm1
6539
6540# qhasm: xmm1 ^= xmm7
6541# asm 1: pxor <xmm7=int6464#8,<xmm1=int6464#5
6542# asm 2: pxor <xmm7=%xmm7,<xmm1=%xmm4
6543pxor %xmm7,%xmm4
6544
6545# qhasm: xmm0 ^= xmm6
6546# asm 1: pxor <xmm6=int6464#6,<xmm0=int6464#7
6547# asm 2: pxor <xmm6=%xmm5,<xmm0=%xmm6
6548pxor %xmm5,%xmm6
6549
6550# qhasm: xmm1 ^= xmm6
6551# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
6552# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
6553pxor %xmm5,%xmm4
6554
6555# qhasm: xmm4 = xmm10
6556# asm 1: movdqa <xmm10=int6464#11,>xmm4=int6464#3
6557# asm 2: movdqa <xmm10=%xmm10,>xmm4=%xmm2
6558movdqa %xmm10,%xmm2
6559
6560# qhasm: xmm5 = xmm12
6561# asm 1: movdqa <xmm12=int6464#13,>xmm5=int6464#4
6562# asm 2: movdqa <xmm12=%xmm12,>xmm5=%xmm3
6563movdqa %xmm12,%xmm3
6564
6565# qhasm: xmm6 = xmm9
6566# asm 1: movdqa <xmm9=int6464#10,>xmm6=int6464#6
6567# asm 2: movdqa <xmm9=%xmm9,>xmm6=%xmm5
6568movdqa %xmm9,%xmm5
6569
6570# qhasm: xmm7 = xmm15
6571# asm 1: movdqa <xmm15=int6464#16,>xmm7=int6464#8
6572# asm 2: movdqa <xmm15=%xmm15,>xmm7=%xmm7
6573movdqa %xmm15,%xmm7
6574
6575# qhasm: xmm4 &= xmm11
6576# asm 1: pand <xmm11=int6464#12,<xmm4=int6464#3
6577# asm 2: pand <xmm11=%xmm11,<xmm4=%xmm2
6578pand %xmm11,%xmm2
6579
6580# qhasm: xmm5 &= xmm8
6581# asm 1: pand <xmm8=int6464#9,<xmm5=int6464#4
6582# asm 2: pand <xmm8=%xmm8,<xmm5=%xmm3
6583pand %xmm8,%xmm3
6584
6585# qhasm: xmm6 &= xmm13
6586# asm 1: pand <xmm13=int6464#14,<xmm6=int6464#6
6587# asm 2: pand <xmm13=%xmm13,<xmm6=%xmm5
6588pand %xmm13,%xmm5
6589
6590# qhasm: xmm7 |= xmm14
6591# asm 1: por <xmm14=int6464#15,<xmm7=int6464#8
6592# asm 2: por <xmm14=%xmm14,<xmm7=%xmm7
6593por %xmm14,%xmm7
6594
6595# qhasm: xmm3 ^= xmm4
6596# asm 1: pxor <xmm4=int6464#3,<xmm3=int6464#1
6597# asm 2: pxor <xmm4=%xmm2,<xmm3=%xmm0
6598pxor %xmm2,%xmm0
6599
6600# qhasm: xmm2 ^= xmm5
6601# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
6602# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
6603pxor %xmm3,%xmm1
6604
6605# qhasm: xmm1 ^= xmm6
6606# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
6607# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
6608pxor %xmm5,%xmm4
6609
6610# qhasm: xmm0 ^= xmm7
6611# asm 1: pxor <xmm7=int6464#8,<xmm0=int6464#7
6612# asm 2: pxor <xmm7=%xmm7,<xmm0=%xmm6
6613pxor %xmm7,%xmm6
6614
6615# qhasm: xmm4 = xmm3
6616# asm 1: movdqa <xmm3=int6464#1,>xmm4=int6464#3
6617# asm 2: movdqa <xmm3=%xmm0,>xmm4=%xmm2
6618movdqa %xmm0,%xmm2
6619
6620# qhasm: xmm4 ^= xmm2
6621# asm 1: pxor <xmm2=int6464#2,<xmm4=int6464#3
6622# asm 2: pxor <xmm2=%xmm1,<xmm4=%xmm2
6623pxor %xmm1,%xmm2
6624
6625# qhasm: xmm3 &= xmm1
6626# asm 1: pand <xmm1=int6464#5,<xmm3=int6464#1
6627# asm 2: pand <xmm1=%xmm4,<xmm3=%xmm0
6628pand %xmm4,%xmm0
6629
6630# qhasm: xmm6 = xmm0
6631# asm 1: movdqa <xmm0=int6464#7,>xmm6=int6464#4
6632# asm 2: movdqa <xmm0=%xmm6,>xmm6=%xmm3
6633movdqa %xmm6,%xmm3
6634
6635# qhasm: xmm6 ^= xmm3
6636# asm 1: pxor <xmm3=int6464#1,<xmm6=int6464#4
6637# asm 2: pxor <xmm3=%xmm0,<xmm6=%xmm3
6638pxor %xmm0,%xmm3
6639
6640# qhasm: xmm7 = xmm4
6641# asm 1: movdqa <xmm4=int6464#3,>xmm7=int6464#6
6642# asm 2: movdqa <xmm4=%xmm2,>xmm7=%xmm5
6643movdqa %xmm2,%xmm5
6644
6645# qhasm: xmm7 &= xmm6
6646# asm 1: pand <xmm6=int6464#4,<xmm7=int6464#6
6647# asm 2: pand <xmm6=%xmm3,<xmm7=%xmm5
6648pand %xmm3,%xmm5
6649
6650# qhasm: xmm7 ^= xmm2
6651# asm 1: pxor <xmm2=int6464#2,<xmm7=int6464#6
6652# asm 2: pxor <xmm2=%xmm1,<xmm7=%xmm5
6653pxor %xmm1,%xmm5
6654
6655# qhasm: xmm5 = xmm1
6656# asm 1: movdqa <xmm1=int6464#5,>xmm5=int6464#8
6657# asm 2: movdqa <xmm1=%xmm4,>xmm5=%xmm7
6658movdqa %xmm4,%xmm7
6659
6660# qhasm: xmm5 ^= xmm0
6661# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
6662# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
6663pxor %xmm6,%xmm7
6664
6665# qhasm: xmm3 ^= xmm2
6666# asm 1: pxor <xmm2=int6464#2,<xmm3=int6464#1
6667# asm 2: pxor <xmm2=%xmm1,<xmm3=%xmm0
6668pxor %xmm1,%xmm0
6669
6670# qhasm: xmm5 &= xmm3
6671# asm 1: pand <xmm3=int6464#1,<xmm5=int6464#8
6672# asm 2: pand <xmm3=%xmm0,<xmm5=%xmm7
6673pand %xmm0,%xmm7
6674
6675# qhasm: xmm5 ^= xmm0
6676# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
6677# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
6678pxor %xmm6,%xmm7
6679
6680# qhasm: xmm1 ^= xmm5
6681# asm 1: pxor <xmm5=int6464#8,<xmm1=int6464#5
6682# asm 2: pxor <xmm5=%xmm7,<xmm1=%xmm4
6683pxor %xmm7,%xmm4
6684
6685# qhasm: xmm2 = xmm6
6686# asm 1: movdqa <xmm6=int6464#4,>xmm2=int6464#1
6687# asm 2: movdqa <xmm6=%xmm3,>xmm2=%xmm0
6688movdqa %xmm3,%xmm0
6689
6690# qhasm: xmm2 ^= xmm5
6691# asm 1: pxor <xmm5=int6464#8,<xmm2=int6464#1
6692# asm 2: pxor <xmm5=%xmm7,<xmm2=%xmm0
6693pxor %xmm7,%xmm0
6694
6695# qhasm: xmm2 &= xmm0
6696# asm 1: pand <xmm0=int6464#7,<xmm2=int6464#1
6697# asm 2: pand <xmm0=%xmm6,<xmm2=%xmm0
6698pand %xmm6,%xmm0
6699
6700# qhasm: xmm1 ^= xmm2
6701# asm 1: pxor <xmm2=int6464#1,<xmm1=int6464#5
6702# asm 2: pxor <xmm2=%xmm0,<xmm1=%xmm4
6703pxor %xmm0,%xmm4
6704
6705# qhasm: xmm6 ^= xmm2
6706# asm 1: pxor <xmm2=int6464#1,<xmm6=int6464#4
6707# asm 2: pxor <xmm2=%xmm0,<xmm6=%xmm3
6708pxor %xmm0,%xmm3
6709
6710# qhasm: xmm6 &= xmm7
6711# asm 1: pand <xmm7=int6464#6,<xmm6=int6464#4
6712# asm 2: pand <xmm7=%xmm5,<xmm6=%xmm3
6713pand %xmm5,%xmm3
6714
6715# qhasm: xmm6 ^= xmm4
6716# asm 1: pxor <xmm4=int6464#3,<xmm6=int6464#4
6717# asm 2: pxor <xmm4=%xmm2,<xmm6=%xmm3
6718pxor %xmm2,%xmm3
6719
6720# qhasm: xmm4 = xmm14
6721# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#1
6722# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm0
6723movdqa %xmm14,%xmm0
6724
6725# qhasm: xmm0 = xmm13
6726# asm 1: movdqa <xmm13=int6464#14,>xmm0=int6464#2
6727# asm 2: movdqa <xmm13=%xmm13,>xmm0=%xmm1
6728movdqa %xmm13,%xmm1
6729
6730# qhasm: xmm2 = xmm7
6731# asm 1: movdqa <xmm7=int6464#6,>xmm2=int6464#3
6732# asm 2: movdqa <xmm7=%xmm5,>xmm2=%xmm2
6733movdqa %xmm5,%xmm2
6734
6735# qhasm: xmm2 ^= xmm6
6736# asm 1: pxor <xmm6=int6464#4,<xmm2=int6464#3
6737# asm 2: pxor <xmm6=%xmm3,<xmm2=%xmm2
6738pxor %xmm3,%xmm2
6739
6740# qhasm: xmm2 &= xmm14
6741# asm 1: pand <xmm14=int6464#15,<xmm2=int6464#3
6742# asm 2: pand <xmm14=%xmm14,<xmm2=%xmm2
6743pand %xmm14,%xmm2
6744
6745# qhasm: xmm14 ^= xmm13
6746# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
6747# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
6748pxor %xmm13,%xmm14
6749
6750# qhasm: xmm14 &= xmm6
6751# asm 1: pand <xmm6=int6464#4,<xmm14=int6464#15
6752# asm 2: pand <xmm6=%xmm3,<xmm14=%xmm14
6753pand %xmm3,%xmm14
6754
6755# qhasm: xmm13 &= xmm7
6756# asm 1: pand <xmm7=int6464#6,<xmm13=int6464#14
6757# asm 2: pand <xmm7=%xmm5,<xmm13=%xmm13
6758pand %xmm5,%xmm13
6759
6760# qhasm: xmm14 ^= xmm13
6761# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
6762# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
6763pxor %xmm13,%xmm14
6764
6765# qhasm: xmm13 ^= xmm2
6766# asm 1: pxor <xmm2=int6464#3,<xmm13=int6464#14
6767# asm 2: pxor <xmm2=%xmm2,<xmm13=%xmm13
6768pxor %xmm2,%xmm13
6769
6770# qhasm: xmm4 ^= xmm8
6771# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#1
6772# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm0
6773pxor %xmm8,%xmm0
6774
6775# qhasm: xmm0 ^= xmm11
6776# asm 1: pxor <xmm11=int6464#12,<xmm0=int6464#2
6777# asm 2: pxor <xmm11=%xmm11,<xmm0=%xmm1
6778pxor %xmm11,%xmm1
6779
6780# qhasm: xmm7 ^= xmm5
6781# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
6782# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
6783pxor %xmm7,%xmm5
6784
6785# qhasm: xmm6 ^= xmm1
6786# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
6787# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
6788pxor %xmm4,%xmm3
6789
6790# qhasm: xmm3 = xmm7
6791# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
6792# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
6793movdqa %xmm5,%xmm2
6794
6795# qhasm: xmm3 ^= xmm6
6796# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
6797# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
6798pxor %xmm3,%xmm2
6799
6800# qhasm: xmm3 &= xmm4
6801# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
6802# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
6803pand %xmm0,%xmm2
6804
6805# qhasm: xmm4 ^= xmm0
6806# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
6807# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
6808pxor %xmm1,%xmm0
6809
6810# qhasm: xmm4 &= xmm6
6811# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
6812# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
6813pand %xmm3,%xmm0
6814
6815# qhasm: xmm0 &= xmm7
6816# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
6817# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
6818pand %xmm5,%xmm1
6819
6820# qhasm: xmm0 ^= xmm4
6821# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
6822# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
6823pxor %xmm0,%xmm1
6824
6825# qhasm: xmm4 ^= xmm3
6826# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
6827# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
6828pxor %xmm2,%xmm0
6829
6830# qhasm: xmm2 = xmm5
6831# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
6832# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
6833movdqa %xmm7,%xmm2
6834
6835# qhasm: xmm2 ^= xmm1
6836# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
6837# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
6838pxor %xmm4,%xmm2
6839
6840# qhasm: xmm2 &= xmm8
6841# asm 1: pand <xmm8=int6464#9,<xmm2=int6464#3
6842# asm 2: pand <xmm8=%xmm8,<xmm2=%xmm2
6843pand %xmm8,%xmm2
6844
6845# qhasm: xmm8 ^= xmm11
6846# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
6847# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
6848pxor %xmm11,%xmm8
6849
6850# qhasm: xmm8 &= xmm1
6851# asm 1: pand <xmm1=int6464#5,<xmm8=int6464#9
6852# asm 2: pand <xmm1=%xmm4,<xmm8=%xmm8
6853pand %xmm4,%xmm8
6854
6855# qhasm: xmm11 &= xmm5
6856# asm 1: pand <xmm5=int6464#8,<xmm11=int6464#12
6857# asm 2: pand <xmm5=%xmm7,<xmm11=%xmm11
6858pand %xmm7,%xmm11
6859
6860# qhasm: xmm8 ^= xmm11
6861# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
6862# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
6863pxor %xmm11,%xmm8
6864
6865# qhasm: xmm11 ^= xmm2
6866# asm 1: pxor <xmm2=int6464#3,<xmm11=int6464#12
6867# asm 2: pxor <xmm2=%xmm2,<xmm11=%xmm11
6868pxor %xmm2,%xmm11
6869
6870# qhasm: xmm14 ^= xmm4
6871# asm 1: pxor <xmm4=int6464#1,<xmm14=int6464#15
6872# asm 2: pxor <xmm4=%xmm0,<xmm14=%xmm14
6873pxor %xmm0,%xmm14
6874
6875# qhasm: xmm8 ^= xmm4
6876# asm 1: pxor <xmm4=int6464#1,<xmm8=int6464#9
6877# asm 2: pxor <xmm4=%xmm0,<xmm8=%xmm8
6878pxor %xmm0,%xmm8
6879
6880# qhasm: xmm13 ^= xmm0
6881# asm 1: pxor <xmm0=int6464#2,<xmm13=int6464#14
6882# asm 2: pxor <xmm0=%xmm1,<xmm13=%xmm13
6883pxor %xmm1,%xmm13
6884
6885# qhasm: xmm11 ^= xmm0
6886# asm 1: pxor <xmm0=int6464#2,<xmm11=int6464#12
6887# asm 2: pxor <xmm0=%xmm1,<xmm11=%xmm11
6888pxor %xmm1,%xmm11
6889
6890# qhasm: xmm4 = xmm15
6891# asm 1: movdqa <xmm15=int6464#16,>xmm4=int6464#1
6892# asm 2: movdqa <xmm15=%xmm15,>xmm4=%xmm0
6893movdqa %xmm15,%xmm0
6894
6895# qhasm: xmm0 = xmm9
6896# asm 1: movdqa <xmm9=int6464#10,>xmm0=int6464#2
6897# asm 2: movdqa <xmm9=%xmm9,>xmm0=%xmm1
6898movdqa %xmm9,%xmm1
6899
6900# qhasm: xmm4 ^= xmm12
6901# asm 1: pxor <xmm12=int6464#13,<xmm4=int6464#1
6902# asm 2: pxor <xmm12=%xmm12,<xmm4=%xmm0
6903pxor %xmm12,%xmm0
6904
6905# qhasm: xmm0 ^= xmm10
6906# asm 1: pxor <xmm10=int6464#11,<xmm0=int6464#2
6907# asm 2: pxor <xmm10=%xmm10,<xmm0=%xmm1
6908pxor %xmm10,%xmm1
6909
6910# qhasm: xmm3 = xmm7
6911# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
6912# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
6913movdqa %xmm5,%xmm2
6914
6915# qhasm: xmm3 ^= xmm6
6916# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
6917# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
6918pxor %xmm3,%xmm2
6919
6920# qhasm: xmm3 &= xmm4
6921# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
6922# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
6923pand %xmm0,%xmm2
6924
6925# qhasm: xmm4 ^= xmm0
6926# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
6927# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
6928pxor %xmm1,%xmm0
6929
6930# qhasm: xmm4 &= xmm6
6931# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
6932# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
6933pand %xmm3,%xmm0
6934
6935# qhasm: xmm0 &= xmm7
6936# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
6937# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
6938pand %xmm5,%xmm1
6939
6940# qhasm: xmm0 ^= xmm4
6941# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
6942# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
6943pxor %xmm0,%xmm1
6944
6945# qhasm: xmm4 ^= xmm3
6946# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
6947# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
6948pxor %xmm2,%xmm0
6949
6950# qhasm: xmm2 = xmm5
6951# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
6952# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
6953movdqa %xmm7,%xmm2
6954
6955# qhasm: xmm2 ^= xmm1
6956# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
6957# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
6958pxor %xmm4,%xmm2
6959
6960# qhasm: xmm2 &= xmm12
6961# asm 1: pand <xmm12=int6464#13,<xmm2=int6464#3
6962# asm 2: pand <xmm12=%xmm12,<xmm2=%xmm2
6963pand %xmm12,%xmm2
6964
6965# qhasm: xmm12 ^= xmm10
6966# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
6967# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
6968pxor %xmm10,%xmm12
6969
6970# qhasm: xmm12 &= xmm1
6971# asm 1: pand <xmm1=int6464#5,<xmm12=int6464#13
6972# asm 2: pand <xmm1=%xmm4,<xmm12=%xmm12
6973pand %xmm4,%xmm12
6974
6975# qhasm: xmm10 &= xmm5
6976# asm 1: pand <xmm5=int6464#8,<xmm10=int6464#11
6977# asm 2: pand <xmm5=%xmm7,<xmm10=%xmm10
6978pand %xmm7,%xmm10
6979
6980# qhasm: xmm12 ^= xmm10
6981# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
6982# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
6983pxor %xmm10,%xmm12
6984
6985# qhasm: xmm10 ^= xmm2
6986# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#11
6987# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm10
6988pxor %xmm2,%xmm10
6989
6990# qhasm: xmm7 ^= xmm5
6991# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
6992# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
6993pxor %xmm7,%xmm5
6994
6995# qhasm: xmm6 ^= xmm1
6996# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
6997# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
6998pxor %xmm4,%xmm3
6999
7000# qhasm: xmm3 = xmm7
7001# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
7002# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
7003movdqa %xmm5,%xmm2
7004
7005# qhasm: xmm3 ^= xmm6
7006# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
7007# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
7008pxor %xmm3,%xmm2
7009
7010# qhasm: xmm3 &= xmm15
7011# asm 1: pand <xmm15=int6464#16,<xmm3=int6464#3
7012# asm 2: pand <xmm15=%xmm15,<xmm3=%xmm2
7013pand %xmm15,%xmm2
7014
7015# qhasm: xmm15 ^= xmm9
7016# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
7017# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
7018pxor %xmm9,%xmm15
7019
7020# qhasm: xmm15 &= xmm6
7021# asm 1: pand <xmm6=int6464#4,<xmm15=int6464#16
7022# asm 2: pand <xmm6=%xmm3,<xmm15=%xmm15
7023pand %xmm3,%xmm15
7024
7025# qhasm: xmm9 &= xmm7
7026# asm 1: pand <xmm7=int6464#6,<xmm9=int6464#10
7027# asm 2: pand <xmm7=%xmm5,<xmm9=%xmm9
7028pand %xmm5,%xmm9
7029
7030# qhasm: xmm15 ^= xmm9
7031# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
7032# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
7033pxor %xmm9,%xmm15
7034
7035# qhasm: xmm9 ^= xmm3
7036# asm 1: pxor <xmm3=int6464#3,<xmm9=int6464#10
7037# asm 2: pxor <xmm3=%xmm2,<xmm9=%xmm9
7038pxor %xmm2,%xmm9
7039
7040# qhasm: xmm15 ^= xmm4
7041# asm 1: pxor <xmm4=int6464#1,<xmm15=int6464#16
7042# asm 2: pxor <xmm4=%xmm0,<xmm15=%xmm15
7043pxor %xmm0,%xmm15
7044
7045# qhasm: xmm12 ^= xmm4
7046# asm 1: pxor <xmm4=int6464#1,<xmm12=int6464#13
7047# asm 2: pxor <xmm4=%xmm0,<xmm12=%xmm12
7048pxor %xmm0,%xmm12
7049
7050# qhasm: xmm9 ^= xmm0
7051# asm 1: pxor <xmm0=int6464#2,<xmm9=int6464#10
7052# asm 2: pxor <xmm0=%xmm1,<xmm9=%xmm9
7053pxor %xmm1,%xmm9
7054
7055# qhasm: xmm10 ^= xmm0
7056# asm 1: pxor <xmm0=int6464#2,<xmm10=int6464#11
7057# asm 2: pxor <xmm0=%xmm1,<xmm10=%xmm10
7058pxor %xmm1,%xmm10
7059
7060# qhasm: xmm15 ^= xmm8
7061# asm 1: pxor <xmm8=int6464#9,<xmm15=int6464#16
7062# asm 2: pxor <xmm8=%xmm8,<xmm15=%xmm15
7063pxor %xmm8,%xmm15
7064
7065# qhasm: xmm9 ^= xmm14
7066# asm 1: pxor <xmm14=int6464#15,<xmm9=int6464#10
7067# asm 2: pxor <xmm14=%xmm14,<xmm9=%xmm9
7068pxor %xmm14,%xmm9
7069
7070# qhasm: xmm12 ^= xmm15
7071# asm 1: pxor <xmm15=int6464#16,<xmm12=int6464#13
7072# asm 2: pxor <xmm15=%xmm15,<xmm12=%xmm12
7073pxor %xmm15,%xmm12
7074
7075# qhasm: xmm14 ^= xmm8
7076# asm 1: pxor <xmm8=int6464#9,<xmm14=int6464#15
7077# asm 2: pxor <xmm8=%xmm8,<xmm14=%xmm14
7078pxor %xmm8,%xmm14
7079
7080# qhasm: xmm8 ^= xmm9
7081# asm 1: pxor <xmm9=int6464#10,<xmm8=int6464#9
7082# asm 2: pxor <xmm9=%xmm9,<xmm8=%xmm8
7083pxor %xmm9,%xmm8
7084
7085# qhasm: xmm9 ^= xmm13
7086# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
7087# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
7088pxor %xmm13,%xmm9
7089
7090# qhasm: xmm13 ^= xmm10
7091# asm 1: pxor <xmm10=int6464#11,<xmm13=int6464#14
7092# asm 2: pxor <xmm10=%xmm10,<xmm13=%xmm13
7093pxor %xmm10,%xmm13
7094
7095# qhasm: xmm12 ^= xmm13
7096# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
7097# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
7098pxor %xmm13,%xmm12
7099
7100# qhasm: xmm10 ^= xmm11
7101# asm 1: pxor <xmm11=int6464#12,<xmm10=int6464#11
7102# asm 2: pxor <xmm11=%xmm11,<xmm10=%xmm10
7103pxor %xmm11,%xmm10
7104
7105# qhasm: xmm11 ^= xmm13
7106# asm 1: pxor <xmm13=int6464#14,<xmm11=int6464#12
7107# asm 2: pxor <xmm13=%xmm13,<xmm11=%xmm11
7108pxor %xmm13,%xmm11
7109
7110# qhasm: xmm14 ^= xmm11
7111# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
7112# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
7113pxor %xmm11,%xmm14
7114
7115# qhasm: xmm0 = shuffle dwords of xmm8 by 0x93
7116# asm 1: pshufd $0x93,<xmm8=int6464#9,>xmm0=int6464#1
7117# asm 2: pshufd $0x93,<xmm8=%xmm8,>xmm0=%xmm0
7118pshufd $0x93,%xmm8,%xmm0
7119
7120# qhasm: xmm1 = shuffle dwords of xmm9 by 0x93
7121# asm 1: pshufd $0x93,<xmm9=int6464#10,>xmm1=int6464#2
7122# asm 2: pshufd $0x93,<xmm9=%xmm9,>xmm1=%xmm1
7123pshufd $0x93,%xmm9,%xmm1
7124
7125# qhasm: xmm2 = shuffle dwords of xmm12 by 0x93
7126# asm 1: pshufd $0x93,<xmm12=int6464#13,>xmm2=int6464#3
7127# asm 2: pshufd $0x93,<xmm12=%xmm12,>xmm2=%xmm2
7128pshufd $0x93,%xmm12,%xmm2
7129
7130# qhasm: xmm3 = shuffle dwords of xmm14 by 0x93
7131# asm 1: pshufd $0x93,<xmm14=int6464#15,>xmm3=int6464#4
7132# asm 2: pshufd $0x93,<xmm14=%xmm14,>xmm3=%xmm3
7133pshufd $0x93,%xmm14,%xmm3
7134
7135# qhasm: xmm4 = shuffle dwords of xmm11 by 0x93
7136# asm 1: pshufd $0x93,<xmm11=int6464#12,>xmm4=int6464#5
7137# asm 2: pshufd $0x93,<xmm11=%xmm11,>xmm4=%xmm4
7138pshufd $0x93,%xmm11,%xmm4
7139
7140# qhasm: xmm5 = shuffle dwords of xmm15 by 0x93
7141# asm 1: pshufd $0x93,<xmm15=int6464#16,>xmm5=int6464#6
7142# asm 2: pshufd $0x93,<xmm15=%xmm15,>xmm5=%xmm5
7143pshufd $0x93,%xmm15,%xmm5
7144
7145# qhasm: xmm6 = shuffle dwords of xmm10 by 0x93
7146# asm 1: pshufd $0x93,<xmm10=int6464#11,>xmm6=int6464#7
7147# asm 2: pshufd $0x93,<xmm10=%xmm10,>xmm6=%xmm6
7148pshufd $0x93,%xmm10,%xmm6
7149
7150# qhasm: xmm7 = shuffle dwords of xmm13 by 0x93
7151# asm 1: pshufd $0x93,<xmm13=int6464#14,>xmm7=int6464#8
7152# asm 2: pshufd $0x93,<xmm13=%xmm13,>xmm7=%xmm7
7153pshufd $0x93,%xmm13,%xmm7
7154
7155# qhasm: xmm8 ^= xmm0
7156# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
7157# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
7158pxor %xmm0,%xmm8
7159
7160# qhasm: xmm9 ^= xmm1
7161# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
7162# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
7163pxor %xmm1,%xmm9
7164
7165# qhasm: xmm12 ^= xmm2
7166# asm 1: pxor <xmm2=int6464#3,<xmm12=int6464#13
7167# asm 2: pxor <xmm2=%xmm2,<xmm12=%xmm12
7168pxor %xmm2,%xmm12
7169
7170# qhasm: xmm14 ^= xmm3
7171# asm 1: pxor <xmm3=int6464#4,<xmm14=int6464#15
7172# asm 2: pxor <xmm3=%xmm3,<xmm14=%xmm14
7173pxor %xmm3,%xmm14
7174
7175# qhasm: xmm11 ^= xmm4
7176# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
7177# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
7178pxor %xmm4,%xmm11
7179
7180# qhasm: xmm15 ^= xmm5
7181# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
7182# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
7183pxor %xmm5,%xmm15
7184
7185# qhasm: xmm10 ^= xmm6
7186# asm 1: pxor <xmm6=int6464#7,<xmm10=int6464#11
7187# asm 2: pxor <xmm6=%xmm6,<xmm10=%xmm10
7188pxor %xmm6,%xmm10
7189
7190# qhasm: xmm13 ^= xmm7
7191# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
7192# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
7193pxor %xmm7,%xmm13
7194
7195# qhasm: xmm0 ^= xmm13
7196# asm 1: pxor <xmm13=int6464#14,<xmm0=int6464#1
7197# asm 2: pxor <xmm13=%xmm13,<xmm0=%xmm0
7198pxor %xmm13,%xmm0
7199
7200# qhasm: xmm1 ^= xmm8
7201# asm 1: pxor <xmm8=int6464#9,<xmm1=int6464#2
7202# asm 2: pxor <xmm8=%xmm8,<xmm1=%xmm1
7203pxor %xmm8,%xmm1
7204
7205# qhasm: xmm2 ^= xmm9
7206# asm 1: pxor <xmm9=int6464#10,<xmm2=int6464#3
7207# asm 2: pxor <xmm9=%xmm9,<xmm2=%xmm2
7208pxor %xmm9,%xmm2
7209
7210# qhasm: xmm1 ^= xmm13
7211# asm 1: pxor <xmm13=int6464#14,<xmm1=int6464#2
7212# asm 2: pxor <xmm13=%xmm13,<xmm1=%xmm1
7213pxor %xmm13,%xmm1
7214
7215# qhasm: xmm3 ^= xmm12
7216# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
7217# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
7218pxor %xmm12,%xmm3
7219
7220# qhasm: xmm4 ^= xmm14
7221# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#5
7222# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm4
7223pxor %xmm14,%xmm4
7224
7225# qhasm: xmm5 ^= xmm11
7226# asm 1: pxor <xmm11=int6464#12,<xmm5=int6464#6
7227# asm 2: pxor <xmm11=%xmm11,<xmm5=%xmm5
7228pxor %xmm11,%xmm5
7229
7230# qhasm: xmm3 ^= xmm13
7231# asm 1: pxor <xmm13=int6464#14,<xmm3=int6464#4
7232# asm 2: pxor <xmm13=%xmm13,<xmm3=%xmm3
7233pxor %xmm13,%xmm3
7234
7235# qhasm: xmm6 ^= xmm15
7236# asm 1: pxor <xmm15=int6464#16,<xmm6=int6464#7
7237# asm 2: pxor <xmm15=%xmm15,<xmm6=%xmm6
7238pxor %xmm15,%xmm6
7239
7240# qhasm: xmm7 ^= xmm10
7241# asm 1: pxor <xmm10=int6464#11,<xmm7=int6464#8
7242# asm 2: pxor <xmm10=%xmm10,<xmm7=%xmm7
7243pxor %xmm10,%xmm7
7244
7245# qhasm: xmm4 ^= xmm13
7246# asm 1: pxor <xmm13=int6464#14,<xmm4=int6464#5
7247# asm 2: pxor <xmm13=%xmm13,<xmm4=%xmm4
7248pxor %xmm13,%xmm4
7249
7250# qhasm: xmm8 = shuffle dwords of xmm8 by 0x4E
7251# asm 1: pshufd $0x4E,<xmm8=int6464#9,>xmm8=int6464#9
7252# asm 2: pshufd $0x4E,<xmm8=%xmm8,>xmm8=%xmm8
7253pshufd $0x4E,%xmm8,%xmm8
7254
7255# qhasm: xmm9 = shuffle dwords of xmm9 by 0x4E
7256# asm 1: pshufd $0x4E,<xmm9=int6464#10,>xmm9=int6464#10
7257# asm 2: pshufd $0x4E,<xmm9=%xmm9,>xmm9=%xmm9
7258pshufd $0x4E,%xmm9,%xmm9
7259
7260# qhasm: xmm12 = shuffle dwords of xmm12 by 0x4E
7261# asm 1: pshufd $0x4E,<xmm12=int6464#13,>xmm12=int6464#13
7262# asm 2: pshufd $0x4E,<xmm12=%xmm12,>xmm12=%xmm12
7263pshufd $0x4E,%xmm12,%xmm12
7264
7265# qhasm: xmm14 = shuffle dwords of xmm14 by 0x4E
7266# asm 1: pshufd $0x4E,<xmm14=int6464#15,>xmm14=int6464#15
7267# asm 2: pshufd $0x4E,<xmm14=%xmm14,>xmm14=%xmm14
7268pshufd $0x4E,%xmm14,%xmm14
7269
7270# qhasm: xmm11 = shuffle dwords of xmm11 by 0x4E
7271# asm 1: pshufd $0x4E,<xmm11=int6464#12,>xmm11=int6464#12
7272# asm 2: pshufd $0x4E,<xmm11=%xmm11,>xmm11=%xmm11
7273pshufd $0x4E,%xmm11,%xmm11
7274
7275# qhasm: xmm15 = shuffle dwords of xmm15 by 0x4E
7276# asm 1: pshufd $0x4E,<xmm15=int6464#16,>xmm15=int6464#16
7277# asm 2: pshufd $0x4E,<xmm15=%xmm15,>xmm15=%xmm15
7278pshufd $0x4E,%xmm15,%xmm15
7279
7280# qhasm: xmm10 = shuffle dwords of xmm10 by 0x4E
7281# asm 1: pshufd $0x4E,<xmm10=int6464#11,>xmm10=int6464#11
7282# asm 2: pshufd $0x4E,<xmm10=%xmm10,>xmm10=%xmm10
7283pshufd $0x4E,%xmm10,%xmm10
7284
7285# qhasm: xmm13 = shuffle dwords of xmm13 by 0x4E
7286# asm 1: pshufd $0x4E,<xmm13=int6464#14,>xmm13=int6464#14
7287# asm 2: pshufd $0x4E,<xmm13=%xmm13,>xmm13=%xmm13
7288pshufd $0x4E,%xmm13,%xmm13
7289
7290# qhasm: xmm0 ^= xmm8
7291# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
7292# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
7293pxor %xmm8,%xmm0
7294
7295# qhasm: xmm1 ^= xmm9
7296# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
7297# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
7298pxor %xmm9,%xmm1
7299
7300# qhasm: xmm2 ^= xmm12
7301# asm 1: pxor <xmm12=int6464#13,<xmm2=int6464#3
7302# asm 2: pxor <xmm12=%xmm12,<xmm2=%xmm2
7303pxor %xmm12,%xmm2
7304
7305# qhasm: xmm3 ^= xmm14
7306# asm 1: pxor <xmm14=int6464#15,<xmm3=int6464#4
7307# asm 2: pxor <xmm14=%xmm14,<xmm3=%xmm3
7308pxor %xmm14,%xmm3
7309
7310# qhasm: xmm4 ^= xmm11
7311# asm 1: pxor <xmm11=int6464#12,<xmm4=int6464#5
7312# asm 2: pxor <xmm11=%xmm11,<xmm4=%xmm4
7313pxor %xmm11,%xmm4
7314
7315# qhasm: xmm5 ^= xmm15
7316# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
7317# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
7318pxor %xmm15,%xmm5
7319
7320# qhasm: xmm6 ^= xmm10
7321# asm 1: pxor <xmm10=int6464#11,<xmm6=int6464#7
7322# asm 2: pxor <xmm10=%xmm10,<xmm6=%xmm6
7323pxor %xmm10,%xmm6
7324
7325# qhasm: xmm7 ^= xmm13
7326# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
7327# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
7328pxor %xmm13,%xmm7
7329
7330# qhasm: xmm0 ^= *(int128 *)(c + 768)
7331# asm 1: pxor 768(<c=int64#4),<xmm0=int6464#1
7332# asm 2: pxor 768(<c=%rcx),<xmm0=%xmm0
7333pxor 768(%rcx),%xmm0
7334
7335# qhasm: shuffle bytes of xmm0 by SR
7336# asm 1: pshufb SR,<xmm0=int6464#1
7337# asm 2: pshufb SR,<xmm0=%xmm0
7338pshufb SR,%xmm0
7339
7340# qhasm: xmm1 ^= *(int128 *)(c + 784)
7341# asm 1: pxor 784(<c=int64#4),<xmm1=int6464#2
7342# asm 2: pxor 784(<c=%rcx),<xmm1=%xmm1
7343pxor 784(%rcx),%xmm1
7344
7345# qhasm: shuffle bytes of xmm1 by SR
7346# asm 1: pshufb SR,<xmm1=int6464#2
7347# asm 2: pshufb SR,<xmm1=%xmm1
7348pshufb SR,%xmm1
7349
7350# qhasm: xmm2 ^= *(int128 *)(c + 800)
7351# asm 1: pxor 800(<c=int64#4),<xmm2=int6464#3
7352# asm 2: pxor 800(<c=%rcx),<xmm2=%xmm2
7353pxor 800(%rcx),%xmm2
7354
7355# qhasm: shuffle bytes of xmm2 by SR
7356# asm 1: pshufb SR,<xmm2=int6464#3
7357# asm 2: pshufb SR,<xmm2=%xmm2
7358pshufb SR,%xmm2
7359
7360# qhasm: xmm3 ^= *(int128 *)(c + 816)
7361# asm 1: pxor 816(<c=int64#4),<xmm3=int6464#4
7362# asm 2: pxor 816(<c=%rcx),<xmm3=%xmm3
7363pxor 816(%rcx),%xmm3
7364
7365# qhasm: shuffle bytes of xmm3 by SR
7366# asm 1: pshufb SR,<xmm3=int6464#4
7367# asm 2: pshufb SR,<xmm3=%xmm3
7368pshufb SR,%xmm3
7369
7370# qhasm: xmm4 ^= *(int128 *)(c + 832)
7371# asm 1: pxor 832(<c=int64#4),<xmm4=int6464#5
7372# asm 2: pxor 832(<c=%rcx),<xmm4=%xmm4
7373pxor 832(%rcx),%xmm4
7374
7375# qhasm: shuffle bytes of xmm4 by SR
7376# asm 1: pshufb SR,<xmm4=int6464#5
7377# asm 2: pshufb SR,<xmm4=%xmm4
7378pshufb SR,%xmm4
7379
7380# qhasm: xmm5 ^= *(int128 *)(c + 848)
7381# asm 1: pxor 848(<c=int64#4),<xmm5=int6464#6
7382# asm 2: pxor 848(<c=%rcx),<xmm5=%xmm5
7383pxor 848(%rcx),%xmm5
7384
7385# qhasm: shuffle bytes of xmm5 by SR
7386# asm 1: pshufb SR,<xmm5=int6464#6
7387# asm 2: pshufb SR,<xmm5=%xmm5
7388pshufb SR,%xmm5
7389
7390# qhasm: xmm6 ^= *(int128 *)(c + 864)
7391# asm 1: pxor 864(<c=int64#4),<xmm6=int6464#7
7392# asm 2: pxor 864(<c=%rcx),<xmm6=%xmm6
7393pxor 864(%rcx),%xmm6
7394
7395# qhasm: shuffle bytes of xmm6 by SR
7396# asm 1: pshufb SR,<xmm6=int6464#7
7397# asm 2: pshufb SR,<xmm6=%xmm6
7398pshufb SR,%xmm6
7399
7400# qhasm: xmm7 ^= *(int128 *)(c + 880)
7401# asm 1: pxor 880(<c=int64#4),<xmm7=int6464#8
7402# asm 2: pxor 880(<c=%rcx),<xmm7=%xmm7
7403pxor 880(%rcx),%xmm7
7404
7405# qhasm: shuffle bytes of xmm7 by SR
7406# asm 1: pshufb SR,<xmm7=int6464#8
7407# asm 2: pshufb SR,<xmm7=%xmm7
7408pshufb SR,%xmm7
7409
7410# qhasm: xmm5 ^= xmm6
7411# asm 1: pxor <xmm6=int6464#7,<xmm5=int6464#6
7412# asm 2: pxor <xmm6=%xmm6,<xmm5=%xmm5
7413pxor %xmm6,%xmm5
7414
7415# qhasm: xmm2 ^= xmm1
7416# asm 1: pxor <xmm1=int6464#2,<xmm2=int6464#3
7417# asm 2: pxor <xmm1=%xmm1,<xmm2=%xmm2
7418pxor %xmm1,%xmm2
7419
7420# qhasm: xmm5 ^= xmm0
7421# asm 1: pxor <xmm0=int6464#1,<xmm5=int6464#6
7422# asm 2: pxor <xmm0=%xmm0,<xmm5=%xmm5
7423pxor %xmm0,%xmm5
7424
7425# qhasm: xmm6 ^= xmm2
7426# asm 1: pxor <xmm2=int6464#3,<xmm6=int6464#7
7427# asm 2: pxor <xmm2=%xmm2,<xmm6=%xmm6
7428pxor %xmm2,%xmm6
7429
7430# qhasm: xmm3 ^= xmm0
7431# asm 1: pxor <xmm0=int6464#1,<xmm3=int6464#4
7432# asm 2: pxor <xmm0=%xmm0,<xmm3=%xmm3
7433pxor %xmm0,%xmm3
7434
7435# qhasm: xmm6 ^= xmm3
7436# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
7437# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
7438pxor %xmm3,%xmm6
7439
7440# qhasm: xmm3 ^= xmm7
7441# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#4
7442# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm3
7443pxor %xmm7,%xmm3
7444
7445# qhasm: xmm3 ^= xmm4
7446# asm 1: pxor <xmm4=int6464#5,<xmm3=int6464#4
7447# asm 2: pxor <xmm4=%xmm4,<xmm3=%xmm3
7448pxor %xmm4,%xmm3
7449
7450# qhasm: xmm7 ^= xmm5
7451# asm 1: pxor <xmm5=int6464#6,<xmm7=int6464#8
7452# asm 2: pxor <xmm5=%xmm5,<xmm7=%xmm7
7453pxor %xmm5,%xmm7
7454
7455# qhasm: xmm3 ^= xmm1
7456# asm 1: pxor <xmm1=int6464#2,<xmm3=int6464#4
7457# asm 2: pxor <xmm1=%xmm1,<xmm3=%xmm3
7458pxor %xmm1,%xmm3
7459
7460# qhasm: xmm4 ^= xmm5
7461# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
7462# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
7463pxor %xmm5,%xmm4
7464
7465# qhasm: xmm2 ^= xmm7
7466# asm 1: pxor <xmm7=int6464#8,<xmm2=int6464#3
7467# asm 2: pxor <xmm7=%xmm7,<xmm2=%xmm2
7468pxor %xmm7,%xmm2
7469
7470# qhasm: xmm1 ^= xmm5
7471# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
7472# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
7473pxor %xmm5,%xmm1
7474
7475# qhasm: xmm11 = xmm7
7476# asm 1: movdqa <xmm7=int6464#8,>xmm11=int6464#9
7477# asm 2: movdqa <xmm7=%xmm7,>xmm11=%xmm8
7478movdqa %xmm7,%xmm8
7479
7480# qhasm: xmm10 = xmm1
7481# asm 1: movdqa <xmm1=int6464#2,>xmm10=int6464#10
7482# asm 2: movdqa <xmm1=%xmm1,>xmm10=%xmm9
7483movdqa %xmm1,%xmm9
7484
7485# qhasm: xmm9 = xmm5
7486# asm 1: movdqa <xmm5=int6464#6,>xmm9=int6464#11
7487# asm 2: movdqa <xmm5=%xmm5,>xmm9=%xmm10
7488movdqa %xmm5,%xmm10
7489
7490# qhasm: xmm13 = xmm2
7491# asm 1: movdqa <xmm2=int6464#3,>xmm13=int6464#12
7492# asm 2: movdqa <xmm2=%xmm2,>xmm13=%xmm11
7493movdqa %xmm2,%xmm11
7494
7495# qhasm: xmm12 = xmm6
7496# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#13
7497# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm12
7498movdqa %xmm6,%xmm12
7499
7500# qhasm: xmm11 ^= xmm4
7501# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#9
7502# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm8
7503pxor %xmm4,%xmm8
7504
7505# qhasm: xmm10 ^= xmm2
7506# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#10
7507# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm9
7508pxor %xmm2,%xmm9
7509
7510# qhasm: xmm9 ^= xmm3
7511# asm 1: pxor <xmm3=int6464#4,<xmm9=int6464#11
7512# asm 2: pxor <xmm3=%xmm3,<xmm9=%xmm10
7513pxor %xmm3,%xmm10
7514
7515# qhasm: xmm13 ^= xmm4
7516# asm 1: pxor <xmm4=int6464#5,<xmm13=int6464#12
7517# asm 2: pxor <xmm4=%xmm4,<xmm13=%xmm11
7518pxor %xmm4,%xmm11
7519
7520# qhasm: xmm12 ^= xmm0
7521# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
7522# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
7523pxor %xmm0,%xmm12
7524
7525# qhasm: xmm14 = xmm11
7526# asm 1: movdqa <xmm11=int6464#9,>xmm14=int6464#14
7527# asm 2: movdqa <xmm11=%xmm8,>xmm14=%xmm13
7528movdqa %xmm8,%xmm13
7529
7530# qhasm: xmm8 = xmm10
7531# asm 1: movdqa <xmm10=int6464#10,>xmm8=int6464#15
7532# asm 2: movdqa <xmm10=%xmm9,>xmm8=%xmm14
7533movdqa %xmm9,%xmm14
7534
7535# qhasm: xmm15 = xmm11
7536# asm 1: movdqa <xmm11=int6464#9,>xmm15=int6464#16
7537# asm 2: movdqa <xmm11=%xmm8,>xmm15=%xmm15
7538movdqa %xmm8,%xmm15
7539
7540# qhasm: xmm10 |= xmm9
7541# asm 1: por <xmm9=int6464#11,<xmm10=int6464#10
7542# asm 2: por <xmm9=%xmm10,<xmm10=%xmm9
7543por %xmm10,%xmm9
7544
7545# qhasm: xmm11 |= xmm12
7546# asm 1: por <xmm12=int6464#13,<xmm11=int6464#9
7547# asm 2: por <xmm12=%xmm12,<xmm11=%xmm8
7548por %xmm12,%xmm8
7549
7550# qhasm: xmm15 ^= xmm8
7551# asm 1: pxor <xmm8=int6464#15,<xmm15=int6464#16
7552# asm 2: pxor <xmm8=%xmm14,<xmm15=%xmm15
7553pxor %xmm14,%xmm15
7554
7555# qhasm: xmm14 &= xmm12
7556# asm 1: pand <xmm12=int6464#13,<xmm14=int6464#14
7557# asm 2: pand <xmm12=%xmm12,<xmm14=%xmm13
7558pand %xmm12,%xmm13
7559
7560# qhasm: xmm8 &= xmm9
7561# asm 1: pand <xmm9=int6464#11,<xmm8=int6464#15
7562# asm 2: pand <xmm9=%xmm10,<xmm8=%xmm14
7563pand %xmm10,%xmm14
7564
7565# qhasm: xmm12 ^= xmm9
7566# asm 1: pxor <xmm9=int6464#11,<xmm12=int6464#13
7567# asm 2: pxor <xmm9=%xmm10,<xmm12=%xmm12
7568pxor %xmm10,%xmm12
7569
7570# qhasm: xmm15 &= xmm12
7571# asm 1: pand <xmm12=int6464#13,<xmm15=int6464#16
7572# asm 2: pand <xmm12=%xmm12,<xmm15=%xmm15
7573pand %xmm12,%xmm15
7574
7575# qhasm: xmm12 = xmm3
7576# asm 1: movdqa <xmm3=int6464#4,>xmm12=int6464#11
7577# asm 2: movdqa <xmm3=%xmm3,>xmm12=%xmm10
7578movdqa %xmm3,%xmm10
7579
7580# qhasm: xmm12 ^= xmm0
7581# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#11
7582# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm10
7583pxor %xmm0,%xmm10
7584
7585# qhasm: xmm13 &= xmm12
7586# asm 1: pand <xmm12=int6464#11,<xmm13=int6464#12
7587# asm 2: pand <xmm12=%xmm10,<xmm13=%xmm11
7588pand %xmm10,%xmm11
7589
7590# qhasm: xmm11 ^= xmm13
7591# asm 1: pxor <xmm13=int6464#12,<xmm11=int6464#9
7592# asm 2: pxor <xmm13=%xmm11,<xmm11=%xmm8
7593pxor %xmm11,%xmm8
7594
7595# qhasm: xmm10 ^= xmm13
7596# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
7597# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
7598pxor %xmm11,%xmm9
7599
7600# qhasm: xmm13 = xmm7
7601# asm 1: movdqa <xmm7=int6464#8,>xmm13=int6464#11
7602# asm 2: movdqa <xmm7=%xmm7,>xmm13=%xmm10
7603movdqa %xmm7,%xmm10
7604
7605# qhasm: xmm13 ^= xmm1
7606# asm 1: pxor <xmm1=int6464#2,<xmm13=int6464#11
7607# asm 2: pxor <xmm1=%xmm1,<xmm13=%xmm10
7608pxor %xmm1,%xmm10
7609
7610# qhasm: xmm12 = xmm5
7611# asm 1: movdqa <xmm5=int6464#6,>xmm12=int6464#12
7612# asm 2: movdqa <xmm5=%xmm5,>xmm12=%xmm11
7613movdqa %xmm5,%xmm11
7614
7615# qhasm: xmm9 = xmm13
7616# asm 1: movdqa <xmm13=int6464#11,>xmm9=int6464#13
7617# asm 2: movdqa <xmm13=%xmm10,>xmm9=%xmm12
7618movdqa %xmm10,%xmm12
7619
7620# qhasm: xmm12 ^= xmm6
7621# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#12
7622# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm11
7623pxor %xmm6,%xmm11
7624
7625# qhasm: xmm9 |= xmm12
7626# asm 1: por <xmm12=int6464#12,<xmm9=int6464#13
7627# asm 2: por <xmm12=%xmm11,<xmm9=%xmm12
7628por %xmm11,%xmm12
7629
7630# qhasm: xmm13 &= xmm12
7631# asm 1: pand <xmm12=int6464#12,<xmm13=int6464#11
7632# asm 2: pand <xmm12=%xmm11,<xmm13=%xmm10
7633pand %xmm11,%xmm10
7634
7635# qhasm: xmm8 ^= xmm13
7636# asm 1: pxor <xmm13=int6464#11,<xmm8=int6464#15
7637# asm 2: pxor <xmm13=%xmm10,<xmm8=%xmm14
7638pxor %xmm10,%xmm14
7639
7640# qhasm: xmm11 ^= xmm15
7641# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#9
7642# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm8
7643pxor %xmm15,%xmm8
7644
7645# qhasm: xmm10 ^= xmm14
7646# asm 1: pxor <xmm14=int6464#14,<xmm10=int6464#10
7647# asm 2: pxor <xmm14=%xmm13,<xmm10=%xmm9
7648pxor %xmm13,%xmm9
7649
7650# qhasm: xmm9 ^= xmm15
7651# asm 1: pxor <xmm15=int6464#16,<xmm9=int6464#13
7652# asm 2: pxor <xmm15=%xmm15,<xmm9=%xmm12
7653pxor %xmm15,%xmm12
7654
7655# qhasm: xmm8 ^= xmm14
7656# asm 1: pxor <xmm14=int6464#14,<xmm8=int6464#15
7657# asm 2: pxor <xmm14=%xmm13,<xmm8=%xmm14
7658pxor %xmm13,%xmm14
7659
7660# qhasm: xmm9 ^= xmm14
7661# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
7662# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
7663pxor %xmm13,%xmm12
7664
7665# qhasm: xmm12 = xmm2
7666# asm 1: movdqa <xmm2=int6464#3,>xmm12=int6464#11
7667# asm 2: movdqa <xmm2=%xmm2,>xmm12=%xmm10
7668movdqa %xmm2,%xmm10
7669
7670# qhasm: xmm13 = xmm4
7671# asm 1: movdqa <xmm4=int6464#5,>xmm13=int6464#12
7672# asm 2: movdqa <xmm4=%xmm4,>xmm13=%xmm11
7673movdqa %xmm4,%xmm11
7674
7675# qhasm: xmm14 = xmm1
7676# asm 1: movdqa <xmm1=int6464#2,>xmm14=int6464#14
7677# asm 2: movdqa <xmm1=%xmm1,>xmm14=%xmm13
7678movdqa %xmm1,%xmm13
7679
7680# qhasm: xmm15 = xmm7
7681# asm 1: movdqa <xmm7=int6464#8,>xmm15=int6464#16
7682# asm 2: movdqa <xmm7=%xmm7,>xmm15=%xmm15
7683movdqa %xmm7,%xmm15
7684
7685# qhasm: xmm12 &= xmm3
7686# asm 1: pand <xmm3=int6464#4,<xmm12=int6464#11
7687# asm 2: pand <xmm3=%xmm3,<xmm12=%xmm10
7688pand %xmm3,%xmm10
7689
7690# qhasm: xmm13 &= xmm0
7691# asm 1: pand <xmm0=int6464#1,<xmm13=int6464#12
7692# asm 2: pand <xmm0=%xmm0,<xmm13=%xmm11
7693pand %xmm0,%xmm11
7694
7695# qhasm: xmm14 &= xmm5
7696# asm 1: pand <xmm5=int6464#6,<xmm14=int6464#14
7697# asm 2: pand <xmm5=%xmm5,<xmm14=%xmm13
7698pand %xmm5,%xmm13
7699
7700# qhasm: xmm15 |= xmm6
7701# asm 1: por <xmm6=int6464#7,<xmm15=int6464#16
7702# asm 2: por <xmm6=%xmm6,<xmm15=%xmm15
7703por %xmm6,%xmm15
7704
7705# qhasm: xmm11 ^= xmm12
7706# asm 1: pxor <xmm12=int6464#11,<xmm11=int6464#9
7707# asm 2: pxor <xmm12=%xmm10,<xmm11=%xmm8
7708pxor %xmm10,%xmm8
7709
7710# qhasm: xmm10 ^= xmm13
7711# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
7712# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
7713pxor %xmm11,%xmm9
7714
7715# qhasm: xmm9 ^= xmm14
7716# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
7717# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
7718pxor %xmm13,%xmm12
7719
7720# qhasm: xmm8 ^= xmm15
7721# asm 1: pxor <xmm15=int6464#16,<xmm8=int6464#15
7722# asm 2: pxor <xmm15=%xmm15,<xmm8=%xmm14
7723pxor %xmm15,%xmm14
7724
7725# qhasm: xmm12 = xmm11
7726# asm 1: movdqa <xmm11=int6464#9,>xmm12=int6464#11
7727# asm 2: movdqa <xmm11=%xmm8,>xmm12=%xmm10
7728movdqa %xmm8,%xmm10
7729
7730# qhasm: xmm12 ^= xmm10
7731# asm 1: pxor <xmm10=int6464#10,<xmm12=int6464#11
7732# asm 2: pxor <xmm10=%xmm9,<xmm12=%xmm10
7733pxor %xmm9,%xmm10
7734
7735# qhasm: xmm11 &= xmm9
7736# asm 1: pand <xmm9=int6464#13,<xmm11=int6464#9
7737# asm 2: pand <xmm9=%xmm12,<xmm11=%xmm8
7738pand %xmm12,%xmm8
7739
7740# qhasm: xmm14 = xmm8
7741# asm 1: movdqa <xmm8=int6464#15,>xmm14=int6464#12
7742# asm 2: movdqa <xmm8=%xmm14,>xmm14=%xmm11
7743movdqa %xmm14,%xmm11
7744
7745# qhasm: xmm14 ^= xmm11
7746# asm 1: pxor <xmm11=int6464#9,<xmm14=int6464#12
7747# asm 2: pxor <xmm11=%xmm8,<xmm14=%xmm11
7748pxor %xmm8,%xmm11
7749
7750# qhasm: xmm15 = xmm12
7751# asm 1: movdqa <xmm12=int6464#11,>xmm15=int6464#14
7752# asm 2: movdqa <xmm12=%xmm10,>xmm15=%xmm13
7753movdqa %xmm10,%xmm13
7754
7755# qhasm: xmm15 &= xmm14
7756# asm 1: pand <xmm14=int6464#12,<xmm15=int6464#14
7757# asm 2: pand <xmm14=%xmm11,<xmm15=%xmm13
7758pand %xmm11,%xmm13
7759
7760# qhasm: xmm15 ^= xmm10
7761# asm 1: pxor <xmm10=int6464#10,<xmm15=int6464#14
7762# asm 2: pxor <xmm10=%xmm9,<xmm15=%xmm13
7763pxor %xmm9,%xmm13
7764
7765# qhasm: xmm13 = xmm9
7766# asm 1: movdqa <xmm9=int6464#13,>xmm13=int6464#16
7767# asm 2: movdqa <xmm9=%xmm12,>xmm13=%xmm15
7768movdqa %xmm12,%xmm15
7769
7770# qhasm: xmm13 ^= xmm8
7771# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
7772# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
7773pxor %xmm14,%xmm15
7774
7775# qhasm: xmm11 ^= xmm10
7776# asm 1: pxor <xmm10=int6464#10,<xmm11=int6464#9
7777# asm 2: pxor <xmm10=%xmm9,<xmm11=%xmm8
7778pxor %xmm9,%xmm8
7779
7780# qhasm: xmm13 &= xmm11
7781# asm 1: pand <xmm11=int6464#9,<xmm13=int6464#16
7782# asm 2: pand <xmm11=%xmm8,<xmm13=%xmm15
7783pand %xmm8,%xmm15
7784
7785# qhasm: xmm13 ^= xmm8
7786# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
7787# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
7788pxor %xmm14,%xmm15
7789
7790# qhasm: xmm9 ^= xmm13
7791# asm 1: pxor <xmm13=int6464#16,<xmm9=int6464#13
7792# asm 2: pxor <xmm13=%xmm15,<xmm9=%xmm12
7793pxor %xmm15,%xmm12
7794
7795# qhasm: xmm10 = xmm14
7796# asm 1: movdqa <xmm14=int6464#12,>xmm10=int6464#9
7797# asm 2: movdqa <xmm14=%xmm11,>xmm10=%xmm8
7798movdqa %xmm11,%xmm8
7799
7800# qhasm: xmm10 ^= xmm13
7801# asm 1: pxor <xmm13=int6464#16,<xmm10=int6464#9
7802# asm 2: pxor <xmm13=%xmm15,<xmm10=%xmm8
7803pxor %xmm15,%xmm8
7804
7805# qhasm: xmm10 &= xmm8
7806# asm 1: pand <xmm8=int6464#15,<xmm10=int6464#9
7807# asm 2: pand <xmm8=%xmm14,<xmm10=%xmm8
7808pand %xmm14,%xmm8
7809
7810# qhasm: xmm9 ^= xmm10
7811# asm 1: pxor <xmm10=int6464#9,<xmm9=int6464#13
7812# asm 2: pxor <xmm10=%xmm8,<xmm9=%xmm12
7813pxor %xmm8,%xmm12
7814
7815# qhasm: xmm14 ^= xmm10
7816# asm 1: pxor <xmm10=int6464#9,<xmm14=int6464#12
7817# asm 2: pxor <xmm10=%xmm8,<xmm14=%xmm11
7818pxor %xmm8,%xmm11
7819
7820# qhasm: xmm14 &= xmm15
7821# asm 1: pand <xmm15=int6464#14,<xmm14=int6464#12
7822# asm 2: pand <xmm15=%xmm13,<xmm14=%xmm11
7823pand %xmm13,%xmm11
7824
7825# qhasm: xmm14 ^= xmm12
7826# asm 1: pxor <xmm12=int6464#11,<xmm14=int6464#12
7827# asm 2: pxor <xmm12=%xmm10,<xmm14=%xmm11
7828pxor %xmm10,%xmm11
7829
7830# qhasm: xmm12 = xmm6
7831# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#9
7832# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm8
7833movdqa %xmm6,%xmm8
7834
7835# qhasm: xmm8 = xmm5
7836# asm 1: movdqa <xmm5=int6464#6,>xmm8=int6464#10
7837# asm 2: movdqa <xmm5=%xmm5,>xmm8=%xmm9
7838movdqa %xmm5,%xmm9
7839
7840# qhasm: xmm10 = xmm15
7841# asm 1: movdqa <xmm15=int6464#14,>xmm10=int6464#11
7842# asm 2: movdqa <xmm15=%xmm13,>xmm10=%xmm10
7843movdqa %xmm13,%xmm10
7844
7845# qhasm: xmm10 ^= xmm14
7846# asm 1: pxor <xmm14=int6464#12,<xmm10=int6464#11
7847# asm 2: pxor <xmm14=%xmm11,<xmm10=%xmm10
7848pxor %xmm11,%xmm10
7849
7850# qhasm: xmm10 &= xmm6
7851# asm 1: pand <xmm6=int6464#7,<xmm10=int6464#11
7852# asm 2: pand <xmm6=%xmm6,<xmm10=%xmm10
7853pand %xmm6,%xmm10
7854
7855# qhasm: xmm6 ^= xmm5
7856# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
7857# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
7858pxor %xmm5,%xmm6
7859
7860# qhasm: xmm6 &= xmm14
7861# asm 1: pand <xmm14=int6464#12,<xmm6=int6464#7
7862# asm 2: pand <xmm14=%xmm11,<xmm6=%xmm6
7863pand %xmm11,%xmm6
7864
7865# qhasm: xmm5 &= xmm15
7866# asm 1: pand <xmm15=int6464#14,<xmm5=int6464#6
7867# asm 2: pand <xmm15=%xmm13,<xmm5=%xmm5
7868pand %xmm13,%xmm5
7869
7870# qhasm: xmm6 ^= xmm5
7871# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
7872# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
7873pxor %xmm5,%xmm6
7874
7875# qhasm: xmm5 ^= xmm10
7876# asm 1: pxor <xmm10=int6464#11,<xmm5=int6464#6
7877# asm 2: pxor <xmm10=%xmm10,<xmm5=%xmm5
7878pxor %xmm10,%xmm5
7879
7880# qhasm: xmm12 ^= xmm0
7881# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#9
7882# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm8
7883pxor %xmm0,%xmm8
7884
7885# qhasm: xmm8 ^= xmm3
7886# asm 1: pxor <xmm3=int6464#4,<xmm8=int6464#10
7887# asm 2: pxor <xmm3=%xmm3,<xmm8=%xmm9
7888pxor %xmm3,%xmm9
7889
7890# qhasm: xmm15 ^= xmm13
7891# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
7892# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
7893pxor %xmm15,%xmm13
7894
7895# qhasm: xmm14 ^= xmm9
7896# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
7897# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
7898pxor %xmm12,%xmm11
7899
7900# qhasm: xmm11 = xmm15
7901# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
7902# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
7903movdqa %xmm13,%xmm10
7904
7905# qhasm: xmm11 ^= xmm14
7906# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
7907# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
7908pxor %xmm11,%xmm10
7909
7910# qhasm: xmm11 &= xmm12
7911# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
7912# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
7913pand %xmm8,%xmm10
7914
7915# qhasm: xmm12 ^= xmm8
7916# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
7917# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
7918pxor %xmm9,%xmm8
7919
7920# qhasm: xmm12 &= xmm14
7921# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
7922# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
7923pand %xmm11,%xmm8
7924
7925# qhasm: xmm8 &= xmm15
7926# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
7927# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
7928pand %xmm13,%xmm9
7929
7930# qhasm: xmm8 ^= xmm12
7931# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
7932# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
7933pxor %xmm8,%xmm9
7934
7935# qhasm: xmm12 ^= xmm11
7936# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
7937# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
7938pxor %xmm10,%xmm8
7939
7940# qhasm: xmm10 = xmm13
7941# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
7942# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
7943movdqa %xmm15,%xmm10
7944
7945# qhasm: xmm10 ^= xmm9
7946# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
7947# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
7948pxor %xmm12,%xmm10
7949
7950# qhasm: xmm10 &= xmm0
7951# asm 1: pand <xmm0=int6464#1,<xmm10=int6464#11
7952# asm 2: pand <xmm0=%xmm0,<xmm10=%xmm10
7953pand %xmm0,%xmm10
7954
7955# qhasm: xmm0 ^= xmm3
7956# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
7957# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
7958pxor %xmm3,%xmm0
7959
7960# qhasm: xmm0 &= xmm9
7961# asm 1: pand <xmm9=int6464#13,<xmm0=int6464#1
7962# asm 2: pand <xmm9=%xmm12,<xmm0=%xmm0
7963pand %xmm12,%xmm0
7964
7965# qhasm: xmm3 &= xmm13
7966# asm 1: pand <xmm13=int6464#16,<xmm3=int6464#4
7967# asm 2: pand <xmm13=%xmm15,<xmm3=%xmm3
7968pand %xmm15,%xmm3
7969
7970# qhasm: xmm0 ^= xmm3
7971# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
7972# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
7973pxor %xmm3,%xmm0
7974
7975# qhasm: xmm3 ^= xmm10
7976# asm 1: pxor <xmm10=int6464#11,<xmm3=int6464#4
7977# asm 2: pxor <xmm10=%xmm10,<xmm3=%xmm3
7978pxor %xmm10,%xmm3
7979
7980# qhasm: xmm6 ^= xmm12
7981# asm 1: pxor <xmm12=int6464#9,<xmm6=int6464#7
7982# asm 2: pxor <xmm12=%xmm8,<xmm6=%xmm6
7983pxor %xmm8,%xmm6
7984
7985# qhasm: xmm0 ^= xmm12
7986# asm 1: pxor <xmm12=int6464#9,<xmm0=int6464#1
7987# asm 2: pxor <xmm12=%xmm8,<xmm0=%xmm0
7988pxor %xmm8,%xmm0
7989
7990# qhasm: xmm5 ^= xmm8
7991# asm 1: pxor <xmm8=int6464#10,<xmm5=int6464#6
7992# asm 2: pxor <xmm8=%xmm9,<xmm5=%xmm5
7993pxor %xmm9,%xmm5
7994
7995# qhasm: xmm3 ^= xmm8
7996# asm 1: pxor <xmm8=int6464#10,<xmm3=int6464#4
7997# asm 2: pxor <xmm8=%xmm9,<xmm3=%xmm3
7998pxor %xmm9,%xmm3
7999
8000# qhasm: xmm12 = xmm7
8001# asm 1: movdqa <xmm7=int6464#8,>xmm12=int6464#9
8002# asm 2: movdqa <xmm7=%xmm7,>xmm12=%xmm8
8003movdqa %xmm7,%xmm8
8004
8005# qhasm: xmm8 = xmm1
8006# asm 1: movdqa <xmm1=int6464#2,>xmm8=int6464#10
8007# asm 2: movdqa <xmm1=%xmm1,>xmm8=%xmm9
8008movdqa %xmm1,%xmm9
8009
8010# qhasm: xmm12 ^= xmm4
8011# asm 1: pxor <xmm4=int6464#5,<xmm12=int6464#9
8012# asm 2: pxor <xmm4=%xmm4,<xmm12=%xmm8
8013pxor %xmm4,%xmm8
8014
8015# qhasm: xmm8 ^= xmm2
8016# asm 1: pxor <xmm2=int6464#3,<xmm8=int6464#10
8017# asm 2: pxor <xmm2=%xmm2,<xmm8=%xmm9
8018pxor %xmm2,%xmm9
8019
8020# qhasm: xmm11 = xmm15
8021# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
8022# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
8023movdqa %xmm13,%xmm10
8024
8025# qhasm: xmm11 ^= xmm14
8026# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
8027# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
8028pxor %xmm11,%xmm10
8029
8030# qhasm: xmm11 &= xmm12
8031# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
8032# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
8033pand %xmm8,%xmm10
8034
8035# qhasm: xmm12 ^= xmm8
8036# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
8037# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
8038pxor %xmm9,%xmm8
8039
8040# qhasm: xmm12 &= xmm14
8041# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
8042# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
8043pand %xmm11,%xmm8
8044
8045# qhasm: xmm8 &= xmm15
8046# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
8047# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
8048pand %xmm13,%xmm9
8049
8050# qhasm: xmm8 ^= xmm12
8051# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
8052# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
8053pxor %xmm8,%xmm9
8054
8055# qhasm: xmm12 ^= xmm11
8056# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
8057# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
8058pxor %xmm10,%xmm8
8059
8060# qhasm: xmm10 = xmm13
8061# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
8062# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
8063movdqa %xmm15,%xmm10
8064
8065# qhasm: xmm10 ^= xmm9
8066# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
8067# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
8068pxor %xmm12,%xmm10
8069
8070# qhasm: xmm10 &= xmm4
8071# asm 1: pand <xmm4=int6464#5,<xmm10=int6464#11
8072# asm 2: pand <xmm4=%xmm4,<xmm10=%xmm10
8073pand %xmm4,%xmm10
8074
8075# qhasm: xmm4 ^= xmm2
8076# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
8077# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
8078pxor %xmm2,%xmm4
8079
8080# qhasm: xmm4 &= xmm9
8081# asm 1: pand <xmm9=int6464#13,<xmm4=int6464#5
8082# asm 2: pand <xmm9=%xmm12,<xmm4=%xmm4
8083pand %xmm12,%xmm4
8084
8085# qhasm: xmm2 &= xmm13
8086# asm 1: pand <xmm13=int6464#16,<xmm2=int6464#3
8087# asm 2: pand <xmm13=%xmm15,<xmm2=%xmm2
8088pand %xmm15,%xmm2
8089
8090# qhasm: xmm4 ^= xmm2
8091# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
8092# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
8093pxor %xmm2,%xmm4
8094
8095# qhasm: xmm2 ^= xmm10
8096# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#3
8097# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm2
8098pxor %xmm10,%xmm2
8099
8100# qhasm: xmm15 ^= xmm13
8101# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
8102# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
8103pxor %xmm15,%xmm13
8104
8105# qhasm: xmm14 ^= xmm9
8106# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
8107# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
8108pxor %xmm12,%xmm11
8109
8110# qhasm: xmm11 = xmm15
8111# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
8112# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
8113movdqa %xmm13,%xmm10
8114
8115# qhasm: xmm11 ^= xmm14
8116# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
8117# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
8118pxor %xmm11,%xmm10
8119
8120# qhasm: xmm11 &= xmm7
8121# asm 1: pand <xmm7=int6464#8,<xmm11=int6464#11
8122# asm 2: pand <xmm7=%xmm7,<xmm11=%xmm10
8123pand %xmm7,%xmm10
8124
8125# qhasm: xmm7 ^= xmm1
8126# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
8127# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
8128pxor %xmm1,%xmm7
8129
8130# qhasm: xmm7 &= xmm14
8131# asm 1: pand <xmm14=int6464#12,<xmm7=int6464#8
8132# asm 2: pand <xmm14=%xmm11,<xmm7=%xmm7
8133pand %xmm11,%xmm7
8134
8135# qhasm: xmm1 &= xmm15
8136# asm 1: pand <xmm15=int6464#14,<xmm1=int6464#2
8137# asm 2: pand <xmm15=%xmm13,<xmm1=%xmm1
8138pand %xmm13,%xmm1
8139
8140# qhasm: xmm7 ^= xmm1
8141# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
8142# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
8143pxor %xmm1,%xmm7
8144
8145# qhasm: xmm1 ^= xmm11
8146# asm 1: pxor <xmm11=int6464#11,<xmm1=int6464#2
8147# asm 2: pxor <xmm11=%xmm10,<xmm1=%xmm1
8148pxor %xmm10,%xmm1
8149
8150# qhasm: xmm7 ^= xmm12
8151# asm 1: pxor <xmm12=int6464#9,<xmm7=int6464#8
8152# asm 2: pxor <xmm12=%xmm8,<xmm7=%xmm7
8153pxor %xmm8,%xmm7
8154
8155# qhasm: xmm4 ^= xmm12
8156# asm 1: pxor <xmm12=int6464#9,<xmm4=int6464#5
8157# asm 2: pxor <xmm12=%xmm8,<xmm4=%xmm4
8158pxor %xmm8,%xmm4
8159
8160# qhasm: xmm1 ^= xmm8
8161# asm 1: pxor <xmm8=int6464#10,<xmm1=int6464#2
8162# asm 2: pxor <xmm8=%xmm9,<xmm1=%xmm1
8163pxor %xmm9,%xmm1
8164
8165# qhasm: xmm2 ^= xmm8
8166# asm 1: pxor <xmm8=int6464#10,<xmm2=int6464#3
8167# asm 2: pxor <xmm8=%xmm9,<xmm2=%xmm2
8168pxor %xmm9,%xmm2
8169
8170# qhasm: xmm7 ^= xmm0
8171# asm 1: pxor <xmm0=int6464#1,<xmm7=int6464#8
8172# asm 2: pxor <xmm0=%xmm0,<xmm7=%xmm7
8173pxor %xmm0,%xmm7
8174
8175# qhasm: xmm1 ^= xmm6
8176# asm 1: pxor <xmm6=int6464#7,<xmm1=int6464#2
8177# asm 2: pxor <xmm6=%xmm6,<xmm1=%xmm1
8178pxor %xmm6,%xmm1
8179
8180# qhasm: xmm4 ^= xmm7
8181# asm 1: pxor <xmm7=int6464#8,<xmm4=int6464#5
8182# asm 2: pxor <xmm7=%xmm7,<xmm4=%xmm4
8183pxor %xmm7,%xmm4
8184
8185# qhasm: xmm6 ^= xmm0
8186# asm 1: pxor <xmm0=int6464#1,<xmm6=int6464#7
8187# asm 2: pxor <xmm0=%xmm0,<xmm6=%xmm6
8188pxor %xmm0,%xmm6
8189
8190# qhasm: xmm0 ^= xmm1
8191# asm 1: pxor <xmm1=int6464#2,<xmm0=int6464#1
8192# asm 2: pxor <xmm1=%xmm1,<xmm0=%xmm0
8193pxor %xmm1,%xmm0
8194
8195# qhasm: xmm1 ^= xmm5
8196# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
8197# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
8198pxor %xmm5,%xmm1
8199
8200# qhasm: xmm5 ^= xmm2
8201# asm 1: pxor <xmm2=int6464#3,<xmm5=int6464#6
8202# asm 2: pxor <xmm2=%xmm2,<xmm5=%xmm5
8203pxor %xmm2,%xmm5
8204
8205# qhasm: xmm4 ^= xmm5
8206# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
8207# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
8208pxor %xmm5,%xmm4
8209
8210# qhasm: xmm2 ^= xmm3
8211# asm 1: pxor <xmm3=int6464#4,<xmm2=int6464#3
8212# asm 2: pxor <xmm3=%xmm3,<xmm2=%xmm2
8213pxor %xmm3,%xmm2
8214
8215# qhasm: xmm3 ^= xmm5
8216# asm 1: pxor <xmm5=int6464#6,<xmm3=int6464#4
8217# asm 2: pxor <xmm5=%xmm5,<xmm3=%xmm3
8218pxor %xmm5,%xmm3
8219
8220# qhasm: xmm6 ^= xmm3
8221# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
8222# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
8223pxor %xmm3,%xmm6
8224
8225# qhasm: xmm8 = shuffle dwords of xmm0 by 0x93
8226# asm 1: pshufd $0x93,<xmm0=int6464#1,>xmm8=int6464#9
8227# asm 2: pshufd $0x93,<xmm0=%xmm0,>xmm8=%xmm8
8228pshufd $0x93,%xmm0,%xmm8
8229
8230# qhasm: xmm9 = shuffle dwords of xmm1 by 0x93
8231# asm 1: pshufd $0x93,<xmm1=int6464#2,>xmm9=int6464#10
8232# asm 2: pshufd $0x93,<xmm1=%xmm1,>xmm9=%xmm9
8233pshufd $0x93,%xmm1,%xmm9
8234
8235# qhasm: xmm10 = shuffle dwords of xmm4 by 0x93
8236# asm 1: pshufd $0x93,<xmm4=int6464#5,>xmm10=int6464#11
8237# asm 2: pshufd $0x93,<xmm4=%xmm4,>xmm10=%xmm10
8238pshufd $0x93,%xmm4,%xmm10
8239
8240# qhasm: xmm11 = shuffle dwords of xmm6 by 0x93
8241# asm 1: pshufd $0x93,<xmm6=int6464#7,>xmm11=int6464#12
8242# asm 2: pshufd $0x93,<xmm6=%xmm6,>xmm11=%xmm11
8243pshufd $0x93,%xmm6,%xmm11
8244
8245# qhasm: xmm12 = shuffle dwords of xmm3 by 0x93
8246# asm 1: pshufd $0x93,<xmm3=int6464#4,>xmm12=int6464#13
8247# asm 2: pshufd $0x93,<xmm3=%xmm3,>xmm12=%xmm12
8248pshufd $0x93,%xmm3,%xmm12
8249
8250# qhasm: xmm13 = shuffle dwords of xmm7 by 0x93
8251# asm 1: pshufd $0x93,<xmm7=int6464#8,>xmm13=int6464#14
8252# asm 2: pshufd $0x93,<xmm7=%xmm7,>xmm13=%xmm13
8253pshufd $0x93,%xmm7,%xmm13
8254
8255# qhasm: xmm14 = shuffle dwords of xmm2 by 0x93
8256# asm 1: pshufd $0x93,<xmm2=int6464#3,>xmm14=int6464#15
8257# asm 2: pshufd $0x93,<xmm2=%xmm2,>xmm14=%xmm14
8258pshufd $0x93,%xmm2,%xmm14
8259
8260# qhasm: xmm15 = shuffle dwords of xmm5 by 0x93
8261# asm 1: pshufd $0x93,<xmm5=int6464#6,>xmm15=int6464#16
8262# asm 2: pshufd $0x93,<xmm5=%xmm5,>xmm15=%xmm15
8263pshufd $0x93,%xmm5,%xmm15
8264
8265# qhasm: xmm0 ^= xmm8
8266# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
8267# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
8268pxor %xmm8,%xmm0
8269
8270# qhasm: xmm1 ^= xmm9
8271# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
8272# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
8273pxor %xmm9,%xmm1
8274
8275# qhasm: xmm4 ^= xmm10
8276# asm 1: pxor <xmm10=int6464#11,<xmm4=int6464#5
8277# asm 2: pxor <xmm10=%xmm10,<xmm4=%xmm4
8278pxor %xmm10,%xmm4
8279
8280# qhasm: xmm6 ^= xmm11
8281# asm 1: pxor <xmm11=int6464#12,<xmm6=int6464#7
8282# asm 2: pxor <xmm11=%xmm11,<xmm6=%xmm6
8283pxor %xmm11,%xmm6
8284
8285# qhasm: xmm3 ^= xmm12
8286# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
8287# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
8288pxor %xmm12,%xmm3
8289
8290# qhasm: xmm7 ^= xmm13
8291# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
8292# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
8293pxor %xmm13,%xmm7
8294
8295# qhasm: xmm2 ^= xmm14
8296# asm 1: pxor <xmm14=int6464#15,<xmm2=int6464#3
8297# asm 2: pxor <xmm14=%xmm14,<xmm2=%xmm2
8298pxor %xmm14,%xmm2
8299
8300# qhasm: xmm5 ^= xmm15
8301# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
8302# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
8303pxor %xmm15,%xmm5
8304
8305# qhasm: xmm8 ^= xmm5
8306# asm 1: pxor <xmm5=int6464#6,<xmm8=int6464#9
8307# asm 2: pxor <xmm5=%xmm5,<xmm8=%xmm8
8308pxor %xmm5,%xmm8
8309
8310# qhasm: xmm9 ^= xmm0
8311# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
8312# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
8313pxor %xmm0,%xmm9
8314
8315# qhasm: xmm10 ^= xmm1
8316# asm 1: pxor <xmm1=int6464#2,<xmm10=int6464#11
8317# asm 2: pxor <xmm1=%xmm1,<xmm10=%xmm10
8318pxor %xmm1,%xmm10
8319
8320# qhasm: xmm9 ^= xmm5
8321# asm 1: pxor <xmm5=int6464#6,<xmm9=int6464#10
8322# asm 2: pxor <xmm5=%xmm5,<xmm9=%xmm9
8323pxor %xmm5,%xmm9
8324
8325# qhasm: xmm11 ^= xmm4
8326# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
8327# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
8328pxor %xmm4,%xmm11
8329
8330# qhasm: xmm12 ^= xmm6
8331# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#13
8332# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm12
8333pxor %xmm6,%xmm12
8334
8335# qhasm: xmm13 ^= xmm3
8336# asm 1: pxor <xmm3=int6464#4,<xmm13=int6464#14
8337# asm 2: pxor <xmm3=%xmm3,<xmm13=%xmm13
8338pxor %xmm3,%xmm13
8339
8340# qhasm: xmm11 ^= xmm5
8341# asm 1: pxor <xmm5=int6464#6,<xmm11=int6464#12
8342# asm 2: pxor <xmm5=%xmm5,<xmm11=%xmm11
8343pxor %xmm5,%xmm11
8344
8345# qhasm: xmm14 ^= xmm7
8346# asm 1: pxor <xmm7=int6464#8,<xmm14=int6464#15
8347# asm 2: pxor <xmm7=%xmm7,<xmm14=%xmm14
8348pxor %xmm7,%xmm14
8349
8350# qhasm: xmm15 ^= xmm2
8351# asm 1: pxor <xmm2=int6464#3,<xmm15=int6464#16
8352# asm 2: pxor <xmm2=%xmm2,<xmm15=%xmm15
8353pxor %xmm2,%xmm15
8354
8355# qhasm: xmm12 ^= xmm5
8356# asm 1: pxor <xmm5=int6464#6,<xmm12=int6464#13
8357# asm 2: pxor <xmm5=%xmm5,<xmm12=%xmm12
8358pxor %xmm5,%xmm12
8359
8360# qhasm: xmm0 = shuffle dwords of xmm0 by 0x4E
8361# asm 1: pshufd $0x4E,<xmm0=int6464#1,>xmm0=int6464#1
8362# asm 2: pshufd $0x4E,<xmm0=%xmm0,>xmm0=%xmm0
8363pshufd $0x4E,%xmm0,%xmm0
8364
8365# qhasm: xmm1 = shuffle dwords of xmm1 by 0x4E
8366# asm 1: pshufd $0x4E,<xmm1=int6464#2,>xmm1=int6464#2
8367# asm 2: pshufd $0x4E,<xmm1=%xmm1,>xmm1=%xmm1
8368pshufd $0x4E,%xmm1,%xmm1
8369
8370# qhasm: xmm4 = shuffle dwords of xmm4 by 0x4E
8371# asm 1: pshufd $0x4E,<xmm4=int6464#5,>xmm4=int6464#5
8372# asm 2: pshufd $0x4E,<xmm4=%xmm4,>xmm4=%xmm4
8373pshufd $0x4E,%xmm4,%xmm4
8374
8375# qhasm: xmm6 = shuffle dwords of xmm6 by 0x4E
8376# asm 1: pshufd $0x4E,<xmm6=int6464#7,>xmm6=int6464#7
8377# asm 2: pshufd $0x4E,<xmm6=%xmm6,>xmm6=%xmm6
8378pshufd $0x4E,%xmm6,%xmm6
8379
8380# qhasm: xmm3 = shuffle dwords of xmm3 by 0x4E
8381# asm 1: pshufd $0x4E,<xmm3=int6464#4,>xmm3=int6464#4
8382# asm 2: pshufd $0x4E,<xmm3=%xmm3,>xmm3=%xmm3
8383pshufd $0x4E,%xmm3,%xmm3
8384
8385# qhasm: xmm7 = shuffle dwords of xmm7 by 0x4E
8386# asm 1: pshufd $0x4E,<xmm7=int6464#8,>xmm7=int6464#8
8387# asm 2: pshufd $0x4E,<xmm7=%xmm7,>xmm7=%xmm7
8388pshufd $0x4E,%xmm7,%xmm7
8389
8390# qhasm: xmm2 = shuffle dwords of xmm2 by 0x4E
8391# asm 1: pshufd $0x4E,<xmm2=int6464#3,>xmm2=int6464#3
8392# asm 2: pshufd $0x4E,<xmm2=%xmm2,>xmm2=%xmm2
8393pshufd $0x4E,%xmm2,%xmm2
8394
8395# qhasm: xmm5 = shuffle dwords of xmm5 by 0x4E
8396# asm 1: pshufd $0x4E,<xmm5=int6464#6,>xmm5=int6464#6
8397# asm 2: pshufd $0x4E,<xmm5=%xmm5,>xmm5=%xmm5
8398pshufd $0x4E,%xmm5,%xmm5
8399
8400# qhasm: xmm8 ^= xmm0
8401# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
8402# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
8403pxor %xmm0,%xmm8
8404
8405# qhasm: xmm9 ^= xmm1
8406# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
8407# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
8408pxor %xmm1,%xmm9
8409
8410# qhasm: xmm10 ^= xmm4
8411# asm 1: pxor <xmm4=int6464#5,<xmm10=int6464#11
8412# asm 2: pxor <xmm4=%xmm4,<xmm10=%xmm10
8413pxor %xmm4,%xmm10
8414
8415# qhasm: xmm11 ^= xmm6
8416# asm 1: pxor <xmm6=int6464#7,<xmm11=int6464#12
8417# asm 2: pxor <xmm6=%xmm6,<xmm11=%xmm11
8418pxor %xmm6,%xmm11
8419
8420# qhasm: xmm12 ^= xmm3
8421# asm 1: pxor <xmm3=int6464#4,<xmm12=int6464#13
8422# asm 2: pxor <xmm3=%xmm3,<xmm12=%xmm12
8423pxor %xmm3,%xmm12
8424
8425# qhasm: xmm13 ^= xmm7
8426# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
8427# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
8428pxor %xmm7,%xmm13
8429
8430# qhasm: xmm14 ^= xmm2
8431# asm 1: pxor <xmm2=int6464#3,<xmm14=int6464#15
8432# asm 2: pxor <xmm2=%xmm2,<xmm14=%xmm14
8433pxor %xmm2,%xmm14
8434
8435# qhasm: xmm15 ^= xmm5
8436# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
8437# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
8438pxor %xmm5,%xmm15
8439
8440# qhasm: xmm8 ^= *(int128 *)(c + 896)
8441# asm 1: pxor 896(<c=int64#4),<xmm8=int6464#9
8442# asm 2: pxor 896(<c=%rcx),<xmm8=%xmm8
8443pxor 896(%rcx),%xmm8
8444
8445# qhasm: shuffle bytes of xmm8 by SR
8446# asm 1: pshufb SR,<xmm8=int6464#9
8447# asm 2: pshufb SR,<xmm8=%xmm8
8448pshufb SR,%xmm8
8449
8450# qhasm: xmm9 ^= *(int128 *)(c + 912)
8451# asm 1: pxor 912(<c=int64#4),<xmm9=int6464#10
8452# asm 2: pxor 912(<c=%rcx),<xmm9=%xmm9
8453pxor 912(%rcx),%xmm9
8454
8455# qhasm: shuffle bytes of xmm9 by SR
8456# asm 1: pshufb SR,<xmm9=int6464#10
8457# asm 2: pshufb SR,<xmm9=%xmm9
8458pshufb SR,%xmm9
8459
8460# qhasm: xmm10 ^= *(int128 *)(c + 928)
8461# asm 1: pxor 928(<c=int64#4),<xmm10=int6464#11
8462# asm 2: pxor 928(<c=%rcx),<xmm10=%xmm10
8463pxor 928(%rcx),%xmm10
8464
8465# qhasm: shuffle bytes of xmm10 by SR
8466# asm 1: pshufb SR,<xmm10=int6464#11
8467# asm 2: pshufb SR,<xmm10=%xmm10
8468pshufb SR,%xmm10
8469
8470# qhasm: xmm11 ^= *(int128 *)(c + 944)
8471# asm 1: pxor 944(<c=int64#4),<xmm11=int6464#12
8472# asm 2: pxor 944(<c=%rcx),<xmm11=%xmm11
8473pxor 944(%rcx),%xmm11
8474
8475# qhasm: shuffle bytes of xmm11 by SR
8476# asm 1: pshufb SR,<xmm11=int6464#12
8477# asm 2: pshufb SR,<xmm11=%xmm11
8478pshufb SR,%xmm11
8479
8480# qhasm: xmm12 ^= *(int128 *)(c + 960)
8481# asm 1: pxor 960(<c=int64#4),<xmm12=int6464#13
8482# asm 2: pxor 960(<c=%rcx),<xmm12=%xmm12
8483pxor 960(%rcx),%xmm12
8484
8485# qhasm: shuffle bytes of xmm12 by SR
8486# asm 1: pshufb SR,<xmm12=int6464#13
8487# asm 2: pshufb SR,<xmm12=%xmm12
8488pshufb SR,%xmm12
8489
8490# qhasm: xmm13 ^= *(int128 *)(c + 976)
8491# asm 1: pxor 976(<c=int64#4),<xmm13=int6464#14
8492# asm 2: pxor 976(<c=%rcx),<xmm13=%xmm13
8493pxor 976(%rcx),%xmm13
8494
8495# qhasm: shuffle bytes of xmm13 by SR
8496# asm 1: pshufb SR,<xmm13=int6464#14
8497# asm 2: pshufb SR,<xmm13=%xmm13
8498pshufb SR,%xmm13
8499
8500# qhasm: xmm14 ^= *(int128 *)(c + 992)
8501# asm 1: pxor 992(<c=int64#4),<xmm14=int6464#15
8502# asm 2: pxor 992(<c=%rcx),<xmm14=%xmm14
8503pxor 992(%rcx),%xmm14
8504
8505# qhasm: shuffle bytes of xmm14 by SR
8506# asm 1: pshufb SR,<xmm14=int6464#15
8507# asm 2: pshufb SR,<xmm14=%xmm14
8508pshufb SR,%xmm14
8509
8510# qhasm: xmm15 ^= *(int128 *)(c + 1008)
8511# asm 1: pxor 1008(<c=int64#4),<xmm15=int6464#16
8512# asm 2: pxor 1008(<c=%rcx),<xmm15=%xmm15
8513pxor 1008(%rcx),%xmm15
8514
8515# qhasm: shuffle bytes of xmm15 by SR
8516# asm 1: pshufb SR,<xmm15=int6464#16
8517# asm 2: pshufb SR,<xmm15=%xmm15
8518pshufb SR,%xmm15
8519
8520# qhasm: xmm13 ^= xmm14
8521# asm 1: pxor <xmm14=int6464#15,<xmm13=int6464#14
8522# asm 2: pxor <xmm14=%xmm14,<xmm13=%xmm13
8523pxor %xmm14,%xmm13
8524
8525# qhasm: xmm10 ^= xmm9
8526# asm 1: pxor <xmm9=int6464#10,<xmm10=int6464#11
8527# asm 2: pxor <xmm9=%xmm9,<xmm10=%xmm10
8528pxor %xmm9,%xmm10
8529
8530# qhasm: xmm13 ^= xmm8
8531# asm 1: pxor <xmm8=int6464#9,<xmm13=int6464#14
8532# asm 2: pxor <xmm8=%xmm8,<xmm13=%xmm13
8533pxor %xmm8,%xmm13
8534
8535# qhasm: xmm14 ^= xmm10
8536# asm 1: pxor <xmm10=int6464#11,<xmm14=int6464#15
8537# asm 2: pxor <xmm10=%xmm10,<xmm14=%xmm14
8538pxor %xmm10,%xmm14
8539
8540# qhasm: xmm11 ^= xmm8
8541# asm 1: pxor <xmm8=int6464#9,<xmm11=int6464#12
8542# asm 2: pxor <xmm8=%xmm8,<xmm11=%xmm11
8543pxor %xmm8,%xmm11
8544
8545# qhasm: xmm14 ^= xmm11
8546# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
8547# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
8548pxor %xmm11,%xmm14
8549
8550# qhasm: xmm11 ^= xmm15
8551# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#12
8552# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm11
8553pxor %xmm15,%xmm11
8554
8555# qhasm: xmm11 ^= xmm12
8556# asm 1: pxor <xmm12=int6464#13,<xmm11=int6464#12
8557# asm 2: pxor <xmm12=%xmm12,<xmm11=%xmm11
8558pxor %xmm12,%xmm11
8559
8560# qhasm: xmm15 ^= xmm13
8561# asm 1: pxor <xmm13=int6464#14,<xmm15=int6464#16
8562# asm 2: pxor <xmm13=%xmm13,<xmm15=%xmm15
8563pxor %xmm13,%xmm15
8564
8565# qhasm: xmm11 ^= xmm9
8566# asm 1: pxor <xmm9=int6464#10,<xmm11=int6464#12
8567# asm 2: pxor <xmm9=%xmm9,<xmm11=%xmm11
8568pxor %xmm9,%xmm11
8569
8570# qhasm: xmm12 ^= xmm13
8571# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
8572# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
8573pxor %xmm13,%xmm12
8574
8575# qhasm: xmm10 ^= xmm15
8576# asm 1: pxor <xmm15=int6464#16,<xmm10=int6464#11
8577# asm 2: pxor <xmm15=%xmm15,<xmm10=%xmm10
8578pxor %xmm15,%xmm10
8579
8580# qhasm: xmm9 ^= xmm13
8581# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
8582# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
8583pxor %xmm13,%xmm9
8584
8585# qhasm: xmm3 = xmm15
8586# asm 1: movdqa <xmm15=int6464#16,>xmm3=int6464#1
8587# asm 2: movdqa <xmm15=%xmm15,>xmm3=%xmm0
8588movdqa %xmm15,%xmm0
8589
8590# qhasm: xmm2 = xmm9
8591# asm 1: movdqa <xmm9=int6464#10,>xmm2=int6464#2
8592# asm 2: movdqa <xmm9=%xmm9,>xmm2=%xmm1
8593movdqa %xmm9,%xmm1
8594
8595# qhasm: xmm1 = xmm13
8596# asm 1: movdqa <xmm13=int6464#14,>xmm1=int6464#3
8597# asm 2: movdqa <xmm13=%xmm13,>xmm1=%xmm2
8598movdqa %xmm13,%xmm2
8599
8600# qhasm: xmm5 = xmm10
8601# asm 1: movdqa <xmm10=int6464#11,>xmm5=int6464#4
8602# asm 2: movdqa <xmm10=%xmm10,>xmm5=%xmm3
8603movdqa %xmm10,%xmm3
8604
8605# qhasm: xmm4 = xmm14
8606# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#5
8607# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm4
8608movdqa %xmm14,%xmm4
8609
8610# qhasm: xmm3 ^= xmm12
8611# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#1
8612# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm0
8613pxor %xmm12,%xmm0
8614
8615# qhasm: xmm2 ^= xmm10
8616# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#2
8617# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm1
8618pxor %xmm10,%xmm1
8619
8620# qhasm: xmm1 ^= xmm11
8621# asm 1: pxor <xmm11=int6464#12,<xmm1=int6464#3
8622# asm 2: pxor <xmm11=%xmm11,<xmm1=%xmm2
8623pxor %xmm11,%xmm2
8624
8625# qhasm: xmm5 ^= xmm12
8626# asm 1: pxor <xmm12=int6464#13,<xmm5=int6464#4
8627# asm 2: pxor <xmm12=%xmm12,<xmm5=%xmm3
8628pxor %xmm12,%xmm3
8629
8630# qhasm: xmm4 ^= xmm8
8631# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
8632# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
8633pxor %xmm8,%xmm4
8634
8635# qhasm: xmm6 = xmm3
8636# asm 1: movdqa <xmm3=int6464#1,>xmm6=int6464#6
8637# asm 2: movdqa <xmm3=%xmm0,>xmm6=%xmm5
8638movdqa %xmm0,%xmm5
8639
8640# qhasm: xmm0 = xmm2
8641# asm 1: movdqa <xmm2=int6464#2,>xmm0=int6464#7
8642# asm 2: movdqa <xmm2=%xmm1,>xmm0=%xmm6
8643movdqa %xmm1,%xmm6
8644
8645# qhasm: xmm7 = xmm3
8646# asm 1: movdqa <xmm3=int6464#1,>xmm7=int6464#8
8647# asm 2: movdqa <xmm3=%xmm0,>xmm7=%xmm7
8648movdqa %xmm0,%xmm7
8649
8650# qhasm: xmm2 |= xmm1
8651# asm 1: por <xmm1=int6464#3,<xmm2=int6464#2
8652# asm 2: por <xmm1=%xmm2,<xmm2=%xmm1
8653por %xmm2,%xmm1
8654
8655# qhasm: xmm3 |= xmm4
8656# asm 1: por <xmm4=int6464#5,<xmm3=int6464#1
8657# asm 2: por <xmm4=%xmm4,<xmm3=%xmm0
8658por %xmm4,%xmm0
8659
8660# qhasm: xmm7 ^= xmm0
8661# asm 1: pxor <xmm0=int6464#7,<xmm7=int6464#8
8662# asm 2: pxor <xmm0=%xmm6,<xmm7=%xmm7
8663pxor %xmm6,%xmm7
8664
8665# qhasm: xmm6 &= xmm4
8666# asm 1: pand <xmm4=int6464#5,<xmm6=int6464#6
8667# asm 2: pand <xmm4=%xmm4,<xmm6=%xmm5
8668pand %xmm4,%xmm5
8669
8670# qhasm: xmm0 &= xmm1
8671# asm 1: pand <xmm1=int6464#3,<xmm0=int6464#7
8672# asm 2: pand <xmm1=%xmm2,<xmm0=%xmm6
8673pand %xmm2,%xmm6
8674
8675# qhasm: xmm4 ^= xmm1
8676# asm 1: pxor <xmm1=int6464#3,<xmm4=int6464#5
8677# asm 2: pxor <xmm1=%xmm2,<xmm4=%xmm4
8678pxor %xmm2,%xmm4
8679
8680# qhasm: xmm7 &= xmm4
8681# asm 1: pand <xmm4=int6464#5,<xmm7=int6464#8
8682# asm 2: pand <xmm4=%xmm4,<xmm7=%xmm7
8683pand %xmm4,%xmm7
8684
8685# qhasm: xmm4 = xmm11
8686# asm 1: movdqa <xmm11=int6464#12,>xmm4=int6464#3
8687# asm 2: movdqa <xmm11=%xmm11,>xmm4=%xmm2
8688movdqa %xmm11,%xmm2
8689
8690# qhasm: xmm4 ^= xmm8
8691# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#3
8692# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm2
8693pxor %xmm8,%xmm2
8694
8695# qhasm: xmm5 &= xmm4
8696# asm 1: pand <xmm4=int6464#3,<xmm5=int6464#4
8697# asm 2: pand <xmm4=%xmm2,<xmm5=%xmm3
8698pand %xmm2,%xmm3
8699
8700# qhasm: xmm3 ^= xmm5
8701# asm 1: pxor <xmm5=int6464#4,<xmm3=int6464#1
8702# asm 2: pxor <xmm5=%xmm3,<xmm3=%xmm0
8703pxor %xmm3,%xmm0
8704
8705# qhasm: xmm2 ^= xmm5
8706# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
8707# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
8708pxor %xmm3,%xmm1
8709
8710# qhasm: xmm5 = xmm15
8711# asm 1: movdqa <xmm15=int6464#16,>xmm5=int6464#3
8712# asm 2: movdqa <xmm15=%xmm15,>xmm5=%xmm2
8713movdqa %xmm15,%xmm2
8714
8715# qhasm: xmm5 ^= xmm9
8716# asm 1: pxor <xmm9=int6464#10,<xmm5=int6464#3
8717# asm 2: pxor <xmm9=%xmm9,<xmm5=%xmm2
8718pxor %xmm9,%xmm2
8719
8720# qhasm: xmm4 = xmm13
8721# asm 1: movdqa <xmm13=int6464#14,>xmm4=int6464#4
8722# asm 2: movdqa <xmm13=%xmm13,>xmm4=%xmm3
8723movdqa %xmm13,%xmm3
8724
8725# qhasm: xmm1 = xmm5
8726# asm 1: movdqa <xmm5=int6464#3,>xmm1=int6464#5
8727# asm 2: movdqa <xmm5=%xmm2,>xmm1=%xmm4
8728movdqa %xmm2,%xmm4
8729
8730# qhasm: xmm4 ^= xmm14
8731# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#4
8732# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm3
8733pxor %xmm14,%xmm3
8734
8735# qhasm: xmm1 |= xmm4
8736# asm 1: por <xmm4=int6464#4,<xmm1=int6464#5
8737# asm 2: por <xmm4=%xmm3,<xmm1=%xmm4
8738por %xmm3,%xmm4
8739
8740# qhasm: xmm5 &= xmm4
8741# asm 1: pand <xmm4=int6464#4,<xmm5=int6464#3
8742# asm 2: pand <xmm4=%xmm3,<xmm5=%xmm2
8743pand %xmm3,%xmm2
8744
8745# qhasm: xmm0 ^= xmm5
8746# asm 1: pxor <xmm5=int6464#3,<xmm0=int6464#7
8747# asm 2: pxor <xmm5=%xmm2,<xmm0=%xmm6
8748pxor %xmm2,%xmm6
8749
8750# qhasm: xmm3 ^= xmm7
8751# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#1
8752# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm0
8753pxor %xmm7,%xmm0
8754
8755# qhasm: xmm2 ^= xmm6
8756# asm 1: pxor <xmm6=int6464#6,<xmm2=int6464#2
8757# asm 2: pxor <xmm6=%xmm5,<xmm2=%xmm1
8758pxor %xmm5,%xmm1
8759
8760# qhasm: xmm1 ^= xmm7
8761# asm 1: pxor <xmm7=int6464#8,<xmm1=int6464#5
8762# asm 2: pxor <xmm7=%xmm7,<xmm1=%xmm4
8763pxor %xmm7,%xmm4
8764
8765# qhasm: xmm0 ^= xmm6
8766# asm 1: pxor <xmm6=int6464#6,<xmm0=int6464#7
8767# asm 2: pxor <xmm6=%xmm5,<xmm0=%xmm6
8768pxor %xmm5,%xmm6
8769
8770# qhasm: xmm1 ^= xmm6
8771# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
8772# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
8773pxor %xmm5,%xmm4
8774
8775# qhasm: xmm4 = xmm10
8776# asm 1: movdqa <xmm10=int6464#11,>xmm4=int6464#3
8777# asm 2: movdqa <xmm10=%xmm10,>xmm4=%xmm2
8778movdqa %xmm10,%xmm2
8779
8780# qhasm: xmm5 = xmm12
8781# asm 1: movdqa <xmm12=int6464#13,>xmm5=int6464#4
8782# asm 2: movdqa <xmm12=%xmm12,>xmm5=%xmm3
8783movdqa %xmm12,%xmm3
8784
8785# qhasm: xmm6 = xmm9
8786# asm 1: movdqa <xmm9=int6464#10,>xmm6=int6464#6
8787# asm 2: movdqa <xmm9=%xmm9,>xmm6=%xmm5
8788movdqa %xmm9,%xmm5
8789
8790# qhasm: xmm7 = xmm15
8791# asm 1: movdqa <xmm15=int6464#16,>xmm7=int6464#8
8792# asm 2: movdqa <xmm15=%xmm15,>xmm7=%xmm7
8793movdqa %xmm15,%xmm7
8794
8795# qhasm: xmm4 &= xmm11
8796# asm 1: pand <xmm11=int6464#12,<xmm4=int6464#3
8797# asm 2: pand <xmm11=%xmm11,<xmm4=%xmm2
8798pand %xmm11,%xmm2
8799
8800# qhasm: xmm5 &= xmm8
8801# asm 1: pand <xmm8=int6464#9,<xmm5=int6464#4
8802# asm 2: pand <xmm8=%xmm8,<xmm5=%xmm3
8803pand %xmm8,%xmm3
8804
8805# qhasm: xmm6 &= xmm13
8806# asm 1: pand <xmm13=int6464#14,<xmm6=int6464#6
8807# asm 2: pand <xmm13=%xmm13,<xmm6=%xmm5
8808pand %xmm13,%xmm5
8809
8810# qhasm: xmm7 |= xmm14
8811# asm 1: por <xmm14=int6464#15,<xmm7=int6464#8
8812# asm 2: por <xmm14=%xmm14,<xmm7=%xmm7
8813por %xmm14,%xmm7
8814
8815# qhasm: xmm3 ^= xmm4
8816# asm 1: pxor <xmm4=int6464#3,<xmm3=int6464#1
8817# asm 2: pxor <xmm4=%xmm2,<xmm3=%xmm0
8818pxor %xmm2,%xmm0
8819
8820# qhasm: xmm2 ^= xmm5
8821# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
8822# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
8823pxor %xmm3,%xmm1
8824
8825# qhasm: xmm1 ^= xmm6
8826# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
8827# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
8828pxor %xmm5,%xmm4
8829
8830# qhasm: xmm0 ^= xmm7
8831# asm 1: pxor <xmm7=int6464#8,<xmm0=int6464#7
8832# asm 2: pxor <xmm7=%xmm7,<xmm0=%xmm6
8833pxor %xmm7,%xmm6
8834
8835# qhasm: xmm4 = xmm3
8836# asm 1: movdqa <xmm3=int6464#1,>xmm4=int6464#3
8837# asm 2: movdqa <xmm3=%xmm0,>xmm4=%xmm2
8838movdqa %xmm0,%xmm2
8839
8840# qhasm: xmm4 ^= xmm2
8841# asm 1: pxor <xmm2=int6464#2,<xmm4=int6464#3
8842# asm 2: pxor <xmm2=%xmm1,<xmm4=%xmm2
8843pxor %xmm1,%xmm2
8844
8845# qhasm: xmm3 &= xmm1
8846# asm 1: pand <xmm1=int6464#5,<xmm3=int6464#1
8847# asm 2: pand <xmm1=%xmm4,<xmm3=%xmm0
8848pand %xmm4,%xmm0
8849
8850# qhasm: xmm6 = xmm0
8851# asm 1: movdqa <xmm0=int6464#7,>xmm6=int6464#4
8852# asm 2: movdqa <xmm0=%xmm6,>xmm6=%xmm3
8853movdqa %xmm6,%xmm3
8854
8855# qhasm: xmm6 ^= xmm3
8856# asm 1: pxor <xmm3=int6464#1,<xmm6=int6464#4
8857# asm 2: pxor <xmm3=%xmm0,<xmm6=%xmm3
8858pxor %xmm0,%xmm3
8859
8860# qhasm: xmm7 = xmm4
8861# asm 1: movdqa <xmm4=int6464#3,>xmm7=int6464#6
8862# asm 2: movdqa <xmm4=%xmm2,>xmm7=%xmm5
8863movdqa %xmm2,%xmm5
8864
8865# qhasm: xmm7 &= xmm6
8866# asm 1: pand <xmm6=int6464#4,<xmm7=int6464#6
8867# asm 2: pand <xmm6=%xmm3,<xmm7=%xmm5
8868pand %xmm3,%xmm5
8869
8870# qhasm: xmm7 ^= xmm2
8871# asm 1: pxor <xmm2=int6464#2,<xmm7=int6464#6
8872# asm 2: pxor <xmm2=%xmm1,<xmm7=%xmm5
8873pxor %xmm1,%xmm5
8874
8875# qhasm: xmm5 = xmm1
8876# asm 1: movdqa <xmm1=int6464#5,>xmm5=int6464#8
8877# asm 2: movdqa <xmm1=%xmm4,>xmm5=%xmm7
8878movdqa %xmm4,%xmm7
8879
8880# qhasm: xmm5 ^= xmm0
8881# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
8882# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
8883pxor %xmm6,%xmm7
8884
8885# qhasm: xmm3 ^= xmm2
8886# asm 1: pxor <xmm2=int6464#2,<xmm3=int6464#1
8887# asm 2: pxor <xmm2=%xmm1,<xmm3=%xmm0
8888pxor %xmm1,%xmm0
8889
8890# qhasm: xmm5 &= xmm3
8891# asm 1: pand <xmm3=int6464#1,<xmm5=int6464#8
8892# asm 2: pand <xmm3=%xmm0,<xmm5=%xmm7
8893pand %xmm0,%xmm7
8894
8895# qhasm: xmm5 ^= xmm0
8896# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
8897# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
8898pxor %xmm6,%xmm7
8899
8900# qhasm: xmm1 ^= xmm5
8901# asm 1: pxor <xmm5=int6464#8,<xmm1=int6464#5
8902# asm 2: pxor <xmm5=%xmm7,<xmm1=%xmm4
8903pxor %xmm7,%xmm4
8904
8905# qhasm: xmm2 = xmm6
8906# asm 1: movdqa <xmm6=int6464#4,>xmm2=int6464#1
8907# asm 2: movdqa <xmm6=%xmm3,>xmm2=%xmm0
8908movdqa %xmm3,%xmm0
8909
8910# qhasm: xmm2 ^= xmm5
8911# asm 1: pxor <xmm5=int6464#8,<xmm2=int6464#1
8912# asm 2: pxor <xmm5=%xmm7,<xmm2=%xmm0
8913pxor %xmm7,%xmm0
8914
8915# qhasm: xmm2 &= xmm0
8916# asm 1: pand <xmm0=int6464#7,<xmm2=int6464#1
8917# asm 2: pand <xmm0=%xmm6,<xmm2=%xmm0
8918pand %xmm6,%xmm0
8919
8920# qhasm: xmm1 ^= xmm2
8921# asm 1: pxor <xmm2=int6464#1,<xmm1=int6464#5
8922# asm 2: pxor <xmm2=%xmm0,<xmm1=%xmm4
8923pxor %xmm0,%xmm4
8924
8925# qhasm: xmm6 ^= xmm2
8926# asm 1: pxor <xmm2=int6464#1,<xmm6=int6464#4
8927# asm 2: pxor <xmm2=%xmm0,<xmm6=%xmm3
8928pxor %xmm0,%xmm3
8929
8930# qhasm: xmm6 &= xmm7
8931# asm 1: pand <xmm7=int6464#6,<xmm6=int6464#4
8932# asm 2: pand <xmm7=%xmm5,<xmm6=%xmm3
8933pand %xmm5,%xmm3
8934
8935# qhasm: xmm6 ^= xmm4
8936# asm 1: pxor <xmm4=int6464#3,<xmm6=int6464#4
8937# asm 2: pxor <xmm4=%xmm2,<xmm6=%xmm3
8938pxor %xmm2,%xmm3
8939
8940# qhasm: xmm4 = xmm14
8941# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#1
8942# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm0
8943movdqa %xmm14,%xmm0
8944
8945# qhasm: xmm0 = xmm13
8946# asm 1: movdqa <xmm13=int6464#14,>xmm0=int6464#2
8947# asm 2: movdqa <xmm13=%xmm13,>xmm0=%xmm1
8948movdqa %xmm13,%xmm1
8949
8950# qhasm: xmm2 = xmm7
8951# asm 1: movdqa <xmm7=int6464#6,>xmm2=int6464#3
8952# asm 2: movdqa <xmm7=%xmm5,>xmm2=%xmm2
8953movdqa %xmm5,%xmm2
8954
8955# qhasm: xmm2 ^= xmm6
8956# asm 1: pxor <xmm6=int6464#4,<xmm2=int6464#3
8957# asm 2: pxor <xmm6=%xmm3,<xmm2=%xmm2
8958pxor %xmm3,%xmm2
8959
8960# qhasm: xmm2 &= xmm14
8961# asm 1: pand <xmm14=int6464#15,<xmm2=int6464#3
8962# asm 2: pand <xmm14=%xmm14,<xmm2=%xmm2
8963pand %xmm14,%xmm2
8964
8965# qhasm: xmm14 ^= xmm13
8966# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
8967# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
8968pxor %xmm13,%xmm14
8969
8970# qhasm: xmm14 &= xmm6
8971# asm 1: pand <xmm6=int6464#4,<xmm14=int6464#15
8972# asm 2: pand <xmm6=%xmm3,<xmm14=%xmm14
8973pand %xmm3,%xmm14
8974
8975# qhasm: xmm13 &= xmm7
8976# asm 1: pand <xmm7=int6464#6,<xmm13=int6464#14
8977# asm 2: pand <xmm7=%xmm5,<xmm13=%xmm13
8978pand %xmm5,%xmm13
8979
8980# qhasm: xmm14 ^= xmm13
8981# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
8982# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
8983pxor %xmm13,%xmm14
8984
8985# qhasm: xmm13 ^= xmm2
8986# asm 1: pxor <xmm2=int6464#3,<xmm13=int6464#14
8987# asm 2: pxor <xmm2=%xmm2,<xmm13=%xmm13
8988pxor %xmm2,%xmm13
8989
8990# qhasm: xmm4 ^= xmm8
8991# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#1
8992# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm0
8993pxor %xmm8,%xmm0
8994
8995# qhasm: xmm0 ^= xmm11
8996# asm 1: pxor <xmm11=int6464#12,<xmm0=int6464#2
8997# asm 2: pxor <xmm11=%xmm11,<xmm0=%xmm1
8998pxor %xmm11,%xmm1
8999
9000# qhasm: xmm7 ^= xmm5
9001# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
9002# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
9003pxor %xmm7,%xmm5
9004
9005# qhasm: xmm6 ^= xmm1
9006# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
9007# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
9008pxor %xmm4,%xmm3
9009
9010# qhasm: xmm3 = xmm7
9011# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
9012# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
9013movdqa %xmm5,%xmm2
9014
9015# qhasm: xmm3 ^= xmm6
9016# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
9017# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
9018pxor %xmm3,%xmm2
9019
9020# qhasm: xmm3 &= xmm4
9021# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
9022# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
9023pand %xmm0,%xmm2
9024
9025# qhasm: xmm4 ^= xmm0
9026# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
9027# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
9028pxor %xmm1,%xmm0
9029
9030# qhasm: xmm4 &= xmm6
9031# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
9032# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
9033pand %xmm3,%xmm0
9034
9035# qhasm: xmm0 &= xmm7
9036# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
9037# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
9038pand %xmm5,%xmm1
9039
9040# qhasm: xmm0 ^= xmm4
9041# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
9042# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
9043pxor %xmm0,%xmm1
9044
9045# qhasm: xmm4 ^= xmm3
9046# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
9047# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
9048pxor %xmm2,%xmm0
9049
9050# qhasm: xmm2 = xmm5
9051# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
9052# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
9053movdqa %xmm7,%xmm2
9054
9055# qhasm: xmm2 ^= xmm1
9056# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
9057# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
9058pxor %xmm4,%xmm2
9059
9060# qhasm: xmm2 &= xmm8
9061# asm 1: pand <xmm8=int6464#9,<xmm2=int6464#3
9062# asm 2: pand <xmm8=%xmm8,<xmm2=%xmm2
9063pand %xmm8,%xmm2
9064
9065# qhasm: xmm8 ^= xmm11
9066# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
9067# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
9068pxor %xmm11,%xmm8
9069
9070# qhasm: xmm8 &= xmm1
9071# asm 1: pand <xmm1=int6464#5,<xmm8=int6464#9
9072# asm 2: pand <xmm1=%xmm4,<xmm8=%xmm8
9073pand %xmm4,%xmm8
9074
9075# qhasm: xmm11 &= xmm5
9076# asm 1: pand <xmm5=int6464#8,<xmm11=int6464#12
9077# asm 2: pand <xmm5=%xmm7,<xmm11=%xmm11
9078pand %xmm7,%xmm11
9079
9080# qhasm: xmm8 ^= xmm11
9081# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
9082# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
9083pxor %xmm11,%xmm8
9084
9085# qhasm: xmm11 ^= xmm2
9086# asm 1: pxor <xmm2=int6464#3,<xmm11=int6464#12
9087# asm 2: pxor <xmm2=%xmm2,<xmm11=%xmm11
9088pxor %xmm2,%xmm11
9089
9090# qhasm: xmm14 ^= xmm4
9091# asm 1: pxor <xmm4=int6464#1,<xmm14=int6464#15
9092# asm 2: pxor <xmm4=%xmm0,<xmm14=%xmm14
9093pxor %xmm0,%xmm14
9094
9095# qhasm: xmm8 ^= xmm4
9096# asm 1: pxor <xmm4=int6464#1,<xmm8=int6464#9
9097# asm 2: pxor <xmm4=%xmm0,<xmm8=%xmm8
9098pxor %xmm0,%xmm8
9099
9100# qhasm: xmm13 ^= xmm0
9101# asm 1: pxor <xmm0=int6464#2,<xmm13=int6464#14
9102# asm 2: pxor <xmm0=%xmm1,<xmm13=%xmm13
9103pxor %xmm1,%xmm13
9104
9105# qhasm: xmm11 ^= xmm0
9106# asm 1: pxor <xmm0=int6464#2,<xmm11=int6464#12
9107# asm 2: pxor <xmm0=%xmm1,<xmm11=%xmm11
9108pxor %xmm1,%xmm11
9109
9110# qhasm: xmm4 = xmm15
9111# asm 1: movdqa <xmm15=int6464#16,>xmm4=int6464#1
9112# asm 2: movdqa <xmm15=%xmm15,>xmm4=%xmm0
9113movdqa %xmm15,%xmm0
9114
9115# qhasm: xmm0 = xmm9
9116# asm 1: movdqa <xmm9=int6464#10,>xmm0=int6464#2
9117# asm 2: movdqa <xmm9=%xmm9,>xmm0=%xmm1
9118movdqa %xmm9,%xmm1
9119
9120# qhasm: xmm4 ^= xmm12
9121# asm 1: pxor <xmm12=int6464#13,<xmm4=int6464#1
9122# asm 2: pxor <xmm12=%xmm12,<xmm4=%xmm0
9123pxor %xmm12,%xmm0
9124
9125# qhasm: xmm0 ^= xmm10
9126# asm 1: pxor <xmm10=int6464#11,<xmm0=int6464#2
9127# asm 2: pxor <xmm10=%xmm10,<xmm0=%xmm1
9128pxor %xmm10,%xmm1
9129
9130# qhasm: xmm3 = xmm7
9131# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
9132# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
9133movdqa %xmm5,%xmm2
9134
9135# qhasm: xmm3 ^= xmm6
9136# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
9137# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
9138pxor %xmm3,%xmm2
9139
9140# qhasm: xmm3 &= xmm4
9141# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
9142# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
9143pand %xmm0,%xmm2
9144
9145# qhasm: xmm4 ^= xmm0
9146# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
9147# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
9148pxor %xmm1,%xmm0
9149
9150# qhasm: xmm4 &= xmm6
9151# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
9152# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
9153pand %xmm3,%xmm0
9154
9155# qhasm: xmm0 &= xmm7
9156# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
9157# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
9158pand %xmm5,%xmm1
9159
9160# qhasm: xmm0 ^= xmm4
9161# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
9162# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
9163pxor %xmm0,%xmm1
9164
9165# qhasm: xmm4 ^= xmm3
9166# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
9167# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
9168pxor %xmm2,%xmm0
9169
9170# qhasm: xmm2 = xmm5
9171# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
9172# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
9173movdqa %xmm7,%xmm2
9174
9175# qhasm: xmm2 ^= xmm1
9176# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
9177# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
9178pxor %xmm4,%xmm2
9179
9180# qhasm: xmm2 &= xmm12
9181# asm 1: pand <xmm12=int6464#13,<xmm2=int6464#3
9182# asm 2: pand <xmm12=%xmm12,<xmm2=%xmm2
9183pand %xmm12,%xmm2
9184
9185# qhasm: xmm12 ^= xmm10
9186# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
9187# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
9188pxor %xmm10,%xmm12
9189
9190# qhasm: xmm12 &= xmm1
9191# asm 1: pand <xmm1=int6464#5,<xmm12=int6464#13
9192# asm 2: pand <xmm1=%xmm4,<xmm12=%xmm12
9193pand %xmm4,%xmm12
9194
9195# qhasm: xmm10 &= xmm5
9196# asm 1: pand <xmm5=int6464#8,<xmm10=int6464#11
9197# asm 2: pand <xmm5=%xmm7,<xmm10=%xmm10
9198pand %xmm7,%xmm10
9199
9200# qhasm: xmm12 ^= xmm10
9201# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
9202# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
9203pxor %xmm10,%xmm12
9204
9205# qhasm: xmm10 ^= xmm2
9206# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#11
9207# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm10
9208pxor %xmm2,%xmm10
9209
9210# qhasm: xmm7 ^= xmm5
9211# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
9212# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
9213pxor %xmm7,%xmm5
9214
9215# qhasm: xmm6 ^= xmm1
9216# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
9217# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
9218pxor %xmm4,%xmm3
9219
9220# qhasm: xmm3 = xmm7
9221# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
9222# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
9223movdqa %xmm5,%xmm2
9224
9225# qhasm: xmm3 ^= xmm6
9226# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
9227# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
9228pxor %xmm3,%xmm2
9229
9230# qhasm: xmm3 &= xmm15
9231# asm 1: pand <xmm15=int6464#16,<xmm3=int6464#3
9232# asm 2: pand <xmm15=%xmm15,<xmm3=%xmm2
9233pand %xmm15,%xmm2
9234
9235# qhasm: xmm15 ^= xmm9
9236# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
9237# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
9238pxor %xmm9,%xmm15
9239
9240# qhasm: xmm15 &= xmm6
9241# asm 1: pand <xmm6=int6464#4,<xmm15=int6464#16
9242# asm 2: pand <xmm6=%xmm3,<xmm15=%xmm15
9243pand %xmm3,%xmm15
9244
9245# qhasm: xmm9 &= xmm7
9246# asm 1: pand <xmm7=int6464#6,<xmm9=int6464#10
9247# asm 2: pand <xmm7=%xmm5,<xmm9=%xmm9
9248pand %xmm5,%xmm9
9249
9250# qhasm: xmm15 ^= xmm9
9251# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
9252# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
9253pxor %xmm9,%xmm15
9254
9255# qhasm: xmm9 ^= xmm3
9256# asm 1: pxor <xmm3=int6464#3,<xmm9=int6464#10
9257# asm 2: pxor <xmm3=%xmm2,<xmm9=%xmm9
9258pxor %xmm2,%xmm9
9259
9260# qhasm: xmm15 ^= xmm4
9261# asm 1: pxor <xmm4=int6464#1,<xmm15=int6464#16
9262# asm 2: pxor <xmm4=%xmm0,<xmm15=%xmm15
9263pxor %xmm0,%xmm15
9264
9265# qhasm: xmm12 ^= xmm4
9266# asm 1: pxor <xmm4=int6464#1,<xmm12=int6464#13
9267# asm 2: pxor <xmm4=%xmm0,<xmm12=%xmm12
9268pxor %xmm0,%xmm12
9269
9270# qhasm: xmm9 ^= xmm0
9271# asm 1: pxor <xmm0=int6464#2,<xmm9=int6464#10
9272# asm 2: pxor <xmm0=%xmm1,<xmm9=%xmm9
9273pxor %xmm1,%xmm9
9274
9275# qhasm: xmm10 ^= xmm0
9276# asm 1: pxor <xmm0=int6464#2,<xmm10=int6464#11
9277# asm 2: pxor <xmm0=%xmm1,<xmm10=%xmm10
9278pxor %xmm1,%xmm10
9279
9280# qhasm: xmm15 ^= xmm8
9281# asm 1: pxor <xmm8=int6464#9,<xmm15=int6464#16
9282# asm 2: pxor <xmm8=%xmm8,<xmm15=%xmm15
9283pxor %xmm8,%xmm15
9284
9285# qhasm: xmm9 ^= xmm14
9286# asm 1: pxor <xmm14=int6464#15,<xmm9=int6464#10
9287# asm 2: pxor <xmm14=%xmm14,<xmm9=%xmm9
9288pxor %xmm14,%xmm9
9289
9290# qhasm: xmm12 ^= xmm15
9291# asm 1: pxor <xmm15=int6464#16,<xmm12=int6464#13
9292# asm 2: pxor <xmm15=%xmm15,<xmm12=%xmm12
9293pxor %xmm15,%xmm12
9294
9295# qhasm: xmm14 ^= xmm8
9296# asm 1: pxor <xmm8=int6464#9,<xmm14=int6464#15
9297# asm 2: pxor <xmm8=%xmm8,<xmm14=%xmm14
9298pxor %xmm8,%xmm14
9299
9300# qhasm: xmm8 ^= xmm9
9301# asm 1: pxor <xmm9=int6464#10,<xmm8=int6464#9
9302# asm 2: pxor <xmm9=%xmm9,<xmm8=%xmm8
9303pxor %xmm9,%xmm8
9304
9305# qhasm: xmm9 ^= xmm13
9306# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
9307# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
9308pxor %xmm13,%xmm9
9309
9310# qhasm: xmm13 ^= xmm10
9311# asm 1: pxor <xmm10=int6464#11,<xmm13=int6464#14
9312# asm 2: pxor <xmm10=%xmm10,<xmm13=%xmm13
9313pxor %xmm10,%xmm13
9314
9315# qhasm: xmm12 ^= xmm13
9316# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
9317# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
9318pxor %xmm13,%xmm12
9319
9320# qhasm: xmm10 ^= xmm11
9321# asm 1: pxor <xmm11=int6464#12,<xmm10=int6464#11
9322# asm 2: pxor <xmm11=%xmm11,<xmm10=%xmm10
9323pxor %xmm11,%xmm10
9324
9325# qhasm: xmm11 ^= xmm13
9326# asm 1: pxor <xmm13=int6464#14,<xmm11=int6464#12
9327# asm 2: pxor <xmm13=%xmm13,<xmm11=%xmm11
9328pxor %xmm13,%xmm11
9329
9330# qhasm: xmm14 ^= xmm11
9331# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
9332# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
9333pxor %xmm11,%xmm14
9334
9335# qhasm: xmm0 = shuffle dwords of xmm8 by 0x93
9336# asm 1: pshufd $0x93,<xmm8=int6464#9,>xmm0=int6464#1
9337# asm 2: pshufd $0x93,<xmm8=%xmm8,>xmm0=%xmm0
9338pshufd $0x93,%xmm8,%xmm0
9339
9340# qhasm: xmm1 = shuffle dwords of xmm9 by 0x93
9341# asm 1: pshufd $0x93,<xmm9=int6464#10,>xmm1=int6464#2
9342# asm 2: pshufd $0x93,<xmm9=%xmm9,>xmm1=%xmm1
9343pshufd $0x93,%xmm9,%xmm1
9344
9345# qhasm: xmm2 = shuffle dwords of xmm12 by 0x93
9346# asm 1: pshufd $0x93,<xmm12=int6464#13,>xmm2=int6464#3
9347# asm 2: pshufd $0x93,<xmm12=%xmm12,>xmm2=%xmm2
9348pshufd $0x93,%xmm12,%xmm2
9349
9350# qhasm: xmm3 = shuffle dwords of xmm14 by 0x93
9351# asm 1: pshufd $0x93,<xmm14=int6464#15,>xmm3=int6464#4
9352# asm 2: pshufd $0x93,<xmm14=%xmm14,>xmm3=%xmm3
9353pshufd $0x93,%xmm14,%xmm3
9354
9355# qhasm: xmm4 = shuffle dwords of xmm11 by 0x93
9356# asm 1: pshufd $0x93,<xmm11=int6464#12,>xmm4=int6464#5
9357# asm 2: pshufd $0x93,<xmm11=%xmm11,>xmm4=%xmm4
9358pshufd $0x93,%xmm11,%xmm4
9359
9360# qhasm: xmm5 = shuffle dwords of xmm15 by 0x93
9361# asm 1: pshufd $0x93,<xmm15=int6464#16,>xmm5=int6464#6
9362# asm 2: pshufd $0x93,<xmm15=%xmm15,>xmm5=%xmm5
9363pshufd $0x93,%xmm15,%xmm5
9364
9365# qhasm: xmm6 = shuffle dwords of xmm10 by 0x93
9366# asm 1: pshufd $0x93,<xmm10=int6464#11,>xmm6=int6464#7
9367# asm 2: pshufd $0x93,<xmm10=%xmm10,>xmm6=%xmm6
9368pshufd $0x93,%xmm10,%xmm6
9369
9370# qhasm: xmm7 = shuffle dwords of xmm13 by 0x93
9371# asm 1: pshufd $0x93,<xmm13=int6464#14,>xmm7=int6464#8
9372# asm 2: pshufd $0x93,<xmm13=%xmm13,>xmm7=%xmm7
9373pshufd $0x93,%xmm13,%xmm7
9374
9375# qhasm: xmm8 ^= xmm0
9376# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
9377# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
9378pxor %xmm0,%xmm8
9379
9380# qhasm: xmm9 ^= xmm1
9381# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
9382# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
9383pxor %xmm1,%xmm9
9384
9385# qhasm: xmm12 ^= xmm2
9386# asm 1: pxor <xmm2=int6464#3,<xmm12=int6464#13
9387# asm 2: pxor <xmm2=%xmm2,<xmm12=%xmm12
9388pxor %xmm2,%xmm12
9389
9390# qhasm: xmm14 ^= xmm3
9391# asm 1: pxor <xmm3=int6464#4,<xmm14=int6464#15
9392# asm 2: pxor <xmm3=%xmm3,<xmm14=%xmm14
9393pxor %xmm3,%xmm14
9394
9395# qhasm: xmm11 ^= xmm4
9396# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
9397# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
9398pxor %xmm4,%xmm11
9399
9400# qhasm: xmm15 ^= xmm5
9401# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
9402# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
9403pxor %xmm5,%xmm15
9404
9405# qhasm: xmm10 ^= xmm6
9406# asm 1: pxor <xmm6=int6464#7,<xmm10=int6464#11
9407# asm 2: pxor <xmm6=%xmm6,<xmm10=%xmm10
9408pxor %xmm6,%xmm10
9409
9410# qhasm: xmm13 ^= xmm7
9411# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
9412# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
9413pxor %xmm7,%xmm13
9414
9415# qhasm: xmm0 ^= xmm13
9416# asm 1: pxor <xmm13=int6464#14,<xmm0=int6464#1
9417# asm 2: pxor <xmm13=%xmm13,<xmm0=%xmm0
9418pxor %xmm13,%xmm0
9419
9420# qhasm: xmm1 ^= xmm8
9421# asm 1: pxor <xmm8=int6464#9,<xmm1=int6464#2
9422# asm 2: pxor <xmm8=%xmm8,<xmm1=%xmm1
9423pxor %xmm8,%xmm1
9424
9425# qhasm: xmm2 ^= xmm9
9426# asm 1: pxor <xmm9=int6464#10,<xmm2=int6464#3
9427# asm 2: pxor <xmm9=%xmm9,<xmm2=%xmm2
9428pxor %xmm9,%xmm2
9429
9430# qhasm: xmm1 ^= xmm13
9431# asm 1: pxor <xmm13=int6464#14,<xmm1=int6464#2
9432# asm 2: pxor <xmm13=%xmm13,<xmm1=%xmm1
9433pxor %xmm13,%xmm1
9434
9435# qhasm: xmm3 ^= xmm12
9436# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
9437# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
9438pxor %xmm12,%xmm3
9439
9440# qhasm: xmm4 ^= xmm14
9441# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#5
9442# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm4
9443pxor %xmm14,%xmm4
9444
9445# qhasm: xmm5 ^= xmm11
9446# asm 1: pxor <xmm11=int6464#12,<xmm5=int6464#6
9447# asm 2: pxor <xmm11=%xmm11,<xmm5=%xmm5
9448pxor %xmm11,%xmm5
9449
9450# qhasm: xmm3 ^= xmm13
9451# asm 1: pxor <xmm13=int6464#14,<xmm3=int6464#4
9452# asm 2: pxor <xmm13=%xmm13,<xmm3=%xmm3
9453pxor %xmm13,%xmm3
9454
9455# qhasm: xmm6 ^= xmm15
9456# asm 1: pxor <xmm15=int6464#16,<xmm6=int6464#7
9457# asm 2: pxor <xmm15=%xmm15,<xmm6=%xmm6
9458pxor %xmm15,%xmm6
9459
9460# qhasm: xmm7 ^= xmm10
9461# asm 1: pxor <xmm10=int6464#11,<xmm7=int6464#8
9462# asm 2: pxor <xmm10=%xmm10,<xmm7=%xmm7
9463pxor %xmm10,%xmm7
9464
9465# qhasm: xmm4 ^= xmm13
9466# asm 1: pxor <xmm13=int6464#14,<xmm4=int6464#5
9467# asm 2: pxor <xmm13=%xmm13,<xmm4=%xmm4
9468pxor %xmm13,%xmm4
9469
9470# qhasm: xmm8 = shuffle dwords of xmm8 by 0x4E
9471# asm 1: pshufd $0x4E,<xmm8=int6464#9,>xmm8=int6464#9
9472# asm 2: pshufd $0x4E,<xmm8=%xmm8,>xmm8=%xmm8
9473pshufd $0x4E,%xmm8,%xmm8
9474
9475# qhasm: xmm9 = shuffle dwords of xmm9 by 0x4E
9476# asm 1: pshufd $0x4E,<xmm9=int6464#10,>xmm9=int6464#10
9477# asm 2: pshufd $0x4E,<xmm9=%xmm9,>xmm9=%xmm9
9478pshufd $0x4E,%xmm9,%xmm9
9479
9480# qhasm: xmm12 = shuffle dwords of xmm12 by 0x4E
9481# asm 1: pshufd $0x4E,<xmm12=int6464#13,>xmm12=int6464#13
9482# asm 2: pshufd $0x4E,<xmm12=%xmm12,>xmm12=%xmm12
9483pshufd $0x4E,%xmm12,%xmm12
9484
9485# qhasm: xmm14 = shuffle dwords of xmm14 by 0x4E
9486# asm 1: pshufd $0x4E,<xmm14=int6464#15,>xmm14=int6464#15
9487# asm 2: pshufd $0x4E,<xmm14=%xmm14,>xmm14=%xmm14
9488pshufd $0x4E,%xmm14,%xmm14
9489
9490# qhasm: xmm11 = shuffle dwords of xmm11 by 0x4E
9491# asm 1: pshufd $0x4E,<xmm11=int6464#12,>xmm11=int6464#12
9492# asm 2: pshufd $0x4E,<xmm11=%xmm11,>xmm11=%xmm11
9493pshufd $0x4E,%xmm11,%xmm11
9494
9495# qhasm: xmm15 = shuffle dwords of xmm15 by 0x4E
9496# asm 1: pshufd $0x4E,<xmm15=int6464#16,>xmm15=int6464#16
9497# asm 2: pshufd $0x4E,<xmm15=%xmm15,>xmm15=%xmm15
9498pshufd $0x4E,%xmm15,%xmm15
9499
9500# qhasm: xmm10 = shuffle dwords of xmm10 by 0x4E
9501# asm 1: pshufd $0x4E,<xmm10=int6464#11,>xmm10=int6464#11
9502# asm 2: pshufd $0x4E,<xmm10=%xmm10,>xmm10=%xmm10
9503pshufd $0x4E,%xmm10,%xmm10
9504
9505# qhasm: xmm13 = shuffle dwords of xmm13 by 0x4E
9506# asm 1: pshufd $0x4E,<xmm13=int6464#14,>xmm13=int6464#14
9507# asm 2: pshufd $0x4E,<xmm13=%xmm13,>xmm13=%xmm13
9508pshufd $0x4E,%xmm13,%xmm13
9509
9510# qhasm: xmm0 ^= xmm8
9511# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
9512# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
9513pxor %xmm8,%xmm0
9514
9515# qhasm: xmm1 ^= xmm9
9516# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
9517# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
9518pxor %xmm9,%xmm1
9519
9520# qhasm: xmm2 ^= xmm12
9521# asm 1: pxor <xmm12=int6464#13,<xmm2=int6464#3
9522# asm 2: pxor <xmm12=%xmm12,<xmm2=%xmm2
9523pxor %xmm12,%xmm2
9524
9525# qhasm: xmm3 ^= xmm14
9526# asm 1: pxor <xmm14=int6464#15,<xmm3=int6464#4
9527# asm 2: pxor <xmm14=%xmm14,<xmm3=%xmm3
9528pxor %xmm14,%xmm3
9529
9530# qhasm: xmm4 ^= xmm11
9531# asm 1: pxor <xmm11=int6464#12,<xmm4=int6464#5
9532# asm 2: pxor <xmm11=%xmm11,<xmm4=%xmm4
9533pxor %xmm11,%xmm4
9534
9535# qhasm: xmm5 ^= xmm15
9536# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
9537# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
9538pxor %xmm15,%xmm5
9539
9540# qhasm: xmm6 ^= xmm10
9541# asm 1: pxor <xmm10=int6464#11,<xmm6=int6464#7
9542# asm 2: pxor <xmm10=%xmm10,<xmm6=%xmm6
9543pxor %xmm10,%xmm6
9544
9545# qhasm: xmm7 ^= xmm13
9546# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
9547# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
9548pxor %xmm13,%xmm7
9549
9550# qhasm: xmm0 ^= *(int128 *)(c + 1024)
9551# asm 1: pxor 1024(<c=int64#4),<xmm0=int6464#1
9552# asm 2: pxor 1024(<c=%rcx),<xmm0=%xmm0
9553pxor 1024(%rcx),%xmm0
9554
9555# qhasm: shuffle bytes of xmm0 by SR
9556# asm 1: pshufb SR,<xmm0=int6464#1
9557# asm 2: pshufb SR,<xmm0=%xmm0
9558pshufb SR,%xmm0
9559
9560# qhasm: xmm1 ^= *(int128 *)(c + 1040)
9561# asm 1: pxor 1040(<c=int64#4),<xmm1=int6464#2
9562# asm 2: pxor 1040(<c=%rcx),<xmm1=%xmm1
9563pxor 1040(%rcx),%xmm1
9564
9565# qhasm: shuffle bytes of xmm1 by SR
9566# asm 1: pshufb SR,<xmm1=int6464#2
9567# asm 2: pshufb SR,<xmm1=%xmm1
9568pshufb SR,%xmm1
9569
9570# qhasm: xmm2 ^= *(int128 *)(c + 1056)
9571# asm 1: pxor 1056(<c=int64#4),<xmm2=int6464#3
9572# asm 2: pxor 1056(<c=%rcx),<xmm2=%xmm2
9573pxor 1056(%rcx),%xmm2
9574
9575# qhasm: shuffle bytes of xmm2 by SR
9576# asm 1: pshufb SR,<xmm2=int6464#3
9577# asm 2: pshufb SR,<xmm2=%xmm2
9578pshufb SR,%xmm2
9579
9580# qhasm: xmm3 ^= *(int128 *)(c + 1072)
9581# asm 1: pxor 1072(<c=int64#4),<xmm3=int6464#4
9582# asm 2: pxor 1072(<c=%rcx),<xmm3=%xmm3
9583pxor 1072(%rcx),%xmm3
9584
9585# qhasm: shuffle bytes of xmm3 by SR
9586# asm 1: pshufb SR,<xmm3=int6464#4
9587# asm 2: pshufb SR,<xmm3=%xmm3
9588pshufb SR,%xmm3
9589
9590# qhasm: xmm4 ^= *(int128 *)(c + 1088)
9591# asm 1: pxor 1088(<c=int64#4),<xmm4=int6464#5
9592# asm 2: pxor 1088(<c=%rcx),<xmm4=%xmm4
9593pxor 1088(%rcx),%xmm4
9594
9595# qhasm: shuffle bytes of xmm4 by SR
9596# asm 1: pshufb SR,<xmm4=int6464#5
9597# asm 2: pshufb SR,<xmm4=%xmm4
9598pshufb SR,%xmm4
9599
9600# qhasm: xmm5 ^= *(int128 *)(c + 1104)
9601# asm 1: pxor 1104(<c=int64#4),<xmm5=int6464#6
9602# asm 2: pxor 1104(<c=%rcx),<xmm5=%xmm5
9603pxor 1104(%rcx),%xmm5
9604
9605# qhasm: shuffle bytes of xmm5 by SR
9606# asm 1: pshufb SR,<xmm5=int6464#6
9607# asm 2: pshufb SR,<xmm5=%xmm5
9608pshufb SR,%xmm5
9609
9610# qhasm: xmm6 ^= *(int128 *)(c + 1120)
9611# asm 1: pxor 1120(<c=int64#4),<xmm6=int6464#7
9612# asm 2: pxor 1120(<c=%rcx),<xmm6=%xmm6
9613pxor 1120(%rcx),%xmm6
9614
9615# qhasm: shuffle bytes of xmm6 by SR
9616# asm 1: pshufb SR,<xmm6=int6464#7
9617# asm 2: pshufb SR,<xmm6=%xmm6
9618pshufb SR,%xmm6
9619
9620# qhasm: xmm7 ^= *(int128 *)(c + 1136)
9621# asm 1: pxor 1136(<c=int64#4),<xmm7=int6464#8
9622# asm 2: pxor 1136(<c=%rcx),<xmm7=%xmm7
9623pxor 1136(%rcx),%xmm7
9624
9625# qhasm: shuffle bytes of xmm7 by SR
9626# asm 1: pshufb SR,<xmm7=int6464#8
9627# asm 2: pshufb SR,<xmm7=%xmm7
9628pshufb SR,%xmm7
9629
9630# qhasm: xmm5 ^= xmm6
9631# asm 1: pxor <xmm6=int6464#7,<xmm5=int6464#6
9632# asm 2: pxor <xmm6=%xmm6,<xmm5=%xmm5
9633pxor %xmm6,%xmm5
9634
9635# qhasm: xmm2 ^= xmm1
9636# asm 1: pxor <xmm1=int6464#2,<xmm2=int6464#3
9637# asm 2: pxor <xmm1=%xmm1,<xmm2=%xmm2
9638pxor %xmm1,%xmm2
9639
9640# qhasm: xmm5 ^= xmm0
9641# asm 1: pxor <xmm0=int6464#1,<xmm5=int6464#6
9642# asm 2: pxor <xmm0=%xmm0,<xmm5=%xmm5
9643pxor %xmm0,%xmm5
9644
9645# qhasm: xmm6 ^= xmm2
9646# asm 1: pxor <xmm2=int6464#3,<xmm6=int6464#7
9647# asm 2: pxor <xmm2=%xmm2,<xmm6=%xmm6
9648pxor %xmm2,%xmm6
9649
9650# qhasm: xmm3 ^= xmm0
9651# asm 1: pxor <xmm0=int6464#1,<xmm3=int6464#4
9652# asm 2: pxor <xmm0=%xmm0,<xmm3=%xmm3
9653pxor %xmm0,%xmm3
9654
9655# qhasm: xmm6 ^= xmm3
9656# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
9657# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
9658pxor %xmm3,%xmm6
9659
9660# qhasm: xmm3 ^= xmm7
9661# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#4
9662# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm3
9663pxor %xmm7,%xmm3
9664
9665# qhasm: xmm3 ^= xmm4
9666# asm 1: pxor <xmm4=int6464#5,<xmm3=int6464#4
9667# asm 2: pxor <xmm4=%xmm4,<xmm3=%xmm3
9668pxor %xmm4,%xmm3
9669
9670# qhasm: xmm7 ^= xmm5
9671# asm 1: pxor <xmm5=int6464#6,<xmm7=int6464#8
9672# asm 2: pxor <xmm5=%xmm5,<xmm7=%xmm7
9673pxor %xmm5,%xmm7
9674
9675# qhasm: xmm3 ^= xmm1
9676# asm 1: pxor <xmm1=int6464#2,<xmm3=int6464#4
9677# asm 2: pxor <xmm1=%xmm1,<xmm3=%xmm3
9678pxor %xmm1,%xmm3
9679
9680# qhasm: xmm4 ^= xmm5
9681# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
9682# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
9683pxor %xmm5,%xmm4
9684
9685# qhasm: xmm2 ^= xmm7
9686# asm 1: pxor <xmm7=int6464#8,<xmm2=int6464#3
9687# asm 2: pxor <xmm7=%xmm7,<xmm2=%xmm2
9688pxor %xmm7,%xmm2
9689
9690# qhasm: xmm1 ^= xmm5
9691# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
9692# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
9693pxor %xmm5,%xmm1
9694
9695# qhasm: xmm11 = xmm7
9696# asm 1: movdqa <xmm7=int6464#8,>xmm11=int6464#9
9697# asm 2: movdqa <xmm7=%xmm7,>xmm11=%xmm8
9698movdqa %xmm7,%xmm8
9699
9700# qhasm: xmm10 = xmm1
9701# asm 1: movdqa <xmm1=int6464#2,>xmm10=int6464#10
9702# asm 2: movdqa <xmm1=%xmm1,>xmm10=%xmm9
9703movdqa %xmm1,%xmm9
9704
9705# qhasm: xmm9 = xmm5
9706# asm 1: movdqa <xmm5=int6464#6,>xmm9=int6464#11
9707# asm 2: movdqa <xmm5=%xmm5,>xmm9=%xmm10
9708movdqa %xmm5,%xmm10
9709
9710# qhasm: xmm13 = xmm2
9711# asm 1: movdqa <xmm2=int6464#3,>xmm13=int6464#12
9712# asm 2: movdqa <xmm2=%xmm2,>xmm13=%xmm11
9713movdqa %xmm2,%xmm11
9714
9715# qhasm: xmm12 = xmm6
9716# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#13
9717# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm12
9718movdqa %xmm6,%xmm12
9719
9720# qhasm: xmm11 ^= xmm4
9721# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#9
9722# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm8
9723pxor %xmm4,%xmm8
9724
9725# qhasm: xmm10 ^= xmm2
9726# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#10
9727# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm9
9728pxor %xmm2,%xmm9
9729
9730# qhasm: xmm9 ^= xmm3
9731# asm 1: pxor <xmm3=int6464#4,<xmm9=int6464#11
9732# asm 2: pxor <xmm3=%xmm3,<xmm9=%xmm10
9733pxor %xmm3,%xmm10
9734
9735# qhasm: xmm13 ^= xmm4
9736# asm 1: pxor <xmm4=int6464#5,<xmm13=int6464#12
9737# asm 2: pxor <xmm4=%xmm4,<xmm13=%xmm11
9738pxor %xmm4,%xmm11
9739
9740# qhasm: xmm12 ^= xmm0
9741# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
9742# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
9743pxor %xmm0,%xmm12
9744
9745# qhasm: xmm14 = xmm11
9746# asm 1: movdqa <xmm11=int6464#9,>xmm14=int6464#14
9747# asm 2: movdqa <xmm11=%xmm8,>xmm14=%xmm13
9748movdqa %xmm8,%xmm13
9749
9750# qhasm: xmm8 = xmm10
9751# asm 1: movdqa <xmm10=int6464#10,>xmm8=int6464#15
9752# asm 2: movdqa <xmm10=%xmm9,>xmm8=%xmm14
9753movdqa %xmm9,%xmm14
9754
9755# qhasm: xmm15 = xmm11
9756# asm 1: movdqa <xmm11=int6464#9,>xmm15=int6464#16
9757# asm 2: movdqa <xmm11=%xmm8,>xmm15=%xmm15
9758movdqa %xmm8,%xmm15
9759
9760# qhasm: xmm10 |= xmm9
9761# asm 1: por <xmm9=int6464#11,<xmm10=int6464#10
9762# asm 2: por <xmm9=%xmm10,<xmm10=%xmm9
9763por %xmm10,%xmm9
9764
9765# qhasm: xmm11 |= xmm12
9766# asm 1: por <xmm12=int6464#13,<xmm11=int6464#9
9767# asm 2: por <xmm12=%xmm12,<xmm11=%xmm8
9768por %xmm12,%xmm8
9769
9770# qhasm: xmm15 ^= xmm8
9771# asm 1: pxor <xmm8=int6464#15,<xmm15=int6464#16
9772# asm 2: pxor <xmm8=%xmm14,<xmm15=%xmm15
9773pxor %xmm14,%xmm15
9774
9775# qhasm: xmm14 &= xmm12
9776# asm 1: pand <xmm12=int6464#13,<xmm14=int6464#14
9777# asm 2: pand <xmm12=%xmm12,<xmm14=%xmm13
9778pand %xmm12,%xmm13
9779
9780# qhasm: xmm8 &= xmm9
9781# asm 1: pand <xmm9=int6464#11,<xmm8=int6464#15
9782# asm 2: pand <xmm9=%xmm10,<xmm8=%xmm14
9783pand %xmm10,%xmm14
9784
9785# qhasm: xmm12 ^= xmm9
9786# asm 1: pxor <xmm9=int6464#11,<xmm12=int6464#13
9787# asm 2: pxor <xmm9=%xmm10,<xmm12=%xmm12
9788pxor %xmm10,%xmm12
9789
9790# qhasm: xmm15 &= xmm12
9791# asm 1: pand <xmm12=int6464#13,<xmm15=int6464#16
9792# asm 2: pand <xmm12=%xmm12,<xmm15=%xmm15
9793pand %xmm12,%xmm15
9794
9795# qhasm: xmm12 = xmm3
9796# asm 1: movdqa <xmm3=int6464#4,>xmm12=int6464#11
9797# asm 2: movdqa <xmm3=%xmm3,>xmm12=%xmm10
9798movdqa %xmm3,%xmm10
9799
9800# qhasm: xmm12 ^= xmm0
9801# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#11
9802# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm10
9803pxor %xmm0,%xmm10
9804
9805# qhasm: xmm13 &= xmm12
9806# asm 1: pand <xmm12=int6464#11,<xmm13=int6464#12
9807# asm 2: pand <xmm12=%xmm10,<xmm13=%xmm11
9808pand %xmm10,%xmm11
9809
9810# qhasm: xmm11 ^= xmm13
9811# asm 1: pxor <xmm13=int6464#12,<xmm11=int6464#9
9812# asm 2: pxor <xmm13=%xmm11,<xmm11=%xmm8
9813pxor %xmm11,%xmm8
9814
9815# qhasm: xmm10 ^= xmm13
9816# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
9817# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
9818pxor %xmm11,%xmm9
9819
9820# qhasm: xmm13 = xmm7
9821# asm 1: movdqa <xmm7=int6464#8,>xmm13=int6464#11
9822# asm 2: movdqa <xmm7=%xmm7,>xmm13=%xmm10
9823movdqa %xmm7,%xmm10
9824
9825# qhasm: xmm13 ^= xmm1
9826# asm 1: pxor <xmm1=int6464#2,<xmm13=int6464#11
9827# asm 2: pxor <xmm1=%xmm1,<xmm13=%xmm10
9828pxor %xmm1,%xmm10
9829
9830# qhasm: xmm12 = xmm5
9831# asm 1: movdqa <xmm5=int6464#6,>xmm12=int6464#12
9832# asm 2: movdqa <xmm5=%xmm5,>xmm12=%xmm11
9833movdqa %xmm5,%xmm11
9834
9835# qhasm: xmm9 = xmm13
9836# asm 1: movdqa <xmm13=int6464#11,>xmm9=int6464#13
9837# asm 2: movdqa <xmm13=%xmm10,>xmm9=%xmm12
9838movdqa %xmm10,%xmm12
9839
9840# qhasm: xmm12 ^= xmm6
9841# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#12
9842# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm11
9843pxor %xmm6,%xmm11
9844
9845# qhasm: xmm9 |= xmm12
9846# asm 1: por <xmm12=int6464#12,<xmm9=int6464#13
9847# asm 2: por <xmm12=%xmm11,<xmm9=%xmm12
9848por %xmm11,%xmm12
9849
9850# qhasm: xmm13 &= xmm12
9851# asm 1: pand <xmm12=int6464#12,<xmm13=int6464#11
9852# asm 2: pand <xmm12=%xmm11,<xmm13=%xmm10
9853pand %xmm11,%xmm10
9854
9855# qhasm: xmm8 ^= xmm13
9856# asm 1: pxor <xmm13=int6464#11,<xmm8=int6464#15
9857# asm 2: pxor <xmm13=%xmm10,<xmm8=%xmm14
9858pxor %xmm10,%xmm14
9859
9860# qhasm: xmm11 ^= xmm15
9861# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#9
9862# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm8
9863pxor %xmm15,%xmm8
9864
9865# qhasm: xmm10 ^= xmm14
9866# asm 1: pxor <xmm14=int6464#14,<xmm10=int6464#10
9867# asm 2: pxor <xmm14=%xmm13,<xmm10=%xmm9
9868pxor %xmm13,%xmm9
9869
9870# qhasm: xmm9 ^= xmm15
9871# asm 1: pxor <xmm15=int6464#16,<xmm9=int6464#13
9872# asm 2: pxor <xmm15=%xmm15,<xmm9=%xmm12
9873pxor %xmm15,%xmm12
9874
9875# qhasm: xmm8 ^= xmm14
9876# asm 1: pxor <xmm14=int6464#14,<xmm8=int6464#15
9877# asm 2: pxor <xmm14=%xmm13,<xmm8=%xmm14
9878pxor %xmm13,%xmm14
9879
9880# qhasm: xmm9 ^= xmm14
9881# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
9882# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
9883pxor %xmm13,%xmm12
9884
9885# qhasm: xmm12 = xmm2
9886# asm 1: movdqa <xmm2=int6464#3,>xmm12=int6464#11
9887# asm 2: movdqa <xmm2=%xmm2,>xmm12=%xmm10
9888movdqa %xmm2,%xmm10
9889
9890# qhasm: xmm13 = xmm4
9891# asm 1: movdqa <xmm4=int6464#5,>xmm13=int6464#12
9892# asm 2: movdqa <xmm4=%xmm4,>xmm13=%xmm11
9893movdqa %xmm4,%xmm11
9894
9895# qhasm: xmm14 = xmm1
9896# asm 1: movdqa <xmm1=int6464#2,>xmm14=int6464#14
9897# asm 2: movdqa <xmm1=%xmm1,>xmm14=%xmm13
9898movdqa %xmm1,%xmm13
9899
9900# qhasm: xmm15 = xmm7
9901# asm 1: movdqa <xmm7=int6464#8,>xmm15=int6464#16
9902# asm 2: movdqa <xmm7=%xmm7,>xmm15=%xmm15
9903movdqa %xmm7,%xmm15
9904
9905# qhasm: xmm12 &= xmm3
9906# asm 1: pand <xmm3=int6464#4,<xmm12=int6464#11
9907# asm 2: pand <xmm3=%xmm3,<xmm12=%xmm10
9908pand %xmm3,%xmm10
9909
9910# qhasm: xmm13 &= xmm0
9911# asm 1: pand <xmm0=int6464#1,<xmm13=int6464#12
9912# asm 2: pand <xmm0=%xmm0,<xmm13=%xmm11
9913pand %xmm0,%xmm11
9914
9915# qhasm: xmm14 &= xmm5
9916# asm 1: pand <xmm5=int6464#6,<xmm14=int6464#14
9917# asm 2: pand <xmm5=%xmm5,<xmm14=%xmm13
9918pand %xmm5,%xmm13
9919
9920# qhasm: xmm15 |= xmm6
9921# asm 1: por <xmm6=int6464#7,<xmm15=int6464#16
9922# asm 2: por <xmm6=%xmm6,<xmm15=%xmm15
9923por %xmm6,%xmm15
9924
9925# qhasm: xmm11 ^= xmm12
9926# asm 1: pxor <xmm12=int6464#11,<xmm11=int6464#9
9927# asm 2: pxor <xmm12=%xmm10,<xmm11=%xmm8
9928pxor %xmm10,%xmm8
9929
9930# qhasm: xmm10 ^= xmm13
9931# asm 1: pxor <xmm13=int6464#12,<xmm10=int6464#10
9932# asm 2: pxor <xmm13=%xmm11,<xmm10=%xmm9
9933pxor %xmm11,%xmm9
9934
9935# qhasm: xmm9 ^= xmm14
9936# asm 1: pxor <xmm14=int6464#14,<xmm9=int6464#13
9937# asm 2: pxor <xmm14=%xmm13,<xmm9=%xmm12
9938pxor %xmm13,%xmm12
9939
9940# qhasm: xmm8 ^= xmm15
9941# asm 1: pxor <xmm15=int6464#16,<xmm8=int6464#15
9942# asm 2: pxor <xmm15=%xmm15,<xmm8=%xmm14
9943pxor %xmm15,%xmm14
9944
9945# qhasm: xmm12 = xmm11
9946# asm 1: movdqa <xmm11=int6464#9,>xmm12=int6464#11
9947# asm 2: movdqa <xmm11=%xmm8,>xmm12=%xmm10
9948movdqa %xmm8,%xmm10
9949
9950# qhasm: xmm12 ^= xmm10
9951# asm 1: pxor <xmm10=int6464#10,<xmm12=int6464#11
9952# asm 2: pxor <xmm10=%xmm9,<xmm12=%xmm10
9953pxor %xmm9,%xmm10
9954
9955# qhasm: xmm11 &= xmm9
9956# asm 1: pand <xmm9=int6464#13,<xmm11=int6464#9
9957# asm 2: pand <xmm9=%xmm12,<xmm11=%xmm8
9958pand %xmm12,%xmm8
9959
9960# qhasm: xmm14 = xmm8
9961# asm 1: movdqa <xmm8=int6464#15,>xmm14=int6464#12
9962# asm 2: movdqa <xmm8=%xmm14,>xmm14=%xmm11
9963movdqa %xmm14,%xmm11
9964
9965# qhasm: xmm14 ^= xmm11
9966# asm 1: pxor <xmm11=int6464#9,<xmm14=int6464#12
9967# asm 2: pxor <xmm11=%xmm8,<xmm14=%xmm11
9968pxor %xmm8,%xmm11
9969
9970# qhasm: xmm15 = xmm12
9971# asm 1: movdqa <xmm12=int6464#11,>xmm15=int6464#14
9972# asm 2: movdqa <xmm12=%xmm10,>xmm15=%xmm13
9973movdqa %xmm10,%xmm13
9974
9975# qhasm: xmm15 &= xmm14
9976# asm 1: pand <xmm14=int6464#12,<xmm15=int6464#14
9977# asm 2: pand <xmm14=%xmm11,<xmm15=%xmm13
9978pand %xmm11,%xmm13
9979
9980# qhasm: xmm15 ^= xmm10
9981# asm 1: pxor <xmm10=int6464#10,<xmm15=int6464#14
9982# asm 2: pxor <xmm10=%xmm9,<xmm15=%xmm13
9983pxor %xmm9,%xmm13
9984
9985# qhasm: xmm13 = xmm9
9986# asm 1: movdqa <xmm9=int6464#13,>xmm13=int6464#16
9987# asm 2: movdqa <xmm9=%xmm12,>xmm13=%xmm15
9988movdqa %xmm12,%xmm15
9989
9990# qhasm: xmm13 ^= xmm8
9991# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
9992# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
9993pxor %xmm14,%xmm15
9994
9995# qhasm: xmm11 ^= xmm10
9996# asm 1: pxor <xmm10=int6464#10,<xmm11=int6464#9
9997# asm 2: pxor <xmm10=%xmm9,<xmm11=%xmm8
9998pxor %xmm9,%xmm8
9999
10000# qhasm: xmm13 &= xmm11
10001# asm 1: pand <xmm11=int6464#9,<xmm13=int6464#16
10002# asm 2: pand <xmm11=%xmm8,<xmm13=%xmm15
10003pand %xmm8,%xmm15
10004
10005# qhasm: xmm13 ^= xmm8
10006# asm 1: pxor <xmm8=int6464#15,<xmm13=int6464#16
10007# asm 2: pxor <xmm8=%xmm14,<xmm13=%xmm15
10008pxor %xmm14,%xmm15
10009
10010# qhasm: xmm9 ^= xmm13
10011# asm 1: pxor <xmm13=int6464#16,<xmm9=int6464#13
10012# asm 2: pxor <xmm13=%xmm15,<xmm9=%xmm12
10013pxor %xmm15,%xmm12
10014
10015# qhasm: xmm10 = xmm14
10016# asm 1: movdqa <xmm14=int6464#12,>xmm10=int6464#9
10017# asm 2: movdqa <xmm14=%xmm11,>xmm10=%xmm8
10018movdqa %xmm11,%xmm8
10019
10020# qhasm: xmm10 ^= xmm13
10021# asm 1: pxor <xmm13=int6464#16,<xmm10=int6464#9
10022# asm 2: pxor <xmm13=%xmm15,<xmm10=%xmm8
10023pxor %xmm15,%xmm8
10024
10025# qhasm: xmm10 &= xmm8
10026# asm 1: pand <xmm8=int6464#15,<xmm10=int6464#9
10027# asm 2: pand <xmm8=%xmm14,<xmm10=%xmm8
10028pand %xmm14,%xmm8
10029
10030# qhasm: xmm9 ^= xmm10
10031# asm 1: pxor <xmm10=int6464#9,<xmm9=int6464#13
10032# asm 2: pxor <xmm10=%xmm8,<xmm9=%xmm12
10033pxor %xmm8,%xmm12
10034
10035# qhasm: xmm14 ^= xmm10
10036# asm 1: pxor <xmm10=int6464#9,<xmm14=int6464#12
10037# asm 2: pxor <xmm10=%xmm8,<xmm14=%xmm11
10038pxor %xmm8,%xmm11
10039
10040# qhasm: xmm14 &= xmm15
10041# asm 1: pand <xmm15=int6464#14,<xmm14=int6464#12
10042# asm 2: pand <xmm15=%xmm13,<xmm14=%xmm11
10043pand %xmm13,%xmm11
10044
10045# qhasm: xmm14 ^= xmm12
10046# asm 1: pxor <xmm12=int6464#11,<xmm14=int6464#12
10047# asm 2: pxor <xmm12=%xmm10,<xmm14=%xmm11
10048pxor %xmm10,%xmm11
10049
10050# qhasm: xmm12 = xmm6
10051# asm 1: movdqa <xmm6=int6464#7,>xmm12=int6464#9
10052# asm 2: movdqa <xmm6=%xmm6,>xmm12=%xmm8
10053movdqa %xmm6,%xmm8
10054
10055# qhasm: xmm8 = xmm5
10056# asm 1: movdqa <xmm5=int6464#6,>xmm8=int6464#10
10057# asm 2: movdqa <xmm5=%xmm5,>xmm8=%xmm9
10058movdqa %xmm5,%xmm9
10059
10060# qhasm: xmm10 = xmm15
10061# asm 1: movdqa <xmm15=int6464#14,>xmm10=int6464#11
10062# asm 2: movdqa <xmm15=%xmm13,>xmm10=%xmm10
10063movdqa %xmm13,%xmm10
10064
10065# qhasm: xmm10 ^= xmm14
10066# asm 1: pxor <xmm14=int6464#12,<xmm10=int6464#11
10067# asm 2: pxor <xmm14=%xmm11,<xmm10=%xmm10
10068pxor %xmm11,%xmm10
10069
10070# qhasm: xmm10 &= xmm6
10071# asm 1: pand <xmm6=int6464#7,<xmm10=int6464#11
10072# asm 2: pand <xmm6=%xmm6,<xmm10=%xmm10
10073pand %xmm6,%xmm10
10074
10075# qhasm: xmm6 ^= xmm5
10076# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
10077# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
10078pxor %xmm5,%xmm6
10079
10080# qhasm: xmm6 &= xmm14
10081# asm 1: pand <xmm14=int6464#12,<xmm6=int6464#7
10082# asm 2: pand <xmm14=%xmm11,<xmm6=%xmm6
10083pand %xmm11,%xmm6
10084
10085# qhasm: xmm5 &= xmm15
10086# asm 1: pand <xmm15=int6464#14,<xmm5=int6464#6
10087# asm 2: pand <xmm15=%xmm13,<xmm5=%xmm5
10088pand %xmm13,%xmm5
10089
10090# qhasm: xmm6 ^= xmm5
10091# asm 1: pxor <xmm5=int6464#6,<xmm6=int6464#7
10092# asm 2: pxor <xmm5=%xmm5,<xmm6=%xmm6
10093pxor %xmm5,%xmm6
10094
10095# qhasm: xmm5 ^= xmm10
10096# asm 1: pxor <xmm10=int6464#11,<xmm5=int6464#6
10097# asm 2: pxor <xmm10=%xmm10,<xmm5=%xmm5
10098pxor %xmm10,%xmm5
10099
10100# qhasm: xmm12 ^= xmm0
10101# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#9
10102# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm8
10103pxor %xmm0,%xmm8
10104
10105# qhasm: xmm8 ^= xmm3
10106# asm 1: pxor <xmm3=int6464#4,<xmm8=int6464#10
10107# asm 2: pxor <xmm3=%xmm3,<xmm8=%xmm9
10108pxor %xmm3,%xmm9
10109
10110# qhasm: xmm15 ^= xmm13
10111# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
10112# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
10113pxor %xmm15,%xmm13
10114
10115# qhasm: xmm14 ^= xmm9
10116# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
10117# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
10118pxor %xmm12,%xmm11
10119
10120# qhasm: xmm11 = xmm15
10121# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
10122# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
10123movdqa %xmm13,%xmm10
10124
10125# qhasm: xmm11 ^= xmm14
10126# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
10127# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
10128pxor %xmm11,%xmm10
10129
10130# qhasm: xmm11 &= xmm12
10131# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
10132# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
10133pand %xmm8,%xmm10
10134
10135# qhasm: xmm12 ^= xmm8
10136# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
10137# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
10138pxor %xmm9,%xmm8
10139
10140# qhasm: xmm12 &= xmm14
10141# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
10142# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
10143pand %xmm11,%xmm8
10144
10145# qhasm: xmm8 &= xmm15
10146# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
10147# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
10148pand %xmm13,%xmm9
10149
10150# qhasm: xmm8 ^= xmm12
10151# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
10152# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
10153pxor %xmm8,%xmm9
10154
10155# qhasm: xmm12 ^= xmm11
10156# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
10157# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
10158pxor %xmm10,%xmm8
10159
10160# qhasm: xmm10 = xmm13
10161# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
10162# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
10163movdqa %xmm15,%xmm10
10164
10165# qhasm: xmm10 ^= xmm9
10166# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
10167# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
10168pxor %xmm12,%xmm10
10169
10170# qhasm: xmm10 &= xmm0
10171# asm 1: pand <xmm0=int6464#1,<xmm10=int6464#11
10172# asm 2: pand <xmm0=%xmm0,<xmm10=%xmm10
10173pand %xmm0,%xmm10
10174
10175# qhasm: xmm0 ^= xmm3
10176# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
10177# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
10178pxor %xmm3,%xmm0
10179
10180# qhasm: xmm0 &= xmm9
10181# asm 1: pand <xmm9=int6464#13,<xmm0=int6464#1
10182# asm 2: pand <xmm9=%xmm12,<xmm0=%xmm0
10183pand %xmm12,%xmm0
10184
10185# qhasm: xmm3 &= xmm13
10186# asm 1: pand <xmm13=int6464#16,<xmm3=int6464#4
10187# asm 2: pand <xmm13=%xmm15,<xmm3=%xmm3
10188pand %xmm15,%xmm3
10189
10190# qhasm: xmm0 ^= xmm3
10191# asm 1: pxor <xmm3=int6464#4,<xmm0=int6464#1
10192# asm 2: pxor <xmm3=%xmm3,<xmm0=%xmm0
10193pxor %xmm3,%xmm0
10194
10195# qhasm: xmm3 ^= xmm10
10196# asm 1: pxor <xmm10=int6464#11,<xmm3=int6464#4
10197# asm 2: pxor <xmm10=%xmm10,<xmm3=%xmm3
10198pxor %xmm10,%xmm3
10199
10200# qhasm: xmm6 ^= xmm12
10201# asm 1: pxor <xmm12=int6464#9,<xmm6=int6464#7
10202# asm 2: pxor <xmm12=%xmm8,<xmm6=%xmm6
10203pxor %xmm8,%xmm6
10204
10205# qhasm: xmm0 ^= xmm12
10206# asm 1: pxor <xmm12=int6464#9,<xmm0=int6464#1
10207# asm 2: pxor <xmm12=%xmm8,<xmm0=%xmm0
10208pxor %xmm8,%xmm0
10209
10210# qhasm: xmm5 ^= xmm8
10211# asm 1: pxor <xmm8=int6464#10,<xmm5=int6464#6
10212# asm 2: pxor <xmm8=%xmm9,<xmm5=%xmm5
10213pxor %xmm9,%xmm5
10214
10215# qhasm: xmm3 ^= xmm8
10216# asm 1: pxor <xmm8=int6464#10,<xmm3=int6464#4
10217# asm 2: pxor <xmm8=%xmm9,<xmm3=%xmm3
10218pxor %xmm9,%xmm3
10219
10220# qhasm: xmm12 = xmm7
10221# asm 1: movdqa <xmm7=int6464#8,>xmm12=int6464#9
10222# asm 2: movdqa <xmm7=%xmm7,>xmm12=%xmm8
10223movdqa %xmm7,%xmm8
10224
10225# qhasm: xmm8 = xmm1
10226# asm 1: movdqa <xmm1=int6464#2,>xmm8=int6464#10
10227# asm 2: movdqa <xmm1=%xmm1,>xmm8=%xmm9
10228movdqa %xmm1,%xmm9
10229
10230# qhasm: xmm12 ^= xmm4
10231# asm 1: pxor <xmm4=int6464#5,<xmm12=int6464#9
10232# asm 2: pxor <xmm4=%xmm4,<xmm12=%xmm8
10233pxor %xmm4,%xmm8
10234
10235# qhasm: xmm8 ^= xmm2
10236# asm 1: pxor <xmm2=int6464#3,<xmm8=int6464#10
10237# asm 2: pxor <xmm2=%xmm2,<xmm8=%xmm9
10238pxor %xmm2,%xmm9
10239
10240# qhasm: xmm11 = xmm15
10241# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
10242# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
10243movdqa %xmm13,%xmm10
10244
10245# qhasm: xmm11 ^= xmm14
10246# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
10247# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
10248pxor %xmm11,%xmm10
10249
10250# qhasm: xmm11 &= xmm12
10251# asm 1: pand <xmm12=int6464#9,<xmm11=int6464#11
10252# asm 2: pand <xmm12=%xmm8,<xmm11=%xmm10
10253pand %xmm8,%xmm10
10254
10255# qhasm: xmm12 ^= xmm8
10256# asm 1: pxor <xmm8=int6464#10,<xmm12=int6464#9
10257# asm 2: pxor <xmm8=%xmm9,<xmm12=%xmm8
10258pxor %xmm9,%xmm8
10259
10260# qhasm: xmm12 &= xmm14
10261# asm 1: pand <xmm14=int6464#12,<xmm12=int6464#9
10262# asm 2: pand <xmm14=%xmm11,<xmm12=%xmm8
10263pand %xmm11,%xmm8
10264
10265# qhasm: xmm8 &= xmm15
10266# asm 1: pand <xmm15=int6464#14,<xmm8=int6464#10
10267# asm 2: pand <xmm15=%xmm13,<xmm8=%xmm9
10268pand %xmm13,%xmm9
10269
10270# qhasm: xmm8 ^= xmm12
10271# asm 1: pxor <xmm12=int6464#9,<xmm8=int6464#10
10272# asm 2: pxor <xmm12=%xmm8,<xmm8=%xmm9
10273pxor %xmm8,%xmm9
10274
10275# qhasm: xmm12 ^= xmm11
10276# asm 1: pxor <xmm11=int6464#11,<xmm12=int6464#9
10277# asm 2: pxor <xmm11=%xmm10,<xmm12=%xmm8
10278pxor %xmm10,%xmm8
10279
10280# qhasm: xmm10 = xmm13
10281# asm 1: movdqa <xmm13=int6464#16,>xmm10=int6464#11
10282# asm 2: movdqa <xmm13=%xmm15,>xmm10=%xmm10
10283movdqa %xmm15,%xmm10
10284
10285# qhasm: xmm10 ^= xmm9
10286# asm 1: pxor <xmm9=int6464#13,<xmm10=int6464#11
10287# asm 2: pxor <xmm9=%xmm12,<xmm10=%xmm10
10288pxor %xmm12,%xmm10
10289
10290# qhasm: xmm10 &= xmm4
10291# asm 1: pand <xmm4=int6464#5,<xmm10=int6464#11
10292# asm 2: pand <xmm4=%xmm4,<xmm10=%xmm10
10293pand %xmm4,%xmm10
10294
10295# qhasm: xmm4 ^= xmm2
10296# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
10297# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
10298pxor %xmm2,%xmm4
10299
10300# qhasm: xmm4 &= xmm9
10301# asm 1: pand <xmm9=int6464#13,<xmm4=int6464#5
10302# asm 2: pand <xmm9=%xmm12,<xmm4=%xmm4
10303pand %xmm12,%xmm4
10304
10305# qhasm: xmm2 &= xmm13
10306# asm 1: pand <xmm13=int6464#16,<xmm2=int6464#3
10307# asm 2: pand <xmm13=%xmm15,<xmm2=%xmm2
10308pand %xmm15,%xmm2
10309
10310# qhasm: xmm4 ^= xmm2
10311# asm 1: pxor <xmm2=int6464#3,<xmm4=int6464#5
10312# asm 2: pxor <xmm2=%xmm2,<xmm4=%xmm4
10313pxor %xmm2,%xmm4
10314
10315# qhasm: xmm2 ^= xmm10
10316# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#3
10317# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm2
10318pxor %xmm10,%xmm2
10319
10320# qhasm: xmm15 ^= xmm13
10321# asm 1: pxor <xmm13=int6464#16,<xmm15=int6464#14
10322# asm 2: pxor <xmm13=%xmm15,<xmm15=%xmm13
10323pxor %xmm15,%xmm13
10324
10325# qhasm: xmm14 ^= xmm9
10326# asm 1: pxor <xmm9=int6464#13,<xmm14=int6464#12
10327# asm 2: pxor <xmm9=%xmm12,<xmm14=%xmm11
10328pxor %xmm12,%xmm11
10329
10330# qhasm: xmm11 = xmm15
10331# asm 1: movdqa <xmm15=int6464#14,>xmm11=int6464#11
10332# asm 2: movdqa <xmm15=%xmm13,>xmm11=%xmm10
10333movdqa %xmm13,%xmm10
10334
10335# qhasm: xmm11 ^= xmm14
10336# asm 1: pxor <xmm14=int6464#12,<xmm11=int6464#11
10337# asm 2: pxor <xmm14=%xmm11,<xmm11=%xmm10
10338pxor %xmm11,%xmm10
10339
10340# qhasm: xmm11 &= xmm7
10341# asm 1: pand <xmm7=int6464#8,<xmm11=int6464#11
10342# asm 2: pand <xmm7=%xmm7,<xmm11=%xmm10
10343pand %xmm7,%xmm10
10344
10345# qhasm: xmm7 ^= xmm1
10346# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
10347# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
10348pxor %xmm1,%xmm7
10349
10350# qhasm: xmm7 &= xmm14
10351# asm 1: pand <xmm14=int6464#12,<xmm7=int6464#8
10352# asm 2: pand <xmm14=%xmm11,<xmm7=%xmm7
10353pand %xmm11,%xmm7
10354
10355# qhasm: xmm1 &= xmm15
10356# asm 1: pand <xmm15=int6464#14,<xmm1=int6464#2
10357# asm 2: pand <xmm15=%xmm13,<xmm1=%xmm1
10358pand %xmm13,%xmm1
10359
10360# qhasm: xmm7 ^= xmm1
10361# asm 1: pxor <xmm1=int6464#2,<xmm7=int6464#8
10362# asm 2: pxor <xmm1=%xmm1,<xmm7=%xmm7
10363pxor %xmm1,%xmm7
10364
10365# qhasm: xmm1 ^= xmm11
10366# asm 1: pxor <xmm11=int6464#11,<xmm1=int6464#2
10367# asm 2: pxor <xmm11=%xmm10,<xmm1=%xmm1
10368pxor %xmm10,%xmm1
10369
10370# qhasm: xmm7 ^= xmm12
10371# asm 1: pxor <xmm12=int6464#9,<xmm7=int6464#8
10372# asm 2: pxor <xmm12=%xmm8,<xmm7=%xmm7
10373pxor %xmm8,%xmm7
10374
10375# qhasm: xmm4 ^= xmm12
10376# asm 1: pxor <xmm12=int6464#9,<xmm4=int6464#5
10377# asm 2: pxor <xmm12=%xmm8,<xmm4=%xmm4
10378pxor %xmm8,%xmm4
10379
10380# qhasm: xmm1 ^= xmm8
10381# asm 1: pxor <xmm8=int6464#10,<xmm1=int6464#2
10382# asm 2: pxor <xmm8=%xmm9,<xmm1=%xmm1
10383pxor %xmm9,%xmm1
10384
10385# qhasm: xmm2 ^= xmm8
10386# asm 1: pxor <xmm8=int6464#10,<xmm2=int6464#3
10387# asm 2: pxor <xmm8=%xmm9,<xmm2=%xmm2
10388pxor %xmm9,%xmm2
10389
10390# qhasm: xmm7 ^= xmm0
10391# asm 1: pxor <xmm0=int6464#1,<xmm7=int6464#8
10392# asm 2: pxor <xmm0=%xmm0,<xmm7=%xmm7
10393pxor %xmm0,%xmm7
10394
10395# qhasm: xmm1 ^= xmm6
10396# asm 1: pxor <xmm6=int6464#7,<xmm1=int6464#2
10397# asm 2: pxor <xmm6=%xmm6,<xmm1=%xmm1
10398pxor %xmm6,%xmm1
10399
10400# qhasm: xmm4 ^= xmm7
10401# asm 1: pxor <xmm7=int6464#8,<xmm4=int6464#5
10402# asm 2: pxor <xmm7=%xmm7,<xmm4=%xmm4
10403pxor %xmm7,%xmm4
10404
10405# qhasm: xmm6 ^= xmm0
10406# asm 1: pxor <xmm0=int6464#1,<xmm6=int6464#7
10407# asm 2: pxor <xmm0=%xmm0,<xmm6=%xmm6
10408pxor %xmm0,%xmm6
10409
10410# qhasm: xmm0 ^= xmm1
10411# asm 1: pxor <xmm1=int6464#2,<xmm0=int6464#1
10412# asm 2: pxor <xmm1=%xmm1,<xmm0=%xmm0
10413pxor %xmm1,%xmm0
10414
10415# qhasm: xmm1 ^= xmm5
10416# asm 1: pxor <xmm5=int6464#6,<xmm1=int6464#2
10417# asm 2: pxor <xmm5=%xmm5,<xmm1=%xmm1
10418pxor %xmm5,%xmm1
10419
10420# qhasm: xmm5 ^= xmm2
10421# asm 1: pxor <xmm2=int6464#3,<xmm5=int6464#6
10422# asm 2: pxor <xmm2=%xmm2,<xmm5=%xmm5
10423pxor %xmm2,%xmm5
10424
10425# qhasm: xmm4 ^= xmm5
10426# asm 1: pxor <xmm5=int6464#6,<xmm4=int6464#5
10427# asm 2: pxor <xmm5=%xmm5,<xmm4=%xmm4
10428pxor %xmm5,%xmm4
10429
10430# qhasm: xmm2 ^= xmm3
10431# asm 1: pxor <xmm3=int6464#4,<xmm2=int6464#3
10432# asm 2: pxor <xmm3=%xmm3,<xmm2=%xmm2
10433pxor %xmm3,%xmm2
10434
10435# qhasm: xmm3 ^= xmm5
10436# asm 1: pxor <xmm5=int6464#6,<xmm3=int6464#4
10437# asm 2: pxor <xmm5=%xmm5,<xmm3=%xmm3
10438pxor %xmm5,%xmm3
10439
10440# qhasm: xmm6 ^= xmm3
10441# asm 1: pxor <xmm3=int6464#4,<xmm6=int6464#7
10442# asm 2: pxor <xmm3=%xmm3,<xmm6=%xmm6
10443pxor %xmm3,%xmm6
10444
10445# qhasm: xmm8 = shuffle dwords of xmm0 by 0x93
10446# asm 1: pshufd $0x93,<xmm0=int6464#1,>xmm8=int6464#9
10447# asm 2: pshufd $0x93,<xmm0=%xmm0,>xmm8=%xmm8
10448pshufd $0x93,%xmm0,%xmm8
10449
10450# qhasm: xmm9 = shuffle dwords of xmm1 by 0x93
10451# asm 1: pshufd $0x93,<xmm1=int6464#2,>xmm9=int6464#10
10452# asm 2: pshufd $0x93,<xmm1=%xmm1,>xmm9=%xmm9
10453pshufd $0x93,%xmm1,%xmm9
10454
10455# qhasm: xmm10 = shuffle dwords of xmm4 by 0x93
10456# asm 1: pshufd $0x93,<xmm4=int6464#5,>xmm10=int6464#11
10457# asm 2: pshufd $0x93,<xmm4=%xmm4,>xmm10=%xmm10
10458pshufd $0x93,%xmm4,%xmm10
10459
10460# qhasm: xmm11 = shuffle dwords of xmm6 by 0x93
10461# asm 1: pshufd $0x93,<xmm6=int6464#7,>xmm11=int6464#12
10462# asm 2: pshufd $0x93,<xmm6=%xmm6,>xmm11=%xmm11
10463pshufd $0x93,%xmm6,%xmm11
10464
10465# qhasm: xmm12 = shuffle dwords of xmm3 by 0x93
10466# asm 1: pshufd $0x93,<xmm3=int6464#4,>xmm12=int6464#13
10467# asm 2: pshufd $0x93,<xmm3=%xmm3,>xmm12=%xmm12
10468pshufd $0x93,%xmm3,%xmm12
10469
10470# qhasm: xmm13 = shuffle dwords of xmm7 by 0x93
10471# asm 1: pshufd $0x93,<xmm7=int6464#8,>xmm13=int6464#14
10472# asm 2: pshufd $0x93,<xmm7=%xmm7,>xmm13=%xmm13
10473pshufd $0x93,%xmm7,%xmm13
10474
10475# qhasm: xmm14 = shuffle dwords of xmm2 by 0x93
10476# asm 1: pshufd $0x93,<xmm2=int6464#3,>xmm14=int6464#15
10477# asm 2: pshufd $0x93,<xmm2=%xmm2,>xmm14=%xmm14
10478pshufd $0x93,%xmm2,%xmm14
10479
10480# qhasm: xmm15 = shuffle dwords of xmm5 by 0x93
10481# asm 1: pshufd $0x93,<xmm5=int6464#6,>xmm15=int6464#16
10482# asm 2: pshufd $0x93,<xmm5=%xmm5,>xmm15=%xmm15
10483pshufd $0x93,%xmm5,%xmm15
10484
10485# qhasm: xmm0 ^= xmm8
10486# asm 1: pxor <xmm8=int6464#9,<xmm0=int6464#1
10487# asm 2: pxor <xmm8=%xmm8,<xmm0=%xmm0
10488pxor %xmm8,%xmm0
10489
10490# qhasm: xmm1 ^= xmm9
10491# asm 1: pxor <xmm9=int6464#10,<xmm1=int6464#2
10492# asm 2: pxor <xmm9=%xmm9,<xmm1=%xmm1
10493pxor %xmm9,%xmm1
10494
10495# qhasm: xmm4 ^= xmm10
10496# asm 1: pxor <xmm10=int6464#11,<xmm4=int6464#5
10497# asm 2: pxor <xmm10=%xmm10,<xmm4=%xmm4
10498pxor %xmm10,%xmm4
10499
10500# qhasm: xmm6 ^= xmm11
10501# asm 1: pxor <xmm11=int6464#12,<xmm6=int6464#7
10502# asm 2: pxor <xmm11=%xmm11,<xmm6=%xmm6
10503pxor %xmm11,%xmm6
10504
10505# qhasm: xmm3 ^= xmm12
10506# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#4
10507# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm3
10508pxor %xmm12,%xmm3
10509
10510# qhasm: xmm7 ^= xmm13
10511# asm 1: pxor <xmm13=int6464#14,<xmm7=int6464#8
10512# asm 2: pxor <xmm13=%xmm13,<xmm7=%xmm7
10513pxor %xmm13,%xmm7
10514
10515# qhasm: xmm2 ^= xmm14
10516# asm 1: pxor <xmm14=int6464#15,<xmm2=int6464#3
10517# asm 2: pxor <xmm14=%xmm14,<xmm2=%xmm2
10518pxor %xmm14,%xmm2
10519
10520# qhasm: xmm5 ^= xmm15
10521# asm 1: pxor <xmm15=int6464#16,<xmm5=int6464#6
10522# asm 2: pxor <xmm15=%xmm15,<xmm5=%xmm5
10523pxor %xmm15,%xmm5
10524
10525# qhasm: xmm8 ^= xmm5
10526# asm 1: pxor <xmm5=int6464#6,<xmm8=int6464#9
10527# asm 2: pxor <xmm5=%xmm5,<xmm8=%xmm8
10528pxor %xmm5,%xmm8
10529
10530# qhasm: xmm9 ^= xmm0
10531# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
10532# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
10533pxor %xmm0,%xmm9
10534
10535# qhasm: xmm10 ^= xmm1
10536# asm 1: pxor <xmm1=int6464#2,<xmm10=int6464#11
10537# asm 2: pxor <xmm1=%xmm1,<xmm10=%xmm10
10538pxor %xmm1,%xmm10
10539
10540# qhasm: xmm9 ^= xmm5
10541# asm 1: pxor <xmm5=int6464#6,<xmm9=int6464#10
10542# asm 2: pxor <xmm5=%xmm5,<xmm9=%xmm9
10543pxor %xmm5,%xmm9
10544
10545# qhasm: xmm11 ^= xmm4
10546# asm 1: pxor <xmm4=int6464#5,<xmm11=int6464#12
10547# asm 2: pxor <xmm4=%xmm4,<xmm11=%xmm11
10548pxor %xmm4,%xmm11
10549
10550# qhasm: xmm12 ^= xmm6
10551# asm 1: pxor <xmm6=int6464#7,<xmm12=int6464#13
10552# asm 2: pxor <xmm6=%xmm6,<xmm12=%xmm12
10553pxor %xmm6,%xmm12
10554
10555# qhasm: xmm13 ^= xmm3
10556# asm 1: pxor <xmm3=int6464#4,<xmm13=int6464#14
10557# asm 2: pxor <xmm3=%xmm3,<xmm13=%xmm13
10558pxor %xmm3,%xmm13
10559
10560# qhasm: xmm11 ^= xmm5
10561# asm 1: pxor <xmm5=int6464#6,<xmm11=int6464#12
10562# asm 2: pxor <xmm5=%xmm5,<xmm11=%xmm11
10563pxor %xmm5,%xmm11
10564
10565# qhasm: xmm14 ^= xmm7
10566# asm 1: pxor <xmm7=int6464#8,<xmm14=int6464#15
10567# asm 2: pxor <xmm7=%xmm7,<xmm14=%xmm14
10568pxor %xmm7,%xmm14
10569
10570# qhasm: xmm15 ^= xmm2
10571# asm 1: pxor <xmm2=int6464#3,<xmm15=int6464#16
10572# asm 2: pxor <xmm2=%xmm2,<xmm15=%xmm15
10573pxor %xmm2,%xmm15
10574
10575# qhasm: xmm12 ^= xmm5
10576# asm 1: pxor <xmm5=int6464#6,<xmm12=int6464#13
10577# asm 2: pxor <xmm5=%xmm5,<xmm12=%xmm12
10578pxor %xmm5,%xmm12
10579
10580# qhasm: xmm0 = shuffle dwords of xmm0 by 0x4E
10581# asm 1: pshufd $0x4E,<xmm0=int6464#1,>xmm0=int6464#1
10582# asm 2: pshufd $0x4E,<xmm0=%xmm0,>xmm0=%xmm0
10583pshufd $0x4E,%xmm0,%xmm0
10584
10585# qhasm: xmm1 = shuffle dwords of xmm1 by 0x4E
10586# asm 1: pshufd $0x4E,<xmm1=int6464#2,>xmm1=int6464#2
10587# asm 2: pshufd $0x4E,<xmm1=%xmm1,>xmm1=%xmm1
10588pshufd $0x4E,%xmm1,%xmm1
10589
10590# qhasm: xmm4 = shuffle dwords of xmm4 by 0x4E
10591# asm 1: pshufd $0x4E,<xmm4=int6464#5,>xmm4=int6464#5
10592# asm 2: pshufd $0x4E,<xmm4=%xmm4,>xmm4=%xmm4
10593pshufd $0x4E,%xmm4,%xmm4
10594
10595# qhasm: xmm6 = shuffle dwords of xmm6 by 0x4E
10596# asm 1: pshufd $0x4E,<xmm6=int6464#7,>xmm6=int6464#7
10597# asm 2: pshufd $0x4E,<xmm6=%xmm6,>xmm6=%xmm6
10598pshufd $0x4E,%xmm6,%xmm6
10599
10600# qhasm: xmm3 = shuffle dwords of xmm3 by 0x4E
10601# asm 1: pshufd $0x4E,<xmm3=int6464#4,>xmm3=int6464#4
10602# asm 2: pshufd $0x4E,<xmm3=%xmm3,>xmm3=%xmm3
10603pshufd $0x4E,%xmm3,%xmm3
10604
10605# qhasm: xmm7 = shuffle dwords of xmm7 by 0x4E
10606# asm 1: pshufd $0x4E,<xmm7=int6464#8,>xmm7=int6464#8
10607# asm 2: pshufd $0x4E,<xmm7=%xmm7,>xmm7=%xmm7
10608pshufd $0x4E,%xmm7,%xmm7
10609
10610# qhasm: xmm2 = shuffle dwords of xmm2 by 0x4E
10611# asm 1: pshufd $0x4E,<xmm2=int6464#3,>xmm2=int6464#3
10612# asm 2: pshufd $0x4E,<xmm2=%xmm2,>xmm2=%xmm2
10613pshufd $0x4E,%xmm2,%xmm2
10614
10615# qhasm: xmm5 = shuffle dwords of xmm5 by 0x4E
10616# asm 1: pshufd $0x4E,<xmm5=int6464#6,>xmm5=int6464#6
10617# asm 2: pshufd $0x4E,<xmm5=%xmm5,>xmm5=%xmm5
10618pshufd $0x4E,%xmm5,%xmm5
10619
10620# qhasm: xmm8 ^= xmm0
10621# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
10622# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
10623pxor %xmm0,%xmm8
10624
10625# qhasm: xmm9 ^= xmm1
10626# asm 1: pxor <xmm1=int6464#2,<xmm9=int6464#10
10627# asm 2: pxor <xmm1=%xmm1,<xmm9=%xmm9
10628pxor %xmm1,%xmm9
10629
10630# qhasm: xmm10 ^= xmm4
10631# asm 1: pxor <xmm4=int6464#5,<xmm10=int6464#11
10632# asm 2: pxor <xmm4=%xmm4,<xmm10=%xmm10
10633pxor %xmm4,%xmm10
10634
10635# qhasm: xmm11 ^= xmm6
10636# asm 1: pxor <xmm6=int6464#7,<xmm11=int6464#12
10637# asm 2: pxor <xmm6=%xmm6,<xmm11=%xmm11
10638pxor %xmm6,%xmm11
10639
10640# qhasm: xmm12 ^= xmm3
10641# asm 1: pxor <xmm3=int6464#4,<xmm12=int6464#13
10642# asm 2: pxor <xmm3=%xmm3,<xmm12=%xmm12
10643pxor %xmm3,%xmm12
10644
10645# qhasm: xmm13 ^= xmm7
10646# asm 1: pxor <xmm7=int6464#8,<xmm13=int6464#14
10647# asm 2: pxor <xmm7=%xmm7,<xmm13=%xmm13
10648pxor %xmm7,%xmm13
10649
10650# qhasm: xmm14 ^= xmm2
10651# asm 1: pxor <xmm2=int6464#3,<xmm14=int6464#15
10652# asm 2: pxor <xmm2=%xmm2,<xmm14=%xmm14
10653pxor %xmm2,%xmm14
10654
10655# qhasm: xmm15 ^= xmm5
10656# asm 1: pxor <xmm5=int6464#6,<xmm15=int6464#16
10657# asm 2: pxor <xmm5=%xmm5,<xmm15=%xmm15
10658pxor %xmm5,%xmm15
10659
10660# qhasm: xmm8 ^= *(int128 *)(c + 1152)
10661# asm 1: pxor 1152(<c=int64#4),<xmm8=int6464#9
10662# asm 2: pxor 1152(<c=%rcx),<xmm8=%xmm8
10663pxor 1152(%rcx),%xmm8
10664
10665# qhasm: shuffle bytes of xmm8 by SRM0
10666# asm 1: pshufb SRM0,<xmm8=int6464#9
10667# asm 2: pshufb SRM0,<xmm8=%xmm8
10668pshufb SRM0,%xmm8
10669
10670# qhasm: xmm9 ^= *(int128 *)(c + 1168)
10671# asm 1: pxor 1168(<c=int64#4),<xmm9=int6464#10
10672# asm 2: pxor 1168(<c=%rcx),<xmm9=%xmm9
10673pxor 1168(%rcx),%xmm9
10674
10675# qhasm: shuffle bytes of xmm9 by SRM0
10676# asm 1: pshufb SRM0,<xmm9=int6464#10
10677# asm 2: pshufb SRM0,<xmm9=%xmm9
10678pshufb SRM0,%xmm9
10679
10680# qhasm: xmm10 ^= *(int128 *)(c + 1184)
10681# asm 1: pxor 1184(<c=int64#4),<xmm10=int6464#11
10682# asm 2: pxor 1184(<c=%rcx),<xmm10=%xmm10
10683pxor 1184(%rcx),%xmm10
10684
10685# qhasm: shuffle bytes of xmm10 by SRM0
10686# asm 1: pshufb SRM0,<xmm10=int6464#11
10687# asm 2: pshufb SRM0,<xmm10=%xmm10
10688pshufb SRM0,%xmm10
10689
10690# qhasm: xmm11 ^= *(int128 *)(c + 1200)
10691# asm 1: pxor 1200(<c=int64#4),<xmm11=int6464#12
10692# asm 2: pxor 1200(<c=%rcx),<xmm11=%xmm11
10693pxor 1200(%rcx),%xmm11
10694
10695# qhasm: shuffle bytes of xmm11 by SRM0
10696# asm 1: pshufb SRM0,<xmm11=int6464#12
10697# asm 2: pshufb SRM0,<xmm11=%xmm11
10698pshufb SRM0,%xmm11
10699
10700# qhasm: xmm12 ^= *(int128 *)(c + 1216)
10701# asm 1: pxor 1216(<c=int64#4),<xmm12=int6464#13
10702# asm 2: pxor 1216(<c=%rcx),<xmm12=%xmm12
10703pxor 1216(%rcx),%xmm12
10704
10705# qhasm: shuffle bytes of xmm12 by SRM0
10706# asm 1: pshufb SRM0,<xmm12=int6464#13
10707# asm 2: pshufb SRM0,<xmm12=%xmm12
10708pshufb SRM0,%xmm12
10709
10710# qhasm: xmm13 ^= *(int128 *)(c + 1232)
10711# asm 1: pxor 1232(<c=int64#4),<xmm13=int6464#14
10712# asm 2: pxor 1232(<c=%rcx),<xmm13=%xmm13
10713pxor 1232(%rcx),%xmm13
10714
10715# qhasm: shuffle bytes of xmm13 by SRM0
10716# asm 1: pshufb SRM0,<xmm13=int6464#14
10717# asm 2: pshufb SRM0,<xmm13=%xmm13
10718pshufb SRM0,%xmm13
10719
10720# qhasm: xmm14 ^= *(int128 *)(c + 1248)
10721# asm 1: pxor 1248(<c=int64#4),<xmm14=int6464#15
10722# asm 2: pxor 1248(<c=%rcx),<xmm14=%xmm14
10723pxor 1248(%rcx),%xmm14
10724
10725# qhasm: shuffle bytes of xmm14 by SRM0
10726# asm 1: pshufb SRM0,<xmm14=int6464#15
10727# asm 2: pshufb SRM0,<xmm14=%xmm14
10728pshufb SRM0,%xmm14
10729
10730# qhasm: xmm15 ^= *(int128 *)(c + 1264)
10731# asm 1: pxor 1264(<c=int64#4),<xmm15=int6464#16
10732# asm 2: pxor 1264(<c=%rcx),<xmm15=%xmm15
10733pxor 1264(%rcx),%xmm15
10734
10735# qhasm: shuffle bytes of xmm15 by SRM0
10736# asm 1: pshufb SRM0,<xmm15=int6464#16
10737# asm 2: pshufb SRM0,<xmm15=%xmm15
10738pshufb SRM0,%xmm15
10739
10740# qhasm: xmm13 ^= xmm14
10741# asm 1: pxor <xmm14=int6464#15,<xmm13=int6464#14
10742# asm 2: pxor <xmm14=%xmm14,<xmm13=%xmm13
10743pxor %xmm14,%xmm13
10744
10745# qhasm: xmm10 ^= xmm9
10746# asm 1: pxor <xmm9=int6464#10,<xmm10=int6464#11
10747# asm 2: pxor <xmm9=%xmm9,<xmm10=%xmm10
10748pxor %xmm9,%xmm10
10749
10750# qhasm: xmm13 ^= xmm8
10751# asm 1: pxor <xmm8=int6464#9,<xmm13=int6464#14
10752# asm 2: pxor <xmm8=%xmm8,<xmm13=%xmm13
10753pxor %xmm8,%xmm13
10754
10755# qhasm: xmm14 ^= xmm10
10756# asm 1: pxor <xmm10=int6464#11,<xmm14=int6464#15
10757# asm 2: pxor <xmm10=%xmm10,<xmm14=%xmm14
10758pxor %xmm10,%xmm14
10759
10760# qhasm: xmm11 ^= xmm8
10761# asm 1: pxor <xmm8=int6464#9,<xmm11=int6464#12
10762# asm 2: pxor <xmm8=%xmm8,<xmm11=%xmm11
10763pxor %xmm8,%xmm11
10764
10765# qhasm: xmm14 ^= xmm11
10766# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
10767# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
10768pxor %xmm11,%xmm14
10769
10770# qhasm: xmm11 ^= xmm15
10771# asm 1: pxor <xmm15=int6464#16,<xmm11=int6464#12
10772# asm 2: pxor <xmm15=%xmm15,<xmm11=%xmm11
10773pxor %xmm15,%xmm11
10774
10775# qhasm: xmm11 ^= xmm12
10776# asm 1: pxor <xmm12=int6464#13,<xmm11=int6464#12
10777# asm 2: pxor <xmm12=%xmm12,<xmm11=%xmm11
10778pxor %xmm12,%xmm11
10779
10780# qhasm: xmm15 ^= xmm13
10781# asm 1: pxor <xmm13=int6464#14,<xmm15=int6464#16
10782# asm 2: pxor <xmm13=%xmm13,<xmm15=%xmm15
10783pxor %xmm13,%xmm15
10784
10785# qhasm: xmm11 ^= xmm9
10786# asm 1: pxor <xmm9=int6464#10,<xmm11=int6464#12
10787# asm 2: pxor <xmm9=%xmm9,<xmm11=%xmm11
10788pxor %xmm9,%xmm11
10789
10790# qhasm: xmm12 ^= xmm13
10791# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
10792# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
10793pxor %xmm13,%xmm12
10794
10795# qhasm: xmm10 ^= xmm15
10796# asm 1: pxor <xmm15=int6464#16,<xmm10=int6464#11
10797# asm 2: pxor <xmm15=%xmm15,<xmm10=%xmm10
10798pxor %xmm15,%xmm10
10799
10800# qhasm: xmm9 ^= xmm13
10801# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
10802# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
10803pxor %xmm13,%xmm9
10804
10805# qhasm: xmm3 = xmm15
10806# asm 1: movdqa <xmm15=int6464#16,>xmm3=int6464#1
10807# asm 2: movdqa <xmm15=%xmm15,>xmm3=%xmm0
10808movdqa %xmm15,%xmm0
10809
10810# qhasm: xmm2 = xmm9
10811# asm 1: movdqa <xmm9=int6464#10,>xmm2=int6464#2
10812# asm 2: movdqa <xmm9=%xmm9,>xmm2=%xmm1
10813movdqa %xmm9,%xmm1
10814
10815# qhasm: xmm1 = xmm13
10816# asm 1: movdqa <xmm13=int6464#14,>xmm1=int6464#3
10817# asm 2: movdqa <xmm13=%xmm13,>xmm1=%xmm2
10818movdqa %xmm13,%xmm2
10819
10820# qhasm: xmm5 = xmm10
10821# asm 1: movdqa <xmm10=int6464#11,>xmm5=int6464#4
10822# asm 2: movdqa <xmm10=%xmm10,>xmm5=%xmm3
10823movdqa %xmm10,%xmm3
10824
10825# qhasm: xmm4 = xmm14
10826# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#5
10827# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm4
10828movdqa %xmm14,%xmm4
10829
10830# qhasm: xmm3 ^= xmm12
10831# asm 1: pxor <xmm12=int6464#13,<xmm3=int6464#1
10832# asm 2: pxor <xmm12=%xmm12,<xmm3=%xmm0
10833pxor %xmm12,%xmm0
10834
10835# qhasm: xmm2 ^= xmm10
10836# asm 1: pxor <xmm10=int6464#11,<xmm2=int6464#2
10837# asm 2: pxor <xmm10=%xmm10,<xmm2=%xmm1
10838pxor %xmm10,%xmm1
10839
10840# qhasm: xmm1 ^= xmm11
10841# asm 1: pxor <xmm11=int6464#12,<xmm1=int6464#3
10842# asm 2: pxor <xmm11=%xmm11,<xmm1=%xmm2
10843pxor %xmm11,%xmm2
10844
10845# qhasm: xmm5 ^= xmm12
10846# asm 1: pxor <xmm12=int6464#13,<xmm5=int6464#4
10847# asm 2: pxor <xmm12=%xmm12,<xmm5=%xmm3
10848pxor %xmm12,%xmm3
10849
10850# qhasm: xmm4 ^= xmm8
10851# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#5
10852# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm4
10853pxor %xmm8,%xmm4
10854
10855# qhasm: xmm6 = xmm3
10856# asm 1: movdqa <xmm3=int6464#1,>xmm6=int6464#6
10857# asm 2: movdqa <xmm3=%xmm0,>xmm6=%xmm5
10858movdqa %xmm0,%xmm5
10859
10860# qhasm: xmm0 = xmm2
10861# asm 1: movdqa <xmm2=int6464#2,>xmm0=int6464#7
10862# asm 2: movdqa <xmm2=%xmm1,>xmm0=%xmm6
10863movdqa %xmm1,%xmm6
10864
10865# qhasm: xmm7 = xmm3
10866# asm 1: movdqa <xmm3=int6464#1,>xmm7=int6464#8
10867# asm 2: movdqa <xmm3=%xmm0,>xmm7=%xmm7
10868movdqa %xmm0,%xmm7
10869
10870# qhasm: xmm2 |= xmm1
10871# asm 1: por <xmm1=int6464#3,<xmm2=int6464#2
10872# asm 2: por <xmm1=%xmm2,<xmm2=%xmm1
10873por %xmm2,%xmm1
10874
10875# qhasm: xmm3 |= xmm4
10876# asm 1: por <xmm4=int6464#5,<xmm3=int6464#1
10877# asm 2: por <xmm4=%xmm4,<xmm3=%xmm0
10878por %xmm4,%xmm0
10879
10880# qhasm: xmm7 ^= xmm0
10881# asm 1: pxor <xmm0=int6464#7,<xmm7=int6464#8
10882# asm 2: pxor <xmm0=%xmm6,<xmm7=%xmm7
10883pxor %xmm6,%xmm7
10884
10885# qhasm: xmm6 &= xmm4
10886# asm 1: pand <xmm4=int6464#5,<xmm6=int6464#6
10887# asm 2: pand <xmm4=%xmm4,<xmm6=%xmm5
10888pand %xmm4,%xmm5
10889
10890# qhasm: xmm0 &= xmm1
10891# asm 1: pand <xmm1=int6464#3,<xmm0=int6464#7
10892# asm 2: pand <xmm1=%xmm2,<xmm0=%xmm6
10893pand %xmm2,%xmm6
10894
10895# qhasm: xmm4 ^= xmm1
10896# asm 1: pxor <xmm1=int6464#3,<xmm4=int6464#5
10897# asm 2: pxor <xmm1=%xmm2,<xmm4=%xmm4
10898pxor %xmm2,%xmm4
10899
10900# qhasm: xmm7 &= xmm4
10901# asm 1: pand <xmm4=int6464#5,<xmm7=int6464#8
10902# asm 2: pand <xmm4=%xmm4,<xmm7=%xmm7
10903pand %xmm4,%xmm7
10904
10905# qhasm: xmm4 = xmm11
10906# asm 1: movdqa <xmm11=int6464#12,>xmm4=int6464#3
10907# asm 2: movdqa <xmm11=%xmm11,>xmm4=%xmm2
10908movdqa %xmm11,%xmm2
10909
10910# qhasm: xmm4 ^= xmm8
10911# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#3
10912# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm2
10913pxor %xmm8,%xmm2
10914
10915# qhasm: xmm5 &= xmm4
10916# asm 1: pand <xmm4=int6464#3,<xmm5=int6464#4
10917# asm 2: pand <xmm4=%xmm2,<xmm5=%xmm3
10918pand %xmm2,%xmm3
10919
10920# qhasm: xmm3 ^= xmm5
10921# asm 1: pxor <xmm5=int6464#4,<xmm3=int6464#1
10922# asm 2: pxor <xmm5=%xmm3,<xmm3=%xmm0
10923pxor %xmm3,%xmm0
10924
10925# qhasm: xmm2 ^= xmm5
10926# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
10927# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
10928pxor %xmm3,%xmm1
10929
10930# qhasm: xmm5 = xmm15
10931# asm 1: movdqa <xmm15=int6464#16,>xmm5=int6464#3
10932# asm 2: movdqa <xmm15=%xmm15,>xmm5=%xmm2
10933movdqa %xmm15,%xmm2
10934
10935# qhasm: xmm5 ^= xmm9
10936# asm 1: pxor <xmm9=int6464#10,<xmm5=int6464#3
10937# asm 2: pxor <xmm9=%xmm9,<xmm5=%xmm2
10938pxor %xmm9,%xmm2
10939
10940# qhasm: xmm4 = xmm13
10941# asm 1: movdqa <xmm13=int6464#14,>xmm4=int6464#4
10942# asm 2: movdqa <xmm13=%xmm13,>xmm4=%xmm3
10943movdqa %xmm13,%xmm3
10944
10945# qhasm: xmm1 = xmm5
10946# asm 1: movdqa <xmm5=int6464#3,>xmm1=int6464#5
10947# asm 2: movdqa <xmm5=%xmm2,>xmm1=%xmm4
10948movdqa %xmm2,%xmm4
10949
10950# qhasm: xmm4 ^= xmm14
10951# asm 1: pxor <xmm14=int6464#15,<xmm4=int6464#4
10952# asm 2: pxor <xmm14=%xmm14,<xmm4=%xmm3
10953pxor %xmm14,%xmm3
10954
10955# qhasm: xmm1 |= xmm4
10956# asm 1: por <xmm4=int6464#4,<xmm1=int6464#5
10957# asm 2: por <xmm4=%xmm3,<xmm1=%xmm4
10958por %xmm3,%xmm4
10959
10960# qhasm: xmm5 &= xmm4
10961# asm 1: pand <xmm4=int6464#4,<xmm5=int6464#3
10962# asm 2: pand <xmm4=%xmm3,<xmm5=%xmm2
10963pand %xmm3,%xmm2
10964
10965# qhasm: xmm0 ^= xmm5
10966# asm 1: pxor <xmm5=int6464#3,<xmm0=int6464#7
10967# asm 2: pxor <xmm5=%xmm2,<xmm0=%xmm6
10968pxor %xmm2,%xmm6
10969
10970# qhasm: xmm3 ^= xmm7
10971# asm 1: pxor <xmm7=int6464#8,<xmm3=int6464#1
10972# asm 2: pxor <xmm7=%xmm7,<xmm3=%xmm0
10973pxor %xmm7,%xmm0
10974
10975# qhasm: xmm2 ^= xmm6
10976# asm 1: pxor <xmm6=int6464#6,<xmm2=int6464#2
10977# asm 2: pxor <xmm6=%xmm5,<xmm2=%xmm1
10978pxor %xmm5,%xmm1
10979
10980# qhasm: xmm1 ^= xmm7
10981# asm 1: pxor <xmm7=int6464#8,<xmm1=int6464#5
10982# asm 2: pxor <xmm7=%xmm7,<xmm1=%xmm4
10983pxor %xmm7,%xmm4
10984
10985# qhasm: xmm0 ^= xmm6
10986# asm 1: pxor <xmm6=int6464#6,<xmm0=int6464#7
10987# asm 2: pxor <xmm6=%xmm5,<xmm0=%xmm6
10988pxor %xmm5,%xmm6
10989
10990# qhasm: xmm1 ^= xmm6
10991# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
10992# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
10993pxor %xmm5,%xmm4
10994
10995# qhasm: xmm4 = xmm10
10996# asm 1: movdqa <xmm10=int6464#11,>xmm4=int6464#3
10997# asm 2: movdqa <xmm10=%xmm10,>xmm4=%xmm2
10998movdqa %xmm10,%xmm2
10999
11000# qhasm: xmm5 = xmm12
11001# asm 1: movdqa <xmm12=int6464#13,>xmm5=int6464#4
11002# asm 2: movdqa <xmm12=%xmm12,>xmm5=%xmm3
11003movdqa %xmm12,%xmm3
11004
11005# qhasm: xmm6 = xmm9
11006# asm 1: movdqa <xmm9=int6464#10,>xmm6=int6464#6
11007# asm 2: movdqa <xmm9=%xmm9,>xmm6=%xmm5
11008movdqa %xmm9,%xmm5
11009
11010# qhasm: xmm7 = xmm15
11011# asm 1: movdqa <xmm15=int6464#16,>xmm7=int6464#8
11012# asm 2: movdqa <xmm15=%xmm15,>xmm7=%xmm7
11013movdqa %xmm15,%xmm7
11014
11015# qhasm: xmm4 &= xmm11
11016# asm 1: pand <xmm11=int6464#12,<xmm4=int6464#3
11017# asm 2: pand <xmm11=%xmm11,<xmm4=%xmm2
11018pand %xmm11,%xmm2
11019
11020# qhasm: xmm5 &= xmm8
11021# asm 1: pand <xmm8=int6464#9,<xmm5=int6464#4
11022# asm 2: pand <xmm8=%xmm8,<xmm5=%xmm3
11023pand %xmm8,%xmm3
11024
11025# qhasm: xmm6 &= xmm13
11026# asm 1: pand <xmm13=int6464#14,<xmm6=int6464#6
11027# asm 2: pand <xmm13=%xmm13,<xmm6=%xmm5
11028pand %xmm13,%xmm5
11029
11030# qhasm: xmm7 |= xmm14
11031# asm 1: por <xmm14=int6464#15,<xmm7=int6464#8
11032# asm 2: por <xmm14=%xmm14,<xmm7=%xmm7
11033por %xmm14,%xmm7
11034
11035# qhasm: xmm3 ^= xmm4
11036# asm 1: pxor <xmm4=int6464#3,<xmm3=int6464#1
11037# asm 2: pxor <xmm4=%xmm2,<xmm3=%xmm0
11038pxor %xmm2,%xmm0
11039
11040# qhasm: xmm2 ^= xmm5
11041# asm 1: pxor <xmm5=int6464#4,<xmm2=int6464#2
11042# asm 2: pxor <xmm5=%xmm3,<xmm2=%xmm1
11043pxor %xmm3,%xmm1
11044
11045# qhasm: xmm1 ^= xmm6
11046# asm 1: pxor <xmm6=int6464#6,<xmm1=int6464#5
11047# asm 2: pxor <xmm6=%xmm5,<xmm1=%xmm4
11048pxor %xmm5,%xmm4
11049
11050# qhasm: xmm0 ^= xmm7
11051# asm 1: pxor <xmm7=int6464#8,<xmm0=int6464#7
11052# asm 2: pxor <xmm7=%xmm7,<xmm0=%xmm6
11053pxor %xmm7,%xmm6
11054
11055# qhasm: xmm4 = xmm3
11056# asm 1: movdqa <xmm3=int6464#1,>xmm4=int6464#3
11057# asm 2: movdqa <xmm3=%xmm0,>xmm4=%xmm2
11058movdqa %xmm0,%xmm2
11059
11060# qhasm: xmm4 ^= xmm2
11061# asm 1: pxor <xmm2=int6464#2,<xmm4=int6464#3
11062# asm 2: pxor <xmm2=%xmm1,<xmm4=%xmm2
11063pxor %xmm1,%xmm2
11064
11065# qhasm: xmm3 &= xmm1
11066# asm 1: pand <xmm1=int6464#5,<xmm3=int6464#1
11067# asm 2: pand <xmm1=%xmm4,<xmm3=%xmm0
11068pand %xmm4,%xmm0
11069
11070# qhasm: xmm6 = xmm0
11071# asm 1: movdqa <xmm0=int6464#7,>xmm6=int6464#4
11072# asm 2: movdqa <xmm0=%xmm6,>xmm6=%xmm3
11073movdqa %xmm6,%xmm3
11074
11075# qhasm: xmm6 ^= xmm3
11076# asm 1: pxor <xmm3=int6464#1,<xmm6=int6464#4
11077# asm 2: pxor <xmm3=%xmm0,<xmm6=%xmm3
11078pxor %xmm0,%xmm3
11079
11080# qhasm: xmm7 = xmm4
11081# asm 1: movdqa <xmm4=int6464#3,>xmm7=int6464#6
11082# asm 2: movdqa <xmm4=%xmm2,>xmm7=%xmm5
11083movdqa %xmm2,%xmm5
11084
11085# qhasm: xmm7 &= xmm6
11086# asm 1: pand <xmm6=int6464#4,<xmm7=int6464#6
11087# asm 2: pand <xmm6=%xmm3,<xmm7=%xmm5
11088pand %xmm3,%xmm5
11089
11090# qhasm: xmm7 ^= xmm2
11091# asm 1: pxor <xmm2=int6464#2,<xmm7=int6464#6
11092# asm 2: pxor <xmm2=%xmm1,<xmm7=%xmm5
11093pxor %xmm1,%xmm5
11094
11095# qhasm: xmm5 = xmm1
11096# asm 1: movdqa <xmm1=int6464#5,>xmm5=int6464#8
11097# asm 2: movdqa <xmm1=%xmm4,>xmm5=%xmm7
11098movdqa %xmm4,%xmm7
11099
11100# qhasm: xmm5 ^= xmm0
11101# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
11102# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
11103pxor %xmm6,%xmm7
11104
11105# qhasm: xmm3 ^= xmm2
11106# asm 1: pxor <xmm2=int6464#2,<xmm3=int6464#1
11107# asm 2: pxor <xmm2=%xmm1,<xmm3=%xmm0
11108pxor %xmm1,%xmm0
11109
11110# qhasm: xmm5 &= xmm3
11111# asm 1: pand <xmm3=int6464#1,<xmm5=int6464#8
11112# asm 2: pand <xmm3=%xmm0,<xmm5=%xmm7
11113pand %xmm0,%xmm7
11114
11115# qhasm: xmm5 ^= xmm0
11116# asm 1: pxor <xmm0=int6464#7,<xmm5=int6464#8
11117# asm 2: pxor <xmm0=%xmm6,<xmm5=%xmm7
11118pxor %xmm6,%xmm7
11119
11120# qhasm: xmm1 ^= xmm5
11121# asm 1: pxor <xmm5=int6464#8,<xmm1=int6464#5
11122# asm 2: pxor <xmm5=%xmm7,<xmm1=%xmm4
11123pxor %xmm7,%xmm4
11124
11125# qhasm: xmm2 = xmm6
11126# asm 1: movdqa <xmm6=int6464#4,>xmm2=int6464#1
11127# asm 2: movdqa <xmm6=%xmm3,>xmm2=%xmm0
11128movdqa %xmm3,%xmm0
11129
11130# qhasm: xmm2 ^= xmm5
11131# asm 1: pxor <xmm5=int6464#8,<xmm2=int6464#1
11132# asm 2: pxor <xmm5=%xmm7,<xmm2=%xmm0
11133pxor %xmm7,%xmm0
11134
11135# qhasm: xmm2 &= xmm0
11136# asm 1: pand <xmm0=int6464#7,<xmm2=int6464#1
11137# asm 2: pand <xmm0=%xmm6,<xmm2=%xmm0
11138pand %xmm6,%xmm0
11139
11140# qhasm: xmm1 ^= xmm2
11141# asm 1: pxor <xmm2=int6464#1,<xmm1=int6464#5
11142# asm 2: pxor <xmm2=%xmm0,<xmm1=%xmm4
11143pxor %xmm0,%xmm4
11144
11145# qhasm: xmm6 ^= xmm2
11146# asm 1: pxor <xmm2=int6464#1,<xmm6=int6464#4
11147# asm 2: pxor <xmm2=%xmm0,<xmm6=%xmm3
11148pxor %xmm0,%xmm3
11149
11150# qhasm: xmm6 &= xmm7
11151# asm 1: pand <xmm7=int6464#6,<xmm6=int6464#4
11152# asm 2: pand <xmm7=%xmm5,<xmm6=%xmm3
11153pand %xmm5,%xmm3
11154
11155# qhasm: xmm6 ^= xmm4
11156# asm 1: pxor <xmm4=int6464#3,<xmm6=int6464#4
11157# asm 2: pxor <xmm4=%xmm2,<xmm6=%xmm3
11158pxor %xmm2,%xmm3
11159
11160# qhasm: xmm4 = xmm14
11161# asm 1: movdqa <xmm14=int6464#15,>xmm4=int6464#1
11162# asm 2: movdqa <xmm14=%xmm14,>xmm4=%xmm0
11163movdqa %xmm14,%xmm0
11164
11165# qhasm: xmm0 = xmm13
11166# asm 1: movdqa <xmm13=int6464#14,>xmm0=int6464#2
11167# asm 2: movdqa <xmm13=%xmm13,>xmm0=%xmm1
11168movdqa %xmm13,%xmm1
11169
11170# qhasm: xmm2 = xmm7
11171# asm 1: movdqa <xmm7=int6464#6,>xmm2=int6464#3
11172# asm 2: movdqa <xmm7=%xmm5,>xmm2=%xmm2
11173movdqa %xmm5,%xmm2
11174
11175# qhasm: xmm2 ^= xmm6
11176# asm 1: pxor <xmm6=int6464#4,<xmm2=int6464#3
11177# asm 2: pxor <xmm6=%xmm3,<xmm2=%xmm2
11178pxor %xmm3,%xmm2
11179
11180# qhasm: xmm2 &= xmm14
11181# asm 1: pand <xmm14=int6464#15,<xmm2=int6464#3
11182# asm 2: pand <xmm14=%xmm14,<xmm2=%xmm2
11183pand %xmm14,%xmm2
11184
11185# qhasm: xmm14 ^= xmm13
11186# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
11187# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
11188pxor %xmm13,%xmm14
11189
11190# qhasm: xmm14 &= xmm6
11191# asm 1: pand <xmm6=int6464#4,<xmm14=int6464#15
11192# asm 2: pand <xmm6=%xmm3,<xmm14=%xmm14
11193pand %xmm3,%xmm14
11194
11195# qhasm: xmm13 &= xmm7
11196# asm 1: pand <xmm7=int6464#6,<xmm13=int6464#14
11197# asm 2: pand <xmm7=%xmm5,<xmm13=%xmm13
11198pand %xmm5,%xmm13
11199
11200# qhasm: xmm14 ^= xmm13
11201# asm 1: pxor <xmm13=int6464#14,<xmm14=int6464#15
11202# asm 2: pxor <xmm13=%xmm13,<xmm14=%xmm14
11203pxor %xmm13,%xmm14
11204
11205# qhasm: xmm13 ^= xmm2
11206# asm 1: pxor <xmm2=int6464#3,<xmm13=int6464#14
11207# asm 2: pxor <xmm2=%xmm2,<xmm13=%xmm13
11208pxor %xmm2,%xmm13
11209
11210# qhasm: xmm4 ^= xmm8
11211# asm 1: pxor <xmm8=int6464#9,<xmm4=int6464#1
11212# asm 2: pxor <xmm8=%xmm8,<xmm4=%xmm0
11213pxor %xmm8,%xmm0
11214
11215# qhasm: xmm0 ^= xmm11
11216# asm 1: pxor <xmm11=int6464#12,<xmm0=int6464#2
11217# asm 2: pxor <xmm11=%xmm11,<xmm0=%xmm1
11218pxor %xmm11,%xmm1
11219
11220# qhasm: xmm7 ^= xmm5
11221# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
11222# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
11223pxor %xmm7,%xmm5
11224
11225# qhasm: xmm6 ^= xmm1
11226# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
11227# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
11228pxor %xmm4,%xmm3
11229
11230# qhasm: xmm3 = xmm7
11231# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
11232# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
11233movdqa %xmm5,%xmm2
11234
11235# qhasm: xmm3 ^= xmm6
11236# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
11237# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
11238pxor %xmm3,%xmm2
11239
11240# qhasm: xmm3 &= xmm4
11241# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
11242# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
11243pand %xmm0,%xmm2
11244
11245# qhasm: xmm4 ^= xmm0
11246# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
11247# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
11248pxor %xmm1,%xmm0
11249
11250# qhasm: xmm4 &= xmm6
11251# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
11252# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
11253pand %xmm3,%xmm0
11254
11255# qhasm: xmm0 &= xmm7
11256# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
11257# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
11258pand %xmm5,%xmm1
11259
11260# qhasm: xmm0 ^= xmm4
11261# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
11262# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
11263pxor %xmm0,%xmm1
11264
11265# qhasm: xmm4 ^= xmm3
11266# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
11267# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
11268pxor %xmm2,%xmm0
11269
11270# qhasm: xmm2 = xmm5
11271# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
11272# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
11273movdqa %xmm7,%xmm2
11274
11275# qhasm: xmm2 ^= xmm1
11276# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
11277# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
11278pxor %xmm4,%xmm2
11279
11280# qhasm: xmm2 &= xmm8
11281# asm 1: pand <xmm8=int6464#9,<xmm2=int6464#3
11282# asm 2: pand <xmm8=%xmm8,<xmm2=%xmm2
11283pand %xmm8,%xmm2
11284
11285# qhasm: xmm8 ^= xmm11
11286# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
11287# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
11288pxor %xmm11,%xmm8
11289
11290# qhasm: xmm8 &= xmm1
11291# asm 1: pand <xmm1=int6464#5,<xmm8=int6464#9
11292# asm 2: pand <xmm1=%xmm4,<xmm8=%xmm8
11293pand %xmm4,%xmm8
11294
11295# qhasm: xmm11 &= xmm5
11296# asm 1: pand <xmm5=int6464#8,<xmm11=int6464#12
11297# asm 2: pand <xmm5=%xmm7,<xmm11=%xmm11
11298pand %xmm7,%xmm11
11299
11300# qhasm: xmm8 ^= xmm11
11301# asm 1: pxor <xmm11=int6464#12,<xmm8=int6464#9
11302# asm 2: pxor <xmm11=%xmm11,<xmm8=%xmm8
11303pxor %xmm11,%xmm8
11304
11305# qhasm: xmm11 ^= xmm2
11306# asm 1: pxor <xmm2=int6464#3,<xmm11=int6464#12
11307# asm 2: pxor <xmm2=%xmm2,<xmm11=%xmm11
11308pxor %xmm2,%xmm11
11309
11310# qhasm: xmm14 ^= xmm4
11311# asm 1: pxor <xmm4=int6464#1,<xmm14=int6464#15
11312# asm 2: pxor <xmm4=%xmm0,<xmm14=%xmm14
11313pxor %xmm0,%xmm14
11314
11315# qhasm: xmm8 ^= xmm4
11316# asm 1: pxor <xmm4=int6464#1,<xmm8=int6464#9
11317# asm 2: pxor <xmm4=%xmm0,<xmm8=%xmm8
11318pxor %xmm0,%xmm8
11319
11320# qhasm: xmm13 ^= xmm0
11321# asm 1: pxor <xmm0=int6464#2,<xmm13=int6464#14
11322# asm 2: pxor <xmm0=%xmm1,<xmm13=%xmm13
11323pxor %xmm1,%xmm13
11324
11325# qhasm: xmm11 ^= xmm0
11326# asm 1: pxor <xmm0=int6464#2,<xmm11=int6464#12
11327# asm 2: pxor <xmm0=%xmm1,<xmm11=%xmm11
11328pxor %xmm1,%xmm11
11329
11330# qhasm: xmm4 = xmm15
11331# asm 1: movdqa <xmm15=int6464#16,>xmm4=int6464#1
11332# asm 2: movdqa <xmm15=%xmm15,>xmm4=%xmm0
11333movdqa %xmm15,%xmm0
11334
11335# qhasm: xmm0 = xmm9
11336# asm 1: movdqa <xmm9=int6464#10,>xmm0=int6464#2
11337# asm 2: movdqa <xmm9=%xmm9,>xmm0=%xmm1
11338movdqa %xmm9,%xmm1
11339
11340# qhasm: xmm4 ^= xmm12
11341# asm 1: pxor <xmm12=int6464#13,<xmm4=int6464#1
11342# asm 2: pxor <xmm12=%xmm12,<xmm4=%xmm0
11343pxor %xmm12,%xmm0
11344
11345# qhasm: xmm0 ^= xmm10
11346# asm 1: pxor <xmm10=int6464#11,<xmm0=int6464#2
11347# asm 2: pxor <xmm10=%xmm10,<xmm0=%xmm1
11348pxor %xmm10,%xmm1
11349
11350# qhasm: xmm3 = xmm7
11351# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
11352# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
11353movdqa %xmm5,%xmm2
11354
11355# qhasm: xmm3 ^= xmm6
11356# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
11357# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
11358pxor %xmm3,%xmm2
11359
11360# qhasm: xmm3 &= xmm4
11361# asm 1: pand <xmm4=int6464#1,<xmm3=int6464#3
11362# asm 2: pand <xmm4=%xmm0,<xmm3=%xmm2
11363pand %xmm0,%xmm2
11364
11365# qhasm: xmm4 ^= xmm0
11366# asm 1: pxor <xmm0=int6464#2,<xmm4=int6464#1
11367# asm 2: pxor <xmm0=%xmm1,<xmm4=%xmm0
11368pxor %xmm1,%xmm0
11369
11370# qhasm: xmm4 &= xmm6
11371# asm 1: pand <xmm6=int6464#4,<xmm4=int6464#1
11372# asm 2: pand <xmm6=%xmm3,<xmm4=%xmm0
11373pand %xmm3,%xmm0
11374
11375# qhasm: xmm0 &= xmm7
11376# asm 1: pand <xmm7=int6464#6,<xmm0=int6464#2
11377# asm 2: pand <xmm7=%xmm5,<xmm0=%xmm1
11378pand %xmm5,%xmm1
11379
11380# qhasm: xmm0 ^= xmm4
11381# asm 1: pxor <xmm4=int6464#1,<xmm0=int6464#2
11382# asm 2: pxor <xmm4=%xmm0,<xmm0=%xmm1
11383pxor %xmm0,%xmm1
11384
11385# qhasm: xmm4 ^= xmm3
11386# asm 1: pxor <xmm3=int6464#3,<xmm4=int6464#1
11387# asm 2: pxor <xmm3=%xmm2,<xmm4=%xmm0
11388pxor %xmm2,%xmm0
11389
11390# qhasm: xmm2 = xmm5
11391# asm 1: movdqa <xmm5=int6464#8,>xmm2=int6464#3
11392# asm 2: movdqa <xmm5=%xmm7,>xmm2=%xmm2
11393movdqa %xmm7,%xmm2
11394
11395# qhasm: xmm2 ^= xmm1
11396# asm 1: pxor <xmm1=int6464#5,<xmm2=int6464#3
11397# asm 2: pxor <xmm1=%xmm4,<xmm2=%xmm2
11398pxor %xmm4,%xmm2
11399
11400# qhasm: xmm2 &= xmm12
11401# asm 1: pand <xmm12=int6464#13,<xmm2=int6464#3
11402# asm 2: pand <xmm12=%xmm12,<xmm2=%xmm2
11403pand %xmm12,%xmm2
11404
11405# qhasm: xmm12 ^= xmm10
11406# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
11407# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
11408pxor %xmm10,%xmm12
11409
11410# qhasm: xmm12 &= xmm1
11411# asm 1: pand <xmm1=int6464#5,<xmm12=int6464#13
11412# asm 2: pand <xmm1=%xmm4,<xmm12=%xmm12
11413pand %xmm4,%xmm12
11414
11415# qhasm: xmm10 &= xmm5
11416# asm 1: pand <xmm5=int6464#8,<xmm10=int6464#11
11417# asm 2: pand <xmm5=%xmm7,<xmm10=%xmm10
11418pand %xmm7,%xmm10
11419
11420# qhasm: xmm12 ^= xmm10
11421# asm 1: pxor <xmm10=int6464#11,<xmm12=int6464#13
11422# asm 2: pxor <xmm10=%xmm10,<xmm12=%xmm12
11423pxor %xmm10,%xmm12
11424
11425# qhasm: xmm10 ^= xmm2
11426# asm 1: pxor <xmm2=int6464#3,<xmm10=int6464#11
11427# asm 2: pxor <xmm2=%xmm2,<xmm10=%xmm10
11428pxor %xmm2,%xmm10
11429
11430# qhasm: xmm7 ^= xmm5
11431# asm 1: pxor <xmm5=int6464#8,<xmm7=int6464#6
11432# asm 2: pxor <xmm5=%xmm7,<xmm7=%xmm5
11433pxor %xmm7,%xmm5
11434
11435# qhasm: xmm6 ^= xmm1
11436# asm 1: pxor <xmm1=int6464#5,<xmm6=int6464#4
11437# asm 2: pxor <xmm1=%xmm4,<xmm6=%xmm3
11438pxor %xmm4,%xmm3
11439
11440# qhasm: xmm3 = xmm7
11441# asm 1: movdqa <xmm7=int6464#6,>xmm3=int6464#3
11442# asm 2: movdqa <xmm7=%xmm5,>xmm3=%xmm2
11443movdqa %xmm5,%xmm2
11444
11445# qhasm: xmm3 ^= xmm6
11446# asm 1: pxor <xmm6=int6464#4,<xmm3=int6464#3
11447# asm 2: pxor <xmm6=%xmm3,<xmm3=%xmm2
11448pxor %xmm3,%xmm2
11449
11450# qhasm: xmm3 &= xmm15
11451# asm 1: pand <xmm15=int6464#16,<xmm3=int6464#3
11452# asm 2: pand <xmm15=%xmm15,<xmm3=%xmm2
11453pand %xmm15,%xmm2
11454
11455# qhasm: xmm15 ^= xmm9
11456# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
11457# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
11458pxor %xmm9,%xmm15
11459
11460# qhasm: xmm15 &= xmm6
11461# asm 1: pand <xmm6=int6464#4,<xmm15=int6464#16
11462# asm 2: pand <xmm6=%xmm3,<xmm15=%xmm15
11463pand %xmm3,%xmm15
11464
11465# qhasm: xmm9 &= xmm7
11466# asm 1: pand <xmm7=int6464#6,<xmm9=int6464#10
11467# asm 2: pand <xmm7=%xmm5,<xmm9=%xmm9
11468pand %xmm5,%xmm9
11469
11470# qhasm: xmm15 ^= xmm9
11471# asm 1: pxor <xmm9=int6464#10,<xmm15=int6464#16
11472# asm 2: pxor <xmm9=%xmm9,<xmm15=%xmm15
11473pxor %xmm9,%xmm15
11474
11475# qhasm: xmm9 ^= xmm3
11476# asm 1: pxor <xmm3=int6464#3,<xmm9=int6464#10
11477# asm 2: pxor <xmm3=%xmm2,<xmm9=%xmm9
11478pxor %xmm2,%xmm9
11479
11480# qhasm: xmm15 ^= xmm4
11481# asm 1: pxor <xmm4=int6464#1,<xmm15=int6464#16
11482# asm 2: pxor <xmm4=%xmm0,<xmm15=%xmm15
11483pxor %xmm0,%xmm15
11484
11485# qhasm: xmm12 ^= xmm4
11486# asm 1: pxor <xmm4=int6464#1,<xmm12=int6464#13
11487# asm 2: pxor <xmm4=%xmm0,<xmm12=%xmm12
11488pxor %xmm0,%xmm12
11489
11490# qhasm: xmm9 ^= xmm0
11491# asm 1: pxor <xmm0=int6464#2,<xmm9=int6464#10
11492# asm 2: pxor <xmm0=%xmm1,<xmm9=%xmm9
11493pxor %xmm1,%xmm9
11494
11495# qhasm: xmm10 ^= xmm0
11496# asm 1: pxor <xmm0=int6464#2,<xmm10=int6464#11
11497# asm 2: pxor <xmm0=%xmm1,<xmm10=%xmm10
11498pxor %xmm1,%xmm10
11499
11500# qhasm: xmm15 ^= xmm8
11501# asm 1: pxor <xmm8=int6464#9,<xmm15=int6464#16
11502# asm 2: pxor <xmm8=%xmm8,<xmm15=%xmm15
11503pxor %xmm8,%xmm15
11504
11505# qhasm: xmm9 ^= xmm14
11506# asm 1: pxor <xmm14=int6464#15,<xmm9=int6464#10
11507# asm 2: pxor <xmm14=%xmm14,<xmm9=%xmm9
11508pxor %xmm14,%xmm9
11509
11510# qhasm: xmm12 ^= xmm15
11511# asm 1: pxor <xmm15=int6464#16,<xmm12=int6464#13
11512# asm 2: pxor <xmm15=%xmm15,<xmm12=%xmm12
11513pxor %xmm15,%xmm12
11514
11515# qhasm: xmm14 ^= xmm8
11516# asm 1: pxor <xmm8=int6464#9,<xmm14=int6464#15
11517# asm 2: pxor <xmm8=%xmm8,<xmm14=%xmm14
11518pxor %xmm8,%xmm14
11519
11520# qhasm: xmm8 ^= xmm9
11521# asm 1: pxor <xmm9=int6464#10,<xmm8=int6464#9
11522# asm 2: pxor <xmm9=%xmm9,<xmm8=%xmm8
11523pxor %xmm9,%xmm8
11524
11525# qhasm: xmm9 ^= xmm13
11526# asm 1: pxor <xmm13=int6464#14,<xmm9=int6464#10
11527# asm 2: pxor <xmm13=%xmm13,<xmm9=%xmm9
11528pxor %xmm13,%xmm9
11529
11530# qhasm: xmm13 ^= xmm10
11531# asm 1: pxor <xmm10=int6464#11,<xmm13=int6464#14
11532# asm 2: pxor <xmm10=%xmm10,<xmm13=%xmm13
11533pxor %xmm10,%xmm13
11534
11535# qhasm: xmm12 ^= xmm13
11536# asm 1: pxor <xmm13=int6464#14,<xmm12=int6464#13
11537# asm 2: pxor <xmm13=%xmm13,<xmm12=%xmm12
11538pxor %xmm13,%xmm12
11539
11540# qhasm: xmm10 ^= xmm11
11541# asm 1: pxor <xmm11=int6464#12,<xmm10=int6464#11
11542# asm 2: pxor <xmm11=%xmm11,<xmm10=%xmm10
11543pxor %xmm11,%xmm10
11544
11545# qhasm: xmm11 ^= xmm13
11546# asm 1: pxor <xmm13=int6464#14,<xmm11=int6464#12
11547# asm 2: pxor <xmm13=%xmm13,<xmm11=%xmm11
11548pxor %xmm13,%xmm11
11549
11550# qhasm: xmm14 ^= xmm11
11551# asm 1: pxor <xmm11=int6464#12,<xmm14=int6464#15
11552# asm 2: pxor <xmm11=%xmm11,<xmm14=%xmm14
11553pxor %xmm11,%xmm14
11554
11555# qhasm: xmm8 ^= *(int128 *)(c + 1280)
11556# asm 1: pxor 1280(<c=int64#4),<xmm8=int6464#9
11557# asm 2: pxor 1280(<c=%rcx),<xmm8=%xmm8
11558pxor 1280(%rcx),%xmm8
11559
11560# qhasm: xmm9 ^= *(int128 *)(c + 1296)
11561# asm 1: pxor 1296(<c=int64#4),<xmm9=int6464#10
11562# asm 2: pxor 1296(<c=%rcx),<xmm9=%xmm9
11563pxor 1296(%rcx),%xmm9
11564
11565# qhasm: xmm12 ^= *(int128 *)(c + 1312)
11566# asm 1: pxor 1312(<c=int64#4),<xmm12=int6464#13
11567# asm 2: pxor 1312(<c=%rcx),<xmm12=%xmm12
11568pxor 1312(%rcx),%xmm12
11569
11570# qhasm: xmm14 ^= *(int128 *)(c + 1328)
11571# asm 1: pxor 1328(<c=int64#4),<xmm14=int6464#15
11572# asm 2: pxor 1328(<c=%rcx),<xmm14=%xmm14
11573pxor 1328(%rcx),%xmm14
11574
11575# qhasm: xmm11 ^= *(int128 *)(c + 1344)
11576# asm 1: pxor 1344(<c=int64#4),<xmm11=int6464#12
11577# asm 2: pxor 1344(<c=%rcx),<xmm11=%xmm11
11578pxor 1344(%rcx),%xmm11
11579
11580# qhasm: xmm15 ^= *(int128 *)(c + 1360)
11581# asm 1: pxor 1360(<c=int64#4),<xmm15=int6464#16
11582# asm 2: pxor 1360(<c=%rcx),<xmm15=%xmm15
11583pxor 1360(%rcx),%xmm15
11584
11585# qhasm: xmm10 ^= *(int128 *)(c + 1376)
11586# asm 1: pxor 1376(<c=int64#4),<xmm10=int6464#11
11587# asm 2: pxor 1376(<c=%rcx),<xmm10=%xmm10
11588pxor 1376(%rcx),%xmm10
11589
11590# qhasm: xmm13 ^= *(int128 *)(c + 1392)
11591# asm 1: pxor 1392(<c=int64#4),<xmm13=int6464#14
11592# asm 2: pxor 1392(<c=%rcx),<xmm13=%xmm13
11593pxor 1392(%rcx),%xmm13
11594
11595# qhasm: xmm0 = xmm10
11596# asm 1: movdqa <xmm10=int6464#11,>xmm0=int6464#1
11597# asm 2: movdqa <xmm10=%xmm10,>xmm0=%xmm0
11598movdqa %xmm10,%xmm0
11599
11600# qhasm: uint6464 xmm0 >>= 1
11601# asm 1: psrlq $1,<xmm0=int6464#1
11602# asm 2: psrlq $1,<xmm0=%xmm0
11603psrlq $1,%xmm0
11604
11605# qhasm: xmm0 ^= xmm13
11606# asm 1: pxor <xmm13=int6464#14,<xmm0=int6464#1
11607# asm 2: pxor <xmm13=%xmm13,<xmm0=%xmm0
11608pxor %xmm13,%xmm0
11609
11610# qhasm: xmm0 &= BS0
11611# asm 1: pand BS0,<xmm0=int6464#1
11612# asm 2: pand BS0,<xmm0=%xmm0
11613pand BS0,%xmm0
11614
11615# qhasm: xmm13 ^= xmm0
11616# asm 1: pxor <xmm0=int6464#1,<xmm13=int6464#14
11617# asm 2: pxor <xmm0=%xmm0,<xmm13=%xmm13
11618pxor %xmm0,%xmm13
11619
11620# qhasm: uint6464 xmm0 <<= 1
11621# asm 1: psllq $1,<xmm0=int6464#1
11622# asm 2: psllq $1,<xmm0=%xmm0
11623psllq $1,%xmm0
11624
11625# qhasm: xmm10 ^= xmm0
11626# asm 1: pxor <xmm0=int6464#1,<xmm10=int6464#11
11627# asm 2: pxor <xmm0=%xmm0,<xmm10=%xmm10
11628pxor %xmm0,%xmm10
11629
11630# qhasm: xmm0 = xmm11
11631# asm 1: movdqa <xmm11=int6464#12,>xmm0=int6464#1
11632# asm 2: movdqa <xmm11=%xmm11,>xmm0=%xmm0
11633movdqa %xmm11,%xmm0
11634
11635# qhasm: uint6464 xmm0 >>= 1
11636# asm 1: psrlq $1,<xmm0=int6464#1
11637# asm 2: psrlq $1,<xmm0=%xmm0
11638psrlq $1,%xmm0
11639
11640# qhasm: xmm0 ^= xmm15
11641# asm 1: pxor <xmm15=int6464#16,<xmm0=int6464#1
11642# asm 2: pxor <xmm15=%xmm15,<xmm0=%xmm0
11643pxor %xmm15,%xmm0
11644
11645# qhasm: xmm0 &= BS0
11646# asm 1: pand BS0,<xmm0=int6464#1
11647# asm 2: pand BS0,<xmm0=%xmm0
11648pand BS0,%xmm0
11649
11650# qhasm: xmm15 ^= xmm0
11651# asm 1: pxor <xmm0=int6464#1,<xmm15=int6464#16
11652# asm 2: pxor <xmm0=%xmm0,<xmm15=%xmm15
11653pxor %xmm0,%xmm15
11654
11655# qhasm: uint6464 xmm0 <<= 1
11656# asm 1: psllq $1,<xmm0=int6464#1
11657# asm 2: psllq $1,<xmm0=%xmm0
11658psllq $1,%xmm0
11659
11660# qhasm: xmm11 ^= xmm0
11661# asm 1: pxor <xmm0=int6464#1,<xmm11=int6464#12
11662# asm 2: pxor <xmm0=%xmm0,<xmm11=%xmm11
11663pxor %xmm0,%xmm11
11664
11665# qhasm: xmm0 = xmm12
11666# asm 1: movdqa <xmm12=int6464#13,>xmm0=int6464#1
11667# asm 2: movdqa <xmm12=%xmm12,>xmm0=%xmm0
11668movdqa %xmm12,%xmm0
11669
11670# qhasm: uint6464 xmm0 >>= 1
11671# asm 1: psrlq $1,<xmm0=int6464#1
11672# asm 2: psrlq $1,<xmm0=%xmm0
11673psrlq $1,%xmm0
11674
11675# qhasm: xmm0 ^= xmm14
11676# asm 1: pxor <xmm14=int6464#15,<xmm0=int6464#1
11677# asm 2: pxor <xmm14=%xmm14,<xmm0=%xmm0
11678pxor %xmm14,%xmm0
11679
11680# qhasm: xmm0 &= BS0
11681# asm 1: pand BS0,<xmm0=int6464#1
11682# asm 2: pand BS0,<xmm0=%xmm0
11683pand BS0,%xmm0
11684
11685# qhasm: xmm14 ^= xmm0
11686# asm 1: pxor <xmm0=int6464#1,<xmm14=int6464#15
11687# asm 2: pxor <xmm0=%xmm0,<xmm14=%xmm14
11688pxor %xmm0,%xmm14
11689
11690# qhasm: uint6464 xmm0 <<= 1
11691# asm 1: psllq $1,<xmm0=int6464#1
11692# asm 2: psllq $1,<xmm0=%xmm0
11693psllq $1,%xmm0
11694
11695# qhasm: xmm12 ^= xmm0
11696# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
11697# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
11698pxor %xmm0,%xmm12
11699
11700# qhasm: xmm0 = xmm8
11701# asm 1: movdqa <xmm8=int6464#9,>xmm0=int6464#1
11702# asm 2: movdqa <xmm8=%xmm8,>xmm0=%xmm0
11703movdqa %xmm8,%xmm0
11704
11705# qhasm: uint6464 xmm0 >>= 1
11706# asm 1: psrlq $1,<xmm0=int6464#1
11707# asm 2: psrlq $1,<xmm0=%xmm0
11708psrlq $1,%xmm0
11709
11710# qhasm: xmm0 ^= xmm9
11711# asm 1: pxor <xmm9=int6464#10,<xmm0=int6464#1
11712# asm 2: pxor <xmm9=%xmm9,<xmm0=%xmm0
11713pxor %xmm9,%xmm0
11714
11715# qhasm: xmm0 &= BS0
11716# asm 1: pand BS0,<xmm0=int6464#1
11717# asm 2: pand BS0,<xmm0=%xmm0
11718pand BS0,%xmm0
11719
11720# qhasm: xmm9 ^= xmm0
11721# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
11722# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
11723pxor %xmm0,%xmm9
11724
11725# qhasm: uint6464 xmm0 <<= 1
11726# asm 1: psllq $1,<xmm0=int6464#1
11727# asm 2: psllq $1,<xmm0=%xmm0
11728psllq $1,%xmm0
11729
11730# qhasm: xmm8 ^= xmm0
11731# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
11732# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
11733pxor %xmm0,%xmm8
11734
11735# qhasm: xmm0 = xmm15
11736# asm 1: movdqa <xmm15=int6464#16,>xmm0=int6464#1
11737# asm 2: movdqa <xmm15=%xmm15,>xmm0=%xmm0
11738movdqa %xmm15,%xmm0
11739
11740# qhasm: uint6464 xmm0 >>= 2
11741# asm 1: psrlq $2,<xmm0=int6464#1
11742# asm 2: psrlq $2,<xmm0=%xmm0
11743psrlq $2,%xmm0
11744
11745# qhasm: xmm0 ^= xmm13
11746# asm 1: pxor <xmm13=int6464#14,<xmm0=int6464#1
11747# asm 2: pxor <xmm13=%xmm13,<xmm0=%xmm0
11748pxor %xmm13,%xmm0
11749
11750# qhasm: xmm0 &= BS1
11751# asm 1: pand BS1,<xmm0=int6464#1
11752# asm 2: pand BS1,<xmm0=%xmm0
11753pand BS1,%xmm0
11754
11755# qhasm: xmm13 ^= xmm0
11756# asm 1: pxor <xmm0=int6464#1,<xmm13=int6464#14
11757# asm 2: pxor <xmm0=%xmm0,<xmm13=%xmm13
11758pxor %xmm0,%xmm13
11759
11760# qhasm: uint6464 xmm0 <<= 2
11761# asm 1: psllq $2,<xmm0=int6464#1
11762# asm 2: psllq $2,<xmm0=%xmm0
11763psllq $2,%xmm0
11764
11765# qhasm: xmm15 ^= xmm0
11766# asm 1: pxor <xmm0=int6464#1,<xmm15=int6464#16
11767# asm 2: pxor <xmm0=%xmm0,<xmm15=%xmm15
11768pxor %xmm0,%xmm15
11769
11770# qhasm: xmm0 = xmm11
11771# asm 1: movdqa <xmm11=int6464#12,>xmm0=int6464#1
11772# asm 2: movdqa <xmm11=%xmm11,>xmm0=%xmm0
11773movdqa %xmm11,%xmm0
11774
11775# qhasm: uint6464 xmm0 >>= 2
11776# asm 1: psrlq $2,<xmm0=int6464#1
11777# asm 2: psrlq $2,<xmm0=%xmm0
11778psrlq $2,%xmm0
11779
11780# qhasm: xmm0 ^= xmm10
11781# asm 1: pxor <xmm10=int6464#11,<xmm0=int6464#1
11782# asm 2: pxor <xmm10=%xmm10,<xmm0=%xmm0
11783pxor %xmm10,%xmm0
11784
11785# qhasm: xmm0 &= BS1
11786# asm 1: pand BS1,<xmm0=int6464#1
11787# asm 2: pand BS1,<xmm0=%xmm0
11788pand BS1,%xmm0
11789
11790# qhasm: xmm10 ^= xmm0
11791# asm 1: pxor <xmm0=int6464#1,<xmm10=int6464#11
11792# asm 2: pxor <xmm0=%xmm0,<xmm10=%xmm10
11793pxor %xmm0,%xmm10
11794
11795# qhasm: uint6464 xmm0 <<= 2
11796# asm 1: psllq $2,<xmm0=int6464#1
11797# asm 2: psllq $2,<xmm0=%xmm0
11798psllq $2,%xmm0
11799
11800# qhasm: xmm11 ^= xmm0
11801# asm 1: pxor <xmm0=int6464#1,<xmm11=int6464#12
11802# asm 2: pxor <xmm0=%xmm0,<xmm11=%xmm11
11803pxor %xmm0,%xmm11
11804
11805# qhasm: xmm0 = xmm9
11806# asm 1: movdqa <xmm9=int6464#10,>xmm0=int6464#1
11807# asm 2: movdqa <xmm9=%xmm9,>xmm0=%xmm0
11808movdqa %xmm9,%xmm0
11809
11810# qhasm: uint6464 xmm0 >>= 2
11811# asm 1: psrlq $2,<xmm0=int6464#1
11812# asm 2: psrlq $2,<xmm0=%xmm0
11813psrlq $2,%xmm0
11814
11815# qhasm: xmm0 ^= xmm14
11816# asm 1: pxor <xmm14=int6464#15,<xmm0=int6464#1
11817# asm 2: pxor <xmm14=%xmm14,<xmm0=%xmm0
11818pxor %xmm14,%xmm0
11819
11820# qhasm: xmm0 &= BS1
11821# asm 1: pand BS1,<xmm0=int6464#1
11822# asm 2: pand BS1,<xmm0=%xmm0
11823pand BS1,%xmm0
11824
11825# qhasm: xmm14 ^= xmm0
11826# asm 1: pxor <xmm0=int6464#1,<xmm14=int6464#15
11827# asm 2: pxor <xmm0=%xmm0,<xmm14=%xmm14
11828pxor %xmm0,%xmm14
11829
11830# qhasm: uint6464 xmm0 <<= 2
11831# asm 1: psllq $2,<xmm0=int6464#1
11832# asm 2: psllq $2,<xmm0=%xmm0
11833psllq $2,%xmm0
11834
11835# qhasm: xmm9 ^= xmm0
11836# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
11837# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
11838pxor %xmm0,%xmm9
11839
11840# qhasm: xmm0 = xmm8
11841# asm 1: movdqa <xmm8=int6464#9,>xmm0=int6464#1
11842# asm 2: movdqa <xmm8=%xmm8,>xmm0=%xmm0
11843movdqa %xmm8,%xmm0
11844
11845# qhasm: uint6464 xmm0 >>= 2
11846# asm 1: psrlq $2,<xmm0=int6464#1
11847# asm 2: psrlq $2,<xmm0=%xmm0
11848psrlq $2,%xmm0
11849
11850# qhasm: xmm0 ^= xmm12
11851# asm 1: pxor <xmm12=int6464#13,<xmm0=int6464#1
11852# asm 2: pxor <xmm12=%xmm12,<xmm0=%xmm0
11853pxor %xmm12,%xmm0
11854
11855# qhasm: xmm0 &= BS1
11856# asm 1: pand BS1,<xmm0=int6464#1
11857# asm 2: pand BS1,<xmm0=%xmm0
11858pand BS1,%xmm0
11859
11860# qhasm: xmm12 ^= xmm0
11861# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
11862# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
11863pxor %xmm0,%xmm12
11864
11865# qhasm: uint6464 xmm0 <<= 2
11866# asm 1: psllq $2,<xmm0=int6464#1
11867# asm 2: psllq $2,<xmm0=%xmm0
11868psllq $2,%xmm0
11869
11870# qhasm: xmm8 ^= xmm0
11871# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
11872# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
11873pxor %xmm0,%xmm8
11874
11875# qhasm: xmm0 = xmm14
11876# asm 1: movdqa <xmm14=int6464#15,>xmm0=int6464#1
11877# asm 2: movdqa <xmm14=%xmm14,>xmm0=%xmm0
11878movdqa %xmm14,%xmm0
11879
11880# qhasm: uint6464 xmm0 >>= 4
11881# asm 1: psrlq $4,<xmm0=int6464#1
11882# asm 2: psrlq $4,<xmm0=%xmm0
11883psrlq $4,%xmm0
11884
11885# qhasm: xmm0 ^= xmm13
11886# asm 1: pxor <xmm13=int6464#14,<xmm0=int6464#1
11887# asm 2: pxor <xmm13=%xmm13,<xmm0=%xmm0
11888pxor %xmm13,%xmm0
11889
11890# qhasm: xmm0 &= BS2
11891# asm 1: pand BS2,<xmm0=int6464#1
11892# asm 2: pand BS2,<xmm0=%xmm0
11893pand BS2,%xmm0
11894
11895# qhasm: xmm13 ^= xmm0
11896# asm 1: pxor <xmm0=int6464#1,<xmm13=int6464#14
11897# asm 2: pxor <xmm0=%xmm0,<xmm13=%xmm13
11898pxor %xmm0,%xmm13
11899
11900# qhasm: uint6464 xmm0 <<= 4
11901# asm 1: psllq $4,<xmm0=int6464#1
11902# asm 2: psllq $4,<xmm0=%xmm0
11903psllq $4,%xmm0
11904
11905# qhasm: xmm14 ^= xmm0
11906# asm 1: pxor <xmm0=int6464#1,<xmm14=int6464#15
11907# asm 2: pxor <xmm0=%xmm0,<xmm14=%xmm14
11908pxor %xmm0,%xmm14
11909
11910# qhasm: xmm0 = xmm12
11911# asm 1: movdqa <xmm12=int6464#13,>xmm0=int6464#1
11912# asm 2: movdqa <xmm12=%xmm12,>xmm0=%xmm0
11913movdqa %xmm12,%xmm0
11914
11915# qhasm: uint6464 xmm0 >>= 4
11916# asm 1: psrlq $4,<xmm0=int6464#1
11917# asm 2: psrlq $4,<xmm0=%xmm0
11918psrlq $4,%xmm0
11919
11920# qhasm: xmm0 ^= xmm10
11921# asm 1: pxor <xmm10=int6464#11,<xmm0=int6464#1
11922# asm 2: pxor <xmm10=%xmm10,<xmm0=%xmm0
11923pxor %xmm10,%xmm0
11924
11925# qhasm: xmm0 &= BS2
11926# asm 1: pand BS2,<xmm0=int6464#1
11927# asm 2: pand BS2,<xmm0=%xmm0
11928pand BS2,%xmm0
11929
11930# qhasm: xmm10 ^= xmm0
11931# asm 1: pxor <xmm0=int6464#1,<xmm10=int6464#11
11932# asm 2: pxor <xmm0=%xmm0,<xmm10=%xmm10
11933pxor %xmm0,%xmm10
11934
11935# qhasm: uint6464 xmm0 <<= 4
11936# asm 1: psllq $4,<xmm0=int6464#1
11937# asm 2: psllq $4,<xmm0=%xmm0
11938psllq $4,%xmm0
11939
11940# qhasm: xmm12 ^= xmm0
11941# asm 1: pxor <xmm0=int6464#1,<xmm12=int6464#13
11942# asm 2: pxor <xmm0=%xmm0,<xmm12=%xmm12
11943pxor %xmm0,%xmm12
11944
11945# qhasm: xmm0 = xmm9
11946# asm 1: movdqa <xmm9=int6464#10,>xmm0=int6464#1
11947# asm 2: movdqa <xmm9=%xmm9,>xmm0=%xmm0
11948movdqa %xmm9,%xmm0
11949
11950# qhasm: uint6464 xmm0 >>= 4
11951# asm 1: psrlq $4,<xmm0=int6464#1
11952# asm 2: psrlq $4,<xmm0=%xmm0
11953psrlq $4,%xmm0
11954
11955# qhasm: xmm0 ^= xmm15
11956# asm 1: pxor <xmm15=int6464#16,<xmm0=int6464#1
11957# asm 2: pxor <xmm15=%xmm15,<xmm0=%xmm0
11958pxor %xmm15,%xmm0
11959
11960# qhasm: xmm0 &= BS2
11961# asm 1: pand BS2,<xmm0=int6464#1
11962# asm 2: pand BS2,<xmm0=%xmm0
11963pand BS2,%xmm0
11964
11965# qhasm: xmm15 ^= xmm0
11966# asm 1: pxor <xmm0=int6464#1,<xmm15=int6464#16
11967# asm 2: pxor <xmm0=%xmm0,<xmm15=%xmm15
11968pxor %xmm0,%xmm15
11969
11970# qhasm: uint6464 xmm0 <<= 4
11971# asm 1: psllq $4,<xmm0=int6464#1
11972# asm 2: psllq $4,<xmm0=%xmm0
11973psllq $4,%xmm0
11974
11975# qhasm: xmm9 ^= xmm0
11976# asm 1: pxor <xmm0=int6464#1,<xmm9=int6464#10
11977# asm 2: pxor <xmm0=%xmm0,<xmm9=%xmm9
11978pxor %xmm0,%xmm9
11979
11980# qhasm: xmm0 = xmm8
11981# asm 1: movdqa <xmm8=int6464#9,>xmm0=int6464#1
11982# asm 2: movdqa <xmm8=%xmm8,>xmm0=%xmm0
11983movdqa %xmm8,%xmm0
11984
11985# qhasm: uint6464 xmm0 >>= 4
11986# asm 1: psrlq $4,<xmm0=int6464#1
11987# asm 2: psrlq $4,<xmm0=%xmm0
11988psrlq $4,%xmm0
11989
11990# qhasm: xmm0 ^= xmm11
11991# asm 1: pxor <xmm11=int6464#12,<xmm0=int6464#1
11992# asm 2: pxor <xmm11=%xmm11,<xmm0=%xmm0
11993pxor %xmm11,%xmm0
11994
11995# qhasm: xmm0 &= BS2
11996# asm 1: pand BS2,<xmm0=int6464#1
11997# asm 2: pand BS2,<xmm0=%xmm0
11998pand BS2,%xmm0
11999
12000# qhasm: xmm11 ^= xmm0
12001# asm 1: pxor <xmm0=int6464#1,<xmm11=int6464#12
12002# asm 2: pxor <xmm0=%xmm0,<xmm11=%xmm11
12003pxor %xmm0,%xmm11
12004
12005# qhasm: uint6464 xmm0 <<= 4
12006# asm 1: psllq $4,<xmm0=int6464#1
12007# asm 2: psllq $4,<xmm0=%xmm0
12008psllq $4,%xmm0
12009
12010# qhasm: xmm8 ^= xmm0
12011# asm 1: pxor <xmm0=int6464#1,<xmm8=int6464#9
12012# asm 2: pxor <xmm0=%xmm0,<xmm8=%xmm8
12013pxor %xmm0,%xmm8
12014
12015# qhasm: unsigned<? =? len-128
12016# asm 1: cmp $128,<len=int64#2
12017# asm 2: cmp $128,<len=%rsi
12018cmp $128,%rsi
12019# comment:fp stack unchanged by jump
12020
12021# qhasm: goto partial if unsigned<
12022jb ._partial
12023# comment:fp stack unchanged by jump
12024
12025# qhasm: goto full if =
12026je ._full
12027
12028# qhasm: tmp = *(uint32 *)(np + 12)
12029# asm 1: movl 12(<np=int64#3),>tmp=int64#5d
12030# asm 2: movl 12(<np=%rdx),>tmp=%r8d
12031movl 12(%rdx),%r8d
12032
12033# qhasm: (uint32) bswap tmp
12034# asm 1: bswap <tmp=int64#5d
12035# asm 2: bswap <tmp=%r8d
12036bswap %r8d
12037
12038# qhasm: tmp += 8
12039# asm 1: add $8,<tmp=int64#5
12040# asm 2: add $8,<tmp=%r8
12041add $8,%r8
12042
12043# qhasm: (uint32) bswap tmp
12044# asm 1: bswap <tmp=int64#5d
12045# asm 2: bswap <tmp=%r8d
12046bswap %r8d
12047
12048# qhasm: *(uint32 *)(np + 12) = tmp
12049# asm 1: movl <tmp=int64#5d,12(<np=int64#3)
12050# asm 2: movl <tmp=%r8d,12(<np=%rdx)
12051movl %r8d,12(%rdx)
12052
12053# qhasm: *(int128 *) (outp + 0) = xmm8
12054# asm 1: movdqa <xmm8=int6464#9,0(<outp=int64#1)
12055# asm 2: movdqa <xmm8=%xmm8,0(<outp=%rdi)
12056movdqa %xmm8,0(%rdi)
12057
12058# qhasm: *(int128 *) (outp + 16) = xmm9
12059# asm 1: movdqa <xmm9=int6464#10,16(<outp=int64#1)
12060# asm 2: movdqa <xmm9=%xmm9,16(<outp=%rdi)
12061movdqa %xmm9,16(%rdi)
12062
12063# qhasm: *(int128 *) (outp + 32) = xmm12
12064# asm 1: movdqa <xmm12=int6464#13,32(<outp=int64#1)
12065# asm 2: movdqa <xmm12=%xmm12,32(<outp=%rdi)
12066movdqa %xmm12,32(%rdi)
12067
12068# qhasm: *(int128 *) (outp + 48) = xmm14
12069# asm 1: movdqa <xmm14=int6464#15,48(<outp=int64#1)
12070# asm 2: movdqa <xmm14=%xmm14,48(<outp=%rdi)
12071movdqa %xmm14,48(%rdi)
12072
12073# qhasm: *(int128 *) (outp + 64) = xmm11
12074# asm 1: movdqa <xmm11=int6464#12,64(<outp=int64#1)
12075# asm 2: movdqa <xmm11=%xmm11,64(<outp=%rdi)
12076movdqa %xmm11,64(%rdi)
12077
12078# qhasm: *(int128 *) (outp + 80) = xmm15
12079# asm 1: movdqa <xmm15=int6464#16,80(<outp=int64#1)
12080# asm 2: movdqa <xmm15=%xmm15,80(<outp=%rdi)
12081movdqa %xmm15,80(%rdi)
12082
12083# qhasm: *(int128 *) (outp + 96) = xmm10
12084# asm 1: movdqa <xmm10=int6464#11,96(<outp=int64#1)
12085# asm 2: movdqa <xmm10=%xmm10,96(<outp=%rdi)
12086movdqa %xmm10,96(%rdi)
12087
12088# qhasm: *(int128 *) (outp + 112) = xmm13
12089# asm 1: movdqa <xmm13=int6464#14,112(<outp=int64#1)
12090# asm 2: movdqa <xmm13=%xmm13,112(<outp=%rdi)
12091movdqa %xmm13,112(%rdi)
12092
12093# qhasm: len -= 128
12094# asm 1: sub $128,<len=int64#2
12095# asm 2: sub $128,<len=%rsi
12096sub $128,%rsi
12097
12098# qhasm: outp += 128
12099# asm 1: add $128,<outp=int64#1
12100# asm 2: add $128,<outp=%rdi
12101add $128,%rdi
12102# comment:fp stack unchanged by jump
12103
12104# qhasm: goto enc_block
12105jmp ._enc_block
12106
12107# qhasm: partial:
12108._partial:
12109
12110# qhasm: lensav = len
12111# asm 1: mov <len=int64#2,>lensav=int64#4
12112# asm 2: mov <len=%rsi,>lensav=%rcx
12113mov %rsi,%rcx
12114
12115# qhasm: (uint32) len >>= 4
12116# asm 1: shr $4,<len=int64#2d
12117# asm 2: shr $4,<len=%esi
12118shr $4,%esi
12119
12120# qhasm: tmp = *(uint32 *)(np + 12)
12121# asm 1: movl 12(<np=int64#3),>tmp=int64#5d
12122# asm 2: movl 12(<np=%rdx),>tmp=%r8d
12123movl 12(%rdx),%r8d
12124
12125# qhasm: (uint32) bswap tmp
12126# asm 1: bswap <tmp=int64#5d
12127# asm 2: bswap <tmp=%r8d
12128bswap %r8d
12129
12130# qhasm: tmp += len
12131# asm 1: add <len=int64#2,<tmp=int64#5
12132# asm 2: add <len=%rsi,<tmp=%r8
12133add %rsi,%r8
12134
12135# qhasm: (uint32) bswap tmp
12136# asm 1: bswap <tmp=int64#5d
12137# asm 2: bswap <tmp=%r8d
12138bswap %r8d
12139
12140# qhasm: *(uint32 *)(np + 12) = tmp
12141# asm 1: movl <tmp=int64#5d,12(<np=int64#3)
12142# asm 2: movl <tmp=%r8d,12(<np=%rdx)
12143movl %r8d,12(%rdx)
12144
12145# qhasm: blp = &bl
12146# asm 1: leaq <bl=stack1024#1,>blp=int64#2
12147# asm 2: leaq <bl=32(%rsp),>blp=%rsi
12148leaq 32(%rsp),%rsi
12149
12150# qhasm: *(int128 *)(blp + 0) = xmm8
12151# asm 1: movdqa <xmm8=int6464#9,0(<blp=int64#2)
12152# asm 2: movdqa <xmm8=%xmm8,0(<blp=%rsi)
12153movdqa %xmm8,0(%rsi)
12154
12155# qhasm: *(int128 *)(blp + 16) = xmm9
12156# asm 1: movdqa <xmm9=int6464#10,16(<blp=int64#2)
12157# asm 2: movdqa <xmm9=%xmm9,16(<blp=%rsi)
12158movdqa %xmm9,16(%rsi)
12159
12160# qhasm: *(int128 *)(blp + 32) = xmm12
12161# asm 1: movdqa <xmm12=int6464#13,32(<blp=int64#2)
12162# asm 2: movdqa <xmm12=%xmm12,32(<blp=%rsi)
12163movdqa %xmm12,32(%rsi)
12164
12165# qhasm: *(int128 *)(blp + 48) = xmm14
12166# asm 1: movdqa <xmm14=int6464#15,48(<blp=int64#2)
12167# asm 2: movdqa <xmm14=%xmm14,48(<blp=%rsi)
12168movdqa %xmm14,48(%rsi)
12169
12170# qhasm: *(int128 *)(blp + 64) = xmm11
12171# asm 1: movdqa <xmm11=int6464#12,64(<blp=int64#2)
12172# asm 2: movdqa <xmm11=%xmm11,64(<blp=%rsi)
12173movdqa %xmm11,64(%rsi)
12174
12175# qhasm: *(int128 *)(blp + 80) = xmm15
12176# asm 1: movdqa <xmm15=int6464#16,80(<blp=int64#2)
12177# asm 2: movdqa <xmm15=%xmm15,80(<blp=%rsi)
12178movdqa %xmm15,80(%rsi)
12179
12180# qhasm: *(int128 *)(blp + 96) = xmm10
12181# asm 1: movdqa <xmm10=int6464#11,96(<blp=int64#2)
12182# asm 2: movdqa <xmm10=%xmm10,96(<blp=%rsi)
12183movdqa %xmm10,96(%rsi)
12184
12185# qhasm: *(int128 *)(blp + 112) = xmm13
12186# asm 1: movdqa <xmm13=int6464#14,112(<blp=int64#2)
12187# asm 2: movdqa <xmm13=%xmm13,112(<blp=%rsi)
12188movdqa %xmm13,112(%rsi)
12189
12190# qhasm: bytes:
12191._bytes:
12192
12193# qhasm: =? lensav-0
12194# asm 1: cmp $0,<lensav=int64#4
12195# asm 2: cmp $0,<lensav=%rcx
12196cmp $0,%rcx
12197# comment:fp stack unchanged by jump
12198
12199# qhasm: goto end if =
12200je ._end
12201
12202# qhasm: b = *(uint8 *)(blp + 0)
12203# asm 1: movzbq 0(<blp=int64#2),>b=int64#3
12204# asm 2: movzbq 0(<blp=%rsi),>b=%rdx
12205movzbq 0(%rsi),%rdx
12206
12207# qhasm: *(uint8 *)(outp + 0) = b
12208# asm 1: movb <b=int64#3b,0(<outp=int64#1)
12209# asm 2: movb <b=%dl,0(<outp=%rdi)
12210movb %dl,0(%rdi)
12211
12212# qhasm: blp += 1
12213# asm 1: add $1,<blp=int64#2
12214# asm 2: add $1,<blp=%rsi
12215add $1,%rsi
12216
12217# qhasm: outp +=1
12218# asm 1: add $1,<outp=int64#1
12219# asm 2: add $1,<outp=%rdi
12220add $1,%rdi
12221
12222# qhasm: lensav -= 1
12223# asm 1: sub $1,<lensav=int64#4
12224# asm 2: sub $1,<lensav=%rcx
12225sub $1,%rcx
12226# comment:fp stack unchanged by jump
12227
12228# qhasm: goto bytes
12229jmp ._bytes
12230
12231# qhasm: full:
12232._full:
12233
12234# qhasm: tmp = *(uint32 *)(np + 12)
12235# asm 1: movl 12(<np=int64#3),>tmp=int64#4d
12236# asm 2: movl 12(<np=%rdx),>tmp=%ecx
12237movl 12(%rdx),%ecx
12238
12239# qhasm: (uint32) bswap tmp
12240# asm 1: bswap <tmp=int64#4d
12241# asm 2: bswap <tmp=%ecx
12242bswap %ecx
12243
12244# qhasm: tmp += len
12245# asm 1: add <len=int64#2,<tmp=int64#4
12246# asm 2: add <len=%rsi,<tmp=%rcx
12247add %rsi,%rcx
12248
12249# qhasm: (uint32) bswap tmp
12250# asm 1: bswap <tmp=int64#4d
12251# asm 2: bswap <tmp=%ecx
12252bswap %ecx
12253
12254# qhasm: *(uint32 *)(np + 12) = tmp
12255# asm 1: movl <tmp=int64#4d,12(<np=int64#3)
12256# asm 2: movl <tmp=%ecx,12(<np=%rdx)
12257movl %ecx,12(%rdx)
12258
12259# qhasm: *(int128 *) (outp + 0) = xmm8
12260# asm 1: movdqa <xmm8=int6464#9,0(<outp=int64#1)
12261# asm 2: movdqa <xmm8=%xmm8,0(<outp=%rdi)
12262movdqa %xmm8,0(%rdi)
12263
12264# qhasm: *(int128 *) (outp + 16) = xmm9
12265# asm 1: movdqa <xmm9=int6464#10,16(<outp=int64#1)
12266# asm 2: movdqa <xmm9=%xmm9,16(<outp=%rdi)
12267movdqa %xmm9,16(%rdi)
12268
12269# qhasm: *(int128 *) (outp + 32) = xmm12
12270# asm 1: movdqa <xmm12=int6464#13,32(<outp=int64#1)
12271# asm 2: movdqa <xmm12=%xmm12,32(<outp=%rdi)
12272movdqa %xmm12,32(%rdi)
12273
12274# qhasm: *(int128 *) (outp + 48) = xmm14
12275# asm 1: movdqa <xmm14=int6464#15,48(<outp=int64#1)
12276# asm 2: movdqa <xmm14=%xmm14,48(<outp=%rdi)
12277movdqa %xmm14,48(%rdi)
12278
12279# qhasm: *(int128 *) (outp + 64) = xmm11
12280# asm 1: movdqa <xmm11=int6464#12,64(<outp=int64#1)
12281# asm 2: movdqa <xmm11=%xmm11,64(<outp=%rdi)
12282movdqa %xmm11,64(%rdi)
12283
12284# qhasm: *(int128 *) (outp + 80) = xmm15
12285# asm 1: movdqa <xmm15=int6464#16,80(<outp=int64#1)
12286# asm 2: movdqa <xmm15=%xmm15,80(<outp=%rdi)
12287movdqa %xmm15,80(%rdi)
12288
12289# qhasm: *(int128 *) (outp + 96) = xmm10
12290# asm 1: movdqa <xmm10=int6464#11,96(<outp=int64#1)
12291# asm 2: movdqa <xmm10=%xmm10,96(<outp=%rdi)
12292movdqa %xmm10,96(%rdi)
12293
12294# qhasm: *(int128 *) (outp + 112) = xmm13
12295# asm 1: movdqa <xmm13=int6464#14,112(<outp=int64#1)
12296# asm 2: movdqa <xmm13=%xmm13,112(<outp=%rdi)
12297movdqa %xmm13,112(%rdi)
12298# comment:fp stack unchanged by fallthrough
12299
12300# qhasm: end:
12301._end:
12302
12303# qhasm: leave
12304add %r11,%rsp
12305mov %rdi,%rax
12306mov %rsi,%rdx
12307xor %rax,%rax
12308ret