Lines Matching refs:rip

373         vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
388 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
392 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
395 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
400 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
423 vpaddd ONE(%rip), %xmm9, %xmm9 # INCR CNT to get Yn
425 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
438 lea SHIFT_MASK+16(%rip), %r12
455 lea SHIFT_MASK+16(%rip), %r12
472 vpshufb SHUF_MASK(%rip), %xmm2, %xmm2
481 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9
485 vpshufb SHUF_MASK(%rip), %xmm9, %xmm9 # shuffle xmm9 back to output as ciphertext
540 vpshufb SHUF_MASK(%rip), %xmm14, %xmm14 # perform a 16Byte swap
609 vpshufb SHUF_MASK(%rip), \T7, \T7
655 vpshufb SHUF_MASK(%rip), \T7, \T7
675 vpshufb SHUF_MASK(%rip), %xmm0, %xmm0
680 vpshufb SHUF_MASK(%rip), %xmm6, %xmm6
691 vpcmpeqd TWOONE(%rip), %xmm2, %xmm2
692 vpand POLY(%rip), %xmm2, %xmm2
762 lea SHIFT_MASK(%rip), %r12
789 vmovdqa SHUF_MASK(%rip), %xmm10
824 vmovdqa SHUF_MASK(%rip), %xmm1
843 vmovdqa SHUF_MASK(%rip), %xmm10
1002 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1004 vpshufb SHUF_MASK(%rip), reg_i, reg_i # perform a 16Byte swap
1053 vpshufb SHUF_MASK(%rip), reg_i, reg_i # prepare ciphertext for GHASH computations
1080 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1082 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1084 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1086 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1088 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1090 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1092 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1094 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1096 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1098 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1100 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1102 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1104 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1106 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1108 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1110 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1206 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1208 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1209 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1210 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1211 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1212 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1213 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1214 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1238 vpaddd ONE(%rip), \CTR, \XMM1 # INCR CNT
1239 vpaddd ONE(%rip), \XMM1, \XMM2
1240 vpaddd ONE(%rip), \XMM2, \XMM3
1241 vpaddd ONE(%rip), \XMM3, \XMM4
1242 vpaddd ONE(%rip), \XMM4, \XMM5
1243 vpaddd ONE(%rip), \XMM5, \XMM6
1244 vpaddd ONE(%rip), \XMM6, \XMM7
1245 vpaddd ONE(%rip), \XMM7, \XMM8
1248 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1249 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1250 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1251 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1252 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1253 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1254 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1255 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1257 vpaddd ONEf(%rip), \CTR, \XMM1 # INCR CNT
1258 vpaddd ONEf(%rip), \XMM1, \XMM2
1259 vpaddd ONEf(%rip), \XMM2, \XMM3
1260 vpaddd ONEf(%rip), \XMM3, \XMM4
1261 vpaddd ONEf(%rip), \XMM4, \XMM5
1262 vpaddd ONEf(%rip), \XMM5, \XMM6
1263 vpaddd ONEf(%rip), \XMM6, \XMM7
1264 vpaddd ONEf(%rip), \XMM7, \XMM8
1575 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
1576 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1577 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
1578 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
1579 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
1580 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
1581 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
1582 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1884 vmovdqa POLY2(%rip), \T3
1950 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
1952 vpshufb SHUF_MASK(%rip), reg_i, reg_i # perform a 16Byte swap
2003 vpshufb SHUF_MASK(%rip), reg_i, reg_i # prepare ciphertext for GHASH computations
2030 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2032 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2034 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2036 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2038 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2040 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2042 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2044 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2046 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2048 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2050 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2052 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2054 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2056 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2058 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0
2060 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2157 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2160 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2161 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2162 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2163 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2164 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2165 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2166 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2193 vpaddd ONE(%rip), \CTR, \XMM1 # INCR CNT
2194 vpaddd ONE(%rip), \XMM1, \XMM2
2195 vpaddd ONE(%rip), \XMM2, \XMM3
2196 vpaddd ONE(%rip), \XMM3, \XMM4
2197 vpaddd ONE(%rip), \XMM4, \XMM5
2198 vpaddd ONE(%rip), \XMM5, \XMM6
2199 vpaddd ONE(%rip), \XMM6, \XMM7
2200 vpaddd ONE(%rip), \XMM7, \XMM8
2203 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2204 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2205 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2206 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2207 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2208 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2209 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2210 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2212 vpaddd ONEf(%rip), \CTR, \XMM1 # INCR CNT
2213 vpaddd ONEf(%rip), \XMM1, \XMM2
2214 vpaddd ONEf(%rip), \XMM2, \XMM3
2215 vpaddd ONEf(%rip), \XMM3, \XMM4
2216 vpaddd ONEf(%rip), \XMM4, \XMM5
2217 vpaddd ONEf(%rip), \XMM5, \XMM6
2218 vpaddd ONEf(%rip), \XMM6, \XMM7
2219 vpaddd ONEf(%rip), \XMM7, \XMM8
2495 vmovdqa POLY2(%rip), \T3
2525 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap
2526 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2527 vpshufb SHUF_MASK(%rip), \XMM3, \XMM3 # perform a 16Byte swap
2528 vpshufb SHUF_MASK(%rip), \XMM4, \XMM4 # perform a 16Byte swap
2529 vpshufb SHUF_MASK(%rip), \XMM5, \XMM5 # perform a 16Byte swap
2530 vpshufb SHUF_MASK(%rip), \XMM6, \XMM6 # perform a 16Byte swap
2531 vpshufb SHUF_MASK(%rip), \XMM7, \XMM7 # perform a 16Byte swap
2532 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2699 vmovdqa POLY2(%rip), \T3