@@ -2,48 +2,48 @@ use crate::core_arch::arm_shared::{uint32x4_t, uint8x16_t};
2
2
3
3
#[ allow( improper_ctypes) ]
4
4
extern "unadjusted" {
5
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.aese" ) ]
5
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.aese" ) ]
6
6
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.aese" ) ]
7
7
fn vaeseq_u8_ ( data : uint8x16_t , key : uint8x16_t ) -> uint8x16_t ;
8
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.aesd" ) ]
8
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.aesd" ) ]
9
9
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.aesd" ) ]
10
10
fn vaesdq_u8_ ( data : uint8x16_t , key : uint8x16_t ) -> uint8x16_t ;
11
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.aesmc" ) ]
11
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.aesmc" ) ]
12
12
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.aesmc" ) ]
13
13
fn vaesmcq_u8_ ( data : uint8x16_t ) -> uint8x16_t ;
14
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.aesimc" ) ]
14
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.aesimc" ) ]
15
15
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.aesimc" ) ]
16
16
fn vaesimcq_u8_ ( data : uint8x16_t ) -> uint8x16_t ;
17
17
18
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha1h" ) ]
18
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha1h" ) ]
19
19
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha1h" ) ]
20
20
fn vsha1h_u32_ ( hash_e : u32 ) -> u32 ;
21
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha1su0" ) ]
21
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha1su0" ) ]
22
22
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha1su0" ) ]
23
23
fn vsha1su0q_u32_ ( w0_3 : uint32x4_t , w4_7 : uint32x4_t , w8_11 : uint32x4_t ) -> uint32x4_t ;
24
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha1su1" ) ]
24
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha1su1" ) ]
25
25
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha1su1" ) ]
26
26
fn vsha1su1q_u32_ ( tw0_3 : uint32x4_t , w12_15 : uint32x4_t ) -> uint32x4_t ;
27
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha1c" ) ]
27
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha1c" ) ]
28
28
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha1c" ) ]
29
29
fn vsha1cq_u32_ ( hash_abcd : uint32x4_t , hash_e : u32 , wk : uint32x4_t ) -> uint32x4_t ;
30
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha1p" ) ]
30
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha1p" ) ]
31
31
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha1p" ) ]
32
32
fn vsha1pq_u32_ ( hash_abcd : uint32x4_t , hash_e : u32 , wk : uint32x4_t ) -> uint32x4_t ;
33
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha1m" ) ]
33
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha1m" ) ]
34
34
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha1m" ) ]
35
35
fn vsha1mq_u32_ ( hash_abcd : uint32x4_t , hash_e : u32 , wk : uint32x4_t ) -> uint32x4_t ;
36
36
37
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha256h" ) ]
37
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha256h" ) ]
38
38
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha256h" ) ]
39
39
fn vsha256hq_u32_ ( hash_abcd : uint32x4_t , hash_efgh : uint32x4_t , wk : uint32x4_t ) -> uint32x4_t ;
40
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha256h2" ) ]
40
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha256h2" ) ]
41
41
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha256h2" ) ]
42
42
fn vsha256h2q_u32_ ( hash_efgh : uint32x4_t , hash_abcd : uint32x4_t , wk : uint32x4_t ) -> uint32x4_t ;
43
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha256su0" ) ]
43
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha256su0" ) ]
44
44
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha256su0" ) ]
45
45
fn vsha256su0q_u32_ ( w0_3 : uint32x4_t , w4_7 : uint32x4_t ) -> uint32x4_t ;
46
- #[ cfg_attr( target_arch = "aarch64" , link_name = "llvm.aarch64.crypto.sha256su1" ) ]
46
+ #[ cfg_attr( any ( target_arch = "aarch64" , target_arch = "arm64ec" ) , link_name = "llvm.aarch64.crypto.sha256su1" ) ]
47
47
#[ cfg_attr( target_arch = "arm" , link_name = "llvm.arm.neon.sha256su1" ) ]
48
48
fn vsha256su1q_u32_ ( tw0_3 : uint32x4_t , w8_11 : uint32x4_t , w12_15 : uint32x4_t ) -> uint32x4_t ;
49
49
}
0 commit comments