@@ -2379,19 +2379,6 @@ def : AMDGPUPat <
2379
2379
let True16Predicate = NotHasTrue16BitInsts in {
2380
2380
def : ROTRPattern <V_ALIGNBIT_B32_e64>;
2381
2381
2382
- def : AMDGPUPat <
2383
- (rotr v2i32:$src0, v2i32:$src1),
2384
- (REG_SEQUENCE VReg_64,
2385
- (V_ALIGNBIT_B32_e64
2386
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2387
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2388
- (i32 (EXTRACT_SUBREG VReg_64:$src1, sub0))), sub0,
2389
- (V_ALIGNBIT_B32_e64
2390
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2391
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2392
- (i32 (EXTRACT_SUBREG VReg_64:$src1, sub1))), sub1)
2393
- >;
2394
-
2395
2382
// Prevents regression in fneg-modifier-casting.ll along with modifications to XorCombine() when v2i32 or is legal.
2396
2383
def : AMDGPUPat <
2397
2384
(fneg (select i1:$src0, (f32 (bitconvert i32:$src1)), (f32 (bitconvert i32:$src2)))),
@@ -2404,20 +2391,6 @@ def : GCNPat<(i32 (trunc (srl i64:$src0, (and i32:$src1, (i32 31))))),
2404
2391
def : GCNPat<(i32 (trunc (srl i64:$src0, (i32 ShiftAmt32Imm:$src1)))),
2405
2392
(V_ALIGNBIT_B32_e64 (i32 (EXTRACT_SUBREG (i64 $src0), sub1)),
2406
2393
(i32 (EXTRACT_SUBREG (i64 $src0), sub0)), $src1)>;
2407
-
2408
- def : GCNPat <
2409
- (rotr v2i32:$src0, v2i32:$src1),
2410
- (REG_SEQUENCE VReg_64,
2411
- (V_ALIGNBIT_B32_e64
2412
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2413
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2414
- (i32 (EXTRACT_SUBREG VReg_64:$src1, sub0))), sub0,
2415
- (V_ALIGNBIT_B32_e64
2416
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2417
- (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2418
- (i32 (EXTRACT_SUBREG VReg_64:$src1, sub1))), sub1)
2419
- >;
2420
-
2421
2394
} // end True16Predicate = NotHasTrue16BitInsts
2422
2395
2423
2396
let True16Predicate = UseRealTrue16Insts in {
@@ -2436,20 +2409,6 @@ def : GCNPat <
2436
2409
/* clamp */ 0, /* op_sel */ 0)
2437
2410
>;
2438
2411
2439
- def : GCNPat <
2440
- (rotr v2i32:$src0, v2i32:$src1),
2441
- (REG_SEQUENCE VReg_64,
2442
- (V_ALIGNBIT_B32_t16_e64
2443
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2444
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2445
- 0, (EXTRACT_SUBREG (i32 (EXTRACT_SUBREG VReg_64:$src1, sub0)) ,lo16),0,0), sub0,
2446
- (V_ALIGNBIT_B32_t16_e64
2447
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2448
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2449
- 0, (EXTRACT_SUBREG (i32 (EXTRACT_SUBREG VReg_64:$src1, sub0)) ,lo16),0,0), sub1)
2450
- >;
2451
-
2452
-
2453
2412
def : GCNPat<(i32 (trunc (srl i64:$src0, (i32 ShiftAmt32Imm:$src1)))),
2454
2413
(V_ALIGNBIT_B32_t16_e64 0, /* src0_modifiers */
2455
2414
(i32 (EXTRACT_SUBREG (i64 $src0), sub1)),
@@ -2476,20 +2435,6 @@ def : GCNPat <
2476
2435
$src1, /* clamp */ 0, /* op_sel */ 0)
2477
2436
>;
2478
2437
2479
- def : GCNPat <
2480
- (rotr v2i32:$src0, v2i32:$src1),
2481
- (REG_SEQUENCE VReg_64,
2482
- (V_ALIGNBIT_B32_fake16_e64
2483
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2484
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub0)),
2485
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src1, sub0)),0,0), sub0,
2486
- (V_ALIGNBIT_B32_fake16_e64
2487
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2488
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src0, sub1)),
2489
- 0, (i32 (EXTRACT_SUBREG VReg_64:$src1, sub1)),0,0), sub1)
2490
- >;
2491
-
2492
-
2493
2438
def : GCNPat<(i32 (trunc (srl i64:$src0, (and i32:$src1, (i32 31))))),
2494
2439
(V_ALIGNBIT_B32_fake16_e64 0, /* src0_modifiers */
2495
2440
(i32 (EXTRACT_SUBREG (i64 $src0), sub1)),
0 commit comments