Lines Matching refs:m0
479 s_mov_b32 s_save_m0, m0
493 s_lshr_b32 m0, s_wave_size, S_WAVE_SIZE
494 s_and_b32 m0, m0, 1
495 s_cmp_eq_u32 m0, 1
562 s_mov_b32 m0, 0x0 //Next lane of v2 to write to
616 s_mov_b32 m0, 0x0 //SGPR initial index value =0
642 s_add_u32 m0, m0, 16 //next sgpr index
643 s_cmp_lt_u32 m0, 96 //scc = (m0 < first 96 SGPR) ? 1 : 0
667 s_lshr_b32 m0, s_wave_size, S_WAVE_SIZE
668 s_and_b32 m0, m0, 1
669 s_cmp_eq_u32 m0, 1
704 s_lshr_b32 m0, s_wave_size, S_WAVE_SIZE
705 s_and_b32 m0, m0, 1
706 s_cmp_eq_u32 m0, 1
707 s_mov_b32 m0, 0x0
721 s_add_u32 m0, m0, 128 //every buffer_store_lds does 128 bytes
723 s_cmp_lt_u32 m0, s_save_alloc_size //scc=(m0 < s_save_alloc_size) ? 1 : 0
740 s_add_u32 m0, m0, s3 //every buffer_store_lds does 128 bytes
743 s_cmp_lt_u32 m0, s_save_alloc_size //scc=(m0 < s_save_alloc_size) ? 1 : 0
759 s_add_u32 m0, m0, 256 //every buffer_store_lds does 256 bytes
761 s_cmp_lt_u32 m0, s_save_alloc_size //scc=(m0 < s_save_alloc_size) ? 1 : 0
778 s_add_u32 m0, m0, s3 //every buffer_store_lds does 256 bytes
781 s_cmp_lt_u32 m0, s_save_alloc_size //scc=(m0 < s_save_alloc_size) ? 1 : 0
789 s_lshr_b32 m0, s_wave_size, S_WAVE_SIZE
790 s_and_b32 m0, m0, 1
791 s_cmp_eq_u32 m0, 1
804 s_lshr_b32 m0, s_wave_size, S_WAVE_SIZE
805 s_and_b32 m0, m0, 1
806 s_cmp_eq_u32 m0, 1
814 s_mov_b32 m0, 0x4 //VGPR initial index value =4
815 s_cmp_lt_u32 m0, s_save_alloc_size
830 s_add_u32 m0, m0, 4
831 s_cmp_lt_u32 m0, s_save_alloc_size
848 s_add_u32 m0, m0, 4 //next vgpr index
850 s_cmp_lt_u32 m0, s_save_alloc_size //scc = (m0 < s_save_alloc_size) ? 1 : 0
859 s_mov_b32 m0, 0x4 //VGPR initial index value =4
860 s_cmp_lt_u32 m0, s_save_alloc_size
875 s_add_u32 m0, m0, 4
876 s_cmp_lt_u32 m0, s_save_alloc_size
893 s_add_u32 m0, m0, 4 //next vgpr index
895 s_cmp_lt_u32 m0, s_save_alloc_size //scc = (m0 < s_save_alloc_size) ? 1 : 0
906 s_add_u32 s_save_alloc_size, s_save_alloc_size, m0
919 s_add_u32 m0, m0, 1
920 s_cmp_lt_u32 m0, s_save_alloc_size
929 s_add_u32 m0, m0, 1 //next vgpr index
931 s_cmp_lt_u32 m0, s_save_alloc_size //scc = (m0 < s_save_alloc_size) ? 1 : 0
954 s_lshr_b32 m0, s_restore_size, S_WAVE_SIZE
955 s_and_b32 m0, m0, 1
956 s_cmp_eq_u32 m0, 1
979 s_lshr_b32 m0, s_restore_size, S_WAVE_SIZE
980 s_and_b32 m0, m0, 1
981 s_cmp_eq_u32 m0, 1
982 s_mov_b32 m0, 0x0
993 s_add_u32 m0, m0, 128 // 128 DW
995 s_cmp_lt_u32 m0, s_restore_alloc_size //scc=(m0 < s_restore_alloc_size) ? 1 : 0
1007 s_add_u32 m0, m0, 256 // 256 DW
1009 s_cmp_lt_u32 m0, s_restore_alloc_size //scc=(m0 < s_restore_alloc_size) ? 1 : 0
1017 s_lshr_b32 m0, s_restore_size, S_WAVE_SIZE
1018 s_and_b32 m0, m0, 1
1019 s_cmp_eq_u32 m0, 1
1030 s_lshr_b32 m0, s_restore_size, S_WAVE_SIZE
1031 s_and_b32 m0, m0, 1
1032 s_cmp_eq_u32 m0, 1
1040 s_mov_b32 m0, 4 //VGPR initial index value = 4
1041 s_cmp_lt_u32 m0, s_restore_alloc_size
1054 s_add_u32 m0, m0, 4 //next vgpr index
1056 s_cmp_lt_u32 m0, s_restore_alloc_size //scc = (m0 < s_restore_alloc_size) ? 1 : 0
1074 s_mov_b32 m0, 4 //VGPR initial index value = 4
1075 s_cmp_lt_u32 m0, s_restore_alloc_size
1088 s_add_u32 m0, m0, 4 //next vgpr index
1090 s_cmp_lt_u32 m0, s_restore_alloc_size //scc = (m0 < s_restore_alloc_size) ? 1 : 0
1101 s_add_u32 s_restore_alloc_size, s_restore_alloc_size, m0
1108 s_add_u32 m0, m0, 1 //next vgpr index
1110 s_cmp_lt_u32 m0, s_restore_alloc_size //scc = (m0 < s_restore_alloc_size) ? 1 : 0
1135 s_mov_b32 m0, s_sgpr_save_num
1140 s_sub_u32 m0, m0, 4 // Restore from S[0] to S[104]
1149 s_sub_u32 m0, m0, 8 // Restore from S[0] to S[96]
1161 s_sub_u32 m0, m0, 16 // Restore from S[n] to S[0]
1173 s_cmp_eq_u32 m0, 0 //scc = (m0 < s_sgpr_save_num) ? 1 : 0
1210 s_mov_b32 m0, s_restore_m0
1276 v_writelane_b32 v2, s, m0
1277 s_add_u32 m0, m0, 0x1
1279 s_mov_b32 exec_lo, m0
1280 s_mov_b32 m0, s_mem_offset
1281 s_buffer_store_dword s, s_rsrc, m0 S_COHERENCE
1283 s_mov_b32 m0, exec_lo