| /arch/arm/tools/ |
| A D | syscallnr.sh | 5 align=1 23 while [ "$(($nr / (256 * $align) ))" -gt 0 ]; do 24 align=$(( $align * 4 )) 26 nr=$(( ($nr + $align - 1) & ~($align - 1) )) 27 echo "/* aligned to $align */"
|
| /arch/x86/lib/ |
| A D | bhi.S | 20 .align 32 25 .align 32 34 .align 32 44 .align 8 51 .align 32 62 .align 32 74 .align 32 87 .align 32 101 .align 8 108 .align 32 [all …]
|
| A D | retpoline.S | 35 .align RETPOLINE_THUNK_SIZE 61 .align RETPOLINE_THUNK_SIZE 68 .align RETPOLINE_THUNK_SIZE 77 .align RETPOLINE_THUNK_SIZE 90 .align RETPOLINE_THUNK_SIZE 197 .align 64 271 .align 64 339 .align 64 395 .align 64, 0xcc 402 .align 64, 0xcc [all …]
|
| /arch/sparc/lib/ |
| A D | bitext.c | 28 int bit_map_string_get(struct bit_map *t, int len, int align) in bit_map_string_get() argument 37 color = align; in bit_map_string_get() 38 align = t->num_colors; in bit_map_string_get() 41 if (align == 0) in bit_map_string_get() 42 align = 1; in bit_map_string_get() 44 align1 = align - 1; in bit_map_string_get() 45 if ((align & align1) != 0) in bit_map_string_get() 47 if (align < 0 || align >= t->size) in bit_map_string_get() 70 t->size, t->used, offset, len, align, count); in bit_map_string_get()
|
| A D | U3copy_from_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| A D | M7copy_from_user.S | 11 .align 4; \ 14 .align 4; 19 .align 4; \ 22 .align 4;
|
| A D | NG4copy_from_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| A D | M7copy_to_user.S | 11 .align 4; \ 14 .align 4; 19 .align 4; \ 22 .align 4;
|
| A D | NG4copy_to_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| A D | U1copy_from_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| A D | U1copy_to_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| A D | U3copy_to_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| A D | NG2copy_from_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| A D | NG2copy_to_user.S | 10 .align 4; \ 13 .align 4; 18 .align 4; \ 21 .align 4;
|
| /arch/xtensa/kernel/ |
| A D | align.S | 182 and a3, a3, a8 # align memory address 307 .align 8 318 mov a9, a3 ; _j .Lexit; .align 8 319 mov a10, a3 ; _j .Lexit; .align 8 320 mov a11, a3 ; _j .Lexit; .align 8 321 mov a12, a3 ; _j .Lexit; .align 8 322 mov a13, a3 ; _j .Lexit; .align 8 323 mov a14, a3 ; _j .Lexit; .align 8 324 mov a15, a3 ; _j .Lexit; .align 8 337 mov a3, a9 ; _j .Lstore_w; .align 8 [all …]
|
| /arch/s390/boot/ |
| A D | kaslr.c | 107 static unsigned long iterate_valid_positions(unsigned long size, unsigned long align, in iterate_valid_positions() argument 117 align = max(align, 8UL); in iterate_valid_positions() 118 _min = round_up(_min, align); in iterate_valid_positions() 122 start = round_up(start, align); in iterate_valid_positions() 143 range_pos = (tmp_end - start - size) / align + 1; in iterate_valid_positions() 148 return start + (find_pos - 1) * align; in iterate_valid_positions() 154 start = round_up(skip_res->end, align); in iterate_valid_positions() 182 unsigned long randomize_within_range(unsigned long size, unsigned long align, in randomize_within_range() argument 192 max_pos = iterate_valid_positions(size, align, min, max, res, ARRAY_SIZE(res), true, 0); in randomize_within_range() 197 return iterate_valid_positions(size, align, min, max, res, ARRAY_SIZE(res), false, pos + 1); in randomize_within_range()
|
| A D | physmem_info.c | 216 size, align, min, max); in die_oom() 284 align = max(align, 8UL); in __physmem_alloc_range() 289 if (round_up(min, align) + size > pos) in __physmem_alloc_range() 291 addr = round_down(pos - size, align); in __physmem_alloc_range() 306 die_oom(size, align, min, max); in __physmem_alloc_range() 311 unsigned long align, unsigned long min, unsigned long max, in physmem_alloc_range() argument 326 unsigned long align, bool die_on_oom) in physmem_alloc() argument 344 addr = __physmem_alloc_range(size, align, 0, addr, ranges_left, in physmem_alloc() 355 addr, addr + size, get_rr_type_name(type), align, !!new_range); in physmem_alloc() 364 unsigned long align) in physmem_alloc_or_die() argument [all …]
|
| /arch/loongarch/lib/ |
| A D | memset.S | 123 .align 4 130 .align 4 133 .align 4 137 .align 4 141 .align 4 146 .align 4 150 .align 4 155 .align 4 160 .align 4 165 .align 4
|
| /arch/alpha/lib/ |
| A D | copy_user.S | 33 .align 4 43 .align 4 64 .align 4 79 .align 4 93 .align 4 96 .align 4
|
| A D | memset.S | 28 .align 5 43 .align 5 70 .align 4 77 .align 3 95 .align 3 113 .align 5
|
| /arch/alpha/boot/ |
| A D | head.S | 20 .align 5 29 .align 5 38 .align 5 88 .align 3 97 .align 3 106 .align 3
|
| /arch/arm64/kernel/ |
| A D | kuser32.S | 19 .align 5 38 .align 5 43 .align 5 54 .align 5
|
| /arch/powerpc/kernel/ptrace/ |
| A D | ptrace-view.c | 597 .size = sizeof(u32), .align = sizeof(u32), 624 .size = sizeof(u64), .align = sizeof(u64), 629 .size = sizeof(u64), .align = sizeof(u64), 634 .size = sizeof(u64), .align = sizeof(u64), 639 .size = sizeof(u64), .align = sizeof(u64), 646 .size = sizeof(u64), .align = sizeof(u64), 651 .size = sizeof(u64), .align = sizeof(u64), 658 .size = sizeof(u64), .align = sizeof(u64), 663 .size = sizeof(u64), .align = sizeof(u64), 668 .size = sizeof(u64), .align = sizeof(u64), [all …]
|
| /arch/m68k/sun3/ |
| A D | sun3dvma.c | 139 static inline unsigned long get_baddr(int len, unsigned long align) in get_baddr() argument 159 if(align > DVMA_PAGE_SIZE) in get_baddr() 160 newlen = len + ((hole->end - len) & (align-1)); in get_baddr() 262 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument 283 if(align == 0) in dvma_map_align() 284 align = DVMA_PAGE_SIZE; in dvma_map_align() 286 align = ((align + (DVMA_PAGE_SIZE-1)) & DVMA_PAGE_MASK); in dvma_map_align() 288 baddr = get_baddr(len, align); in dvma_map_align() 317 void *dvma_malloc_align(unsigned long len, unsigned long align) in dvma_malloc_align() argument 332 if((baddr = (unsigned long)dvma_map_align(kaddr, len, align)) == 0) { in dvma_malloc_align()
|
| /arch/sh/include/asm/ |
| A D | romimage-macros.h | 16 .align 2 29 .align 2 42 .align 2 58 .align 2 69 .align 2
|