/linux-6.14.4/arch/s390/boot/ |
D | startup.c | 260 unsigned long kernel_start; in setup_kernel_memory_layout() local 316 kernel_start = round_down(kernel_end - kernel_size, THREAD_SIZE); in setup_kernel_memory_layout() 318 boot_debug("kernel image: 0x%016lx-0x%016lx (kaslr)\n", kernel_start, in setup_kernel_memory_layout() 321 kernel_start = round_down(vmax - kernel_size, THREAD_SIZE); in setup_kernel_memory_layout() 322 boot_debug("kernel image: 0x%016lx-0x%016lx (constrained)\n", kernel_start, in setup_kernel_memory_layout() 323 kernel_start + kernel_size); in setup_kernel_memory_layout() 325 kernel_start = __NO_KASLR_START_KERNEL; in setup_kernel_memory_layout() 326 boot_debug("kernel image: 0x%016lx-0x%016lx (nokaslr)\n", kernel_start, in setup_kernel_memory_layout() 327 kernel_start + kernel_size); in setup_kernel_memory_layout() 329 __kaslr_offset = kernel_start; in setup_kernel_memory_layout() [all …]
|
D | vmem.c | 95 static void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 125 kasan_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KASAN_MAP_SHADOW); in kasan_populate_shadow() 190 static inline void kasan_populate_shadow(unsigned long kernel_start, unsigned long kernel_end) in kasan_populate_shadow() argument 479 void setup_vmem(unsigned long kernel_start, unsigned long kernel_end, unsigned long asce_limit) in setup_vmem() argument 545 pgtable_populate(kernel_start + TEXT_OFFSET, kernel_end, POPULATE_KERNEL); in setup_vmem() 553 kasan_populate_shadow(kernel_start, kernel_end); in setup_vmem()
|
D | boot.h | 75 void setup_vmem(unsigned long kernel_start, unsigned long kernel_end, unsigned long asce_limit);
|
/linux-6.14.4/tools/perf/util/ |
D | thread-stack.c | 96 u64 kernel_start; member 162 ts->kernel_start = machine__kernel_start(machine); in thread_stack__init() 166 ts->kernel_start = 1ULL << 63; in thread_stack__init() 516 static inline u64 callchain_context(u64 ip, u64 kernel_start) in callchain_context() argument 518 return ip < kernel_start ? PERF_CONTEXT_USER : PERF_CONTEXT_KERNEL; in callchain_context() 523 size_t sz, u64 ip, u64 kernel_start) in thread_stack__sample() argument 526 u64 context = callchain_context(ip, kernel_start); in thread_stack__sample() 547 context = callchain_context(ip, kernel_start); in thread_stack__sample() 566 u64 sample_ip, u64 kernel_start) in thread_stack__sample_late() argument 569 u64 sample_context = callchain_context(sample_ip, kernel_start); in thread_stack__sample_late() [all …]
|
D | thread-stack.h | 87 size_t sz, u64 ip, u64 kernel_start); 90 u64 kernel_start); 95 u64 sample_ip, u64 kernel_start);
|
D | machine.h | 48 u64 kernel_start; member 85 if (!machine->kernel_start) in machine__kernel_start() 87 return machine->kernel_start; in machine__kernel_start() 92 u64 kernel_start = machine__kernel_start(machine); in machine__kernel_ip() local 94 return ip >= kernel_start; in machine__kernel_ip()
|
D | db-export.c | 215 u64 kernel_start = machine__kernel_start(machine); in call_path_from_sample() local 268 kernel_start); in call_path_from_sample()
|
D | intel-pt.c | 88 u64 kernel_start; member 681 return ip >= ptq->pt->kernel_start ? in intel_pt_nr_cpumode() 997 } else if (ip >= ptq->pt->kernel_start) { in __intel_pt_pgd_ip() 1244 pt->kernel_start); in intel_pt_add_callchain() 1282 pt->kernel_start); in intel_pt_add_br_stack() 1823 sample->ip, pt->kernel_start); in intel_pt_prep_sample() 2285 pt->kernel_start); in intel_pt_do_synth_pebs_sample() 2921 if (!pt->kernel_start) { in intel_pt_run_decoder() 2922 pt->kernel_start = machine__kernel_start(pt->machine); in intel_pt_run_decoder() 2948 state->from_ip >= pt->kernel_start) { in intel_pt_run_decoder() [all …]
|
D | arm-spe.c | 78 u64 kernel_start; member 252 return ip >= spe->kernel_start ? in arm_spe_cpumode() 797 if (!spe->kernel_start) in arm_spe_run_decoder() 798 spe->kernel_start = machine__kernel_start(spe->machine); in arm_spe_run_decoder()
|
D | machine.c | 95 machine->kernel_start = 0; in machine__init() 3109 machine->kernel_start = 1ULL << 63; in machine__get_kernel_start() 3118 machine->kernel_start = map__start(map); in machine__get_kernel_start()
|
/linux-6.14.4/arch/x86/mm/ |
D | mem_encrypt_identity.c | 290 unsigned long kernel_start, kernel_end, kernel_len; in sme_encrypt_kernel() local 321 kernel_start = (unsigned long)RIP_REL_REF(_text); in sme_encrypt_kernel() 323 kernel_len = kernel_end - kernel_start; in sme_encrypt_kernel() 357 pgtable_area_len += sme_pgtable_calc(execute_end - kernel_start) * 2; in sme_encrypt_kernel() 422 ppd.paddr = kernel_start; in sme_encrypt_kernel() 423 ppd.vaddr = kernel_start; in sme_encrypt_kernel() 428 ppd.paddr = kernel_start; in sme_encrypt_kernel() 429 ppd.vaddr = kernel_start + decrypted_base; in sme_encrypt_kernel() 460 sme_encrypt_execute(kernel_start, kernel_start + decrypted_base, in sme_encrypt_kernel() 473 ppd.vaddr = kernel_start + decrypted_base; in sme_encrypt_kernel()
|
/linux-6.14.4/arch/riscv/mm/ |
D | physaddr.c | 24 unsigned long kernel_start = kernel_map.virt_addr; in __phys_addr_symbol() local 25 unsigned long kernel_end = kernel_start + kernel_map.size; in __phys_addr_symbol() 31 VIRTUAL_BUG_ON(x < kernel_start || x > kernel_end); in __phys_addr_symbol()
|
/linux-6.14.4/arch/arm/boot/bootp/ |
D | kernel.S | 2 .globl kernel_start symbol 3 kernel_start: label
|
D | init.S | 58 b kernel_start @ call kernel
|
/linux-6.14.4/arch/riscv/kernel/ |
D | elf_kexec.c | 206 unsigned long kernel_start; in elf_kexec_load() local 220 kernel_start = image->start; in elf_kexec_load() 286 &kernel_start, in elf_kexec_load() 287 sizeof(kernel_start), 0); in elf_kexec_load()
|
/linux-6.14.4/arch/arm64/mm/ |
D | init.c | 493 u64 kernel_start = (u64)_text; in module_init_limits() local 494 u64 kernel_size = kernel_end - kernel_start; in module_init_limits() 509 u64 min = kernel_start; in module_init_limits()
|
D | mmu.c | 626 phys_addr_t kernel_start = __pa_symbol(_stext); in map_mem() local 657 memblock_mark_nomap(kernel_start, kernel_end - kernel_start); in map_mem() 682 __map_memblock(pgdp, kernel_start, kernel_end, in map_mem() 684 memblock_clear_nomap(kernel_start, kernel_end - kernel_start); in map_mem()
|
/linux-6.14.4/arch/parisc/mm/ |
D | init.c | 358 unsigned long kernel_start, kernel_end; in map_pages() local 362 kernel_start = __pa((unsigned long)&__init_begin); in map_pages() 401 } else if (address < kernel_start || address >= kernel_end) { in map_pages()
|