/linux-6.14.4/tools/perf/tests/ |
D | topology.c | 1 // SPDX-License-Identifier: GPL-2.0 14 #define TEMPL "/tmp/perf-test-XXXXXX" 26 return -1; in get_temp() 44 session->evlist = evlist__new_default(); in session_write_header() 45 TEST_ASSERT_VAL("can't get evlist", session->evlist); in session_write_header() 47 perf_header__set_feat(&session->header, HEADER_CPU_TOPOLOGY); in session_write_header() 48 perf_header__set_feat(&session->header, HEADER_NRCPUS); in session_write_header() 49 perf_header__set_feat(&session->header, HEADER_ARCH); in session_write_header() 51 session->header.data_size += DATA_SIZE; in session_write_header() 54 !perf_session__write_header(session, session->evlist, data.file.fd, true)); in session_write_header() [all …]
|
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0 6 #include "util/synthetic-events.h" 19 struct perf_record_cpu_map *map_event = &event->cpu_map; in process_event_mask() 21 struct perf_cpu_map *map; in process_event_mask() local 24 data = &map_event->data; in process_event_mask() 26 TEST_ASSERT_VAL("wrong type", data->type == PERF_CPU_MAP__MASK); in process_event_mask() 28 long_size = data->mask32_data.long_size; in process_event_mask() 32 TEST_ASSERT_VAL("wrong nr", data->mask32_data.nr == 1); in process_event_mask() 34 TEST_ASSERT_VAL("wrong cpu", perf_record_cpu_map_data__test_bit(0, data)); in process_event_mask() 35 TEST_ASSERT_VAL("wrong cpu", !perf_record_cpu_map_data__test_bit(1, data)); in process_event_mask() [all …]
|
/linux-6.14.4/arch/powerpc/platforms/cell/ |
D | cbe_regs.c | 1 // SPDX-License-Identifier: GPL-2.0-only 19 #include <asm/cell-regs.h> 22 * Current implementation uses "cpu" nodes. We build our own mapping 23 * array of cpu numbers to cpu nodes locally for now to allow interrupt 25 * we implement cpu hotplug, we'll have to install an appropriate notifier 26 * in order to release references to the cpu going away 48 static cpumask_t cbe_local_mask[MAX_CBE] = { [0 ... MAX_CBE-1] = {CPU_BITS_NONE} }; 64 if (np->data) in cbe_find_map() 65 return np->data; in cbe_find_map() 67 /* walk up path until cpu or be node was found */ in cbe_find_map() [all …]
|
/linux-6.14.4/tools/perf/util/ |
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0 22 * CPU number. 34 return (data->mask32_data.long_size == 4) in perf_record_cpu_map_data__test_bit() 35 ? (bit_word32 < data->mask32_data.nr) && in perf_record_cpu_map_data__test_bit() 36 (data->mask32_data.mask[bit_word32] & bit_mask32) != 0 in perf_record_cpu_map_data__test_bit() 37 : (bit_word64 < data->mask64_data.nr) && in perf_record_cpu_map_data__test_bit() 38 (data->mask64_data.mask[bit_word64] & bit_mask64) != 0; in perf_record_cpu_map_data__test_bit() 41 /* Read ith mask value from data into the given 64-bit sized bitmap */ 46 if (data->mask32_data.long_size == 4) in perf_record_cpu_map_data__read_one_mask() 47 bitmap[0] = data->mask32_data.mask[i]; in perf_record_cpu_map_data__read_one_mask() [all …]
|
D | mmap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 3 * Copyright (C) 2011-2017, Red Hat Inc, Arnaldo Carvalho de Melo <[email protected]> 5 * Parts came from evlist.c builtin-{top,stat,record}.c, see those files for further 34 len = bitmap_scnprintf(mask->bits, mask->nbits, buf, MASK_SIZE); in mmap_cpu_mask__scnprintf() 36 pr_debug("%p: %s mask[%zd]: %s\n", mask, tag, mask->nbits, buf); in mmap_cpu_mask__scnprintf() 39 size_t mmap__mmap_len(struct mmap *map) in mmap__mmap_len() argument 41 return perf_mmap__mmap_len(&map->core); in mmap__mmap_len() 71 static int perf_mmap__aio_enabled(struct mmap *map) in perf_mmap__aio_enabled() argument 73 return map->aio.nr_cblocks > 0; in perf_mmap__aio_enabled() 77 static int perf_mmap__aio_alloc(struct mmap *map, int idx) in perf_mmap__aio_alloc() argument [all …]
|
/linux-6.14.4/tools/lib/perf/ |
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 18 void perf_cpu_map__set_nr(struct perf_cpu_map *map, int nr_cpus) in perf_cpu_map__set_nr() argument 20 RC_CHK_ACCESS(map)->nr = nr_cpus; in perf_cpu_map__set_nr() 33 cpus->nr = nr_cpus; in perf_cpu_map__alloc() 34 refcount_set(&cpus->refcnt, 1); in perf_cpu_map__alloc() 44 RC_CHK_ACCESS(cpus)->map[0].cpu = -1; in perf_cpu_map__new_any_cpu() 49 static void cpu_map__delete(struct perf_cpu_map *map) in cpu_map__delete() argument 51 if (map) { in cpu_map__delete() 52 WARN_ONCE(refcount_read(perf_cpu_map__refcnt(map)) != 0, in cpu_map__delete() 54 RC_CHK_FREE(map); in cpu_map__delete() [all …]
|
D | evlist.c | 1 // SPDX-License-Identifier: GPL-2.0 30 INIT_LIST_HEAD(&evlist->entries); in perf_evlist__init() 31 evlist->nr_entries = 0; in perf_evlist__init() 32 fdarray__init(&evlist->pollfd, 64); in perf_evlist__init() 39 if (evsel->system_wide) { in __perf_evlist__propagate_maps() 40 /* System wide: set the cpu map of the evsel to all online CPUs. */ in __perf_evlist__propagate_maps() 41 perf_cpu_map__put(evsel->cpus); in __perf_evlist__propagate_maps() 42 evsel->cpus = perf_cpu_map__new_online_cpus(); in __perf_evlist__propagate_maps() 43 } else if (evlist->has_user_cpus && evsel->is_pmu_core) { in __perf_evlist__propagate_maps() 48 perf_cpu_map__put(evsel->cpus); in __perf_evlist__propagate_maps() [all …]
|
/linux-6.14.4/tools/lib/perf/include/perf/ |
D | cpumap.h | 1 /* SPDX-License-Identifier: GPL-2.0 */ 8 /** A wrapper around a CPU to avoid confusion with the perf_cpu_map's map's indices. */ 10 int cpu; member 21 * perf_cpu_map__new_any_cpu - a map with a singular "any CPU"/dummy -1 value. 25 * perf_cpu_map__new_online_cpus - a map read from 26 * /sys/devices/system/cpu/online if 27 * available. If reading wasn't possible a map 34 * perf_cpu_map__new - create a map from the given cpu_list such as "0-7". If no 39 LIBPERF_API struct perf_cpu_map *perf_cpu_map__get(struct perf_cpu_map *map); 44 LIBPERF_API void perf_cpu_map__put(struct perf_cpu_map *map); [all …]
|
/linux-6.14.4/Documentation/devicetree/bindings/cpu/ |
D | cpu-topology.txt | 2 CPU topology binding description 6 1 - Introduction 12 - socket 13 - cluster 14 - core 15 - thread 18 symmetric multi-threading (SMT) is supported or not. 20 For instance in a system where CPUs support SMT, "cpu" nodes represent all 21 threads existing in the system and map to the hierarchy level "thread" above. 22 In systems where SMT is not supported "cpu" nodes represent all cores present [all …]
|
/linux-6.14.4/kernel/bpf/ |
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 8 * DOC: cpu map 9 * The 'cpumap' is primarily used as a backend map for XDP BPF helper 13 * this map type redirects raw XDP frames to another CPU. The remote 14 * CPU will do SKB-allocation and call the normal network stack. 20 * basically allows for 10G wirespeed pre-filtering via bpf. 39 /* General idea: XDP packets getting XDP redirected to another CPU, 40 * will maximum be stored/queued for one driver ->poll() call. It is 42 * same CPU. Thus, cpu_map_flush operation can deduct via this_cpu_ptr() 46 #define CPU_MAP_BULK_SIZE 8 /* 8 == one cacheline on 64-bit archs */ [all …]
|
D | hashtab.c | 1 // SPDX-License-Identifier: GPL-2.0-only 2 /* Copyright (c) 2011-2014 PLUMgrid, http://plumgrid.com 47 * from sys_bpf(). BPF recursion is prevented by incrementing the per CPU 51 * by pinning the task to the current CPU and incrementing the recursion 52 * protection across the map operation. 73 * it is only safe to use raw spinlock for preallocated hash map on a RT kernel, 75 * after hash map was fully converted to use bpf_mem_alloc, there will be 76 * non-synchronous memory allocation for non-preallocated hash map, so it is 85 #define HASHTAB_MAP_LOCK_MASK (HASHTAB_MAP_LOCK_COUNT - 1) 88 struct bpf_map map; member [all …]
|
D | local_storage.c | 1 // SPDX-License-Identifier: GPL-2.0 2 #include <linux/bpf-cgroup.h> 16 #include "../cgroup/cgroup-internal.h" 22 struct bpf_map map; member 29 static struct bpf_cgroup_storage_map *map_to_storage(struct bpf_map *map) in map_to_storage() argument 31 return container_of(map, struct bpf_cgroup_storage_map, map); in map_to_storage() 34 static bool attach_type_isolated(const struct bpf_map *map) in attach_type_isolated() argument 36 return map->key_size == sizeof(struct bpf_cgroup_storage_key); in attach_type_isolated() 39 static int bpf_cgroup_storage_key_cmp(const struct bpf_cgroup_storage_map *map, in bpf_cgroup_storage_key_cmp() argument 42 if (attach_type_isolated(&map->map)) { in bpf_cgroup_storage_key_cmp() [all …]
|
D | arraymap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 2 /* Copyright (c) 2011-2014 PLUMgrid, http://plumgrid.com 26 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu() 27 free_percpu(array->pptrs[i]); in bpf_array_free_percpu() 37 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu() 38 ptr = bpf_map_alloc_percpu(&array->map, array->elem_size, 8, in bpf_array_alloc_percpu() 42 return -ENOMEM; in bpf_array_alloc_percpu() 44 array->pptrs[i] = ptr; in bpf_array_alloc_percpu() 54 bool percpu = attr->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; in array_map_alloc_check() 58 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check() [all …]
|
/linux-6.14.4/drivers/clocksource/ |
D | ingenic-timer.c | 1 // SPDX-License-Identifier: GPL-2.0 14 #include <linux/mfd/ingenic-tcu.h> 23 #include <dt-bindings/clock/ingenic,tcu.h> 32 unsigned int cpu; member 40 struct regmap *map; member 56 regmap_read(tcu->map, TCU_REG_TCNTc(tcu->cs_channel), &count); in ingenic_tcu_timer_read() 69 return container_of(timer, struct ingenic_tcu, timers[timer->cpu]); in to_ingenic_tcu() 83 regmap_write(tcu->map, TCU_REG_TECR, BIT(timer->channel)); in ingenic_tcu_cevt_set_state_shutdown() 95 return -EINVAL; in ingenic_tcu_cevt_set_next() 97 regmap_write(tcu->map, TCU_REG_TDFRc(timer->channel), next); in ingenic_tcu_cevt_set_next() [all …]
|
/linux-6.14.4/lib/ |
D | cpu_rmap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 3 * cpu_rmap.c: CPU affinity reverse-map support 13 * objects with CPU affinities. This can be seen as a reverse-map of 14 * CPU affinity. However, we do not assume that the object affinities 17 * CPU topology. 21 * alloc_cpu_rmap - allocate CPU affinity reverse-map 28 unsigned int cpu; in alloc_cpu_rmap() local 39 rmap = kzalloc(obj_offset + size * sizeof(rmap->obj[0]), flags); in alloc_cpu_rmap() 43 kref_init(&rmap->refcount); in alloc_cpu_rmap() 44 rmap->obj = (void **)((char *)rmap + obj_offset); in alloc_cpu_rmap() [all …]
|
/linux-6.14.4/samples/bpf/ |
D | map_perf_test_user.c | 1 // SPDX-License-Identifier: GPL-2.0-only 82 static void test_hash_prealloc(int cpu) in test_hash_prealloc() argument 90 printf("%d:hash_map_perf pre-alloc %lld events per sec\n", in test_hash_prealloc() 91 cpu, max_cnt * 1000000000ll / (time_get_ns() - start_time)); in test_hash_prealloc() 106 * It is fine that the user requests for a map with in pre_test_lru_hash_lookup() 108 * may return not found. For LRU map, we are not interested in pre_test_lru_hash_lookup() 109 * in such small map performance. in pre_test_lru_hash_lookup() 120 static void do_test_lru(enum test_type test, int cpu) in do_test_lru() argument 129 if (test == INNER_LRU_HASH_PREALLOC && cpu) { in do_test_lru() 130 /* If CPU is not 0, create inner_lru hash map and insert the fd in do_test_lru() [all …]
|
/linux-6.14.4/Documentation/bpf/ |
D | map_hash.rst | 1 .. SPDX-License-Identifier: GPL-2.0-only 3 .. Copyright (C) 2022-2023 Isovalent, Inc. 10 - ``BPF_MAP_TYPE_HASH`` was introduced in kernel version 3.19 11 - ``BPF_MAP_TYPE_PERCPU_HASH`` was introduced in version 4.6 12 - Both ``BPF_MAP_TYPE_LRU_HASH`` and ``BPF_MAP_TYPE_LRU_PERCPU_HASH`` 16 purpose hash map storage. Both the key and the value can be structs, 20 to the max_entries limit that you specify. Hash maps use pre-allocation 22 used to disable pre-allocation when it is too memory expensive. 25 CPU. The per-cpu values are stored internally in an array. 32 shared across CPUs but it is possible to request a per CPU LRU list with [all …]
|
D | map_cpumap.rst | 1 .. SPDX-License-Identifier: GPL-2.0-only 9 - ``BPF_MAP_TYPE_CPUMAP`` was introduced in kernel version 4.15 11 .. kernel-doc:: kernel/bpf/cpumap.c 12 :doc: cpu map 14 An example use-case for this map type is software based Receive Side Scaling (RSS). 16 The CPUMAP represents the CPUs in the system indexed as the map-key, and the 17 map-value is the config setting (per CPUMAP entry). Each CPUMAP entry has a dedicated 18 kernel thread bound to the given CPU to represent the remote CPU execution unit. 21 on the remote CPU. This allows an XDP program to split its processing across 22 multiple CPUs. For example, a scenario where the initial CPU (that sees/receives [all …]
|
D | map_cgroup_storage.rst | 1 .. SPDX-License-Identifier: GPL-2.0-only 8 The ``BPF_MAP_TYPE_CGROUP_STORAGE`` map type represents a local fix-sized 13 The map provide a local storage at the cgroup that the BPF program is attached 19 ``BPF_MAP_TYPE_CGROUP_STORAGE`` map type. Some of its behaviors was changed in 25 The map uses key of type of either ``__u64 cgroup_inode_id`` or 38 map will share the same storage. Otherwise, if the type is 44 void *bpf_get_local_storage(void *map, u64 flags) 75 Userspace accessing map declared above:: 80 __u32 map_lookup(struct bpf_map *map, __u64 cgrp, enum bpf_attach_type type) 87 bpf_map_lookup_elem(bpf_map__fd(map), &key, &value); [all …]
|
/linux-6.14.4/drivers/gpu/drm/i915/gem/selftests/ |
D | i915_gem_coherency.c | 2 * SPDX-License-Identifier: MIT 27 u32 *cpu; in cpu_set() local 30 i915_gem_object_lock(ctx->obj, NULL); in cpu_set() 31 err = i915_gem_object_prepare_write(ctx->obj, &needs_clflush); in cpu_set() 35 page = i915_gem_object_get_page(ctx->obj, offset >> PAGE_SHIFT); in cpu_set() 36 cpu = kmap_local_page(page) + offset_in_page(offset); in cpu_set() 39 drm_clflush_virt_range(cpu, sizeof(*cpu)); in cpu_set() 41 *cpu = v; in cpu_set() 44 drm_clflush_virt_range(cpu, sizeof(*cpu)); in cpu_set() 46 kunmap_local(cpu); in cpu_set() [all …]
|
/linux-6.14.4/arch/sh/kernel/cpu/sh4/ |
D | sq.c | 1 // SPDX-License-Identifier: GPL-2.0 3 * arch/sh/kernel/cpu/sh4/sq.c 5 * General management API for SH-4 integrated Store Queues 7 * Copyright (C) 2001 - 2006 Paul Mundt 11 #include <linux/cpu.h> 23 #include <cpu/sq.h> 50 * sq_flush_range - Flush (prefetch) a specific SQ range 62 for (len >>= 5; len--; sq += 8) in sq_flush_range() 70 static inline void sq_mapping_list_add(struct sq_mapping *map) in sq_mapping_list_add() argument 78 p = &tmp->next; in sq_mapping_list_add() [all …]
|
/linux-6.14.4/arch/x86/kernel/cpu/ |
D | topology.c | 1 // SPDX-License-Identifier: GPL-2.0-only 3 * CPU/APIC topology 24 #define pr_fmt(fmt) "CPU topo: " fmt 25 #include <linux/cpu.h> 35 #include "cpu.h" 38 * Map cpu index to physical APIC ID 48 /* Used for CPU number allocation and parallel CPU bringup */ 49 u32 cpuid_to_apicid[] __ro_after_init = { [0 ... NR_CPUS - 1] = BAD_APICID, }; 52 static struct { DECLARE_BITMAP(map, MAX_LOCAL_APIC); } apic_maps[TOPO_MAX_DOMAIN] __ro_after_init; 56 * with 1 as CPU #0 is reserved for the boot CPU. [all …]
|
/linux-6.14.4/arch/arm/kernel/ |
D | devtree.c | 1 // SPDX-License-Identifier: GPL-2.0-only 24 #include <asm/mach-types.h> 39 if (of_property_read_string(node, "enable-method", &method)) in set_smp_ops_by_method() 42 for (; m->method; m++) in set_smp_ops_by_method() 43 if (!strcmp(m->method, method)) { in set_smp_ops_by_method() 44 smp_set_ops(m->ops); in set_smp_ops_by_method() 59 * arm_dt_init_cpu_maps - Function retrieves cpu nodes from the device tree 60 * and builds the cpu logical map array containing MPIDR values related to 63 * Updates the cpu possible mask with the number of parsed cpu nodes 68 * Temp logical map is initialized with UINT_MAX values that are in arm_dt_init_cpu_maps() [all …]
|
/linux-6.14.4/Documentation/ABI/stable/ |
D | sysfs-devices-system-cpu | 1 What: /sys/devices/system/cpu/dscr_default 2 Date: 13-May-2014 6 /sys/devices/system/cpu/cpuN/dscr on all CPUs. 9 all per-CPU defaults at the same time. 12 What: /sys/devices/system/cpu/cpu[0-9]+/dscr 13 Date: 13-May-2014 17 a CPU. 22 on any CPU where it executes (overriding the value described 27 What: /sys/devices/system/cpu/cpuX/topology/physical_package_id 33 What: /sys/devices/system/cpu/cpuX/topology/die_id [all …]
|
/linux-6.14.4/arch/arc/plat-axs10x/ |
D | axs10x.c | 1 // SPDX-License-Identifier: GPL-2.0-only 5 * Copyright (C) 2013-15 Synopsys, Inc. (www.synopsys.com) 11 #include <asm/asm-offsets.h> 30 * Peripherals on CPU Card and Mother Board are wired to cpu intc via in axs10x_enable_gpio_intc_wire() 33 * --------------------- in axs10x_enable_gpio_intc_wire() 34 * | snps,arc700-intc | in axs10x_enable_gpio_intc_wire() 35 * --------------------- in axs10x_enable_gpio_intc_wire() 37 * ------------------- ------------------- in axs10x_enable_gpio_intc_wire() 38 * | snps,dw-apb-gpio | | snps,dw-apb-gpio | in axs10x_enable_gpio_intc_wire() 39 * ------------------- ------------------- in axs10x_enable_gpio_intc_wire() [all …]
|