Lines Matching full:ff

101 static int __do_write_fd(struct feat_fd *ff, const void *buf, size_t size)  in __do_write_fd()  argument
103 ssize_t ret = writen(ff->fd, buf, size); in __do_write_fd()
110 static int __do_write_buf(struct feat_fd *ff, const void *buf, size_t size) in __do_write_buf() argument
114 size_t new_size = ff->size; in __do_write_buf()
117 if (size + ff->offset > max_size) in __do_write_buf()
120 while (size > (new_size - ff->offset)) in __do_write_buf()
124 if (ff->size < new_size) { in __do_write_buf()
125 addr = realloc(ff->buf, new_size); in __do_write_buf()
128 ff->buf = addr; in __do_write_buf()
129 ff->size = new_size; in __do_write_buf()
132 memcpy(ff->buf + ff->offset, buf, size); in __do_write_buf()
133 ff->offset += size; in __do_write_buf()
139 int do_write(struct feat_fd *ff, const void *buf, size_t size) in do_write() argument
141 if (!ff->buf) in do_write()
142 return __do_write_fd(ff, buf, size); in do_write()
143 return __do_write_buf(ff, buf, size); in do_write()
147 static int do_write_bitmap(struct feat_fd *ff, unsigned long *set, u64 size) in do_write_bitmap() argument
152 ret = do_write(ff, &size, sizeof(size)); in do_write_bitmap()
157 ret = do_write(ff, p + i, sizeof(*p)); in do_write_bitmap()
166 int write_padded(struct feat_fd *ff, const void *bf, in write_padded() argument
170 int err = do_write(ff, bf, count); in write_padded()
173 err = do_write(ff, zero_buf, count_aligned - count); in write_padded()
182 static int do_write_string(struct feat_fd *ff, const char *str) in do_write_string() argument
191 ret = do_write(ff, &len, sizeof(len)); in do_write_string()
195 return write_padded(ff, str, olen, len); in do_write_string()
198 static int __do_read_fd(struct feat_fd *ff, void *addr, ssize_t size) in __do_read_fd() argument
200 ssize_t ret = readn(ff->fd, addr, size); in __do_read_fd()
207 static int __do_read_buf(struct feat_fd *ff, void *addr, ssize_t size) in __do_read_buf() argument
209 if (size > (ssize_t)ff->size - ff->offset) in __do_read_buf()
212 memcpy(addr, ff->buf + ff->offset, size); in __do_read_buf()
213 ff->offset += size; in __do_read_buf()
219 static int __do_read(struct feat_fd *ff, void *addr, ssize_t size) in __do_read() argument
221 if (!ff->buf) in __do_read()
222 return __do_read_fd(ff, addr, size); in __do_read()
223 return __do_read_buf(ff, addr, size); in __do_read()
226 static int do_read_u32(struct feat_fd *ff, u32 *addr) in do_read_u32() argument
230 ret = __do_read(ff, addr, sizeof(*addr)); in do_read_u32()
234 if (ff->ph->needs_swap) in do_read_u32()
239 static int do_read_u64(struct feat_fd *ff, u64 *addr) in do_read_u64() argument
243 ret = __do_read(ff, addr, sizeof(*addr)); in do_read_u64()
247 if (ff->ph->needs_swap) in do_read_u64()
252 static char *do_read_string(struct feat_fd *ff) in do_read_string() argument
257 if (do_read_u32(ff, &len)) in do_read_string()
264 if (!__do_read(ff, buf, len)) { in do_read_string()
278 static int do_read_bitmap(struct feat_fd *ff, unsigned long **pset, u64 *psize) in do_read_bitmap() argument
284 ret = do_read_u64(ff, &size); in do_read_bitmap()
295 ret = do_read_u64(ff, p + i); in do_read_bitmap()
308 static int write_tracing_data(struct feat_fd *ff, in write_tracing_data() argument
311 if (WARN(ff->buf, "Error: calling %s in pipe-mode.\n", __func__)) in write_tracing_data()
314 return read_tracing_data(ff->fd, &evlist->core.entries); in write_tracing_data()
318 static int write_build_id(struct feat_fd *ff, in write_build_id() argument
324 session = container_of(ff->ph, struct perf_session, header); in write_build_id()
329 if (WARN(ff->buf, "Error: calling %s in pipe-mode.\n", __func__)) in write_build_id()
332 err = perf_session__write_buildid_table(session, ff); in write_build_id()
342 static int write_hostname(struct feat_fd *ff, in write_hostname() argument
352 return do_write_string(ff, uts.nodename); in write_hostname()
355 static int write_osrelease(struct feat_fd *ff, in write_osrelease() argument
365 return do_write_string(ff, uts.release); in write_osrelease()
368 static int write_arch(struct feat_fd *ff, in write_arch() argument
378 return do_write_string(ff, uts.machine); in write_arch()
381 static int write_version(struct feat_fd *ff, in write_version() argument
384 return do_write_string(ff, perf_version_string); in write_version()
387 static int __write_cpudesc(struct feat_fd *ff, const char *cpuinfo_proc) in __write_cpudesc() argument
435 ret = do_write_string(ff, s); in __write_cpudesc()
442 static int write_cpudesc(struct feat_fd *ff, in write_cpudesc() argument
470 ret = __write_cpudesc(ff, cpuinfo_procs[i]); in write_cpudesc()
478 static int write_nrcpus(struct feat_fd *ff, in write_nrcpus() argument
493 ret = do_write(ff, &nrc, sizeof(nrc)); in write_nrcpus()
497 return do_write(ff, &nra, sizeof(nra)); in write_nrcpus()
500 static int write_event_desc(struct feat_fd *ff, in write_event_desc() argument
512 ret = do_write(ff, &nre, sizeof(nre)); in write_event_desc()
520 ret = do_write(ff, &sz, sizeof(sz)); in write_event_desc()
525 ret = do_write(ff, &evsel->core.attr, sz); in write_event_desc()
536 ret = do_write(ff, &nri, sizeof(nri)); in write_event_desc()
543 ret = do_write_string(ff, evsel__name(evsel)); in write_event_desc()
549 ret = do_write(ff, evsel->core.id, evsel->core.ids * sizeof(u64)); in write_event_desc()
556 static int write_cmdline(struct feat_fd *ff, in write_cmdline() argument
568 ret = do_write(ff, &n, sizeof(n)); in write_cmdline()
572 ret = do_write_string(ff, buf); in write_cmdline()
577 ret = do_write_string(ff, perf_env.cmdline_argv[i]); in write_cmdline()
585 static int write_cpu_topology(struct feat_fd *ff, in write_cpu_topology() argument
596 ret = do_write(ff, &tp->package_cpus_lists, sizeof(tp->package_cpus_lists)); in write_cpu_topology()
601 ret = do_write_string(ff, tp->package_cpus_list[i]); in write_cpu_topology()
605 ret = do_write(ff, &tp->core_cpus_lists, sizeof(tp->core_cpus_lists)); in write_cpu_topology()
610 ret = do_write_string(ff, tp->core_cpus_list[i]); in write_cpu_topology()
620 ret = do_write(ff, &perf_env.cpu[j].core_id, in write_cpu_topology()
624 ret = do_write(ff, &perf_env.cpu[j].socket_id, in write_cpu_topology()
633 ret = do_write(ff, &tp->die_cpus_lists, sizeof(tp->die_cpus_lists)); in write_cpu_topology()
638 ret = do_write_string(ff, tp->die_cpus_list[i]); in write_cpu_topology()
644 ret = do_write(ff, &perf_env.cpu[j].die_id, in write_cpu_topology()
657 static int write_total_mem(struct feat_fd *ff, in write_total_mem() argument
678 ret = do_write(ff, &mem, sizeof(mem)); in write_total_mem()
686 static int write_numa_topology(struct feat_fd *ff, in write_numa_topology() argument
697 ret = do_write(ff, &tp->nr, sizeof(u32)); in write_numa_topology()
704 ret = do_write(ff, &n->node, sizeof(u32)); in write_numa_topology()
708 ret = do_write(ff, &n->mem_total, sizeof(u64)); in write_numa_topology()
712 ret = do_write(ff, &n->mem_free, sizeof(u64)); in write_numa_topology()
716 ret = do_write_string(ff, n->cpus); in write_numa_topology()
740 static int write_pmu_mappings(struct feat_fd *ff, in write_pmu_mappings() argument
754 ret = do_write(ff, &pmu_num, sizeof(pmu_num)); in write_pmu_mappings()
759 ret = do_write(ff, &pmu->type, sizeof(pmu->type)); in write_pmu_mappings()
763 ret = do_write_string(ff, pmu->name); in write_pmu_mappings()
783 static int write_group_desc(struct feat_fd *ff, in write_group_desc() argument
790 ret = do_write(ff, &nr_groups, sizeof(nr_groups)); in write_group_desc()
800 ret = do_write_string(ff, name); in write_group_desc()
804 ret = do_write(ff, &leader_idx, sizeof(leader_idx)); in write_group_desc()
808 ret = do_write(ff, &nr_members, sizeof(nr_members)); in write_group_desc()
885 static int write_cpuid(struct feat_fd *ff, struct evlist *evlist) in write_cpuid() argument
895 return do_write_string(ff, buffer); in write_cpuid()
898 static int write_branch_stack(struct feat_fd *ff __maybe_unused, in write_branch_stack()
904 static int write_auxtrace(struct feat_fd *ff, in write_auxtrace() argument
910 if (WARN(ff->buf, "Error: calling %s in pipe-mode.\n", __func__)) in write_auxtrace()
913 session = container_of(ff->ph, struct perf_session, header); in write_auxtrace()
915 err = auxtrace_index__write(ff->fd, &session->auxtrace_index); in write_auxtrace()
921 static int write_clockid(struct feat_fd *ff, in write_clockid() argument
924 return do_write(ff, &ff->ph->env.clock.clockid_res_ns, in write_clockid()
925 sizeof(ff->ph->env.clock.clockid_res_ns)); in write_clockid()
928 static int write_clock_data(struct feat_fd *ff, in write_clock_data() argument
938 ret = do_write(ff, &data32, sizeof(data32)); in write_clock_data()
943 data32 = ff->ph->env.clock.clockid; in write_clock_data()
945 ret = do_write(ff, &data32, sizeof(data32)); in write_clock_data()
950 data64 = &ff->ph->env.clock.tod_ns; in write_clock_data()
952 ret = do_write(ff, data64, sizeof(*data64)); in write_clock_data()
957 data64 = &ff->ph->env.clock.clockid_ns; in write_clock_data()
959 return do_write(ff, data64, sizeof(*data64)); in write_clock_data()
962 static int write_hybrid_topology(struct feat_fd *ff, in write_hybrid_topology() argument
973 ret = do_write(ff, &tp->nr, sizeof(u32)); in write_hybrid_topology()
980 ret = do_write_string(ff, n->pmu_name); in write_hybrid_topology()
984 ret = do_write_string(ff, n->cpus); in write_hybrid_topology()
996 static int write_dir_format(struct feat_fd *ff, in write_dir_format() argument
1002 session = container_of(ff->ph, struct perf_session, header); in write_dir_format()
1008 return do_write(ff, &data->dir.version, sizeof(data->dir.version)); in write_dir_format()
1012 static int write_bpf_prog_info(struct feat_fd *ff, in write_bpf_prog_info() argument
1015 struct perf_env *env = &ff->ph->env; in write_bpf_prog_info()
1022 ret = do_write(ff, &env->bpf_progs.infos_cnt, in write_bpf_prog_info()
1040 ret = do_write(ff, node->info_linear, len); in write_bpf_prog_info()
1054 static int write_bpf_btf(struct feat_fd *ff, in write_bpf_btf() argument
1057 struct perf_env *env = &ff->ph->env; in write_bpf_btf()
1064 ret = do_write(ff, &env->bpf_progs.btfs_cnt, in write_bpf_btf()
1077 ret = do_write(ff, &node->id, in write_bpf_btf()
1236 static int write_cache(struct feat_fd *ff, in write_cache() argument
1250 ret = do_write(ff, &version, sizeof(u32)); in write_cache()
1254 ret = do_write(ff, &cnt, sizeof(u32)); in write_cache()
1262 ret = do_write(ff, &c->v, sizeof(u32)); \ in write_cache()
1273 ret = do_write_string(ff, (const char *) c->v); \ in write_cache()
1289 static int write_stat(struct feat_fd *ff __maybe_unused, in write_stat()
1295 static int write_sample_time(struct feat_fd *ff, in write_sample_time() argument
1300 ret = do_write(ff, &evlist->first_sample_time, in write_sample_time()
1305 return do_write(ff, &evlist->last_sample_time, in write_sample_time()
1448 static int write_mem_topology(struct feat_fd *ff __maybe_unused, in write_mem_topology()
1464 ret = do_write(ff, &version, sizeof(version)); in write_mem_topology()
1468 ret = do_write(ff, &bsize, sizeof(bsize)); in write_mem_topology()
1472 ret = do_write(ff, &nr, sizeof(nr)); in write_mem_topology()
1480 ret = do_write(ff, &n->v, sizeof(n->v)); \ in write_mem_topology()
1489 ret = do_write_bitmap(ff, n->set, n->size); in write_mem_topology()
1499 static int write_compressed(struct feat_fd *ff __maybe_unused, in write_compressed()
1504 ret = do_write(ff, &(ff->ph->env.comp_ver), sizeof(ff->ph->env.comp_ver)); in write_compressed()
1508 ret = do_write(ff, &(ff->ph->env.comp_type), sizeof(ff->ph->env.comp_type)); in write_compressed()
1512 ret = do_write(ff, &(ff->ph->env.comp_level), sizeof(ff->ph->env.comp_level)); in write_compressed()
1516 ret = do_write(ff, &(ff->ph->env.comp_ratio), sizeof(ff->ph->env.comp_ratio)); in write_compressed()
1520 return do_write(ff, &(ff->ph->env.comp_mmap_len), sizeof(ff->ph->env.comp_mmap_len)); in write_compressed()
1523 static int __write_pmu_caps(struct feat_fd *ff, struct perf_pmu *pmu, in __write_pmu_caps() argument
1529 ret = do_write(ff, &pmu->nr_caps, sizeof(pmu->nr_caps)); in __write_pmu_caps()
1534 ret = do_write_string(ff, caps->name); in __write_pmu_caps()
1538 ret = do_write_string(ff, caps->value); in __write_pmu_caps()
1544 ret = do_write_string(ff, pmu->name); in __write_pmu_caps()
1552 static int write_cpu_pmu_caps(struct feat_fd *ff, in write_cpu_pmu_caps() argument
1565 return __write_pmu_caps(ff, cpu_pmu, false); in write_cpu_pmu_caps()
1568 static int write_pmu_caps(struct feat_fd *ff, in write_pmu_caps() argument
1589 ret = do_write(ff, &nr_pmu, sizeof(nr_pmu)); in write_pmu_caps()
1608 ret = __write_pmu_caps(ff, pmu, true); in write_pmu_caps()
1615 static void print_hostname(struct feat_fd *ff, FILE *fp) in print_hostname() argument
1617 fprintf(fp, "# hostname : %s\n", ff->ph->env.hostname); in print_hostname()
1620 static void print_osrelease(struct feat_fd *ff, FILE *fp) in print_osrelease() argument
1622 fprintf(fp, "# os release : %s\n", ff->ph->env.os_release); in print_osrelease()
1625 static void print_arch(struct feat_fd *ff, FILE *fp) in print_arch() argument
1627 fprintf(fp, "# arch : %s\n", ff->ph->env.arch); in print_arch()
1630 static void print_cpudesc(struct feat_fd *ff, FILE *fp) in print_cpudesc() argument
1632 fprintf(fp, "# cpudesc : %s\n", ff->ph->env.cpu_desc); in print_cpudesc()
1635 static void print_nrcpus(struct feat_fd *ff, FILE *fp) in print_nrcpus() argument
1637 fprintf(fp, "# nrcpus online : %u\n", ff->ph->env.nr_cpus_online); in print_nrcpus()
1638 fprintf(fp, "# nrcpus avail : %u\n", ff->ph->env.nr_cpus_avail); in print_nrcpus()
1641 static void print_version(struct feat_fd *ff, FILE *fp) in print_version() argument
1643 fprintf(fp, "# perf version : %s\n", ff->ph->env.version); in print_version()
1646 static void print_cmdline(struct feat_fd *ff, FILE *fp) in print_cmdline() argument
1650 nr = ff->ph->env.nr_cmdline; in print_cmdline()
1655 char *argv_i = strdup(ff->ph->env.cmdline_argv[i]); in print_cmdline()
1657 fprintf(fp, "%s ", ff->ph->env.cmdline_argv[i]); in print_cmdline()
1675 static void print_cpu_topology(struct feat_fd *ff, FILE *fp) in print_cpu_topology() argument
1677 struct perf_header *ph = ff->ph; in print_cpu_topology()
1732 static void print_clockid(struct feat_fd *ff, FILE *fp) in print_clockid() argument
1735 ff->ph->env.clock.clockid_res_ns * 1000); in print_clockid()
1738 static void print_clock_data(struct feat_fd *ff, FILE *fp) in print_clock_data() argument
1747 if (!ff->ph->env.clock.enabled) { in print_clock_data()
1753 ref = ff->ph->env.clock.tod_ns; in print_clock_data()
1759 ref = ff->ph->env.clock.clockid_ns; in print_clock_data()
1764 clockid = ff->ph->env.clock.clockid; in print_clock_data()
1781 static void print_hybrid_topology(struct feat_fd *ff, FILE *fp) in print_hybrid_topology() argument
1787 for (i = 0; i < ff->ph->env.nr_hybrid_nodes; i++) { in print_hybrid_topology()
1788 n = &ff->ph->env.hybrid_nodes[i]; in print_hybrid_topology()
1793 static void print_dir_format(struct feat_fd *ff, FILE *fp) in print_dir_format() argument
1798 session = container_of(ff->ph, struct perf_session, header); in print_dir_format()
1805 static void print_bpf_prog_info(struct feat_fd *ff, FILE *fp) in print_bpf_prog_info() argument
1807 struct perf_env *env = &ff->ph->env; in print_bpf_prog_info()
1829 static void print_bpf_btf(struct feat_fd *ff, FILE *fp) in print_bpf_btf() argument
1831 struct perf_env *env = &ff->ph->env; in print_bpf_btf()
1901 static struct evsel *read_event_desc(struct feat_fd *ff) in read_event_desc() argument
1910 if (do_read_u32(ff, &nre)) in read_event_desc()
1913 if (do_read_u32(ff, &sz)) in read_event_desc()
1937 if (__do_read(ff, buf, sz)) in read_event_desc()
1940 if (ff->ph->needs_swap) in read_event_desc()
1948 if (do_read_u32(ff, &nr)) in read_event_desc()
1951 if (ff->ph->needs_swap) in read_event_desc()
1954 evsel->name = do_read_string(ff); in read_event_desc()
1968 if (do_read_u64(ff, id)) in read_event_desc()
1988 static void print_event_desc(struct feat_fd *ff, FILE *fp) in print_event_desc() argument
1994 if (ff->events) in print_event_desc()
1995 events = ff->events; in print_event_desc()
1997 events = read_event_desc(ff); in print_event_desc()
2023 ff->events = NULL; in print_event_desc()
2026 static void print_total_mem(struct feat_fd *ff, FILE *fp) in print_total_mem() argument
2028 fprintf(fp, "# total memory : %llu kB\n", ff->ph->env.total_mem); in print_total_mem()
2031 static void print_numa_topology(struct feat_fd *ff, FILE *fp) in print_numa_topology() argument
2036 for (i = 0; i < ff->ph->env.nr_numa_nodes; i++) { in print_numa_topology()
2037 n = &ff->ph->env.numa_nodes[i]; in print_numa_topology()
2048 static void print_cpuid(struct feat_fd *ff, FILE *fp) in print_cpuid() argument
2050 fprintf(fp, "# cpuid : %s\n", ff->ph->env.cpuid); in print_cpuid()
2053 static void print_branch_stack(struct feat_fd *ff __maybe_unused, FILE *fp) in print_branch_stack()
2058 static void print_auxtrace(struct feat_fd *ff __maybe_unused, FILE *fp) in print_auxtrace()
2063 static void print_stat(struct feat_fd *ff __maybe_unused, FILE *fp) in print_stat()
2068 static void print_cache(struct feat_fd *ff, FILE *fp __maybe_unused) in print_cache() argument
2073 for (i = 0; i < ff->ph->env.caches_cnt; i++) { in print_cache()
2075 cpu_cache_level__fprintf(fp, &ff->ph->env.caches[i]); in print_cache()
2079 static void print_compressed(struct feat_fd *ff, FILE *fp) in print_compressed() argument
2082 ff->ph->env.comp_type == PERF_COMP_ZSTD ? "Zstd" : "Unknown", in print_compressed()
2083 ff->ph->env.comp_level, ff->ph->env.comp_ratio); in print_compressed()
2105 static void print_cpu_pmu_caps(struct feat_fd *ff, FILE *fp) in print_cpu_pmu_caps() argument
2107 __print_pmu_caps(fp, ff->ph->env.nr_cpu_pmu_caps, in print_cpu_pmu_caps()
2108 ff->ph->env.cpu_pmu_caps, (char *)"cpu"); in print_cpu_pmu_caps()
2111 static void print_pmu_caps(struct feat_fd *ff, FILE *fp) in print_pmu_caps() argument
2115 for (int i = 0; i < ff->ph->env.nr_pmus_with_caps; i++) { in print_pmu_caps()
2116 pmu_caps = &ff->ph->env.pmu_caps[i]; in print_pmu_caps()
2121 if (strcmp(perf_env__arch(&ff->ph->env), "x86") == 0 && in print_pmu_caps()
2122 perf_env__has_pmu_mapping(&ff->ph->env, "ibs_op")) { in print_pmu_caps()
2123 char *max_precise = perf_env__find_pmu_cap(&ff->ph->env, "cpu", "max_precise"); in print_pmu_caps()
2130 static void print_pmu_mappings(struct feat_fd *ff, FILE *fp) in print_pmu_mappings() argument
2137 pmu_num = ff->ph->env.nr_pmu_mappings; in print_pmu_mappings()
2143 str = ff->ph->env.pmu_mappings; in print_pmu_mappings()
2166 static void print_group_desc(struct feat_fd *ff, FILE *fp) in print_group_desc() argument
2172 session = container_of(ff->ph, struct perf_session, header); in print_group_desc()
2188 static void print_sample_time(struct feat_fd *ff, FILE *fp) in print_sample_time() argument
2194 session = container_of(ff->ph, struct perf_session, header); in print_sample_time()
2223 static void print_mem_topology(struct feat_fd *ff, FILE *fp) in print_mem_topology() argument
2228 nodes = ff->ph->env.memory_nodes; in print_mem_topology()
2229 nr = ff->ph->env.nr_memory_nodes; in print_mem_topology()
2232 nr, ff->ph->env.memory_bsize); in print_mem_topology()
2235 memory_node__fprintf(&nodes[i], ff->ph->env.memory_bsize, fp); in print_mem_topology()
2401 static int process_##__feat(struct feat_fd *ff, void *data __maybe_unused) \
2403 free(ff->ph->env.__feat_env); \
2404 ff->ph->env.__feat_env = do_read_string(ff); \
2405 return ff->ph->env.__feat_env ? 0 : -ENOMEM; \
2416 static int process_tracing_data(struct feat_fd *ff, void *data) in process_tracing_data() argument
2418 ssize_t ret = trace_report(ff->fd, data, false); in process_tracing_data()
2424 static int process_build_id(struct feat_fd *ff, void *data __maybe_unused) in process_build_id() argument
2426 if (perf_header__read_build_ids(ff->ph, ff->fd, ff->offset, ff->size)) in process_build_id()
2431 static int process_nrcpus(struct feat_fd *ff, void *data __maybe_unused) in process_nrcpus() argument
2436 ret = do_read_u32(ff, &nr_cpus_avail); in process_nrcpus()
2440 ret = do_read_u32(ff, &nr_cpus_online); in process_nrcpus()
2443 ff->ph->env.nr_cpus_avail = (int)nr_cpus_avail; in process_nrcpus()
2444 ff->ph->env.nr_cpus_online = (int)nr_cpus_online; in process_nrcpus()
2448 static int process_total_mem(struct feat_fd *ff, void *data __maybe_unused) in process_total_mem() argument
2453 ret = do_read_u64(ff, &total_mem); in process_total_mem()
2456 ff->ph->env.total_mem = (unsigned long long)total_mem; in process_total_mem()
2490 process_event_desc(struct feat_fd *ff, void *data __maybe_unused) in process_event_desc() argument
2493 struct evsel *evsel, *events = read_event_desc(ff); in process_event_desc()
2498 session = container_of(ff->ph, struct perf_session, header); in process_event_desc()
2503 ff->events = events; in process_event_desc()
2515 static int process_cmdline(struct feat_fd *ff, void *data __maybe_unused) in process_cmdline() argument
2520 if (do_read_u32(ff, &nr)) in process_cmdline()
2523 ff->ph->env.nr_cmdline = nr; in process_cmdline()
2525 cmdline = zalloc(ff->size + nr + 1); in process_cmdline()
2534 str = do_read_string(ff); in process_cmdline()
2543 ff->ph->env.cmdline = cmdline; in process_cmdline()
2544 ff->ph->env.cmdline_argv = (const char **) argv; in process_cmdline()
2553 static int process_cpu_topology(struct feat_fd *ff, void *data __maybe_unused) in process_cpu_topology() argument
2558 int cpu_nr = ff->ph->env.nr_cpus_avail; in process_cpu_topology()
2560 struct perf_header *ph = ff->ph; in process_cpu_topology()
2567 if (do_read_u32(ff, &nr)) in process_cpu_topology()
2576 str = do_read_string(ff); in process_cpu_topology()
2588 if (do_read_u32(ff, &nr)) in process_cpu_topology()
2595 str = do_read_string(ff); in process_cpu_topology()
2611 if (ff->size <= size) { in process_cpu_topology()
2626 if (do_read_u32(ff, &nr)) in process_cpu_topology()
2632 if (do_read_u32(ff, &nr)) in process_cpu_topology()
2649 if (ff->size <= size) in process_cpu_topology()
2652 if (do_read_u32(ff, &nr)) in process_cpu_topology()
2659 str = do_read_string(ff); in process_cpu_topology()
2672 if (do_read_u32(ff, &nr)) in process_cpu_topology()
2688 static int process_numa_topology(struct feat_fd *ff, void *data __maybe_unused) in process_numa_topology() argument
2695 if (do_read_u32(ff, &nr)) in process_numa_topology()
2706 if (do_read_u32(ff, &n->node)) in process_numa_topology()
2709 if (do_read_u64(ff, &n->mem_total)) in process_numa_topology()
2712 if (do_read_u64(ff, &n->mem_free)) in process_numa_topology()
2715 str = do_read_string(ff); in process_numa_topology()
2724 ff->ph->env.nr_numa_nodes = nr; in process_numa_topology()
2725 ff->ph->env.numa_nodes = nodes; in process_numa_topology()
2733 static int process_pmu_mappings(struct feat_fd *ff, void *data __maybe_unused) in process_pmu_mappings() argument
2740 if (do_read_u32(ff, &pmu_num)) in process_pmu_mappings()
2748 ff->ph->env.nr_pmu_mappings = pmu_num; in process_pmu_mappings()
2753 if (do_read_u32(ff, &type)) in process_pmu_mappings()
2756 name = do_read_string(ff); in process_pmu_mappings()
2767 ff->ph->env.msr_pmu_type = type; in process_pmu_mappings()
2772 ff->ph->env.pmu_mappings = strbuf_detach(&sb, NULL); in process_pmu_mappings()
2780 static int process_group_desc(struct feat_fd *ff, void *data __maybe_unused) in process_group_desc() argument
2792 if (do_read_u32(ff, &nr_groups)) in process_group_desc()
2795 ff->ph->env.nr_groups = nr_groups; in process_group_desc()
2806 desc[i].name = do_read_string(ff); in process_group_desc()
2810 if (do_read_u32(ff, &desc[i].leader_idx)) in process_group_desc()
2813 if (do_read_u32(ff, &desc[i].nr_members)) in process_group_desc()
2820 session = container_of(ff->ph, struct perf_session, header); in process_group_desc()
2863 static int process_auxtrace(struct feat_fd *ff, void *data __maybe_unused) in process_auxtrace() argument
2868 session = container_of(ff->ph, struct perf_session, header); in process_auxtrace()
2870 err = auxtrace_index__process(ff->fd, ff->size, session, in process_auxtrace()
2871 ff->ph->needs_swap); in process_auxtrace()
2877 static int process_cache(struct feat_fd *ff, void *data __maybe_unused) in process_cache() argument
2882 if (do_read_u32(ff, &version)) in process_cache()
2888 if (do_read_u32(ff, &cnt)) in process_cache()
2899 if (do_read_u32(ff, &c->v)) \ in process_cache()
2909 c->v = do_read_string(ff); \ in process_cache()
2919 ff->ph->env.caches = caches; in process_cache()
2920 ff->ph->env.caches_cnt = cnt; in process_cache()
2932 static int process_sample_time(struct feat_fd *ff, void *data __maybe_unused) in process_sample_time() argument
2938 session = container_of(ff->ph, struct perf_session, header); in process_sample_time()
2940 ret = do_read_u64(ff, &first_sample_time); in process_sample_time()
2944 ret = do_read_u64(ff, &last_sample_time); in process_sample_time()
2953 static int process_mem_topology(struct feat_fd *ff, in process_mem_topology() argument
2960 if (do_read_u64(ff, &version)) in process_mem_topology()
2966 if (do_read_u64(ff, &bsize)) in process_mem_topology()
2969 if (do_read_u64(ff, &nr)) in process_mem_topology()
2980 if (do_read_u64(ff, &n.v)) \ in process_mem_topology()
2988 if (do_read_bitmap(ff, &n.set, &n.size)) in process_mem_topology()
2994 ff->ph->env.memory_bsize = bsize; in process_mem_topology()
2995 ff->ph->env.memory_nodes = nodes; in process_mem_topology()
2996 ff->ph->env.nr_memory_nodes = nr; in process_mem_topology()
3005 static int process_clockid(struct feat_fd *ff, in process_clockid() argument
3008 if (do_read_u64(ff, &ff->ph->env.clock.clockid_res_ns)) in process_clockid()
3014 static int process_clock_data(struct feat_fd *ff, in process_clock_data() argument
3021 if (do_read_u32(ff, &data32)) in process_clock_data()
3028 if (do_read_u32(ff, &data32)) in process_clock_data()
3031 ff->ph->env.clock.clockid = data32; in process_clock_data()
3034 if (do_read_u64(ff, &data64)) in process_clock_data()
3037 ff->ph->env.clock.tod_ns = data64; in process_clock_data()
3040 if (do_read_u64(ff, &data64)) in process_clock_data()
3043 ff->ph->env.clock.clockid_ns = data64; in process_clock_data()
3044 ff->ph->env.clock.enabled = true; in process_clock_data()
3048 static int process_hybrid_topology(struct feat_fd *ff, in process_hybrid_topology() argument
3055 if (do_read_u32(ff, &nr)) in process_hybrid_topology()
3065 n->pmu_name = do_read_string(ff); in process_hybrid_topology()
3069 n->cpus = do_read_string(ff); in process_hybrid_topology()
3074 ff->ph->env.nr_hybrid_nodes = nr; in process_hybrid_topology()
3075 ff->ph->env.hybrid_nodes = nodes; in process_hybrid_topology()
3088 static int process_dir_format(struct feat_fd *ff, in process_dir_format() argument
3094 session = container_of(ff->ph, struct perf_session, header); in process_dir_format()
3100 return do_read_u64(ff, &data->dir.version); in process_dir_format()
3104 static int process_bpf_prog_info(struct feat_fd *ff, void *data __maybe_unused) in process_bpf_prog_info() argument
3107 struct perf_env *env = &ff->ph->env; in process_bpf_prog_info()
3112 if (ff->ph->needs_swap) { in process_bpf_prog_info()
3117 if (do_read_u32(ff, &count)) in process_bpf_prog_info()
3127 if (do_read_u32(ff, &info_len)) in process_bpf_prog_info()
3129 if (do_read_u32(ff, &data_len)) in process_bpf_prog_info()
3143 if (do_read_u64(ff, (u64 *)(&info_linear->arrays))) in process_bpf_prog_info()
3145 if (__do_read(ff, &info_linear->info, info_len)) in process_bpf_prog_info()
3151 if (__do_read(ff, info_linear->data, data_len)) in process_bpf_prog_info()
3176 static int process_bpf_btf(struct feat_fd *ff, void *data __maybe_unused) in process_bpf_btf() argument
3178 struct perf_env *env = &ff->ph->env; in process_bpf_btf()
3183 if (ff->ph->needs_swap) { in process_bpf_btf()
3188 if (do_read_u32(ff, &count)) in process_bpf_btf()
3196 if (do_read_u32(ff, &id)) in process_bpf_btf()
3198 if (do_read_u32(ff, &data_size)) in process_bpf_btf()
3208 if (__do_read(ff, node->data, data_size)) in process_bpf_btf()
3224 static int process_compressed(struct feat_fd *ff, in process_compressed() argument
3227 if (do_read_u32(ff, &(ff->ph->env.comp_ver))) in process_compressed()
3230 if (do_read_u32(ff, &(ff->ph->env.comp_type))) in process_compressed()
3233 if (do_read_u32(ff, &(ff->ph->env.comp_level))) in process_compressed()
3236 if (do_read_u32(ff, &(ff->ph->env.comp_ratio))) in process_compressed()
3239 if (do_read_u32(ff, &(ff->ph->env.comp_mmap_len))) in process_compressed()
3245 static int __process_pmu_caps(struct feat_fd *ff, int *nr_caps, in __process_pmu_caps() argument
3256 if (do_read_u32(ff, &nr_pmu_caps)) in __process_pmu_caps()
3267 name = do_read_string(ff); in __process_pmu_caps()
3271 value = do_read_string(ff); in __process_pmu_caps()
3308 static int process_cpu_pmu_caps(struct feat_fd *ff, in process_cpu_pmu_caps() argument
3311 int ret = __process_pmu_caps(ff, &ff->ph->env.nr_cpu_pmu_caps, in process_cpu_pmu_caps()
3312 &ff->ph->env.cpu_pmu_caps, in process_cpu_pmu_caps()
3313 &ff->ph->env.max_branches, in process_cpu_pmu_caps()
3314 &ff->ph->env.br_cntr_nr, in process_cpu_pmu_caps()
3315 &ff->ph->env.br_cntr_width); in process_cpu_pmu_caps()
3317 if (!ret && !ff->ph->env.cpu_pmu_caps) in process_cpu_pmu_caps()
3322 static int process_pmu_caps(struct feat_fd *ff, void *data __maybe_unused) in process_pmu_caps() argument
3329 if (do_read_u32(ff, &nr_pmu)) in process_pmu_caps()
3342 ret = __process_pmu_caps(ff, &pmu_caps[i].nr_caps, in process_pmu_caps()
3350 pmu_caps[i].pmu_name = do_read_string(ff); in process_pmu_caps()
3361 ff->ph->env.nr_pmus_with_caps = nr_pmu; in process_pmu_caps()
3362 ff->ph->env.pmu_caps = pmu_caps; in process_pmu_caps()
3454 struct feat_fd ff; in perf_file_section__fprintf_info() local
3468 ff = (struct feat_fd) { in perf_file_section__fprintf_info()
3474 feat_ops[feat].print(&ff, hd->fp); in perf_file_section__fprintf_info()
3524 struct feat_fd *ff; member
3531 return do_write(h->ff, buf, sz); in feat_writer_cb()
3534 static int do_write_feat(struct feat_fd *ff, int type, in do_write_feat() argument
3542 if (perf_header__has_feat(ff->ph, type)) { in do_write_feat()
3546 if (WARN(ff->buf, "Error: calling %s in pipe-mode.\n", __func__)) in do_write_feat()
3549 (*p)->offset = lseek(ff->fd, 0, SEEK_CUR); in do_write_feat()
3558 .ff = ff, in do_write_feat()
3567 err = feat_ops[type].write(ff, evlist); in do_write_feat()
3572 lseek(ff->fd, (*p)->offset, SEEK_SET); in do_write_feat()
3576 (*p)->size = lseek(ff->fd, 0, SEEK_CUR) - (*p)->offset; in do_write_feat()
3587 struct feat_fd ff = { in perf_header__adds_write() local
3611 if (do_write_feat(&ff, feat, &p, evlist, fc)) in perf_header__adds_write()
3620 err = do_write(&ff, feat_sec, sec_size); in perf_header__adds_write()
3623 free(ff.buf); /* TODO: added to silence clang-tidy. */ in perf_header__adds_write()
3631 struct feat_fd ff = { in perf_header__write_pipe() local
3641 err = do_write(&ff, &f_header, sizeof(f_header)); in perf_header__write_pipe()
3646 free(ff.buf); in perf_header__write_pipe()
3659 struct feat_fd ff = { in perf_session__do_write_header() local
3688 err = do_write(&ff, evsel->core.id, evsel->core.ids * sizeof(u64)); in perf_session__do_write_header()
3715 err = do_write(&ff, &f_attr, sizeof(f_attr)); in perf_session__do_write_header()
3757 err = do_write(&ff, &f_header, sizeof(f_header)); in perf_session__do_write_header()
3766 free(ff.buf); in perf_session__do_write_header()
4342 struct feat_fd ff = { .fd = 0 }; in perf_event__process_feature() local
4360 ff.buf = (void *)fe->data; in perf_event__process_feature()
4361 ff.size = event->header.size - sizeof(*fe); in perf_event__process_feature()
4362 ff.ph = &session->header; in perf_event__process_feature()
4364 if (feat_ops[feat].process(&ff, NULL)) { in perf_event__process_feature()
4374 feat_ops[feat].print(&ff, stdout); in perf_event__process_feature()
4380 free_event_desc(ff.events); in perf_event__process_feature()