/Linux-v5.4/tools/perf/util/ |
D | intel-pt.c | 138 struct intel_pt *pt; member 179 static void intel_pt_dump(struct intel_pt *pt __maybe_unused, in intel_pt_dump() 219 static void intel_pt_dump_event(struct intel_pt *pt, unsigned char *buf, in intel_pt_dump_event() argument 223 intel_pt_dump(pt, buf, len); in intel_pt_dump_event() 236 static int intel_pt_do_fix_overlap(struct intel_pt *pt, struct auxtrace_buffer *a, in intel_pt_do_fix_overlap() argument 243 pt->have_tsc, &consecutive); in intel_pt_do_fix_overlap() 261 int fd = perf_data__fd(ptq->pt->session->data); in intel_pt_get_buffer() 268 might_overlap = ptq->pt->snapshot_mode || ptq->pt->sampling_mode; in intel_pt_get_buffer() 270 intel_pt_do_fix_overlap(ptq->pt, old_buffer, buffer)) in intel_pt_get_buffer() 312 queue = &ptq->pt->queues.queue_array[ptq->queue_nr]; in intel_pt_lookahead() [all …]
|
/Linux-v5.4/sound/i2c/other/ |
D | pt2258.c | 32 int snd_pt2258_reset(struct snd_pt2258 *pt) in snd_pt2258_reset() argument 39 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() 40 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) in snd_pt2258_reset() 42 snd_i2c_unlock(pt->i2c_bus); in snd_pt2258_reset() 45 pt->mute = 1; in snd_pt2258_reset() 47 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() 48 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) in snd_pt2258_reset() 50 snd_i2c_unlock(pt->i2c_bus); in snd_pt2258_reset() 54 pt->volume[i] = 0; in snd_pt2258_reset() 57 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() [all …]
|
/Linux-v5.4/arch/ia64/kernel/ |
D | signal.c | 59 err |= __get_user(scr->pt.ar_unat, &sc->sc_ar_unat); in restore_sigcontext() 60 err |= __get_user(scr->pt.ar_fpsr, &sc->sc_ar_fpsr); in restore_sigcontext() 61 err |= __get_user(scr->pt.ar_pfs, &sc->sc_ar_pfs); in restore_sigcontext() 62 err |= __get_user(scr->pt.pr, &sc->sc_pr); /* predicates */ in restore_sigcontext() 63 err |= __get_user(scr->pt.b0, &sc->sc_br[0]); /* b0 (rp) */ in restore_sigcontext() 64 err |= __get_user(scr->pt.b6, &sc->sc_br[6]); /* b6 */ in restore_sigcontext() 65 err |= __copy_from_user(&scr->pt.r1, &sc->sc_gr[1], 8); /* r1 */ in restore_sigcontext() 66 err |= __copy_from_user(&scr->pt.r8, &sc->sc_gr[8], 4*8); /* r8-r11 */ in restore_sigcontext() 67 err |= __copy_from_user(&scr->pt.r12, &sc->sc_gr[12], 2*8); /* r12-r13 */ in restore_sigcontext() 68 err |= __copy_from_user(&scr->pt.r15, &sc->sc_gr[15], 8); /* r15 */ in restore_sigcontext() [all …]
|
D | ptrace.c | 67 in_syscall (struct pt_regs *pt) in in_syscall() argument 69 return (long) pt->cr_ifs >= 0; in in_syscall() 77 ia64_get_scratch_nat_bits (struct pt_regs *pt, unsigned long scratch_unat) in ia64_get_scratch_nat_bits() argument 81 unsigned long bit = ia64_unat_pos(&pt->r##first); \ in ia64_get_scratch_nat_bits() 116 ia64_put_scratch_nat_bits (struct pt_regs *pt, unsigned long nat) in ia64_put_scratch_nat_bits() argument 120 unsigned long bit = ia64_unat_pos(&pt->r##first); \ in ia64_put_scratch_nat_bits() 260 struct pt_regs *pt; in get_rnat() local 262 pt = task_pt_regs(task); in get_rnat() 264 ubspstore = (unsigned long *) pt->ar_bspstore; in get_rnat() 286 urnat = (pt->ar_rnat & umask); in get_rnat() [all …]
|
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
D | base.c | 34 struct nvkm_mmu_pt *pt; member 42 nvkm_mmu_ptp_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt *pt) in nvkm_mmu_ptp_put() argument 44 const int slot = pt->base >> pt->ptp->shift; in nvkm_mmu_ptp_put() 45 struct nvkm_mmu_ptp *ptp = pt->ptp; in nvkm_mmu_ptp_put() 56 nvkm_mmu_ptc_put(mmu, force, &ptp->pt); in nvkm_mmu_ptp_put() 61 kfree(pt); in nvkm_mmu_ptp_put() 67 struct nvkm_mmu_pt *pt; in nvkm_mmu_ptp_get() local 71 if (!(pt = kzalloc(sizeof(*pt), GFP_KERNEL))) in nvkm_mmu_ptp_get() 78 kfree(pt); in nvkm_mmu_ptp_get() 82 ptp->pt = nvkm_mmu_ptc_get(mmu, 0x1000, 0x1000, false); in nvkm_mmu_ptp_get() [all …]
|
D | vmmnv44.c | 27 nv44_vmm_pgt_fill(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv44_vmm_pgt_fill() argument 33 tmp[0] = nvkm_ro32(pt->memory, pteo + 0x0); in nv44_vmm_pgt_fill() 34 tmp[1] = nvkm_ro32(pt->memory, pteo + 0x4); in nv44_vmm_pgt_fill() 35 tmp[2] = nvkm_ro32(pt->memory, pteo + 0x8); in nv44_vmm_pgt_fill() 36 tmp[3] = nvkm_ro32(pt->memory, pteo + 0xc); in nv44_vmm_pgt_fill() 66 VMM_WO032(pt, vmm, pteo + 0x0, tmp[0]); in nv44_vmm_pgt_fill() 67 VMM_WO032(pt, vmm, pteo + 0x4, tmp[1]); in nv44_vmm_pgt_fill() 68 VMM_WO032(pt, vmm, pteo + 0x8, tmp[2]); in nv44_vmm_pgt_fill() 69 VMM_WO032(pt, vmm, pteo + 0xc, tmp[3] | 0x40000000); in nv44_vmm_pgt_fill() 73 nv44_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv44_vmm_pgt_pte() argument [all …]
|
D | vmmgp100.c | 35 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in gp100_vmm_pfn_unmap() argument 40 nvkm_kmap(pt->memory); in gp100_vmm_pfn_unmap() 42 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0); in gp100_vmm_pfn_unmap() 43 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4); in gp100_vmm_pfn_unmap() 51 nvkm_done(pt->memory); in gp100_vmm_pfn_unmap() 56 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in gp100_vmm_pfn_clear() argument 59 nvkm_kmap(pt->memory); in gp100_vmm_pfn_clear() 61 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0); in gp100_vmm_pfn_clear() 62 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4); in gp100_vmm_pfn_clear() 65 VMM_WO064(pt, vmm, ptei * 8, data & ~BIT_ULL(0)); in gp100_vmm_pfn_clear() [all …]
|
D | vmmnv41.c | 27 nv41_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_pte() argument 32 VMM_WO032(pt, vmm, ptei++ * 4, data); in nv41_vmm_pgt_pte() 38 nv41_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_sgl() argument 41 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte); in nv41_vmm_pgt_sgl() 45 nv41_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_dma() argument 49 nvkm_kmap(pt->memory); in nv41_vmm_pgt_dma() 52 VMM_WO032(pt, vmm, ptei++ * 4, data); in nv41_vmm_pgt_dma() 54 nvkm_done(pt->memory); in nv41_vmm_pgt_dma() 56 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte); in nv41_vmm_pgt_dma() 62 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in nv41_vmm_pgt_unmap() argument [all …]
|
D | vmmgf100.c | 32 gf100_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in gf100_vmm_pgt_pte() argument 44 VMM_WO064(pt, vmm, ptei++ * 8, data); in gf100_vmm_pgt_pte() 51 VMM_WO064(pt, vmm, ptei++ * 8, data); in gf100_vmm_pgt_pte() 58 gf100_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in gf100_vmm_pgt_sgl() argument 61 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte); in gf100_vmm_pgt_sgl() 65 gf100_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in gf100_vmm_pgt_dma() argument 70 nvkm_kmap(pt->memory); in gf100_vmm_pgt_dma() 73 VMM_WO064(pt, vmm, ptei++ * 8, data); in gf100_vmm_pgt_dma() 76 nvkm_done(pt->memory); in gf100_vmm_pgt_dma() 80 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte); in gf100_vmm_pgt_dma() [all …]
|
D | vmmnv04.c | 28 nv04_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv04_vmm_pgt_pte() argument 33 VMM_WO032(pt, vmm, 8 + ptei++ * 4, data); in nv04_vmm_pgt_pte() 39 nv04_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv04_vmm_pgt_sgl() argument 42 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv04_vmm_pgt_pte); in nv04_vmm_pgt_sgl() 46 nv04_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv04_vmm_pgt_dma() argument 50 nvkm_kmap(pt->memory); in nv04_vmm_pgt_dma() 52 VMM_WO032(pt, vmm, 8 + (ptei++ * 4), *map->dma++ | 0x00000003); in nv04_vmm_pgt_dma() 53 nvkm_done(pt->memory); in nv04_vmm_pgt_dma() 55 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv04_vmm_pgt_pte); in nv04_vmm_pgt_dma() 61 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in nv04_vmm_pgt_unmap() argument [all …]
|
/Linux-v5.4/arch/sparc/lib/ |
D | Memcpy_utils.S | 20 ba,pt %xcc, __restore_asi 24 ba,pt %xcc, __restore_asi 28 ba,pt %xcc, __restore_asi 32 ba,pt %xcc, __restore_asi 36 ba,pt %xcc, __restore_asi 40 ba,pt %xcc, __restore_asi 44 ba,pt %xcc, __restore_asi 48 ba,pt %xcc, __restore_asi 52 ba,pt %xcc, __restore_asi 57 ba,pt %xcc, __restore_asi [all …]
|
D | NGmemcpy.S | 87 ba,pt %xcc, __restore_asi 91 ba,pt %xcc, __restore_asi 96 ba,pt %xcc, __restore_asi 101 ba,pt %xcc, __restore_asi 106 ba,pt %xcc, __restore_asi 111 ba,pt %xcc, __restore_asi 116 ba,pt %xcc, __restore_asi 121 ba,pt %xcc, __restore_asi 126 ba,pt %xcc, __restore_asi 130 ba,pt %xcc, __restore_asi [all …]
|
D | NG2memcpy.S | 188 ba,pt %xcc, __restore_asi 192 ba,pt %xcc, __restore_asi 196 ba,pt %xcc, __restore_asi 200 ba,pt %xcc, __restore_asi 205 ba,pt %xcc, __restore_asi 210 ba,pt %xcc, __restore_asi 215 ba,pt %xcc, __restore_asi 219 ba,pt %xcc, __restore_fp 224 ba,pt %xcc, __restore_fp 229 ba,pt %xcc, __restore_asi [all …]
|
/Linux-v5.4/drivers/gpu/drm/i915/selftests/ |
D | scatterlist.c | 41 static noinline int expect_pfn_sg(struct pfn_table *pt, in expect_pfn_sg() argument 50 pfn = pt->start; in expect_pfn_sg() 51 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { in expect_pfn_sg() 53 unsigned int npages = npages_fn(n, pt->st.nents, rnd); in expect_pfn_sg() 72 if (pfn != pt->end) { in expect_pfn_sg() 74 __func__, who, pt->end, pfn); in expect_pfn_sg() 81 static noinline int expect_pfn_sg_page_iter(struct pfn_table *pt, in expect_pfn_sg_page_iter() argument 88 pfn = pt->start; in expect_pfn_sg_page_iter() 89 for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) { in expect_pfn_sg_page_iter() 103 if (pfn != pt->end) { in expect_pfn_sg_page_iter() [all …]
|
/Linux-v5.4/drivers/gpu/drm/gma500/ |
D | mmu.c | 231 static void psb_mmu_free_pt(struct psb_mmu_pt *pt) in psb_mmu_free_pt() argument 233 __free_page(pt->p); in psb_mmu_free_pt() 234 kfree(pt); in psb_mmu_free_pt() 242 struct psb_mmu_pt *pt; in psb_mmu_free_pagedir() local 255 pt = pd->tables[i]; in psb_mmu_free_pagedir() 256 if (pt) in psb_mmu_free_pagedir() 257 psb_mmu_free_pt(pt); in psb_mmu_free_pagedir() 270 struct psb_mmu_pt *pt = kmalloc(sizeof(*pt), GFP_KERNEL); in psb_mmu_alloc_pt() local 279 if (!pt) in psb_mmu_alloc_pt() 282 pt->p = alloc_page(GFP_DMA32); in psb_mmu_alloc_pt() [all …]
|
/Linux-v5.4/drivers/dma-buf/ |
D | sw_sync.c | 133 struct sync_pt *pt = dma_fence_to_sync_pt(fence); in timeline_fence_release() local 138 if (!list_empty(&pt->link)) { in timeline_fence_release() 139 list_del(&pt->link); in timeline_fence_release() 140 rb_erase(&pt->node, &parent->pt_tree); in timeline_fence_release() 194 struct sync_pt *pt, *next; in sync_timeline_signal() local 202 list_for_each_entry_safe(pt, next, &obj->pt_list, link) { in sync_timeline_signal() 203 if (!timeline_fence_signaled(&pt->base)) in sync_timeline_signal() 206 list_del_init(&pt->link); in sync_timeline_signal() 207 rb_erase(&pt->node, &obj->pt_tree); in sync_timeline_signal() 217 dma_fence_signal_locked(&pt->base); in sync_timeline_signal() [all …]
|
/Linux-v5.4/arch/alpha/kernel/ |
D | process.c | 289 dump_elf_thread(elf_greg_t *dest, struct pt_regs *pt, struct thread_info *ti) in dump_elf_thread() argument 292 struct switch_stack * sw = ((struct switch_stack *) pt) - 1; in dump_elf_thread() 294 dest[ 0] = pt->r0; in dump_elf_thread() 295 dest[ 1] = pt->r1; in dump_elf_thread() 296 dest[ 2] = pt->r2; in dump_elf_thread() 297 dest[ 3] = pt->r3; in dump_elf_thread() 298 dest[ 4] = pt->r4; in dump_elf_thread() 299 dest[ 5] = pt->r5; in dump_elf_thread() 300 dest[ 6] = pt->r6; in dump_elf_thread() 301 dest[ 7] = pt->r7; in dump_elf_thread() [all …]
|
/Linux-v5.4/lib/ |
D | nlattr.c | 99 static int nla_validate_int_range(const struct nla_policy *pt, in nla_validate_int_range() argument 106 validate_min = pt->validation_type == NLA_VALIDATE_RANGE || in nla_validate_int_range() 107 pt->validation_type == NLA_VALIDATE_MIN; in nla_validate_int_range() 108 validate_max = pt->validation_type == NLA_VALIDATE_RANGE || in nla_validate_int_range() 109 pt->validation_type == NLA_VALIDATE_MAX; in nla_validate_int_range() 111 switch (pt->type) { in nla_validate_int_range() 135 if ((validate_min && nla_get_u64(nla) < pt->min) || in nla_validate_int_range() 136 (validate_max && nla_get_u64(nla) > pt->max)) { in nla_validate_int_range() 147 if ((validate_min && value < pt->min) || in nla_validate_int_range() 148 (validate_max && value > pt->max)) { in nla_validate_int_range() [all …]
|
/Linux-v5.4/tools/perf/util/intel-pt-decoder/ |
D | Build | 1 perf-$(CONFIG_AUXTRACE) += intel-pt-pkt-decoder.o intel-pt-insn-decoder.o intel-pt-log.o intel-pt-d… 6 $(OUTPUT)util/intel-pt-decoder/inat-tables.c: $(inat_tables_script) $(inat_tables_maps) 12 $(OUTPUT)util/intel-pt-decoder/intel-pt-insn-decoder.o: util/intel-pt-decoder/intel-pt-insn-decoder… 16 CFLAGS_intel-pt-insn-decoder.o += -I$(OUTPUT)util/intel-pt-decoder 19 CFLAGS_intel-pt-insn-decoder.o += -Wno-override-init
|
/Linux-v5.4/arch/x86/kernel/ |
D | vm86_32.c | 79 #define AL(regs) (((unsigned char *)&((regs)->pt.ax))[0]) 80 #define AH(regs) (((unsigned char *)&((regs)->pt.ax))[1]) 81 #define IP(regs) (*(unsigned short *)&((regs)->pt.ip)) 82 #define SP(regs) (*(unsigned short *)&((regs)->pt.sp)) 114 set_flags(regs->pt.flags, VEFLAGS, X86_EFLAGS_VIF | vm86->veflags_mask); in save_v86_state() 125 put_user_ex(regs->pt.bx, &user->regs.ebx); in save_v86_state() 126 put_user_ex(regs->pt.cx, &user->regs.ecx); in save_v86_state() 127 put_user_ex(regs->pt.dx, &user->regs.edx); in save_v86_state() 128 put_user_ex(regs->pt.si, &user->regs.esi); in save_v86_state() 129 put_user_ex(regs->pt.di, &user->regs.edi); in save_v86_state() [all …]
|
/Linux-v5.4/drivers/media/common/saa7146/ |
D | saa7146_core.c | 164 void *saa7146_vmalloc_build_pgtable(struct pci_dev *pci, long length, struct saa7146_pgtable *pt) in saa7146_vmalloc_build_pgtable() argument 173 if (!(pt->slist = vmalloc_to_sg(mem, pages))) in saa7146_vmalloc_build_pgtable() 176 if (saa7146_pgtable_alloc(pci, pt)) in saa7146_vmalloc_build_pgtable() 179 pt->nents = pages; in saa7146_vmalloc_build_pgtable() 180 slen = pci_map_sg(pci,pt->slist,pt->nents,PCI_DMA_FROMDEVICE); in saa7146_vmalloc_build_pgtable() 184 if (0 != saa7146_pgtable_build_single(pci, pt, pt->slist, slen)) in saa7146_vmalloc_build_pgtable() 190 pci_unmap_sg(pci, pt->slist, pt->nents, PCI_DMA_FROMDEVICE); in saa7146_vmalloc_build_pgtable() 192 saa7146_pgtable_free(pci, pt); in saa7146_vmalloc_build_pgtable() 194 kfree(pt->slist); in saa7146_vmalloc_build_pgtable() 195 pt->slist = NULL; in saa7146_vmalloc_build_pgtable() [all …]
|
/Linux-v5.4/arch/x86/events/intel/ |
D | pt.c | 28 static DEFINE_PER_CPU(struct pt, pt_ctx); 102 PMU_FORMAT_ATTR(pt, "config:0" ); 431 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_filters() local 453 if (pt->filters.filter[range].msr_a != filter->msr_a) { in pt_config_filters() 455 pt->filters.filter[range].msr_a = filter->msr_a; in pt_config_filters() 458 if (pt->filters.filter[range].msr_b != filter->msr_b) { in pt_config_filters() 460 pt->filters.filter[range].msr_b = filter->msr_b; in pt_config_filters() 471 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config() local 504 if (READ_ONCE(pt->vmx_on)) in pt_config() 505 perf_aux_output_flag(&pt->handle, PERF_AUX_FLAG_PARTIAL); in pt_config() [all …]
|
/Linux-v5.4/arch/sparc/include/asm/ |
D | ttable.h | 27 ba,pt %xcc, etrap; \ 31 ba,pt %xcc, rtrap; \ 37 ba,pt %xcc, etrap; \ 41 ba,pt %xcc, rtrap; \ 46 ba,pt %xcc, do_fptrap; \ 50 ba,pt %xcc, rtrap; \ 55 ba,pt %xcc, routine; \ 60 ba,pt %xcc, routine; \ 66 ba,pt %xcc, etraptl1; \ 70 ba,pt %xcc, rtrap; \ [all …]
|
/Linux-v5.4/net/core/ |
D | net-procfs.c | 187 struct packet_type *pt = NULL; in ptype_get_idx() local 191 list_for_each_entry_rcu(pt, &ptype_all, list) { in ptype_get_idx() 193 return pt; in ptype_get_idx() 198 list_for_each_entry_rcu(pt, &ptype_base[t], list) { in ptype_get_idx() 200 return pt; in ptype_get_idx() 216 struct packet_type *pt; in ptype_seq_next() local 224 pt = v; in ptype_seq_next() 225 nxt = pt->list.next; in ptype_seq_next() 226 if (pt->type == htons(ETH_P_ALL)) { in ptype_seq_next() 232 hash = ntohs(pt->type) & PTYPE_HASH_MASK; in ptype_seq_next() [all …]
|
/Linux-v5.4/tools/lib/bpf/ |
D | nlattr.c | 47 struct libbpf_nla_policy *pt; in validate_nla() local 54 pt = &policy[type]; in validate_nla() 56 if (pt->type > LIBBPF_NLA_TYPE_MAX) in validate_nla() 59 if (pt->minlen) in validate_nla() 60 minlen = pt->minlen; in validate_nla() 61 else if (pt->type != LIBBPF_NLA_UNSPEC) in validate_nla() 62 minlen = nla_attr_minlen[pt->type]; in validate_nla() 67 if (pt->maxlen && libbpf_nla_len(nla) > pt->maxlen) in validate_nla() 70 if (pt->type == LIBBPF_NLA_STRING) { in validate_nla()
|