Home
last modified time | relevance | path

Searched refs:pt (Results 1 – 25 of 346) sorted by relevance

12345678910>>...14

/kernel/linux/linux-5.10/tools/perf/util/
Dintel-pt.c146 struct intel_pt *pt; member
185 static void intel_pt_dump(struct intel_pt *pt __maybe_unused, in intel_pt_dump()
225 static void intel_pt_dump_event(struct intel_pt *pt, unsigned char *buf, in intel_pt_dump_event() argument
229 intel_pt_dump(pt, buf, len); in intel_pt_dump_event()
245 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, in intel_pt_dump_sample() local
249 intel_pt_dump(pt, sample->aux_sample.data, sample->aux_sample.size); in intel_pt_dump_sample()
252 static bool intel_pt_log_events(struct intel_pt *pt, u64 tm) in intel_pt_log_events() argument
254 struct perf_time_interval *range = pt->synth_opts.ptime_range; in intel_pt_log_events()
255 int n = pt->synth_opts.range_num; in intel_pt_log_events()
257 if (pt->synth_opts.log_plus_flags & AUXTRACE_LOG_FLG_ALL_PERF_EVTS) in intel_pt_log_events()
[all …]
/kernel/linux/linux-5.10/sound/i2c/other/
Dpt2258.c32 int snd_pt2258_reset(struct snd_pt2258 *pt) in snd_pt2258_reset() argument
39 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset()
40 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) in snd_pt2258_reset()
42 snd_i2c_unlock(pt->i2c_bus); in snd_pt2258_reset()
45 pt->mute = 1; in snd_pt2258_reset()
47 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset()
48 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) in snd_pt2258_reset()
50 snd_i2c_unlock(pt->i2c_bus); in snd_pt2258_reset()
54 pt->volume[i] = 0; in snd_pt2258_reset()
57 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset()
[all …]
/kernel/linux/linux-5.10/arch/ia64/kernel/
Dsignal.c59 err |= __get_user(scr->pt.ar_unat, &sc->sc_ar_unat); in restore_sigcontext()
60 err |= __get_user(scr->pt.ar_fpsr, &sc->sc_ar_fpsr); in restore_sigcontext()
61 err |= __get_user(scr->pt.ar_pfs, &sc->sc_ar_pfs); in restore_sigcontext()
62 err |= __get_user(scr->pt.pr, &sc->sc_pr); /* predicates */ in restore_sigcontext()
63 err |= __get_user(scr->pt.b0, &sc->sc_br[0]); /* b0 (rp) */ in restore_sigcontext()
64 err |= __get_user(scr->pt.b6, &sc->sc_br[6]); /* b6 */ in restore_sigcontext()
65 err |= __copy_from_user(&scr->pt.r1, &sc->sc_gr[1], 8); /* r1 */ in restore_sigcontext()
66 err |= __copy_from_user(&scr->pt.r8, &sc->sc_gr[8], 4*8); /* r8-r11 */ in restore_sigcontext()
67 err |= __copy_from_user(&scr->pt.r12, &sc->sc_gr[12], 2*8); /* r12-r13 */ in restore_sigcontext()
68 err |= __copy_from_user(&scr->pt.r15, &sc->sc_gr[15], 8); /* r15 */ in restore_sigcontext()
[all …]
Dptrace.c63 in_syscall (struct pt_regs *pt) in in_syscall() argument
65 return (long) pt->cr_ifs >= 0; in in_syscall()
73 ia64_get_scratch_nat_bits (struct pt_regs *pt, unsigned long scratch_unat) in ia64_get_scratch_nat_bits() argument
77 unsigned long bit = ia64_unat_pos(&pt->r##first); \ in ia64_get_scratch_nat_bits()
112 ia64_put_scratch_nat_bits (struct pt_regs *pt, unsigned long nat) in ia64_put_scratch_nat_bits() argument
116 unsigned long bit = ia64_unat_pos(&pt->r##first); \ in ia64_put_scratch_nat_bits()
256 struct pt_regs *pt; in get_rnat() local
258 pt = task_pt_regs(task); in get_rnat()
260 ubspstore = (unsigned long *) pt->ar_bspstore; in get_rnat()
282 urnat = (pt->ar_rnat & umask); in get_rnat()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
Dintel_ppgtt.c15 struct i915_page_table *pt; in alloc_pt() local
17 pt = kmalloc(sizeof(*pt), I915_GFP_ALLOW_FAIL); in alloc_pt()
18 if (unlikely(!pt)) in alloc_pt()
21 pt->base = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); in alloc_pt()
22 if (IS_ERR(pt->base)) { in alloc_pt()
23 kfree(pt); in alloc_pt()
27 atomic_set(&pt->used, 0); in alloc_pt()
28 return pt; in alloc_pt()
57 pd->pt.base = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); in alloc_pd()
58 if (IS_ERR(pd->pt.base)) { in alloc_pd()
[all …]
Dgen6_ppgtt.c17 const struct i915_page_table *pt) in gen6_write_pde() argument
19 dma_addr_t addr = pt ? px_dma(pt) : px_dma(ppgtt->base.vm.scratch[1]); in gen6_write_pde()
90 struct i915_page_table * const pt = in gen6_ppgtt_clear_range() local
97 GEM_BUG_ON(count > atomic_read(&pt->used)); in gen6_ppgtt_clear_range()
98 if (!atomic_sub_return(count, &pt->used)) in gen6_ppgtt_clear_range()
108 vaddr = kmap_atomic_px(pt); in gen6_ppgtt_clear_range()
161 struct i915_page_table *pt; in gen6_flush_pd() local
169 gen6_for_each_pde(pt, pd, start, end, pde) in gen6_flush_pd()
170 gen6_write_pde(ppgtt, pde, pt); in gen6_flush_pd()
186 struct i915_page_table *pt; in gen6_alloc_va_range() local
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
Dbase.c34 struct nvkm_mmu_pt *pt; member
42 nvkm_mmu_ptp_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt *pt) in nvkm_mmu_ptp_put() argument
44 const int slot = pt->base >> pt->ptp->shift; in nvkm_mmu_ptp_put()
45 struct nvkm_mmu_ptp *ptp = pt->ptp; in nvkm_mmu_ptp_put()
56 nvkm_mmu_ptc_put(mmu, force, &ptp->pt); in nvkm_mmu_ptp_put()
61 kfree(pt); in nvkm_mmu_ptp_put()
67 struct nvkm_mmu_pt *pt; in nvkm_mmu_ptp_get() local
71 if (!(pt = kzalloc(sizeof(*pt), GFP_KERNEL))) in nvkm_mmu_ptp_get()
78 kfree(pt); in nvkm_mmu_ptp_get()
82 ptp->pt = nvkm_mmu_ptc_get(mmu, 0x1000, 0x1000, false); in nvkm_mmu_ptp_get()
[all …]
Dvmmnv44.c27 nv44_vmm_pgt_fill(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv44_vmm_pgt_fill() argument
33 tmp[0] = nvkm_ro32(pt->memory, pteo + 0x0); in nv44_vmm_pgt_fill()
34 tmp[1] = nvkm_ro32(pt->memory, pteo + 0x4); in nv44_vmm_pgt_fill()
35 tmp[2] = nvkm_ro32(pt->memory, pteo + 0x8); in nv44_vmm_pgt_fill()
36 tmp[3] = nvkm_ro32(pt->memory, pteo + 0xc); in nv44_vmm_pgt_fill()
66 VMM_WO032(pt, vmm, pteo + 0x0, tmp[0]); in nv44_vmm_pgt_fill()
67 VMM_WO032(pt, vmm, pteo + 0x4, tmp[1]); in nv44_vmm_pgt_fill()
68 VMM_WO032(pt, vmm, pteo + 0x8, tmp[2]); in nv44_vmm_pgt_fill()
69 VMM_WO032(pt, vmm, pteo + 0xc, tmp[3] | 0x40000000); in nv44_vmm_pgt_fill()
73 nv44_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv44_vmm_pgt_pte() argument
[all …]
Dvmmgp100.c35 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in gp100_vmm_pfn_unmap() argument
40 nvkm_kmap(pt->memory); in gp100_vmm_pfn_unmap()
42 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0); in gp100_vmm_pfn_unmap()
43 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4); in gp100_vmm_pfn_unmap()
51 nvkm_done(pt->memory); in gp100_vmm_pfn_unmap()
56 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in gp100_vmm_pfn_clear() argument
59 nvkm_kmap(pt->memory); in gp100_vmm_pfn_clear()
61 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0); in gp100_vmm_pfn_clear()
62 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4); in gp100_vmm_pfn_clear()
65 VMM_WO064(pt, vmm, ptei * 8, data & ~BIT_ULL(0)); in gp100_vmm_pfn_clear()
[all …]
Dvmmnv41.c27 nv41_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_pte() argument
32 VMM_WO032(pt, vmm, ptei++ * 4, data); in nv41_vmm_pgt_pte()
38 nv41_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_sgl() argument
41 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte); in nv41_vmm_pgt_sgl()
45 nv41_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_dma() argument
49 nvkm_kmap(pt->memory); in nv41_vmm_pgt_dma()
52 VMM_WO032(pt, vmm, ptei++ * 4, data); in nv41_vmm_pgt_dma()
54 nvkm_done(pt->memory); in nv41_vmm_pgt_dma()
56 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte); in nv41_vmm_pgt_dma()
62 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in nv41_vmm_pgt_unmap() argument
[all …]
Dvmmgf100.c32 gf100_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in gf100_vmm_pgt_pte() argument
44 VMM_WO064(pt, vmm, ptei++ * 8, data); in gf100_vmm_pgt_pte()
51 VMM_WO064(pt, vmm, ptei++ * 8, data); in gf100_vmm_pgt_pte()
58 gf100_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in gf100_vmm_pgt_sgl() argument
61 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte); in gf100_vmm_pgt_sgl()
65 gf100_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in gf100_vmm_pgt_dma() argument
70 nvkm_kmap(pt->memory); in gf100_vmm_pgt_dma()
73 VMM_WO064(pt, vmm, ptei++ * 8, data); in gf100_vmm_pgt_dma()
76 nvkm_done(pt->memory); in gf100_vmm_pgt_dma()
80 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte); in gf100_vmm_pgt_dma()
[all …]
Dvmmnv04.c28 nv04_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv04_vmm_pgt_pte() argument
33 VMM_WO032(pt, vmm, 8 + ptei++ * 4, data); in nv04_vmm_pgt_pte()
39 nv04_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv04_vmm_pgt_sgl() argument
42 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv04_vmm_pgt_pte); in nv04_vmm_pgt_sgl()
46 nv04_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv04_vmm_pgt_dma() argument
50 nvkm_kmap(pt->memory); in nv04_vmm_pgt_dma()
52 VMM_WO032(pt, vmm, 8 + (ptei++ * 4), *map->dma++ | 0x00000003); in nv04_vmm_pgt_dma()
53 nvkm_done(pt->memory); in nv04_vmm_pgt_dma()
55 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv04_vmm_pgt_pte); in nv04_vmm_pgt_dma()
61 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in nv04_vmm_pgt_unmap() argument
[all …]
/kernel/linux/linux-5.10/arch/sparc/lib/
DMemcpy_utils.S20 ba,pt %xcc, __restore_asi
24 ba,pt %xcc, __restore_asi
28 ba,pt %xcc, __restore_asi
32 ba,pt %xcc, __restore_asi
36 ba,pt %xcc, __restore_asi
40 ba,pt %xcc, __restore_asi
44 ba,pt %xcc, __restore_asi
48 ba,pt %xcc, __restore_asi
52 ba,pt %xcc, __restore_asi
57 ba,pt %xcc, __restore_asi
[all …]
DNGmemcpy.S87 ba,pt %xcc, __restore_asi
91 ba,pt %xcc, __restore_asi
96 ba,pt %xcc, __restore_asi
101 ba,pt %xcc, __restore_asi
106 ba,pt %xcc, __restore_asi
111 ba,pt %xcc, __restore_asi
116 ba,pt %xcc, __restore_asi
121 ba,pt %xcc, __restore_asi
126 ba,pt %xcc, __restore_asi
130 ba,pt %xcc, __restore_asi
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/
Dscatterlist.c41 static noinline int expect_pfn_sg(struct pfn_table *pt, in expect_pfn_sg() argument
50 pfn = pt->start; in expect_pfn_sg()
51 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { in expect_pfn_sg()
53 unsigned int npages = npages_fn(n, pt->st.nents, rnd); in expect_pfn_sg()
72 if (pfn != pt->end) { in expect_pfn_sg()
74 __func__, who, pt->end, pfn); in expect_pfn_sg()
81 static noinline int expect_pfn_sg_page_iter(struct pfn_table *pt, in expect_pfn_sg_page_iter() argument
88 pfn = pt->start; in expect_pfn_sg_page_iter()
89 for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) { in expect_pfn_sg_page_iter()
103 if (pfn != pt->end) { in expect_pfn_sg_page_iter()
[all …]
/kernel/linux/linux-5.10/lib/
Dnlattr.c114 void nla_get_range_unsigned(const struct nla_policy *pt, in nla_get_range_unsigned() argument
117 WARN_ON_ONCE(pt->validation_type != NLA_VALIDATE_RANGE_PTR && in nla_get_range_unsigned()
118 (pt->min < 0 || pt->max < 0)); in nla_get_range_unsigned()
122 switch (pt->type) { in nla_get_range_unsigned()
142 switch (pt->validation_type) { in nla_get_range_unsigned()
145 range->min = pt->min; in nla_get_range_unsigned()
146 range->max = pt->max; in nla_get_range_unsigned()
149 *range = *pt->range; in nla_get_range_unsigned()
152 range->min = pt->min; in nla_get_range_unsigned()
155 range->max = pt->max; in nla_get_range_unsigned()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/gma500/
Dmmu.c231 static void psb_mmu_free_pt(struct psb_mmu_pt *pt) in psb_mmu_free_pt() argument
233 __free_page(pt->p); in psb_mmu_free_pt()
234 kfree(pt); in psb_mmu_free_pt()
242 struct psb_mmu_pt *pt; in psb_mmu_free_pagedir() local
255 pt = pd->tables[i]; in psb_mmu_free_pagedir()
256 if (pt) in psb_mmu_free_pagedir()
257 psb_mmu_free_pt(pt); in psb_mmu_free_pagedir()
270 struct psb_mmu_pt *pt = kmalloc(sizeof(*pt), GFP_KERNEL); in psb_mmu_alloc_pt() local
279 if (!pt) in psb_mmu_alloc_pt()
282 pt->p = alloc_page(GFP_DMA32); in psb_mmu_alloc_pt()
[all …]
/kernel/linux/linux-5.10/drivers/dma-buf/
Dsw_sync.c133 struct sync_pt *pt = dma_fence_to_sync_pt(fence); in timeline_fence_release() local
138 if (!list_empty(&pt->link)) { in timeline_fence_release()
139 list_del(&pt->link); in timeline_fence_release()
140 rb_erase(&pt->node, &parent->pt_tree); in timeline_fence_release()
194 struct sync_pt *pt, *next; in sync_timeline_signal() local
202 list_for_each_entry_safe(pt, next, &obj->pt_list, link) { in sync_timeline_signal()
203 if (!timeline_fence_signaled(&pt->base)) in sync_timeline_signal()
206 list_del_init(&pt->link); in sync_timeline_signal()
207 rb_erase(&pt->node, &obj->pt_tree); in sync_timeline_signal()
217 dma_fence_signal_locked(&pt->base); in sync_timeline_signal()
[all …]
/kernel/linux/linux-5.10/net/netlink/
Dpolicy.c199 int netlink_policy_dump_attr_size_estimate(const struct nla_policy *pt) in netlink_policy_dump_attr_size_estimate() argument
204 switch (pt->type) { in netlink_policy_dump_attr_size_estimate()
243 const struct nla_policy *pt, in __netlink_policy_dump_write_attr() argument
246 int estimate = netlink_policy_dump_attr_size_estimate(pt); in __netlink_policy_dump_write_attr()
254 switch (pt->type) { in __netlink_policy_dump_write_attr()
265 if (pt->type == NLA_NESTED_ARRAY) in __netlink_policy_dump_write_attr()
267 if (state && pt->nested_policy && pt->len && in __netlink_policy_dump_write_attr()
270 pt->nested_policy, in __netlink_policy_dump_write_attr()
271 pt->len)) || in __netlink_policy_dump_write_attr()
273 pt->len))) in __netlink_policy_dump_write_attr()
[all …]
/kernel/linux/linux-5.10/arch/alpha/kernel/
Dprocess.c287 dump_elf_thread(elf_greg_t *dest, struct pt_regs *pt, struct thread_info *ti) in dump_elf_thread() argument
290 struct switch_stack * sw = ((struct switch_stack *) pt) - 1; in dump_elf_thread()
292 dest[ 0] = pt->r0; in dump_elf_thread()
293 dest[ 1] = pt->r1; in dump_elf_thread()
294 dest[ 2] = pt->r2; in dump_elf_thread()
295 dest[ 3] = pt->r3; in dump_elf_thread()
296 dest[ 4] = pt->r4; in dump_elf_thread()
297 dest[ 5] = pt->r5; in dump_elf_thread()
298 dest[ 6] = pt->r6; in dump_elf_thread()
299 dest[ 7] = pt->r7; in dump_elf_thread()
[all …]
/kernel/linux/linux-5.10/tools/perf/util/intel-pt-decoder/
DBuild1 perf-$(CONFIG_AUXTRACE) += intel-pt-pkt-decoder.o intel-pt-insn-decoder.o intel-pt-log.o intel-pt-d…
6 $(OUTPUT)util/intel-pt-decoder/inat-tables.c: $(inat_tables_script) $(inat_tables_maps)
12 $(OUTPUT)util/intel-pt-decoder/intel-pt-insn-decoder.o: util/intel-pt-decoder/intel-pt-insn-decoder…
16 CFLAGS_intel-pt-insn-decoder.o += -I$(OUTPUT)util/intel-pt-decoder
19 CFLAGS_intel-pt-insn-decoder.o += -Wno-override-init
/kernel/linux/linux-5.10/arch/x86/events/intel/
Dpt.c28 static DEFINE_PER_CPU(struct pt, pt_ctx);
102 PMU_FORMAT_ATTR(pt, "config:0" );
400 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_start() local
404 if (READ_ONCE(pt->vmx_on)) in pt_config_start()
405 perf_aux_output_flag(&pt->handle, PERF_AUX_FLAG_PARTIAL); in pt_config_start()
443 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_filters() local
465 if (pt->filters.filter[range].msr_a != filter->msr_a) { in pt_config_filters()
467 pt->filters.filter[range].msr_a = filter->msr_a; in pt_config_filters()
470 if (pt->filters.filter[range].msr_b != filter->msr_b) { in pt_config_filters()
472 pt->filters.filter[range].msr_b = filter->msr_b; in pt_config_filters()
[all …]
/kernel/linux/linux-5.10/arch/x86/kernel/
Dvm86_32.c79 #define AL(regs) (((unsigned char *)&((regs)->pt.ax))[0])
80 #define AH(regs) (((unsigned char *)&((regs)->pt.ax))[1])
81 #define IP(regs) (*(unsigned short *)&((regs)->pt.ip))
82 #define SP(regs) (*(unsigned short *)&((regs)->pt.sp))
113 set_flags(regs->pt.flags, VEFLAGS, X86_EFLAGS_VIF | vm86->veflags_mask); in save_v86_state()
121 unsafe_put_user(regs->pt.bx, &user->regs.ebx, Efault_end); in save_v86_state()
122 unsafe_put_user(regs->pt.cx, &user->regs.ecx, Efault_end); in save_v86_state()
123 unsafe_put_user(regs->pt.dx, &user->regs.edx, Efault_end); in save_v86_state()
124 unsafe_put_user(regs->pt.si, &user->regs.esi, Efault_end); in save_v86_state()
125 unsafe_put_user(regs->pt.di, &user->regs.edi, Efault_end); in save_v86_state()
[all …]
/kernel/linux/linux-5.10/net/core/
Dnet-procfs.c199 struct packet_type *pt = NULL; in ptype_get_idx() local
206 list_for_each_entry_rcu(pt, ptype_list, list) { in ptype_get_idx()
208 return pt; in ptype_get_idx()
213 list_for_each_entry_rcu(pt, &ptype_all, list) { in ptype_get_idx()
215 return pt; in ptype_get_idx()
220 list_for_each_entry_rcu(pt, &ptype_base[t], list) { in ptype_get_idx()
222 return pt; in ptype_get_idx()
239 struct packet_type *pt; in ptype_seq_next() local
247 pt = v; in ptype_seq_next()
248 nxt = pt->list.next; in ptype_seq_next()
[all …]
/kernel/linux/linux-5.10/drivers/media/common/saa7146/
Dsaa7146_core.c164 void *saa7146_vmalloc_build_pgtable(struct pci_dev *pci, long length, struct saa7146_pgtable *pt) in saa7146_vmalloc_build_pgtable() argument
173 if (!(pt->slist = vmalloc_to_sg(mem, pages))) in saa7146_vmalloc_build_pgtable()
176 if (saa7146_pgtable_alloc(pci, pt)) in saa7146_vmalloc_build_pgtable()
179 pt->nents = pages; in saa7146_vmalloc_build_pgtable()
180 slen = pci_map_sg(pci,pt->slist,pt->nents,PCI_DMA_FROMDEVICE); in saa7146_vmalloc_build_pgtable()
184 if (0 != saa7146_pgtable_build_single(pci, pt, pt->slist, slen)) in saa7146_vmalloc_build_pgtable()
190 pci_unmap_sg(pci, pt->slist, pt->nents, PCI_DMA_FROMDEVICE); in saa7146_vmalloc_build_pgtable()
192 saa7146_pgtable_free(pci, pt); in saa7146_vmalloc_build_pgtable()
194 kfree(pt->slist); in saa7146_vmalloc_build_pgtable()
195 pt->slist = NULL; in saa7146_vmalloc_build_pgtable()
[all …]

12345678910>>...14