Home
last modified time | relevance | path

Searched refs:sg (Results 1 – 25 of 42) sorted by relevance

12

/include/linux/
Dscatterlist.h31 #define sg_dma_address(sg) ((sg)->dma_address) argument
34 #define sg_dma_len(sg) ((sg)->dma_length) argument
36 #define sg_dma_len(sg) ((sg)->length) argument
77 static inline unsigned int __sg_flags(struct scatterlist *sg) in __sg_flags() argument
79 return sg->page_link & SG_PAGE_LINK_MASK; in __sg_flags()
82 static inline struct scatterlist *sg_chain_ptr(struct scatterlist *sg) in sg_chain_ptr() argument
84 return (struct scatterlist *)(sg->page_link & ~SG_PAGE_LINK_MASK); in sg_chain_ptr()
87 static inline bool sg_is_chain(struct scatterlist *sg) in sg_is_chain() argument
89 return __sg_flags(sg) & SG_CHAIN; in sg_is_chain()
92 static inline bool sg_is_last(struct scatterlist *sg) in sg_is_last() argument
[all …]
Dskmsg.h44 struct sk_msg_sg sg; member
148 WARN_ON(i == msg->sg.end && bytes); in sk_msg_check_to_free()
182 sk_msg_iter_var_prev(msg->sg.which)
185 sk_msg_iter_var_next(msg->sg.which)
189 BUILD_BUG_ON(ARRAY_SIZE(msg->sg.data) - 1 != NR_MSG_FRAG_IDS); in sk_msg_init()
191 sg_init_marker(msg->sg.data, NR_MSG_FRAG_IDS); in sk_msg_init()
197 dst->sg.data[which] = src->sg.data[which]; in sk_msg_xfer()
198 dst->sg.data[which].length = size; in sk_msg_xfer()
199 dst->sg.size += size; in sk_msg_xfer()
200 src->sg.size -= size; in sk_msg_xfer()
[all …]
Dpage_reporting.h14 struct scatterlist *sg, unsigned int nents);
Ddma-mapping.h125 unsigned int dma_map_sg_attrs(struct device *dev, struct scatterlist *sg,
127 void dma_unmap_sg_attrs(struct device *dev, struct scatterlist *sg,
180 struct scatterlist *sg, int nents, enum dma_data_direction dir, in dma_map_sg_attrs() argument
186 struct scatterlist *sg, int nents, enum dma_data_direction dir, in dma_unmap_sg_attrs() argument
369 void __dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg,
371 void __dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg,
396 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument
399 __dma_sync_sg_for_cpu(dev, sg, nelems, dir); in dma_sync_sg_for_cpu()
403 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument
406 __dma_sync_sg_for_device(dev, sg, nelems, dir); in dma_sync_sg_for_device()
[all …]
Diommu-dma.h29 int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg, int nents,
31 void iommu_dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nents,
Dtifm.h144 int tifm_map_sg(struct tifm_dev *sock, struct scatterlist *sg, int nents,
146 void tifm_unmap_sg(struct tifm_dev *sock, struct scatterlist *sg, int nents,
Ddma-map-ops.h46 int (*map_sg)(struct device *dev, struct scatterlist *sg, int nents,
48 void (*unmap_sg)(struct device *dev, struct scatterlist *sg, int nents,
61 void (*sync_sg_for_cpu)(struct device *dev, struct scatterlist *sg,
63 void (*sync_sg_for_device)(struct device *dev, struct scatterlist *sg,
406 bool arch_dma_map_sg_direct(struct device *dev, struct scatterlist *sg,
408 bool arch_dma_unmap_sg_direct(struct device *dev, struct scatterlist *sg,
Dvirtio.h44 struct scatterlist sg[], unsigned int num,
49 struct scatterlist sg[], unsigned int num,
54 struct scatterlist sg[], unsigned int num,
Dkmsan.h209 void kmsan_handle_dma_sg(struct scatterlist *sg, int nents,
380 static inline void kmsan_handle_dma_sg(struct scatterlist *sg, int nents, in kmsan_handle_dma_sg() argument
Dmemstick.h243 struct scatterlist sg; member
320 const struct scatterlist *sg);
Dmpi.h58 int mpi_write_to_sgl(MPI a, struct scatterlist *sg, unsigned nbytes,
/include/crypto/
Dscatterwalk.h21 struct scatterlist *sg, int num) in scatterwalk_crypto_chain() argument
23 if (sg) in scatterwalk_crypto_chain()
24 sg_chain(head, num, sg); in scatterwalk_crypto_chain()
31 unsigned int len = walk->sg->offset + walk->sg->length - walk->offset; in scatterwalk_pagelen()
51 return sg_page(walk->sg) + (walk->offset >> PAGE_SHIFT); in scatterwalk_page()
60 struct scatterlist *sg) in scatterwalk_start() argument
62 walk->sg = sg; in scatterwalk_start()
63 walk->offset = sg->offset; in scatterwalk_start()
78 page = sg_page(walk->sg) + ((walk->offset - 1) >> PAGE_SHIFT); in scatterwalk_pagedone()
82 if (more && walk->offset >= walk->sg->offset + walk->sg->length) in scatterwalk_pagedone()
[all …]
/include/soc/fsl/
Ddpaa2-fd.h310 static inline dma_addr_t dpaa2_sg_get_addr(const struct dpaa2_sg_entry *sg) in dpaa2_sg_get_addr() argument
312 return (dma_addr_t)le64_to_cpu(sg->addr); in dpaa2_sg_get_addr()
320 static inline void dpaa2_sg_set_addr(struct dpaa2_sg_entry *sg, dma_addr_t addr) in dpaa2_sg_set_addr() argument
322 sg->addr = cpu_to_le64(addr); in dpaa2_sg_set_addr()
325 static inline bool dpaa2_sg_short_len(const struct dpaa2_sg_entry *sg) in dpaa2_sg_short_len() argument
327 return !!((le16_to_cpu(sg->format_offset) >> SG_SHORT_LEN_FLAG_SHIFT) in dpaa2_sg_short_len()
337 static inline u32 dpaa2_sg_get_len(const struct dpaa2_sg_entry *sg) in dpaa2_sg_get_len() argument
339 if (dpaa2_sg_short_len(sg)) in dpaa2_sg_get_len()
340 return le32_to_cpu(sg->len) & SG_SHORT_LEN_MASK; in dpaa2_sg_get_len()
342 return le32_to_cpu(sg->len); in dpaa2_sg_get_len()
[all …]
Dqman.h202 static inline dma_addr_t qm_sg_addr(const struct qm_sg_entry *sg) in qm_sg_addr() argument
204 return be64_to_cpu(sg->data) & 0xffffffffffLLU; in qm_sg_addr()
207 static inline u64 qm_sg_entry_get64(const struct qm_sg_entry *sg) in qm_sg_entry_get64() argument
209 return be64_to_cpu(sg->data) & 0xffffffffffLLU; in qm_sg_entry_get64()
212 static inline void qm_sg_entry_set64(struct qm_sg_entry *sg, u64 addr) in qm_sg_entry_set64() argument
214 sg->addr_hi = upper_32_bits(addr); in qm_sg_entry_set64()
215 sg->addr_lo = cpu_to_be32(lower_32_bits(addr)); in qm_sg_entry_set64()
218 static inline bool qm_sg_entry_is_final(const struct qm_sg_entry *sg) in qm_sg_entry_is_final() argument
220 return be32_to_cpu(sg->cfg) & QM_SG_FIN; in qm_sg_entry_is_final()
223 static inline bool qm_sg_entry_is_ext(const struct qm_sg_entry *sg) in qm_sg_entry_is_ext() argument
[all …]
/include/trace/events/
Ddma.h179 struct scatterlist *sg;
183 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i)
184 ((u64 *)__get_dynamic_array(phys_addrs))[i] = sg_phys(sg);
259 struct scatterlist *sg;
263 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i)
264 ((u64 *)__get_dynamic_array(phys_addrs))[i] = sg_phys(sg);
295 struct scatterlist *sg;
299 for_each_sg(sgl, sg, nents, i)
300 ((u64 *)__get_dynamic_array(phys_addrs))[i] = sg_phys(sg);
301 for_each_sg(sgl, sg, ents, i) {
[all …]
/include/rdma/
Drw.h46 struct scatterlist *sg, u32 sg_cnt, u32 sg_offset,
49 u32 port_num, struct scatterlist *sg, u32 sg_cnt,
53 u32 port_num, struct scatterlist *sg, u32 sg_cnt,
58 u32 port_num, struct scatterlist *sg, u32 sg_cnt,
Dib_umem.h138 struct scatterlist *sg = umem->sgt_append.sgt.sgl; in ib_umem_find_best_pgoff() local
141 dma_addr = sg_dma_address(sg) + (umem->address & ~PAGE_MASK); in ib_umem_find_best_pgoff()
/include/scsi/
Dscsi_cmnd.h172 extern void *scsi_kmap_atomic_sg(struct scatterlist *sg, int sg_count,
212 #define scsi_for_each_sg(cmd, sg, nseg, __i) \ argument
213 for_each_sg(scsi_sglist(cmd), sg, nseg, __i)
337 #define scsi_for_each_prot_sg(cmd, sg, nseg, __i) \ argument
338 for_each_sg(scsi_prot_sglist(cmd), sg, nseg, __i)
Dlibiscsi_tcp.h36 struct scatterlist *sg; member
/include/linux/spi/
Dspi-mem.h335 struct sg_table *sg);
339 struct sg_table *sg);
347 struct sg_table *sg) in spi_controller_dma_map_mem_op_data() argument
355 struct sg_table *sg) in spi_controller_dma_unmap_mem_op_data() argument
/include/drm/ttm/
Dttm_bo.h137 struct sg_table *sg; member
403 struct sg_table *sg, struct dma_resv *resv,
408 struct sg_table *sg, struct dma_resv *resv,
Dttm_tt.h103 struct sg_table *sg; member
/include/drm/
Ddrm_prime.h109 void drm_prime_gem_destroy(struct drm_gem_object *obj, struct sg_table *sg);
/include/linux/mtd/
Dubi.h129 struct scatterlist sg[UBI_MAX_SG_COUNT]; member
/include/linux/mmc/
Dcore.h126 struct scatterlist *sg; /* I/O scatter list */ member

12