Linux cpanel.rrshost.in 5.15.0-25-generic #25-Ubuntu SMP Wed Mar 30 15:54:22 UTC 2022 x86_64
Apache
: 109.123.238.221 | : 172.70.80.18
128 Domain
8.2.28
aev999
Terminal
AUTO ROOT
Adminer
Backdoor Destroyer
Linux Exploit
Lock Shell
Lock File
Create User
CREATE RDP
PHP Mailer
BACKCONNECT
HASH IDENTIFIER
README
+ Create Folder
+ Create File
/
usr /
src /
linux-headers-5.15.0-25 /
include /
rdma /
[ HOME SHELL ]
Name
Size
Permission
Action
ib.h
2.09
KB
-rw-r--r--
ib_addr.h
7.66
KB
-rw-r--r--
ib_cache.h
4.23
KB
-rw-r--r--
ib_cm.h
16.92
KB
-rw-r--r--
ib_hdrs.h
6.79
KB
-rw-r--r--
ib_mad.h
24.48
KB
-rw-r--r--
ib_marshall.h
778
B
-rw-r--r--
ib_pack.h
7.71
KB
-rw-r--r--
ib_pma.h
4.17
KB
-rw-r--r--
ib_sa.h
19.5
KB
-rw-r--r--
ib_smi.h
4.21
KB
-rw-r--r--
ib_sysfs.h
1.23
KB
-rw-r--r--
ib_umem.h
6.64
KB
-rw-r--r--
ib_umem_odp.h
3.39
KB
-rw-r--r--
ib_verbs.h
138.24
KB
-rw-r--r--
iba.h
5.51
KB
-rw-r--r--
ibta_vol1_c12.h
12.94
KB
-rw-r--r--
iw_cm.h
7.12
KB
-rw-r--r--
iw_portmap.h
2.19
KB
-rw-r--r--
lag.h
530
B
-rw-r--r--
mr_pool.h
568
B
-rw-r--r--
opa_addr.h
2.39
KB
-rw-r--r--
opa_port_info.h
13.43
KB
-rw-r--r--
opa_smi.h
3
KB
-rw-r--r--
opa_vnic.h
2
KB
-rw-r--r--
peer_mem.h
6.19
KB
-rw-r--r--
rdma_cm.h
12.81
KB
-rw-r--r--
rdma_cm_ib.h
797
B
-rw-r--r--
rdma_counter.h
1.78
KB
-rw-r--r--
rdma_netlink.h
3.97
KB
-rw-r--r--
rdma_vt.h
14.22
KB
-rw-r--r--
rdmavt_cq.h
1.74
KB
-rw-r--r--
rdmavt_mr.h
3.94
KB
-rw-r--r--
rdmavt_qp.h
28.01
KB
-rw-r--r--
restrack.h
4.2
KB
-rw-r--r--
rw.h
2.22
KB
-rw-r--r--
signature.h
2.9
KB
-rw-r--r--
tid_rdma_defs.h
2.05
KB
-rw-r--r--
uverbs_ioctl.h
32.22
KB
-rw-r--r--
uverbs_named_ioctl.h
4.72
KB
-rw-r--r--
uverbs_std_types.h
5.23
KB
-rw-r--r--
uverbs_types.h
6.06
KB
-rw-r--r--
Delete
Unzip
Zip
${this.title}
Close
Code Editor : ib_umem.h
/* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */ /* * Copyright (c) 2007 Cisco Systems. All rights reserved. * Copyright (c) 2020 Intel Corporation. All rights reserved. */ #ifndef IB_UMEM_H #define IB_UMEM_H #include <linux/list.h> #include <linux/scatterlist.h> #include <linux/workqueue.h> #include <rdma/ib_verbs.h> struct ib_ucontext; struct ib_umem_odp; struct dma_buf_attach_ops; struct ib_umem { struct ib_device *ibdev; struct mm_struct *owning_mm; u64 iova; size_t length; unsigned long address; u32 writable : 1; u32 is_odp : 1; u32 is_dmabuf : 1; /* Placing at the end of the bitfield list is ABI preserving on LE */ u32 is_peer : 1; struct work_struct work; struct sg_append_table sgt_append; }; struct ib_umem_dmabuf { struct ib_umem umem; struct dma_buf_attachment *attach; struct sg_table *sgt; struct scatterlist *first_sg; struct scatterlist *last_sg; unsigned long first_sg_offset; unsigned long last_sg_trim; void *private; }; static inline struct ib_umem_dmabuf *to_ib_umem_dmabuf(struct ib_umem *umem) { return container_of(umem, struct ib_umem_dmabuf, umem); } typedef void (*umem_invalidate_func_t)(struct ib_umem *umem, void *priv); enum ib_peer_mem_flags { IB_PEER_MEM_ALLOW = 1 << 0, IB_PEER_MEM_INVAL_SUPP = 1 << 1, }; /* Returns the offset of the umem start relative to the first page. */ static inline int ib_umem_offset(struct ib_umem *umem) { return umem->address & ~PAGE_MASK; } static inline unsigned long ib_umem_dma_offset(struct ib_umem *umem, unsigned long pgsz) { return (sg_dma_address(umem->sgt_append.sgt.sgl) + ib_umem_offset(umem)) & (pgsz - 1); } static inline size_t ib_umem_num_dma_blocks(struct ib_umem *umem, unsigned long pgsz) { return (size_t)((ALIGN(umem->iova + umem->length, pgsz) - ALIGN_DOWN(umem->iova, pgsz))) / pgsz; } static inline size_t ib_umem_num_pages(struct ib_umem *umem) { return ib_umem_num_dma_blocks(umem, PAGE_SIZE); } static inline void __rdma_umem_block_iter_start(struct ib_block_iter *biter, struct ib_umem *umem, unsigned long pgsz) { __rdma_block_iter_start(biter, umem->sgt_append.sgt.sgl, umem->sgt_append.sgt.nents, pgsz); } /** * rdma_umem_for_each_dma_block - iterate over contiguous DMA blocks of the umem * @umem: umem to iterate over * @pgsz: Page size to split the list into * * pgsz must be <= PAGE_SIZE or computed by ib_umem_find_best_pgsz(). The * returned DMA blocks will be aligned to pgsz and span the range: * ALIGN_DOWN(umem->address, pgsz) to ALIGN(umem->address + umem->length, pgsz) * * Performs exactly ib_umem_num_dma_blocks() iterations. */ #define rdma_umem_for_each_dma_block(umem, biter, pgsz) \ for (__rdma_umem_block_iter_start(biter, umem, pgsz); \ __rdma_block_iter_next(biter);) #ifdef CONFIG_INFINIBAND_USER_MEM struct ib_umem *ib_umem_get(struct ib_device *device, unsigned long addr, size_t size, int access); void ib_umem_release(struct ib_umem *umem); int ib_umem_copy_from(void *dst, struct ib_umem *umem, size_t offset, size_t length); unsigned long ib_umem_find_best_pgsz(struct ib_umem *umem, unsigned long pgsz_bitmap, unsigned long virt); /** * ib_umem_find_best_pgoff - Find best HW page size * * @umem: umem struct * @pgsz_bitmap bitmap of HW supported page sizes * @pgoff_bitmask: Mask of bits that can be represented with an offset * * This is very similar to ib_umem_find_best_pgsz() except instead of accepting * an IOVA it accepts a bitmask specifying what address bits can be represented * with a page offset. * * For instance if the HW has multiple page sizes, requires 64 byte alignemnt, * and can support aligned offsets up to 4032 then pgoff_bitmask would be * "111111000000". * * If the pgoff_bitmask requires either alignment in the low bit or an * unavailable page size for the high bits, this function returns 0. */ static inline unsigned long ib_umem_find_best_pgoff(struct ib_umem *umem, unsigned long pgsz_bitmap, u64 pgoff_bitmask) { struct scatterlist *sg = umem->sgt_append.sgt.sgl; dma_addr_t dma_addr; dma_addr = sg_dma_address(sg) + (umem->address & ~PAGE_MASK); return ib_umem_find_best_pgsz(umem, pgsz_bitmap, dma_addr & pgoff_bitmask); } struct ib_umem_dmabuf *ib_umem_dmabuf_get(struct ib_device *device, unsigned long offset, size_t size, int fd, int access, const struct dma_buf_attach_ops *ops); int ib_umem_dmabuf_map_pages(struct ib_umem_dmabuf *umem_dmabuf); void ib_umem_dmabuf_unmap_pages(struct ib_umem_dmabuf *umem_dmabuf); void ib_umem_dmabuf_release(struct ib_umem_dmabuf *umem_dmabuf); struct ib_umem *ib_umem_get_peer(struct ib_device *device, unsigned long addr, size_t size, int access, unsigned long peer_mem_flags); void ib_umem_activate_invalidation_notifier(struct ib_umem *umem, umem_invalidate_func_t func, void *cookie); void ib_umem_stop_invalidation_notifier(struct ib_umem *umem); #else /* CONFIG_INFINIBAND_USER_MEM */ #include <linux/err.h> static inline struct ib_umem *ib_umem_get(struct ib_device *device, unsigned long addr, size_t size, int access) { return ERR_PTR(-EOPNOTSUPP); } static inline void ib_umem_release(struct ib_umem *umem) { } static inline int ib_umem_copy_from(void *dst, struct ib_umem *umem, size_t offset, size_t length) { return -EOPNOTSUPP; } static inline unsigned long ib_umem_find_best_pgsz(struct ib_umem *umem, unsigned long pgsz_bitmap, unsigned long virt) { return 0; } static inline unsigned long ib_umem_find_best_pgoff(struct ib_umem *umem, unsigned long pgsz_bitmap, u64 pgoff_bitmask) { return 0; } static inline struct ib_umem_dmabuf *ib_umem_dmabuf_get(struct ib_device *device, unsigned long offset, size_t size, int fd, int access, struct dma_buf_attach_ops *ops) { return ERR_PTR(-EOPNOTSUPP); } static inline int ib_umem_dmabuf_map_pages(struct ib_umem_dmabuf *umem_dmabuf) { return -EOPNOTSUPP; } static inline void ib_umem_dmabuf_unmap_pages(struct ib_umem_dmabuf *umem_dmabuf) { } static inline void ib_umem_dmabuf_release(struct ib_umem_dmabuf *umem_dmabuf) { } static inline struct ib_umem *ib_umem_get_peer(struct ib_device *device, unsigned long addr, size_t size, int access, unsigned long peer_mem_flags) { return ERR_PTR(-EINVAL); } static inline void ib_umem_activate_invalidation_notifier( struct ib_umem *umem, umem_invalidate_func_t func, void *cookie) { } static inline void ib_umem_stop_invalidation_notifier(struct ib_umem *umem) { } #endif /* CONFIG_INFINIBAND_USER_MEM */ #endif /* IB_UMEM_H */
Close