|
@@ -53,6 +53,67 @@
|
|
|
#include "iowait.h"
|
|
|
#include "user_exp_rcv.h"
|
|
|
|
|
|
+/* The maximum number of Data io vectors per message/request */
|
|
|
+#define MAX_VECTORS_PER_REQ 8
|
|
|
+/*
|
|
|
+ * Maximum number of packet to send from each message/request
|
|
|
+ * before moving to the next one.
|
|
|
+ */
|
|
|
+#define MAX_PKTS_PER_QUEUE 16
|
|
|
+
|
|
|
+#define num_pages(x) (1 + ((((x) - 1) & PAGE_MASK) >> PAGE_SHIFT))
|
|
|
+
|
|
|
+#define req_opcode(x) \
|
|
|
+ (((x) >> HFI1_SDMA_REQ_OPCODE_SHIFT) & HFI1_SDMA_REQ_OPCODE_MASK)
|
|
|
+#define req_version(x) \
|
|
|
+ (((x) >> HFI1_SDMA_REQ_VERSION_SHIFT) & HFI1_SDMA_REQ_OPCODE_MASK)
|
|
|
+#define req_iovcnt(x) \
|
|
|
+ (((x) >> HFI1_SDMA_REQ_IOVCNT_SHIFT) & HFI1_SDMA_REQ_IOVCNT_MASK)
|
|
|
+
|
|
|
+/* Number of BTH.PSN bits used for sequence number in expected rcvs */
|
|
|
+#define BTH_SEQ_MASK 0x7ffull
|
|
|
+
|
|
|
+#define AHG_KDETH_INTR_SHIFT 12
|
|
|
+#define AHG_KDETH_SH_SHIFT 13
|
|
|
+#define AHG_KDETH_ARRAY_SIZE 9
|
|
|
+
|
|
|
+#define PBC2LRH(x) ((((x) & 0xfff) << 2) - 4)
|
|
|
+#define LRH2PBC(x) ((((x) >> 2) + 1) & 0xfff)
|
|
|
+
|
|
|
+#define AHG_HEADER_SET(arr, idx, dw, bit, width, value) \
|
|
|
+ do { \
|
|
|
+ if ((idx) < ARRAY_SIZE((arr))) \
|
|
|
+ (arr)[(idx++)] = sdma_build_ahg_descriptor( \
|
|
|
+ (__force u16)(value), (dw), (bit), \
|
|
|
+ (width)); \
|
|
|
+ else \
|
|
|
+ return -ERANGE; \
|
|
|
+ } while (0)
|
|
|
+
|
|
|
+/* Tx request flag bits */
|
|
|
+#define TXREQ_FLAGS_REQ_ACK BIT(0) /* Set the ACK bit in the header */
|
|
|
+#define TXREQ_FLAGS_REQ_DISABLE_SH BIT(1) /* Disable header suppression */
|
|
|
+
|
|
|
+#define SDMA_PKT_Q_INACTIVE BIT(0)
|
|
|
+#define SDMA_PKT_Q_ACTIVE BIT(1)
|
|
|
+#define SDMA_PKT_Q_DEFERRED BIT(2)
|
|
|
+
|
|
|
+/*
|
|
|
+ * Maximum retry attempts to submit a TX request
|
|
|
+ * before putting the process to sleep.
|
|
|
+ */
|
|
|
+#define MAX_DEFER_RETRY_COUNT 1
|
|
|
+
|
|
|
+#define SDMA_IOWAIT_TIMEOUT 1000 /* in milliseconds */
|
|
|
+
|
|
|
+#define SDMA_DBG(req, fmt, ...) \
|
|
|
+ hfi1_cdbg(SDMA, "[%u:%u:%u:%u] " fmt, (req)->pq->dd->unit, \
|
|
|
+ (req)->pq->ctxt, (req)->pq->subctxt, (req)->info.comp_idx, \
|
|
|
+ ##__VA_ARGS__)
|
|
|
+#define SDMA_Q_DBG(pq, fmt, ...) \
|
|
|
+ hfi1_cdbg(SDMA, "[%u:%u:%u] " fmt, (pq)->dd->unit, (pq)->ctxt, \
|
|
|
+ (pq)->subctxt, ##__VA_ARGS__)
|
|
|
+
|
|
|
extern uint extended_psn;
|
|
|
|
|
|
struct hfi1_user_sdma_pkt_q {
|
|
@@ -79,6 +140,111 @@ struct hfi1_user_sdma_comp_q {
|
|
|
struct hfi1_sdma_comp_entry *comps;
|
|
|
};
|
|
|
|
|
|
+struct sdma_mmu_node {
|
|
|
+ struct mmu_rb_node rb;
|
|
|
+ struct hfi1_user_sdma_pkt_q *pq;
|
|
|
+ atomic_t refcount;
|
|
|
+ struct page **pages;
|
|
|
+ unsigned int npages;
|
|
|
+};
|
|
|
+
|
|
|
+struct user_sdma_iovec {
|
|
|
+ struct list_head list;
|
|
|
+ struct iovec iov;
|
|
|
+ /* number of pages in this vector */
|
|
|
+ unsigned int npages;
|
|
|
+ /* array of pinned pages for this vector */
|
|
|
+ struct page **pages;
|
|
|
+ /*
|
|
|
+ * offset into the virtual address space of the vector at
|
|
|
+ * which we last left off.
|
|
|
+ */
|
|
|
+ u64 offset;
|
|
|
+ struct sdma_mmu_node *node;
|
|
|
+};
|
|
|
+
|
|
|
+/* evict operation argument */
|
|
|
+struct evict_data {
|
|
|
+ u32 cleared; /* count evicted so far */
|
|
|
+ u32 target; /* target count to evict */
|
|
|
+};
|
|
|
+
|
|
|
+struct user_sdma_request {
|
|
|
+ /* This is the original header from user space */
|
|
|
+ struct hfi1_pkt_header hdr;
|
|
|
+
|
|
|
+ /* Read mostly fields */
|
|
|
+ struct hfi1_user_sdma_pkt_q *pq ____cacheline_aligned_in_smp;
|
|
|
+ struct hfi1_user_sdma_comp_q *cq;
|
|
|
+ /*
|
|
|
+ * Pointer to the SDMA engine for this request.
|
|
|
+ * Since different request could be on different VLs,
|
|
|
+ * each request will need it's own engine pointer.
|
|
|
+ */
|
|
|
+ struct sdma_engine *sde;
|
|
|
+ struct sdma_req_info info;
|
|
|
+ /* TID array values copied from the tid_iov vector */
|
|
|
+ u32 *tids;
|
|
|
+ /* total length of the data in the request */
|
|
|
+ u32 data_len;
|
|
|
+ /* number of elements copied to the tids array */
|
|
|
+ u16 n_tids;
|
|
|
+ /*
|
|
|
+ * We copy the iovs for this request (based on
|
|
|
+ * info.iovcnt). These are only the data vectors
|
|
|
+ */
|
|
|
+ u8 data_iovs;
|
|
|
+ s8 ahg_idx;
|
|
|
+
|
|
|
+ /* Writeable fields shared with interrupt */
|
|
|
+ u64 seqcomp ____cacheline_aligned_in_smp;
|
|
|
+ u64 seqsubmitted;
|
|
|
+ /* status of the last txreq completed */
|
|
|
+ int status;
|
|
|
+
|
|
|
+ /* Send side fields */
|
|
|
+ struct list_head txps ____cacheline_aligned_in_smp;
|
|
|
+ u64 seqnum;
|
|
|
+ /*
|
|
|
+ * KDETH.OFFSET (TID) field
|
|
|
+ * The offset can cover multiple packets, depending on the
|
|
|
+ * size of the TID entry.
|
|
|
+ */
|
|
|
+ u32 tidoffset;
|
|
|
+ /*
|
|
|
+ * KDETH.Offset (Eager) field
|
|
|
+ * We need to remember the initial value so the headers
|
|
|
+ * can be updated properly.
|
|
|
+ */
|
|
|
+ u32 koffset;
|
|
|
+ u32 sent;
|
|
|
+ /* TID index copied from the tid_iov vector */
|
|
|
+ u16 tididx;
|
|
|
+ /* progress index moving along the iovs array */
|
|
|
+ u8 iov_idx;
|
|
|
+ u8 done;
|
|
|
+ u8 has_error;
|
|
|
+
|
|
|
+ struct user_sdma_iovec iovs[MAX_VECTORS_PER_REQ];
|
|
|
+} ____cacheline_aligned_in_smp;
|
|
|
+
|
|
|
+/*
|
|
|
+ * A single txreq could span up to 3 physical pages when the MTU
|
|
|
+ * is sufficiently large (> 4K). Each of the IOV pointers also
|
|
|
+ * needs it's own set of flags so the vector has been handled
|
|
|
+ * independently of each other.
|
|
|
+ */
|
|
|
+struct user_sdma_txreq {
|
|
|
+ /* Packet header for the txreq */
|
|
|
+ struct hfi1_pkt_header hdr;
|
|
|
+ struct sdma_txreq txreq;
|
|
|
+ struct list_head list;
|
|
|
+ struct user_sdma_request *req;
|
|
|
+ u16 flags;
|
|
|
+ unsigned int busycount;
|
|
|
+ u64 seqnum;
|
|
|
+};
|
|
|
+
|
|
|
int hfi1_user_sdma_alloc_queues(struct hfi1_ctxtdata *uctxt,
|
|
|
struct hfi1_filedata *fd);
|
|
|
int hfi1_user_sdma_free_queues(struct hfi1_filedata *fd,
|