[ofw] [PATCH 1/10] DAPL v2.0: dat: add definitions for MPI offloaded collectives in IB transport extensions

Davis, Arlin R arlin.r.davis at intel.com
Wed Aug 10 17:41:36 PDT 2011


New definitions for offloaded MPI collectives. Roll IB extensions to v2.0.6
and clean up some counter definitions for consistency. 

Signed-off-by: Arlin Davis <arlin.r.davis at intel.com>
---
 dat/include/dat2/dat_ib_extensions.h |  685 +++++++++++++++++++++++++++++++---
 1 files changed, 624 insertions(+), 61 deletions(-)

diff --git a/dat/include/dat2/dat_ib_extensions.h b/dat/include/dat2/dat_ib_extensions.h
index a32a4ed..ac69fed 100755
--- a/dat/include/dat2/dat_ib_extensions.h
+++ b/dat/include/dat2/dat_ib_extensions.h
@@ -1,5 +1,5 @@
 /*
- * Copyright (c) 2007 Intel Corporation.  All rights reserved.
+ * Copyright (c) 2007-2011 Intel Corporation.  All rights reserved.
  * 
  * This Software is licensed under one of the following licenses:
  * 
@@ -72,15 +72,36 @@
  *
  * 2.0.4 - Add DAT_IB_UD_CONNECTION_REJECT_EVENT extended UD event
  * 2.0.5 - Add DAT_IB_UD extended UD connection error events
+ * 2.0.6 - Add MPI over IB collective extensions
  *
  */
-#define DAT_IB_EXTENSION_VERSION	205	/* 2.0.5 */
-#define DAT_ATTR_COUNTERS		"DAT_COUNTERS"
+#define DAT_IB_EXTENSION_VERSION	206	/* 2.0.6 */
+#define DAT_IB_ATTR_COUNTERS		"DAT_COUNTERS"
 #define DAT_IB_ATTR_FETCH_AND_ADD	"DAT_IB_FETCH_AND_ADD"
 #define DAT_IB_ATTR_CMP_AND_SWAP	"DAT_IB_CMP_AND_SWAP"
 #define DAT_IB_ATTR_IMMED_DATA		"DAT_IB_IMMED_DATA"
 #define DAT_IB_ATTR_UD			"DAT_IB_UD"
 
+#define DAT_IB_COLL_SET_CLOCK		"DAT_COLL_SET_CLOCK"
+#define DAT_IB_COLL_READ_CLOCK		"DAT_COLL_READ_CLOCK"
+#define DAT_IB_COLL_BROADCAST		"DAT_COLL_BROADCAST"
+#define DAT_IB_COLL_BARRIER		"DAT_COLL_BARRIER"
+#define DAT_IB_COLL_SCATTER		"DAT_COLL_SCATTER"
+#define DAT_IB_COLL_SCATTERV		"DAT_COLL_SCATTERV"
+#define DAT_IB_COLL_GATHER		"DAT_COLL_GATHER"
+#define DAT_IB_COLL_GATHERV		"DAT_COLL_GATHERV"
+#define DAT_IB_COLL_ALLGATHER		"DAT_COLL_ALLGATHER"
+#define DAT_IB_COLL_ALLGATHERV		"DAT_COLL_ALLGATHERV"
+#define DAT_IB_COLL_ALLTOALL		"DAT_COLL_ALLTOALL"
+#define DAT_IB_COLL_ALLTOALLV		"DAT_COLL_ALLTOALLV"
+#define DAT_IB_COLL_REDUCE		"DAT_COLL_REDUCE"
+#define DAT_IB_COLL_ALLREDUCE		"DAT_COLL_ALLREDUCE"
+#define DAT_IB_COLL_REDUCE_SCATTER	"DAT_COLL_REDUCE_SCATTER"
+#define DAT_IB_COLL_SCAN		"DAT_COLL_SCAN"
+
+/* Collective handle */
+typedef	DAT_HANDLE	DAT_IB_COLLECTIVE_HANDLE;
+
 /* 
  * Definition for extended EVENT numbers, DAT_IB_EXTENSION_BASE_RANGE
  * is used by these extensions as a starting point for extended event numbers 
@@ -94,7 +115,8 @@ typedef enum dat_ib_event_number
 	DAT_IB_UD_CONNECTION_REQUEST_EVENT,
 	DAT_IB_UD_CONNECTION_EVENT_ESTABLISHED,
 	DAT_IB_UD_CONNECTION_REJECT_EVENT,
-	DAT_IB_UD_CONNECTION_ERROR_EVENT
+	DAT_IB_UD_CONNECTION_ERROR_EVENT,
+	DAT_IB_COLLECTIVE_EVENT,
 
 } DAT_IB_EVENT_NUMBER;
 
@@ -107,8 +129,28 @@ typedef enum dat_ib_op
 	DAT_IB_CMP_AND_SWAP_OP,
 	DAT_IB_RDMA_WRITE_IMMED_OP,
 	DAT_IB_UD_SEND_OP,
-	DAT_QUERY_COUNTERS_OP,
-	DAT_PRINT_COUNTERS_OP
+	DAT_IB_QUERY_COUNTERS_OP,
+	DAT_IB_PRINT_COUNTERS_OP,
+	DAT_IB_COLLECTIVE_CREATE_MEMBER_OP,
+	DAT_IB_COLLECTIVE_FREE_MEMBER_OP,
+	DAT_IB_COLLECTIVE_CREATE_GROUP_OP,
+	DAT_IB_COLLECTIVE_FREE_GROUP_OP,
+	DAT_IB_COLLECTIVE_SET_CLOCK_OP,
+	DAT_IB_COLLECTIVE_READ_CLOCK_OP,
+	DAT_IB_COLLECTIVE_SCATTER_OP,
+	DAT_IB_COLLECTIVE_SCATTERV_OP,
+	DAT_IB_COLLECTIVE_GATHER_OP,
+	DAT_IB_COLLECTIVE_GATHERV_OP,
+	DAT_IB_COLLECTIVE_ALLGATHER_OP,
+	DAT_IB_COLLECTIVE_ALLGATHERV_OP,
+	DAT_IB_COLLECTIVE_ALLTOALL_OP,
+	DAT_IB_COLLECTIVE_ALLTOALLV_OP,
+	DAT_IB_COLLECTIVE_REDUCE_OP,
+	DAT_IB_COLLECTIVE_ALLREDUCE_OP,
+	DAT_IB_COLLECTIVE_REDUCE_SCATTER_OP,
+	DAT_IB_COLLECTIVE_SCAN_OP,
+	DAT_IB_COLLECTIVE_BROADCAST_OP,
+	DAT_IB_COLLECTIVE_BARRIER_OP,
 	
 } DAT_IB_OP;
 
@@ -135,6 +177,24 @@ typedef enum dat_ib_ext_type
 	DAT_IB_UD_CONNECT_REJECT,	// 10
 	DAT_IB_UD_CONNECT_ERROR,	// 11
 
+	DAT_IB_COLLECTIVE_CREATE_STATUS,	// 12
+	DAT_IB_COLLECTIVE_CREATE_DATA,		// 13
+	DAT_IB_COLLECTIVE_CLOCK_SET_STATUS,	// 14
+	DAT_IB_COLLECTIVE_SCATTER_STATUS,	// 15
+	DAT_IB_COLLECTIVE_SCATTERV_STATUS,	// 16
+	DAT_IB_COLLECTIVE_GATHER_STATUS,	// 17
+	DAT_IB_COLLECTIVE_GATHERV_STATUS,	// 18
+	DAT_IB_COLLECTIVE_ALLGATHER_STATUS,	// 19
+	DAT_IB_COLLECTIVE_ALLGATHERV_STATUS,	// 20
+	DAT_IB_COLLECTIVE_ALLTOALL_STATUS,	// 21
+	DAT_IB_COLLECTIVE_ALLTOALLV_STATUS,	// 22
+	DAT_IB_COLLECTIVE_REDUCE_STATUS,	// 23
+	DAT_IB_COLLECTIVE_ALLREDUCE_STATUS,	// 24
+	DAT_IB_COLLECTIVE_REDUCE_SCATTER_STATUS,// 25
+	DAT_IB_COLLECTIVE_SCAN_STATUS,		// 26
+	DAT_IB_COLLECTIVE_BROADCAST_STATUS,	// 27
+	DAT_IB_COLLECTIVE_BARRIER_STATUS,	// 28
+
 } DAT_IB_EXT_TYPE;
 
 /* 
@@ -144,10 +204,10 @@ typedef enum dat_ib_status
 {
 	DAT_OP_SUCCESS = DAT_SUCCESS,
 	DAT_IB_OP_ERR,
+	DAT_IB_COLL_COMP_ERR,
 
 } DAT_IB_STATUS;
 
-
 /* 
  * Definitions for additional extension type RETURN codes above
  * standard DAT types. Included with standard DAT_TYPE_STATUS 
@@ -156,6 +216,7 @@ typedef enum dat_ib_status
 typedef enum dat_ib_return
 {
 	DAT_IB_ERR = DAT_EXTENSION_BASE,
+	DAT_IB_COLLECTIVE_ERR
 
 } DAT_IB_RETURN;
 
@@ -173,7 +234,8 @@ typedef enum dat_ib_dtos
 	DAT_IB_DTO_SEND_UD,
 	DAT_IB_DTO_RECV_UD,
 	DAT_IB_DTO_RECV_UD_IMMED,	
-
+	DAT_IB_DTO_COLLECTIVES,
+	
 } DAT_IB_DTOS;
 
 /* 
@@ -184,6 +246,7 @@ typedef enum dat_ib_dtos
 typedef enum dat_ib_handle_type
 {
     DAT_IB_HANDLE_TYPE_EXT = DAT_HANDLE_TYPE_EXTENSION_BASE,
+    DAT_IB_HANDLE_TYPE_COLLECTIVE
 
 } DAT_IB_HANDLE_TYPE;
 
@@ -221,14 +284,8 @@ typedef struct dat_ib_addr_handle
    
 } DAT_IB_ADDR_HANDLE;
 
-/* 
- * Definitions for extended event data:
- *	When dat_event->event_number >= DAT_IB_EXTENSION_BASE_RANGE
- *	then dat_event->extension_data == DAT_IB_EXT_EVENT_DATA type
- *	and ((DAT_IB_EXT_EVENT_DATA*)dat_event->extension_data)->type
- *	specifies extension data values. 
- * NOTE: DAT_IB_EXT_EVENT_DATA cannot exceed 64 bytes as defined by 
- *	 "DAT_UINT64 extension_data[8]" in DAT_EVENT (dat.h)
+/*
+ * Definition for the value filed of extended event that contains immediate data
  */
 typedef struct dat_ib_immed_data 
 {
@@ -236,13 +293,21 @@ typedef struct dat_ib_immed_data
 
 } DAT_IB_IMMED_DATA;
 
+/* definition for IB collective event data */
+typedef struct dat_ib_collective_event_data
+{
+    DAT_HANDLE  handle;
+    DAT_CONTEXT context;
+
+} DAT_IB_COLLECTIVE_EVENT_DATA;
+
 /* 
  * Definitions for extended event data:
  *	When dat_event->event_number >= DAT_IB_EXTENSION_BASE_RANGE
- *	then dat_event->extension_data == DAT_EXTENSION_EVENT_DATA type
- *	and ((DAT_EXTENSION_EVENT_DATA*)dat_event->extension_data)->type
+ *	then dat_event->extension_data == DAT_IB_EXTENSION_EVENT_DATA type
+ *	and ((DAT_IB_EXTENSION_EVENT_DATA*)dat_event->extension_data)->type
  *	specifies extension data values. 
- * NOTE: DAT_EXTENSION_EVENT_DATA cannot exceed 64 bytes as defined by 
+ * NOTE: DAT_IB_EXTENSION_EVENT_DATA cannot exceed 64 bytes as defined by 
  *	 "DAT_UINT64 extension_data[8]" in DAT_EVENT (dat.h)
  *
  *  Provide UD address handles via extended connect establishment. 
@@ -255,7 +320,10 @@ typedef struct dat_ib_extension_event_data
     union {
 		DAT_IB_IMMED_DATA	immed;
     } val;
-    DAT_IB_ADDR_HANDLE	remote_ah; 
+    union {
+	    DAT_IB_ADDR_HANDLE			remote_ah;
+	    DAT_IB_COLLECTIVE_EVENT_DATA	coll;
+    };
 
 } DAT_IB_EXTENSION_EVENT_DATA;
 
@@ -357,6 +425,71 @@ typedef enum dat_evd_counters
 
 } DAT_EVD_COUNTERS;
 
+/*
+ * Data type for reduce operations
+ */
+typedef enum dat_ib_collective_data_type
+{
+        DAT_IB_COLLECTIVE_TYPE_INT8,
+        DAT_IB_COLLECTIVE_TYPE_UINT8,
+        DAT_IB_COLLECTIVE_TYPE_INT16,
+        DAT_IB_COLLECTIVE_TYPE_UINT16,
+        DAT_IB_COLLECTIVE_TYPE_INT32,
+        DAT_IB_COLLECTIVE_TYPE_UINT32,
+        DAT_IB_COLLECTIVE_TYPE_INT64,
+        DAT_IB_COLLECTIVE_TYPE_UINT64,
+        DAT_IB_COLLECTIVE_TYPE_FLOAT,
+        DAT_IB_COLLECTIVE_TYPE_DOUBLE,
+        DAT_IB_COLLECTIVE_TYPE_LONG_DOUBLE,
+        DAT_IB_COLLECTIVE_TYPE_SHORT_INT,
+        DAT_IB_COLLECTIVE_TYPE_2INT,
+        DAT_IB_COLLECTIVE_TYPE_FLOAT_INT,
+        DAT_IB_COLLECTIVE_TYPE_LONG_INT,
+        DAT_IB_COLLECTIVE_TYPE_DOUBLE_INT,
+        
+} DAT_IB_COLLECTIVE_DATA_TYPE;
+
+/*
+ * Opcode for reduce operations
+ */
+typedef enum dat_ib_collective_reduce_data_op
+{
+        DAT_IB_COLLECTIVE_REDUCE_OP_MAX, 
+        DAT_IB_COLLECTIVE_REDUCE_OP_MIN,
+        DAT_IB_COLLECTIVE_REDUCE_OP_SUM,
+        DAT_IB_COLLECTIVE_REDUCE_OP_PROD,
+        DAT_IB_COLLECTIVE_REDUCE_OP_LAND,
+        DAT_IB_COLLECTIVE_REDUCE_OP_BAND,
+        DAT_IB_COLLECTIVE_REDUCE_OP_LOR,
+        DAT_IB_COLLECTIVE_REDUCE_OP_BOR,
+        DAT_IB_COLLECTIVE_REDUCE_OP_LXOR,
+        DAT_IB_COLLECTIVE_REDUCE_OP_BXOR,
+        DAT_IB_COLLECTIVE_REDUCE_OP_MAXLOC,
+        DAT_IB_COLLECTIVE_REDUCE_OP_MINLOC
+
+} DAT_IB_COLLECTIVE_REDUCE_DATA_OP;
+
+/*
+ * For group creation
+ */
+typedef unsigned int DAT_IB_COLLECTIVE_RANK;
+typedef unsigned int DAT_IB_COLLECTIVE_ID;
+typedef	void * DAT_IB_COLLECTIVE_MEMBER;
+
+typedef struct dat_ib_collective_group
+{
+	int	local_size;       /* # of processes on this node */
+	int	local_rank;       /* my rank within the node */
+	int 	*local_ranks;     /* global rank for each local process */
+	int	external_size;    /* # of nodes, each node has exactly one external process (local root) */
+	int	external_rank;    /* my rank among all external processes if one of them, otherwise -1 */
+	int	*external_ranks;  /* global rank for each external process */
+	int	*intranode_table; /* mapping from global rank to local rank. -1 if the process is on a different node */
+	int	*internode_table; /* mapping from global rank to external rank. -1 if the process is >not external */
+	int	is_comm_world;
+
+} DAT_IB_COLLECTIVE_GROUP;
+
 /* Extended RETURN and EVENT STATUS string helper functions */
 
 /* DAT_EXT_RETURN error to string */
@@ -397,6 +530,9 @@ dat_strerror_ext_status (
 /* 
  * Extended IB transport specific APIs
  *  redirection via DAT extension function
+ *  va_arg function: DAT_HANDLE and OP type MUST be first 2 parameters
+ *
+ *  RETURN VALUE: DAT_RETURN
  */
 
 /*
@@ -406,13 +542,14 @@ dat_strerror_ext_status (
  * and the result is stored in the local_iov.  
  */
 #define dat_ib_post_fetch_and_add(ep, add_val, lbuf, cookie, rbuf, flgs) \
-	     dat_extension_op(  ep, \
-				DAT_IB_FETCH_AND_ADD_OP, \
-				(add_val), \
-				(lbuf), \
-				(cookie), \
-				(rbuf), \
-				(flgs))
+	dat_extension_op(\
+		IN (DAT_EP_HANDLE) (ep), \
+		IN (DAT_IB_OP) DAT_IB_FETCH_AND_ADD_OP, \
+		IN (DAT_UINT64) (add_val), \
+		IN (DAT_LMR_TRIPLET *) (lbuf), \
+		IN (cookie), \
+		IN (DAT_RMR_TRIPLET *) (rbuf), \
+		IN (DAT_COMPLETION_FLAGS) (flgs))
 				
 /*
  * This asynchronous call is modeled after the InfiniBand atomic 
@@ -423,14 +560,15 @@ dat_strerror_ext_status (
  * value stored in the remote memory location is copied to the local_iov.
  */
 #define dat_ib_post_cmp_and_swap(ep, cmp_val, swap_val, lbuf, cookie, rbuf, flgs) \
-	     dat_extension_op(  ep, \
-				DAT_IB_CMP_AND_SWAP_OP, \
-				(cmp_val), \
-				(swap_val), \
-				(lbuf), \
-				(cookie), \
-				(rbuf), \
-				(flgs))
+	dat_extension_op(\
+		IN (DAT_EP_HANDLE) (ep), \
+		IN (DAT_IB_OP) DAT_IB_CMP_AND_SWAP_OP, \
+		IN (DAT_UINT64) (cmp_val), \
+		IN (DAT_UINT64) (swap_val), \
+		IN (DAT_LMR_TRIPLET *) (lbuf), \
+		IN (cookie), \
+		IN (DAT_RMR_TRIPLET *) (rbuf), \
+		IN (DAT_COMPLETION_FLAGS) (flgs))
 
 /* 
  * RDMA Write with IMMEDIATE:
@@ -449,14 +587,15 @@ dat_strerror_ext_status (
  *	n/a
  */
 #define dat_ib_post_rdma_write_immed(ep, size, lbuf, cookie, rbuf, idata, flgs) \
-	     dat_extension_op(  ep, \
-				DAT_IB_RDMA_WRITE_IMMED_OP, \
-				(size), \
-				(lbuf), \
-				(cookie), \
-				(rbuf), \
-				(idata), \
-				(flgs))
+	dat_extension_op(\
+		IN (DAT_EP_HANDLE) (ep), \
+		IN (DAT_IB_OP) DAT_IB_RDMA_WRITE_IMMED_OP, \
+		IN (DAT_COUNT) (size), \
+		IN (DAT_LMR_TRIPLET *) (lbuf), \
+		IN (cookie), \
+		IN (DAT_RMR_TRIPLET *) (rbuf), \
+		IN (DAT_UINT32) (idata), \
+		IN (DAT_COMPLETION_FLAGS) (flgs))
 
 /* 
  * Unreliable datagram: msg send 
@@ -471,14 +610,21 @@ dat_strerror_ext_status (
  *	n/a
  */
 #define dat_ib_post_send_ud(ep, segments, lbuf, ah_ptr, cookie, flgs) \
-	     dat_extension_op(  ep, \
-				DAT_IB_UD_SEND_OP, \
-				(segments), \
-				(lbuf), \
-				(ah_ptr), \
-				(cookie), \
-				(flgs))
+	dat_extension_op(\
+		IN (DAT_EP_HANDLE) (ep), \
+		IN (DAT_IB_OP) DAT_IB_UD_SEND_OP, \
+		IN (DAT_COUNT) (segments), \
+		IN (DAT_LMR_TRIPLET *) (lbuf), \
+		IN (DAT_IB_ADDR_HANDLE *) (ah_ptr), \
+		IN (cookie), \
+		IN (DAT_COMPLETION_FLAGS) (flgs))
 
+/* 
+ * Unreliable datagram: msg recv 
+ *
+ * Mapping to standard EP post call.
+ */
+#define dat_ib_post_recv_ud	dat_ep_post_recv
 
 /* 
  * Query counter(s):  
@@ -487,12 +633,13 @@ dat_strerror_ext_status (
  *
  * use _ALL_COUNTERS to query all
  */
-#define dat_query_counters(dat_handle, cntr, p_cntrs_out, reset) \
-	     dat_extension_op(  dat_handle, \
-				DAT_QUERY_COUNTERS_OP, \
-				(cntr), \
-				(p_cntrs_out), \
-				(reset))
+#define dat_ib_query_counters(dat_handle, cntr, p_cntrs_out, reset) \
+	dat_extension_op(\
+		IN (DAT_HANDLE) dat_handle, \
+		IN (DAT_IB_OP) DAT_QUERY_COUNTERS_OP, \
+		IN (int) (cntr), \
+		IN (DAT_UINT64 *) (p_cntrs_out), \
+		IN (int) (reset))
 /* 
  * Print counter(s):  
  * Provide IA, EP, or EVD and call will print appropriate counters
@@ -500,11 +647,427 @@ dat_strerror_ext_status (
  * 
  * use _ALL_COUNTERS to print all
  */
-#define dat_print_counters(dat_handle, cntr, reset) \
-	     dat_extension_op(  dat_handle, \
-				DAT_PRINT_COUNTERS_OP, \
-				(cntr), \
-				(reset))
+#define dat_ib_print_counters(dat_handle, cntr, reset) \
+	dat_extension_op(\
+		IN (DAT_HANDLE) dat_handle, \
+		IN (DAT_IB_OP) DAT_PRINT_COUNTERS_OP, \
+		IN (int) (cntr), \
+		IN (int) (reset))
+
+/*
+ ************************ MPI IB Collective Functions ***********************
+ */
+
+/* MPI collective member and group setup functions */
+
+/*
+ * This synchronous call creates and returns local member
+ * address information for a collective device or provider
+ * for each rank. The size of the member address information
+ * is dependent on the collective device or provider.
+ * This address information, for each rank, must be exchanged
+ * and used for group creation on all ranks.
+ */
+#define dat_ib_collective_create_member(ia_handle, progress_func, member, member_size) \
+	dat_extension_op(\
+		IN  (DAT_IA_HANDLE) (ia_handle), \
+		IN  (DAT_IB_OP) DAT_IB_COLLECTIVE_CREATE_MEMBER_OP, \
+		IN  (void *) (progress_func), \
+		OUT (DAT_IB_COLLECTIVE_MEMBER *) (member), \
+		OUT (DAT_UINT32 *) (member_size))
+
+/*
+ * This synchronous call destroys a previously created member
+ * information associated with the this device ia_handle argument.
+ */
+#define dat_ib_collective_free_member(ia_handle, member) \
+	dat_extension_op(\
+		IN (DAT_IA_HANDLE) (ia_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_FREE_MEMBER_OP, \
+		IN (DAT_IB_COLLECTIVE_MEMBER) (member))
+
+/*
+ * This asynchronous call initiates the process of creating a collective
+ * group and must be called by all group members. The collective_group
+ * argument points to an array of address/connection qualifier pairs that
+ * identify the members of the group in rank order. The group_size argument
+ * specifies the size of the group and therefore the size of the coll_group
+ * array. The self argument identifies the rank of the caller.
+ * The group_id argument specifies a network-unique identifier for this
+ * instance of the collective group. The group_info provides global and local
+ * rank and process information. All members of the group must specify
+ * the same group_id value for the same collective instance. The evd_handle
+ * argument specifies the EVD used for all asynchronous collective completions
+ * including this call. The user_context argument will be returned in the
+ * DAT_EXT_COLLECTIVE_CREATE_DATA event.
+ *
+ * On a successful completion, each group member will receive a
+ * DAT_EXT_COLLECTIVE_CREATE_DATA event on the EVD specified by evd_handle.
+ * The event contains the collective handle, the rank of the receiving
+ * Endpoint within the collective group, the size of the group, and the
+ * caller specified user_context. The returned collective handle can be used
+ * in network clock, Multicast, and other collective operations.
+ *
+ * RETURN VALUE: DAT_RETURN
+ */
+#define dat_ib_collective_create_group(members, group_size, self, group_id, group_info, evd, pd, user_context) \
+	dat_extension_op(\
+		IN (DAT_EVD_HANDLE) (evd), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_CREATE_GROUP_OP, \
+		IN (DAT_IB_COLLECTIVE_MEMBER *) (members), \
+		IN (DAT_COUNT) (group_size), \
+		IN (DAT_IB_COLLECTIVE_RANK) (self), \
+		IN (DAT_IB_COLLECTIVE_ID) (group_id), \
+		IN (DAT_IB_COLLECTIVE_GROUP *) (group_info), \
+		IN (DAT_PZ_HANDLE) (pd), \
+		IN (DAT_CONTEXT) (user_context))
+
+/*
+ * This synchronous call destroys a previously created collective group
+ * associated with the collective_handle argument. Any pending or
+ * in-process requests associated with the collective group will be
+ * terminated and be posted to the appropriate EVD.
+ *
+ * RETURN VALUE: DAT_RETURN
+ */
+#define dat_ib_collective_free_group(coll_handle) \
+	dat_extension_op(\
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_FREE_GROUP_OP)
+
+
+/* MPI collective data operations */
+
+/*
+ * This call sets the network clock associated with
+ * collective_handle. A provider implementation may keep a single
+ * global clock for all collective handles. When this is the case,
+ * this call sets an adjustment for the given handle so that
+ * subsequent calls to read the clock will be relative to the value
+ * specified by clock_value. This is an asynchronous call that
+ * completes on the collective EVD. The network clock will not be
+ * synchronized until the request is completed. Any member of the
+ * collective can set the clock and only one member should make
+ * this call on behave of the entire collective.
+ */
+#define dat_ib_collective_set_clock(coll_handle, clock_value, user_context ) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_READ_CLOCK_OP, \
+		IN (DAT_UINT64) (clock_value), \
+		IN (DAT_CONTEXT) (user_contex))
+
+/*
+ * This synchronous call returns the current value of the network clock
+ * associated with the given collective handle. This is a light weight
+ * call to minimize skew
+ */
+#define dat_ib_collective_read_clock(coll_handle, clock_value ) \
+	dat_extension_op( \
+		IN  (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN  (DAT_IB_OP) DAT_IB_COLLECTIVE_READ_CLOCK_OP, \
+		OUT (DAT_UINT64 *) clock_value))
+
+/*
+ * This call performs a scatter of the data specified by the
+ * send_buffer argument to the collective group specified by coll_handle.
+ * Data is received in the buffer specified by the recv_buffer argument.
+ * The recv_byte_count argument specifies the size of the receive buffer.
+ * Data from the root send_buffer will be divided by the number of members
+ * in the collective group to form equal and contiguous memory partitions.
+ * Each member of the collective group will receive its rank relative
+ * partition. An error is returned if the send_byte_count does not describe
+ * memory that can be evenly divided by the size of the collective group.
+ * An "in place" transfer for the root rank can be indicated by passing NULL
+ * as the recv_buffer argument. The send_buffer and send_byte_count
+ * arguments are ignored on non-root members. The operation is completed on
+ * the collective EVD unless completions are suppressed through the
+ * completion flags.
+ */
+#define dat_ib_collective_scatter(coll_handle, sendbuf, sendsize, recvbuf, recvsize, root, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_SCATTER_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_IB_COLLECTIVE_RANK) (root), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call performs a non-uniform scatter of the data
+ * specified by the send_buffers array argument to the collective group
+ * specified by coll_handle. The send_buffers array contains one buffer
+ * pointer for each member of the collective group, in rank order.
+ * The send_byte_counts array contains a byte count for each corresponding
+ * send buffer pointer. The recv_buffer and recev_byte_count arguments
+ * specify where received portions of the scatter are to be received.
+ * An "in place" transfer for the root rank can be indicated by passing
+ * NULL as the recv_buffer argument. The send_buffers and send_byte_counts
+ * arguments are ignored on non-root members. The operation is completed
+ * on the collective EVD unless completions are suppressed through the
+ * completion flags.
+ *
+ */
+#define dat_ib_collective_scatterv(coll_handle, sendbuf, sendsizes, displs, recvbuf, recvsize, root, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_SCATTERV_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT *) (sendsizes), \
+		IN (DAT_COUNT *) (displs), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_IB_COLLECTIVE_RANK) (root), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call performs a gather of the data sent by all
+ * members of the collective specified by the collective_handle argument.
+ * The data to be sent is specified by the send_buffer and send_byte_count
+ * arguments. Data is received by the collective member specified by the
+ * root argument in the buffer specified by the recv_buffer and
+ * recv_byte_count arguments.  Data is placed into the receive buffer in
+ * collective rank order.  An "in place" transfer for the root rank can
+ * be indicated by passing NULL as the send_buffer argument.
+ * The recv_buffer and recv_byte_count arguments are ignored on non-root
+ * members.  The operation is completed on the collective EVD unless
+ * completions are suppressed through the completion flags.
+ */
+#define dat_ib_collective_gather(coll_handle, sendbuf, sendsize, recvbuf, recvsize, root, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_GATHER_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_IB_COLLECTIVE_RANK) (root), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS)(flags))
+
+/*
+ * This call performs a non-uniform gather of the data sent by
+ * all members of the collective specified by the collective_handle argument.
+ * The data to be sent is specified by the send_buffer and send_byte_count
+ * arguments.  Data is received by the collective member specified by the
+ * root argument into the buffers specified by the recv_buffers and
+ * recv_byte_counts array arguments.  Data is placed into the receive buffer
+ * associated with the rank that sent it. An "in place" transfer for the root
+ * rank can be indicated by passing NULL as the send_buffer argument.
+ * The recv_buffers and recv_byte_counts arguments are ignored on non-root
+ * members.  The operation is completed on the collective EVD unless
+ * completions are suppressed through the completion flags.
+ */
+#define dat_ib_collective_gatherv(coll_handle, sendbuf, sendsize, recvbufs, recvsizes, displs, root, user_context, flags) \
+	dat_extension_op( \
+		(DAT_IB_COLLECTIVE_HANDLE)(coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_GATHERV_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT *) (recvsizes), \
+		IN (DAT_COUNT *) (displs), \
+		IN (DAT_IB_COLLECTIVE_RANK) (root), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call is equivalent to having all members of a collective
+ * group perform a dat_collective_gather() as the root.  This results in all
+ * members of the collective having identical contents in their receive buffer
+ */
+#define dat_ib_collective_allgather(coll_handle, sendbuf, sendsize, recvbuf, recvsize, user_context, flags) \
+	dat_extension_op( \
+		(DAT_IB_COLLECTIVE_HANDLE)(coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_ALLGATHER_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call performs a non-uniform dat_collective_allgather()
+ * operation.  It is equivalent to having all members of a collective group
+ * perform a dat_collective_gatherv() as the root.  This results in all
+ * members of the collective having identical contents in their receive
+ * buffer.
+ */
+#define dat_ib_collective_allgatherv(coll_handle, sendbuf, sendsize, recvbuf, recvsizes, displs, user_context, flags) \
+	dat_extension_op( \
+		(DAT_IB_COLLECTIVE_HANDLE)(coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_ALLGATHERV_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT *) (recvsizes), \
+		IN (DAT_COUNT *) (displs), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call is an extension of dat_collective_allgather()
+ * to the case where each member sends distinct data specified by send_buffer
+ * to each of the other members. The jth block sent from rank i is received
+ * by rank j and is placed in the ith block of recv_buffer.
+ */
+#define dat_ib_collective_alltoall(coll_handle, sendbuf, sendsize, recvbuf, recvsize, user_context, flags) \
+	dat_extension_op( \
+		(DAT_IB_COLLECTIVE_HANDLE)(coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_ALLTOALL_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call performs a non-uniform dat_collective_alltoall() operation
+ */
+#define dat_ib_collective_alltoallv(coll_handle, sendbuf, sendsizes, senddspls, recvbuf, recvsizes, recvdispls, user_context, flags) \
+	dat_extension_op( \
+		(DAT_IB_COLLECTIVE_HANDLE)(coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_ALLTOALLV_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT *) (sendsizes), \
+		IN (DAT_COUNT *) (senddispls), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT *) (recvsizes), \
+		IN (DAT_COUNT *) (recvdispls), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call combines the elements of the data type specified
+ * by data_type from the buffer specified by send_buffer of all members of
+ * the collective by performing the operation specified by reduce_operation
+ * and placing the result into the buffer of the root member specified by
+ * recv_buffer. It is an error to specify a floating point type with
+ * any of the logical reduction operators.When using the REDUCE_OP_MINLOC
+ * and REDUCE_OP _MAXLOC operations, it is assumed that the input and output
+ * buffers contain pair values where the first member of the pair is of the
+ * type specified by data_type followed by a COLLECTIVE_TYPE_UINT32 type.
+ * When the reduction is complete, the receive buffer will contain the
+ * MIN/MAX value in the first member of the pair with the first member rank
+ * that contained it in the second member of the pair.  The tables below
+ * show the result of a REDUCE_OP_SUM reduce operation.
+ */
+#define dat_ib_collective_reduce(coll_handle, sendbuf, sendsize, recvbuf, recvsize, op, type, root, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE)(coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_REDUCE_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_IB_COLLECTIVE_REDUCE_DATA_OP) (op), \
+		IN (DAT_IB_COLLECTIVE_DATA_TYPE) (type), \
+		IN (DAT_IB_COLLECTIVE_RANK) (root), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call is identical to the dat_collective_reduce()
+ * call with the exception that the recv_buffer and recv_byte_count arguments
+ * are valid for all members of the collective and all members of will
+ * receive the reduction results.
+ */
+#define dat_ib_collective_allreduce(coll_handle, sendbuf, sendsize, recvbuf, recvsize, op, type, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_ALLREDUCE_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_IB_COLLECTIVE_REDUCE_DATA_OP) (op), \
+		IN (DAT_IB_COLLECTIVE_DATA_TYPE) (type), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+/*
+ * This call is identical to rank 0 of the collective calling
+ * this dat_collective_reduce() followed by dat_collective_scatterv().
+ * The number of bytes received in the scatter for each rank is determined
+ * by rank offset into the recv_byte_counts array.
+ */
+#define dat_ib_collective_reduce_scatter(coll_handle, sendbuf, sendsize, recvbuf, recvsizes, op, type, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_REDUCE_SCATTER_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT *) (recvsizes), \
+		IN (DAT_IB_COLLECTIVE_REDUCE_DATA_OP) (op), \
+		IN (DAT_IB_COLLECTIVE_DATA_TYPE) (type), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call is used to perform a prefix reduction on data
+ * distributed across the group. The operation returns, in recv_buffer of
+ * the member with rank i, the reduction of the values in send_buffer of
+ * members with ranks 0,...,i (inclusive). The tables below show the
+ * result of a REDUCE_OP_SUM scan operation.
+ */
+#define dat_ib_collective_scan(coll_handle, sendbuf, sendsize, recvbuf, recvsize, op, type, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_SCAN_OP, \
+		IN (DAT_PVOID) (sendbuf), \
+		IN (DAT_COUNT) (sendsize), \
+		IN (DAT_PVOID) (recvbuf), \
+		IN (DAT_COUNT) (recvsize), \
+		IN (DAT_IB_COLLECTIVE_REDUCE_DATA_OP) (op), \
+		IN (DAT_IB_COLLECTIVE_DATA_TYPE) (type), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call performs a broadcast send operation that transfers
+ * data specified by the buffer argument of the root into the buffer argument
+ * of all other Endpoints in the collective group specified by coll_handle.
+ * The operation is completed on the collective EVD unless completions are
+ * suppressed through the completion flags.  All broadcasts are considered
+ * o?=in placeo?= transfers.  The tables below show the result of a broadcast
+ * operation.
+ */
+#define dat_ib_collective_broadcast(coll_handle, buf, size, root, user_context, flags) \
+	dat_extension_op(\
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_BROADCAST_OP, \
+		IN (DAT_PVOID) (buf), \
+		IN (DAT_COUNT) (size), \
+		IN (DAT_IB_COLLECTIVE_RANK) (root), \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+/*
+ * This call will synchronize all endpoints of the collective
+ * group specified by coll_handle. This is an asynchronous call that
+ * will post a completion to the collective EVD when all endpoints
+ * have synchronized.
+ */
+#define dat_ib_collective_barrier(coll_handle, user_context, flags) \
+	dat_extension_op( \
+		IN (DAT_IB_COLLECTIVE_HANDLE) (coll_handle), \
+		IN (DAT_IB_OP) DAT_IB_COLLECTIVE_BARRIER_OP, \
+		IN (DAT_CONTEXT) (user_context), \
+		IN (DAT_COMPLETION_FLAGS) (flags))
+
+
+/* Backward compatibility */
+#define DAT_ATTR_COUNTERS DAT_IB_ATTR_COUNTERS
+#define dat_query_counters dat_ib_query_counters	
+#define dat_print_counters dat_ib_print_counters
+#define DAT_QUERY_COUNTERS_OP DAT_IB_QUERY_COUNTERS_OP
+#define DAT_PRINT_COUNTERS_OP DAT_IB_PRINT_COUNTERS_OP
 
 #endif /* _DAT_IB_EXTENSIONS_H_ */
 
-- 
1.7.3






More information about the ofw mailing list