Mooncake Transfer Engine is a high-performance, zero-copy data transfer library designed around two core abstractions: Segment and BatchTransfer.
-
Segment represents a contiguous address space that can be remotely read and written, which can be either non-persistent storage provided by DRAM or VRAM, known as RAM Segment, or persistent storage provided by NVMeof, known as NVMeof Segment.
-
BatchTransfer encapsulates operation requests, specifically responsible for synchronizing data between a set of non-contiguous data spaces in one Segment and the corresponding spaces in another set of Segments, supporting Read/Write in both directions, thus acting like an asynchronous and more flexible AllScatter/AllGather.
As shown in the diagram, each specific client corresponds to a TransferEngine
, which not only includes a RAM Segment but also integrates management for high-speed transfers across multiple threads and network cards. The RAM Segment, in principle, corresponds to the entire virtual address space of this TransferEngine
, but in reality, only parts of it (known as a Buffer
) are registered for (GPUDirect) RDMA Read/Write. Each Buffer can have separate permissions (corresponding to RDMA rkey
, etc.) and network card affinity (e.g., preferred NICs for different types of memory).
Mooncake Transfer Engine provides interfaces through the TransferEngine
class (located in mooncake-transfer-engine/include/transfer_engine.h
), where the specific data transfer functions for different backends are implemented by the Transport
class, currently supporting TcpTransport
, RdmaTransport
, and NVMeoFTransport
.
Segment represents a collection of source address ranges and target address ranges available during the data transfer process in Transfer Engine. That is, all local and remote addresses involved in BatchTransfer
requests must be within the valid segment range. Transfer Engine supports the following two types of Segments.
When each process starts, Transfer Engine automatically creates a segment named after its own local_hostname
(see the initialization function of TransferEngine
, which needs to be globally unique), which logically covers the entire memory address space, including storage media such as DRAM/VRAM. When using the BatchTransfer
interface for transfer tasks, Transfer Engine automatically determines the corresponding hardware information to choose the best transfer method. Each process has and only has one segment. Other processes can reference the segment and complete read/write operations by calling the openSegment
interface and passing the correct name.
In actual deployment, application systems usually only use part of the memory address space for data transfer, so Transfer Engine further divides the segment into multiple buffers internally. Each Buffer represents a contiguous address space located on the same device, and when users use the BatchTransfer
interface to complete read/write operations, if referencing a RAM Segment, each read/write task must be within one of the valid buffers.
The memory range within a segment does not need to be contiguous, which means that multiple DRAM/VRAM address spaces can be allocated and included in the same segment.
In addition, Transfer Engine also supports registering some local DRAM areas, which are merely used as the local side storage space for data operations, such as the DRAM PageCache area of vLLM. It is also considered a part of the effective RAM Segment in the current process but cannot be referenced by other processes by calling the openSegment
interface.
Transfer Engine also leverages the NVMeof protocol to support direct data transfer from files on NVMe to DRAM/VRAM via PCIe, without going through the CPU and achieving zero-copy. Users need to follow the instructions to mount remote storage nodes locally and use the openSegment
interface for reference to complete data read/write operations.
With the help of Transfer Engine, Mooncake Store can achieve local DRAM/VRAM reading and writing of specified parts in valid segments through TCP, (GPUDirect) RDMA, NVMe-of protocols, etc.
Remote ↓ Local → | DRAM | VRAM |
---|---|---|
DRAM | ✓ | ✓ |
VRAM | ✓ | ✓ |
NVMe-of | ✓ | ✓ |
- Local memcpy: If the target Segment is actually in the local DRAM/VRAM, direct data copy interfaces such as memcpy, cudaMemcpy are used.
- TCP: Supports data transfer between local DRAM and remote DRAM.
- RDMA: Supports data transfer between local DRAM/VRAM and remote DRAM. It supports multi-network card pooling and retry functions in implementation.
- cuFile (GPUDirect Storage): Implements data transfer between local DRAM/VRAM and Local/Remote NVMeof.
The BatchTransfer API uses an array of requests, which specify the operation type (READ or WRITE), data length, and local and remote memory addresses. The transfer operation is applicable to DRAM and GPU VRAM. The completion of these operations can be asynchronously monitored through the getTransferStatus
API.
Modern inference servers often consist of multiple CPU sockets, DRAM, GPUs, and RDMA NIC devices. Although it's technically possible to transfer data from local DRAM or VRAM to a remote location using any RDMA NIC, these transfers can be limited by the bandwidth constraints of the Ultra Path Interconnect (UPI) or PCIe Switch. To overcome these limitations, Transfer Engine implements a topology-aware path selection algorithm.
Before processing requests, each server generates a topology matrix and broadcasts it across the cluster. This matrix categorizes network interface cards (NICs) into preferred and secondary lists for various types of memory, which types are specified during memory registration. Under normal conditions, a NIC from the preferred list is selected for transfers, facilitating RDMA operations within the local NUMA or GPU Direct RDMA through the local PCIe switch only. In case of failures, NICs from both lists may be utilized. The process involves identifying the appropriate local and target NICs based on the memory addresses, establishing a connection, and executing the data transfer.
For instance, as illustrated in figure above, to transfer data from buffer 0 (assigned to cpu:0) in the local node to buffer 1 (assigned to cpu:1) in the target node, the engine first identifies the preferred NICs for cpu:0 using the local server's topology matrix and selects one, such as mlx5_1, as the local NIC. Similarly, the target NIC, such as mlx5_3, is selected based on the target memory address. This setup enables establishing an RDMA connection from mlx5_1@local to mlx5_3@target to carry out RDMA read and write operations.
To further maximize bandwidth utilization, if a single request's transfer is internally divided into multiple slices if its length exeeds 16KB. Each slice might use a different path, enabling collaborative work among all RDMA NICs.
If you do not want to manually configure the topology matrix, we also provide a function (mooncake::discoverTopologyMatrix
in topology.h
) to automatically discover the toplogy between CPU/CUDA and RDMA devices. Supports for more device types are working in progress. The automatic discovery mechanism might not always be accurate, and we welcome your feedbacks and improvement ideas!
Mooncake Store employs a pair of end- points to represent the connection between a local RDMA NIC and a remote RDMA NIC. In practice, each endpoint includes one or more RDMA queue pair objects. Connec- tions in Mooncake Store are established in an on demand manner; endpoints remain unpaired until the first request is made. To prevent a large number of endpoints from slowing down request processing, Mooncake Store employs endpoint pooling, which caps the maximum number of active connections. We use the SIEVE algorithm to manage endpoint eviction. If a connection fails due to link errors, it is removed from the endpoint pools on both sides and re-established during the next data transfer attempt.
In a multi-NIC environment, one common failure scenario is the temporary unavailability of a specific NIC, while other routes may still connect two nodes. Mooncake Store is designed to adeptly manage such temporary failures effectively. If a connection is identified as unavailable, Mooncake Store automatically identifies an alternative, reachable path and resubmits the request to a different RDMA NIC device. Furthermore, Mooncake Store is capable of detecting prob- lems with other RDMA resources, including RDMA contexts and completion queues. It temporarily avoids using these resources until the issue, such as a downed link, is resolved.
The sample program provided in mooncake-transfer-engine/example/transfer_engine_bench.cpp
demonstrates the basic usage of Transfer Engine by repeatedly reading/writing data blocks from the DRAM of the target node to the initiator node through the Transfer Engine interface. It can also be used to measure read and write throughput. Currently, the Transfer Engine Bench tool supports RDMA and TCP protocols.
After successfully compiling Transfer Engine, the test program transfer_engine_bench
can be found in the build/mooncake-transfer-engine/example
directory.
-
Start the
metadata
service. This service is used for the centralized highly available management of various metadata for Mooncake, including the internal connection status of Transfer Engine. It is necessary to ensure that both the initiator and target nodes can smoothly access this metadata service, so pay attention to:- The listening IP of the metadata service should not be 127.0.0.1; it should be determined in conjunction with the network environment. In the experimental environment, 0.0.0.0 can be used.
- On some platforms, if the initiator and target nodes have set the
http_proxy
orhttps_proxy
environment variables, it will also affect the communication between Transfer Engine and the metadata service.
Transfer Engine support multiple kinds of metadata services, including
etcd
,redis
, andhttp
. The following describes how to start the metadata service usingetcd
andhttp
as examples.1.1.
etcd
For example, the following command line can be used to start the etcd service:
# This is 10.0.0.1 etcd --listen-client-urls http://0.0.0.0:2379 --advertise-client-urls http://10.0.0.1:2379
1.2.
http
For example, you can use the
http
service in themooncake-transfer-engine/example/http-metadata-server
example:# This is 10.0.0.1 # cd mooncake-transfer-engine/example/http-metadata-server go run . --addr=:8080
-
Start the target node.
# This is 10.0.0.2 export MC_GID_INDEX=n ./transfer_engine_bench --mode=target \ --metadata_server=10.0.0.1:2379 \ --local_server_name=10.0.0.2:12345 \ --device_name=erdma_0
The meanings of the various parameters are as follows:
- The default value of the parameter corresponding to the environment variable
MC_GID_INDEX
is 0, which means that the Transfer Engine selects a GID that is most likely to be connected. If the connection is hung, the user still needs to set the value of such a environment variable manually. --mode=target
indicates the start of the target node. The target node does not initiate read/write requests; it passively supplies or writes data as required by the initiator node.Note: In actual applications, there is no need to distinguish between target nodes and initiator nodes; each node can freely initiate read/write requests to other nodes in the cluster.
--metadata_server
is the address of the metadata server (the full address of the etcd service).Change
--metadata_server
to--metadata_server=http://10.0.0.1:8080/metadata
and add--metadata_type=http
when usinghttp
as themetadata
service.--local_server_name
represents the address of this machine, which does not need to be set in most cases. If this option is not set, the value is equivalent to the hostname of this machine (i.e.,hostname(2)
). Other nodes in the cluster will use this address to attempt out-of-band communication with this node to establish RDMA connections.Note: If out-of-band communication fails, the connection cannot be established. Therefore, if necessary, you need to modify the
/etc/hosts
file on all nodes in the cluster to locate the correct node through the hostname.--device_name
indicates the name of the RDMA network card used in the transfer process.Tip: Advanced users can also pass in a JSON file of the network card priority matrix through
--nic_priority_matrix
, for details, refer to the developer manual of Transfer Engine.- In network environments that only support TCP, the
--protocol=tcp
parameter can be used; in this case, there is no need to specify the--device_name
parameter.
- The default value of the parameter corresponding to the environment variable
-
Start the initiator node.
# This is 10.0.0.3 export MC_GID_INDEX=n ./transfer_engine_bench --metadata_server=10.0.0.1:2379 \ --segment_id=10.0.0.2:12345 \ --local_server_name=10.0.0.3:12346 \ --device_name=erdma_1
The meanings of the various parameters are as follows (the rest are the same as before):
--segment_id
can be simply understood as the hostname of the target node and needs to be consistent with the value passed to--local_server_name
when starting the target node (if any).
Under normal circumstances, the initiator node will start the transfer operation, wait for 10 seconds, and then display the "Test completed" message, indicating that the test is complete.
The initiator node can also configure the following test parameters:
--operation
(can be"read"
or"write"
),batch_size
,block_size
,duration
,threads
, etc.
If an exception occurs during execution, it is usually due to incorrect parameter settings. It is recommended to refer to the troubleshooting document for preliminary troubleshooting.
The following video shows a normal run as described above, with the Target on the right and the Initiator on the left, at the end of the test the Initiator reports the test duration (10 seconds), IOPS (379008 requests/s), and throughput (19.87 GiB/s). The throughput here exceeds the maximum throughput supported by a single card on the host computer used.
Transfer Engine provides interfaces through the TransferEngine
class (located in mooncake-transfer-engine/include/transfer_engine.h
), where the specific data transfer functions for different backends are implemented by the Transport
class, currently supporting TcpTransport
, RdmaTransport
and NVMeoFTransport
.
The core API provided by Mooncake Transfer Engine is submitting a group of asynchronous Transport::TransferRequest
tasks through the Transport::submitTransfer
interface, and querying their status through the Transport::getTransferStatus
interface. Each Transport::TransferRequest
specifies reading or writing a continuous data space of length
starting from the local starting address source
, to the position starting at target_offset
in the segment corresponding to target_id
.
The Transport::TransferRequest
structure is defined as follows:
using SegmentID = int32_t;
struct TransferRequest
{
enum OpCode { READ, WRITE };
OpCode opcode;
void *source;
SegmentID target_id; // The ID of the target segment, which may correspond to local or remote DRAM/VRAM/NVMeof, with the specific routing logic hidden
size_t target_offset;
size_t length;
};
opcode
takes the valuesREAD
orWRITE
.READ
indicates that data is copied from the target address indicated by<target_id, target_offset>
to the local starting addresssource
;WRITE
indicates that data is copied fromsource
to the address indicated by<target_id, target_offset>
.source
represents the DRAM/VRAM buffer managed by the currentTransferEngine
, which must have been registered in advance by theregisterLocalMemory
interface.target_id
represents the segment ID of the transfer target. The segment ID is obtained using theopenSegment
interface. Segments are divided into the following types:- RAM space type, covering DRAM/VRAM. As mentioned earlier, there is only one segment under the same process (or
TransferEngine
instance), which contains various types of Buffers (DRAM/VRAM). In this case, the segment name passed to theopenSegment
interface is equivalent to the server hostname.target_offset
is the virtual address of the target server. - NVMeOF space type, where each file corresponds to a segment. In this case, the segment name passed to the
openSegment
interface is equivalent to the unique identifier of the file.target_offset
is the offset of the target file.
- RAM space type, covering DRAM/VRAM. As mentioned earlier, there is only one segment under the same process (or
length
represents the amount of data transferred. TransferEngine may further split this into multiple read/write requests internally.
BatchID allocateBatchID(size_t batch_size);
Allocates a BatchID
. A maximum of batch_size
TransferRequest
s can be submitted under the same BatchID
.
batch_size
: The maximum number ofTransferRequest
s that can be submitted under the sameBatchID
;- Return value: If successful, returns
BatchID
(non-negative); otherwise, returns a negative value.
int submitTransfer(BatchID batch_id, const std::vector<TransferRequest> &entries);
Submits new TransferRequest
tasks to batch_id
. The task is asynchronously submitted to the background thread pool. The total number of entries
accumulated under the same batch_id
should not exceed the batch_size
defined at creation.
batch_id
: TheBatchID
it belongs to;entries
: Array ofTransferRequest
;- Return value: If successful, returns 0; otherwise, returns a negative value.
enum TaskStatus
{
WAITING, // In the transfer phase
PENDING, // Not supported
INVALID, // Ilvalid parameters
CANNELED, // Not supported
COMPLETED, // Transfer completed
TIMEOUT, // Not supported
FAILED // Transfer failed even after retries
};
struct TransferStatus {
TaskStatus s;
size_t transferred; // How much data has been successfully transferred (not necessarily an accurate value, but it is a lower bound)
};
int getTransferStatus(BatchID batch_id, size_t task_id, TransferStatus &status)
Obtains the running status of the TransferRequest
with task_id
in batch_id
.
batch_id
: TheBatchID
it belongs to;task_id
: The sequence number of theTransferRequest
to query;status
: Output Transfer status;- Return value: If successful, returns 0; otherwise, returns a negative value.
int freeBatchID(BatchID batch_id);
Recycles BatchID
, and subsequent operations on submitTransfer
and getTransferStatus
are undefined. If there are still TransferRequest
s pending completion in the BatchID
, the operation is refused.
batch_id
: TheBatchID
it belongs to;- Return value: If successful, returns 0; otherwise, returns a negative value.
The TransferEngine
class internally manages multiple backend Transport
classes, and users can load or unload Transport
for different backends in TransferEngine
.
Transport* installOrGetTransport(const std::string& proto, void** args);
Registers Transport
in TransferEngine
. If a Transport
for a certain protocol already exists, it returns that Transport
.
proto
: The name of the transport protocol used byTransport
, currently supportingtcp
,rdma
,nvmeof
.args
: Additional parameters required forTransport
initialization, presented as a variable-length array, with the last member beingnullptr
.- Return value: If
proto
is within the determined range, returns theTransport
corresponding toproto
; otherwise, returns a null pointer.
For TCP transfer mode, there is no need to pass args
objects when registering the Transport
object.
engine->installOrGetTransport("tcp", nullptr);
For RDMA transfer mode, the network card priority marrix must be specified through args
during the registration of Transport
.
void** args = (void**) malloc(2 * sizeof(void*));
args[0] = /* topology matrix */;
args[1] = nullptr;
engine->installOrGetTransport("rdma", args);
The network card priority marrix is a JSON string indicating the storage medium name and the list of network cards to be used preferentially, as shown in the example below:
{
"cpu:0": [["mlx0", "mlx1"], ["mlx2", "mlx3"]],
"cuda:0": [["mlx1", "mlx0"]],
...
}
Each key
represents the device name corresponding to a CPU socket or a GPU device.
Each value
is a tuple of (preferred_nic_list
, accessable_nic_list
), each of which is a list of NIC names.
preferred_nic_list
indicates the preferred NICs, such as NICs directly connected to the CPU rather than across NUMA, or NICs under the same PCIe Switch for GPUs.accessable_nic_list
indicates NICs that are not preferred but can theoretically connect, used for fault retry scenarios.
For NVMeOF transfer mode, the file path must be specified through args
during the registration of Transport
.
void** args = (void**) malloc(2 * sizeof(void*));
args[0] = /* topology matrix */;
args[1] = nullptr;
engine->installOrGetTransport("nvmeof", args);
int uninstallTransport(const std::string& proto);
Unloads Transport
from TransferEngine
.
proto
: The name of the transport protocol used byTransport
, currently supportingrdma
,nvmeof
.- Return value: If successful, returns 0; otherwise, returns a negative value.
For the RDMA transfer process, the source pointer TransferRequest::source
must be registered in advance as an RDMA readable/writable Memory Region space, that is, included as part of the RAM Segment of the current process. Therefore, the following functions are needed:
int registerLocalMemory(void *addr, size_t size, string location, bool remote_accessible);
Registers a space starting at address addr
with a length of size
on the local DRAM/VRAM.
addr
: The starting address of the registration space;size
: The length of the registration space;location
: Thedevice
corresponding to this memory segment, such ascuda:0
indicating the GPU device,cpu:0
indicating the CPU socket, by matching with the network card priority order table (seeinstallOrGetTransport
), the preferred network card is identified.remote_accessible
: Indicates whether this memory can be accessed by remote nodes.- Return value: If successful, returns 0; otherwise, returns a negative value.
int unregisterLocalMemory(void *addr);
Unregisters the region.
addr
: The starting address of the registration space;- Return value: If successful, returns 0; otherwise, returns a negative value.
TransferEngine provides the openSegment
function, which obtains a SegmentHandle
for subsequent Transport
transfers.
SegmentHandle openSegment(const std::string& segment_name);
segment_name
: The unique identifier of the segment. For RAM Segment, this needs to be consistent with theserver_name
filled in by the peer process when initializing the TransferEngine object.- Return value: If successful, returns the corresponding
SegmentHandle
; otherwise, returns a negative value.
int closeSegment(SegmentHandle segment_id);
segment_id
: The unique identifier of the segment.- Return value: If successful, returns 0; otherwise, returns a negative value.
Metadata Format
// Used to find the communicable address and exposed rpc port based on server_name.
// Created: when calling TransferEngine::init().
// Deleted: when TransferEngine is destructed.
Key = mooncake/rpc_meta/[server_name]
Value = {
'ip_or_host_name': 'node01'
'rpc_port': 12345
}
// For segments, the key naming method of mooncake/[proto]/[segment_name] is used, and the segment name can use the Server Name.
// A segment corresponds to a machine, and a buffer corresponds to different segments of memory or different files or different disks on the machine. Different buffers of the same segment are in the same fault domain.
// RAM Segment, used by RDMA Transport to obtain transfer information.
// Created: command line tool register.py, at this time buffers are empty, only fill in the information that can be known in advance.
// Modified: TransferEngine at runtime through register / unregister to add or delete Buffer.
Key = mooncake/ram/[segment_name]
Value = {
'server_name': server_name,
'protocol': rdma,
'devices': [
{ 'name': 'mlx5_2', 'lid': 17, 'gid': 'fe:00:...' },
{ 'name': 'mlx5_3', 'lid': 22, 'gid': 'fe:00:...' }
],
'priority_matrix': {
"cpu:0": [["mlx5_2"], ["mlx5_3"]],
"cpu:1": [["mlx5_3"], ["mlx5_2"]],
"cuda:0": [["mlx5_2"], ["mlx5_3"]],
},
'buffers': [
{
'name': 'cpu:0',
'addr': 0x7fa16bdf5000,
'length': 1073741824,
'rkey': [1fe000, 1fdf00, ...], // The length is the same as the number of elements in the 'devices' field
},
],
}
// Created: command line tool register.py, determine the file path that can be mounted.
// Modified: command line tool mount.py, add a mapping of the machine mounting the file to the file path on the mounting machine to the buffers.local_path_map.
Key = mooncake/nvmeof/[segment_name]
Value = {
'server_name': server_name,
'protocol': nvmeof,
'buffers':[
{
'length': 1073741824,
'file_path': "/mnt/nvme0" // The file path on this machine
'local_path_map': {
"node01": "/mnt/transfer_engine/node01/nvme0", // The machine mounting the file -> The file path on the mounting machine
.....
},
},
{
'length': 1073741824,
'file_path': "/mnt/nvme1",
'local_path_map': {
"node02": "/mnt/transfer_engine/node02/nvme1",
.....
},
}
]
}
The HTTP server should implement three following RESTful APIs, while the metadata server configured to http://host:port/metadata
as an example:
GET /metadata?key=$KEY
: Get the metadata corresponding to$KEY
.PUT /metadata?key=$KEY
: Update the metadata corresponding to$KEY
to the value of the request body.DELETE /metadata?key=$KEY
: Delete the metadata corresponding to$KEY
.
For specific implementation, refer to the demo service implemented in Golang at mooncake-transfer-engine/example/http-metadata-server.
TransferEngine(std::unique_ptr<TransferMetadata> metadata_client);
TransferMetadata(const std::string &metadata_server, const std::string &protocol = "etcd");
- Pointer to a
TransferMetadata
object, which abstracts the communication logic between the TransferEngine framework and the metadata server. We currently supportetcd
,redis
andhttp
protocols, whilemetadata_server
represents the IP address or hostname of the etcd or redis server, or the base HTTP URI of http server.
For easy exception handling, TransferEngine needs to call the init function for secondary construction after construction:
int init(std::string& server_name, std::string& connectable_name, uint64_t rpc_port = 12345);
server_name
: The local server name, ensuring uniqueness within the cluster. It also serves as the name of the RAM Segment that other nodes refer to the current instance (i.e., Segment Name).connectable_name
: The name used for other clients to connect, which can be a hostname or IP address.rpc_port
: The rpc port used for interaction with other clients.- Return value: If successful, returns 0; if TransferEngine has already been init, returns -1.
~TransferEngine();
Reclaims all allocated resources and also deletes the global meta data server information.
After compiling Mooncake Store, you can move the compiled static library file libtransfer_engine.a
and the C header file transfer_engine_c.h
into your own project. There is no need to reference other files under src/transfer_engine
.
During the project build phase, you need to configure the following options for your application:
-I/path/to/include
-L/path/to/lib -ltransfer_engine
-lnuma -lglog -libverbs -ljsoncpp -letcd-cpp-api -lprotobuf -lgrpc++ -lgrpc
To support the operational needs of P2P Store, Transfer Engine provides a Golang interface wrapper, see mooncake-p2p-store/src/p2pstore/transfer_engine.go
.
When compiling the project, enable the -DWITH_P2P_STORE=ON
option to compile the P2P Store example program at the same time.
Under mooncake-transfer-engine/rust
, the Rust interface implementation of TransferEngine is provided, and a Rust version of the benchmark is implemented based on the interface, similar to transfer_engine_bench.cpp. To compile the rust example, you need to install the Rust SDK and add -DWITH_RUST_EXAMPLE=ON
in the cmake command.
For advanced users, TransferEngine provides the following advanced runtime options, all of which can be passed in through environment variables.
MC_NUM_CQ_PER_CTX
The number of CQs created per device instance, default value 1MC_NUM_COMP_CHANNELS_PER_CTX
The number of Completion Channel created per device instance, default value 1MC_IB_PORT
The IB port number used per device instance, default value 1MC_GID_INDEX
The GID index used per device instance, default value 3 (or the maximum value supported by the platform)MC_MAX_CQE_PER_CTX
The CQ buffer size per device instance, default value 4096MC_MAX_EP_PER_CTX
The maximum number of active EndPoint per device instance, default value 256MC_NUM_QP_PER_EP
The number of QPs per EndPoint, the more the number, the better the fine-grained I/O performance, default value 2MC_MAX_SGE
The maximum number of SGEs supported per QP, default value 4 (or the highest value supported by the platform)MC_MAX_WR
The maximum number of Work Request supported per QP, default value 256 (or the highest value supported by the platform)MC_MAX_INLINE
The maximum Inline write data volume (bytes) supported per QP, default value 64 (or the highest value supported by the platform)MC_MTU
The MTU length used per device instance, can be 512, 1024, 2048, 4096, default value 4096 (or the maximum length supported by the platform)MC_WORKERS_PER_CTX
The number of asynchronous worker threads corresponding to each device instanceMC_SLICE_SIZE
The segmentation granularity of user requests in Transfer EngineMC_RETRY_CNT
The maximum number of retries in Transfer EngineMC_VERBOSE
If this option is set, more detailed logs will be output during runtime