Commit b954b568 authored by Paras Garg's avatar Paras Garg

Added code for follower

parent 72250328
No preview for this file type
No preview for this file type
......@@ -3,12 +3,59 @@
#include <queue>
#include <mutex>
#include <condition_variable>
#include <set>
#include <string>
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include "Logger.hpp"
class Comparator
{
public:
inline bool operator()(const struct ibv_wc *c1, const struct ibv_wc *c2)
const
{
struct MessageHeader *req1 = (struct MessageHeader *)c1->wr_id;
struct MessageHeader *req2 = (struct MessageHeader *)c2->wr_id;
if (req1->keySize != req2->keySize)
return true;
char *key1 = (char *)req1 + MessageHeaderSize;
char *key2 = (char *)req2 + MessageHeaderSize;
for (uint32_t i = 0; i < req1->keySize; i++)
{
if (key1[i] != key2[i])
return true;
}
return false;
}
};
class ConcurrentQueue
{
private:
std::queue<struct ibv_wc *> queue1;
std::queue<struct ibv_wc *> queue2;
std::mutex queueMutex;
std::set<struct ibv_wc *, Comparator> runningRequests;
std::condition_variable queueCv;
public:
void push(struct ibv_wc *const &data);
bool empty();
struct ibv_wc *try_pop();
void removeFromSet(struct ibv_wc *data);
void wait_and_pop(struct ibv_wc *&popped_value);
};
#endif
/*
template <typename Data>
class ConcurrentQueue
{
private:
std::queue<Data> queue;
std::queue1<Data> queue;
std::queue2<Data> queue;
std::mutex queueMutex;
std::condition_variable queueCv;
......@@ -16,7 +63,7 @@ public:
void push(Data const &data)
{
std::unique_lock<std::mutex> lock(queueMutex);
queue.push(data);
queue1.push(data);
lock.unlock();
queueCv.notify_one();
}
......@@ -30,7 +77,7 @@ public:
bool try_pop(Data &popped_value)
{
std::unique_lock<std::mutex> lock(queueMutex);
if (queue.empty())
if (queue2.empty())
{
return false;
}
......@@ -48,4 +95,5 @@ public:
queue.pop();
}
};
#endif
\ No newline at end of file
#endif
*/
\ No newline at end of file
#ifndef __Executor__
#define __Executor__
#include <vector>
#include <map>
#include "CqEventData.hpp"
#include "ConcurrentQueue.hpp"
#include "RdmaEndpointGroup.hpp"
#include "TaskThread.hpp"
#include "RdmaSalEndpoint.hpp"
#include "Logger.hpp"
class Executor
{
int _size{0};
std::vector<TaskThread *> *_taskThreads{NULL};
ConcurrentQueue<struct ibv_wc *> *_taskQueue{NULL};
std::unordered_map<uint32_t, RdmaSalEndpoint *> *_qpSalEndpointMap{NULL};
std::vector<TaskThread *> _taskThreads{NULL};
ConcurrentQueue *_taskQueue{NULL};
public:
Executor(int size,std::unordered_map<uint32_t, RdmaSalEndpoint *> *_qpSalEndpointMap);
Executor(int size);
void createThreads(std::unordered_map<uint32_t, RdmaRepEndpoint *> *clientRepMap,
std::unordered_map<uint32_t, RdmaRepEndpoint *> *serverRepMap,
std::unordered_map<uint32_t, RdmaSalEndpoint *> *salMap);
void submit(struct ibv_wc *task);
void getTask();
void stop();
};
//long affinities[]
// long affinities[]
#endif
\ No newline at end of file
#ifndef __MessageFormats__
#define __MessageFormats__
enum RequestType
enum MessageType
{
GET,
PUT,
DELETE,
INVALIDATE
};
enum ResponseStatus
{
SUCCESS,
FAILURE
GET = (1u << 0),
PUT = (1u << 1),
DELETE = (1u << 2),
INVALIDATE = (1u << 3),
SUCCESS = (1u << 4),
FAILURE = (1u <<5)
};
struct __attribute__ ((__packed__)) SalRequestHeader
struct __attribute__ ((__packed__)) MessageHeader
{
uint32_t id;
enum RequestType type;
enum MessageType type;
uint32_t keySize;
uint32_t valueSize;
};
static const uint32_t MessageHeaderSize = sizeof(MessageHeader);
/*
struct __attribute__ ((__packed__)) SalResponseHeader
{
uint32_t id;
enum ResponseStatus status;
/*
* Note value will be present only in case of response status is success
*/
enum MessageType status;
//Note value will be present only in case of response status is success
uint32_t valueSize;
};
struct __attribute__ ((__packed__)) InvRequestHeader
{
uint32_t id;
enum RequestType type;
enum MessageType type;
uint32_t keySize;
};
......@@ -42,11 +40,12 @@ struct __attribute__ ((__packed__)) InvRequestHeader
struct __attribute__ ((__packed__)) InvResponseHeader
{
uint32_t id;
enum ResponseStatus status;
enum MessageType status; MessageHeader
};
static uint32_t SalRequestHeaderSize = sizeof(SalRequestHeader);
*/
/*
static uint32_t SalResponseHeaderSize = sizeof(SalResponseHeader);
static uint32_t InvRequestHeaderSize = sizeof(InvRequestHeader);
static uint32_t InvResponseHeaderSize = sizeof(InvResponseHeader);
*/
#endif
\ No newline at end of file
#ifndef __Properties__
#define __Properties__
#include <string>
#include <iostream>
#include <fstream>
#include <map>
class Properties
{
private:
std::map<std::string, std::string> _props;
const std::string _WHITESPACE = " \n\r\t\f\v";
std::string ltrim(const std::string &s);
std::string rtrim(const std::string &s);
std::string trim(const std::string &);
public:
Properties(std::string filename);
std::string getValue(std::string key);
};
#endif
\ No newline at end of file
#ifndef __RDMACLIENTREPENDPOINT__
#define __RDMACLIENTREPENDPOINT__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
......@@ -5,16 +8,18 @@
#include <iostream>
#include <errno.h>
#include <arpa/inet.h>
#ifndef __RDMACLIENTENDPOINT__
#define __RDMACLIENTENDPOINT__
#include <mutex>
#include <atomic>
#include <shared_mutex>
#include <queue>
#include <rocksdb/db.h>
#include "RdmaRepEndpoint.hpp"
#include "Logger.hpp"
#include "MessageFormats.hpp"
#include <boost/lockfree/queue.hpp>
class RdmaClientEndpoint
class RdmaClientRepEndpoint : public RdmaRepEndpoint
{
static int CONN_STATE_INITIALIZED;
static int CONN_STATE_ADDR_RESOLVED;
static int CONN_STATE_ROUTE_RESOLVED;
......@@ -22,42 +27,28 @@ class RdmaClientEndpoint
static int CONN_STATE_CONNECTED;
static int CONN_STATE_PARTIAL_CLOSED;
static int CONN_STATE_CLOSED;
int _state{0};
int _maxInLine{0};
int _timeoutMs{0};
int _retryCount{0};
int _maxRetryCount{2};
struct rdma_cm_id *_cm_id = NULL;
struct ibv_pd *_protectionDomain;
int _sendQueueSize;
int _recvQueueSize;
int _sendMsgSize;
int _recvMsgSize;
int _state;
int _timeoutMs;
int _maxInLine;
const char *_connData;
void *_sendBuff = NULL;
void *_recvBuff = NULL;
struct ibv_mr *_sendMr = NULL;
struct ibv_mr *_recvMr = NULL;
boost::lockfree::queue<void *> *_sendBuffers;
void completeClose();
void connect();
void registerMemory();
public:
std::atomic<uint64_t> _requestId{12};
RdmaClientEndpoint(struct rdma_cm_id *id, int sendQueueSize, int recvQueueSize,
int sendMsgSize, int recvMsgSize, int maxInLine, int timeout);
RdmaClientRepEndpoint(struct rdma_cm_id *id, int sendQueueSize, int recvQueueSize,
int sendMsgSize, int recvMsgSize, int maxInLine, int timeout, rocksdb::DB *_db);
void connect();
void connect(const char *ip, const char *port, const char *connData);
bool isConnected();
void processCmEvent(struct rdma_cm_event *event);
void createResources(struct ibv_cq *cq);
void close();
int sendMessage(const char *buffer, int size);
void processSendComp(struct ibv_wc);
void processRecvComp(struct ibv_wc);
};
#endif
\ No newline at end of file
#ifndef __RDMACMPROCESSOR__
#define __RDMACMPROCESSOR__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
#include <thread>
#include <iostream>
#include "Logger.hpp"
#include "RdmaEndpointGroup.hpp"
class RdmaCmProcessor
{
struct rdma_event_channel *_eventChannel{NULL};
std::thread *_cmEventThread{NULL};
RdmaEndpointGroup *_endpointGroup{NULL};
bool _stop{false};
public:
RdmaCmProcessor(RdmaEndpointGroup *group);
struct rdma_cm_id *createId();
void processCmEvent();
void start(bool newThread);
void close();
};
#endif
\ No newline at end of file
#ifndef __RDMACQPROCESSOR__
#define __RDMACQPROCESSOR__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
......@@ -5,30 +8,23 @@
#include <thread>
#include <unordered_map>
#ifndef __RDMACQPROCESSOR__
#define __RDMACQPROCESSOR__
#include "RdmaClientEndpoint.hpp"
#include "Executor.hpp"
#include "Logger.hpp"
class RdmaCqProcessor
{
public:
struct ibv_comp_channel *_compChannel;
struct ibv_cq *_completionQueue;
std::thread *_compQueueThread;
std::unordered_map<uint32_t, RdmaClientEndpoint *> *_qpEndpointMap{NULL};
struct ibv_comp_channel *_compChannel{NULL};
struct ibv_cq *_completionQueue{NULL};
std::thread *_compQueueThread{NULL};
bool _stop{false};
Executor *_executor{NULL};
RdmaCqProcessor(ibv_context *verbs, int compQueueSize);
struct ibv_cq *getCq();
void start();
void processCQEvents();
void dispatchCqEvents(ibv_wc *wc_array, int size);
void close();
void registerEp(uint64_t qpum, RdmaClientEndpoint* ep);
RdmaCqProcessor(Executor *ex, ibv_context *verbs, int compQueueSize);
struct ibv_cq *getCq();
void start();
void processCQEvents();
void close();
};
#endif
// https://www.rdmamojo.com/2013/03/09/ibv_get_cq_event/
// https://man7.org/linux/man-pages/man3/ibv_poll_cq.3.html
// https://man7.org/linux/man-pages/man3/ibv_get_cq_event.3.html
\ No newline at end of file
#ifndef __RDMAENDPOINT__
#define __RDMAENDPOINT__
#include <iostream>
#include <boost/lockfree/queue.hpp>
#include <queue>
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
......@@ -12,7 +12,9 @@
#include <vector>
#include <mutex>
#include <shared_mutex>
#include "CqEventData.hpp"
#include "Logger.hpp"
class RdmaEndpoint
{
......@@ -36,15 +38,17 @@ public:
char *_recvBuff{NULL};
struct ibv_mr *_sendMr{NULL};
struct ibv_mr *_recvMr{NULL};
boost::lockfree::queue<void *> *_sendBuffers{NULL};
std::queue<char*> _sendBuffers;
std::mutex _sendBuffersM;
RdmaEndpoint(struct rdma_cm_id *id, struct ibv_cq *completionQueue, int sendQueueSize,
int recvQueueSize, int sendMsgSize, int recvMsgSize);
void createResources();
void processCmEvent(struct rdma_cm_event *event);
void clientClose();
void close();
virtual void processSendCompletion(struct ibv_wc *data) = 0;
virtual void processRecvCompletion(struct ibv_wc *data) = 0;
virtual ~RdmaEndpoint();
};
#endif
\ No newline at end of file
#ifndef __RDMA_ENDPOINT_GROUP__
#define __RDMA_ENDPOINT_GROUP__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
#include <thread>
#include <netdb.h>
#include <arpa/inet.h>
#include <map>
#include <vector>
#include <mutex>
#include <shared_mutex>
#include "RdmaSalEndpoint.hpp"
#include "RdmaRepEndpoint.hpp"
#include "Logger.hpp"
class RdmaEndpointGroup
{
public:
std::vector<RdmaSalEndpoint *> _salEps;
std::unordered_map<uint32_t, RdmaSalEndpoint *> _qpSalEndpointMap;
virtual void processCmEvent(struct rdma_cm_event *event) = 0;
};
#endif
\ No newline at end of file
#ifndef __RDMAREPENDPOINT__
#define __RDMAREPENDPOINT__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
#include <netdb.h>
#include <iostream>
#include <errno.h>
#include <arpa/inet.h>
#include <mutex>
#include <atomic>
#include <shared_mutex>
#include <queue>
#include <rocksdb/db.h>
#include "Logger.hpp"
#include "MessageFormats.hpp"
class RdmaRepEndpoint
{
public:
struct rdma_cm_id *_cm_id{NULL};
struct ibv_pd *_protectionDomain;
int _sendQueueSize;
int _recvQueueSize;
int _sendMsgSize;
int _recvMsgSize;
char *_sendBuff = NULL;
char *_recvBuff = NULL;
struct ibv_mr *_sendMr = NULL;
struct ibv_mr *_recvMr = NULL;
std::queue<char *> _sendBuffers;
std::mutex _sendBuffersM;
rocksdb::DB *_db;
void connect();
void registerMemory();
public:
std::atomic<uint64_t> _requestId{12};
RdmaRepEndpoint(struct rdma_cm_id *id, int sendQueueSize, int recvQueueSize,
int sendMsgSize, int recvMsgSize, rocksdb::DB *_db);
void processSendCompletion(struct ibv_wc *data);
void processRecvCompletion(struct ibv_wc *data);
void processDelete(struct MessageHeader *req);
void processGet(struct MessageHeader *req);
void processPut(struct MessageHeader *req);
int sendMessage(const char *buffer, uint32_t size);
};
#endif
\ No newline at end of file
#ifndef __RDMAREPENDPOINTGROUP__
#define __RDMAREPENDPOINTGROUP__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
#include <errno.h>
#include <unordered_map>
#ifndef __RDMACLIENTENDPOINTGROUP__
#define __RDMACLIENTENDPOINTGROUP__
#include <rocksdb/db.h>
#include "RdmaCqProcessor.hpp"
#include "RdmaClientEndpoint.hpp"
#include "RdmaRepEndpoint.hpp"
#include "RdmaClientRepEndpoint.hpp"
#include "Executor.hpp"
class RdmaClientEndpointGroup
class RdmaRepEndpointGroup
{
RdmaCqProcessor *_cqProcessor = NULL;
//struct rdma_cm_id *_cm_id;
// struct rdma_cm_id *_cm_id;
int _sendQueueSize;
int _recvQueueSize;
int _compQueueSize;
int _sendMsgSize;
int _recvMsgSize;
int _timeoutMs;
int _maxInLine;
int _timeoutMs;
struct rdma_event_channel *_eventChannel;
std::thread *_cmEventThread;
bool _stop;
bool _stopCMThread;
Executor *_executor;
rocksdb::DB *_db;
public:
RdmaClientEndpointGroup(int sendQueueSize, int recvQueueSize, int compQueueSize, int sendMsgSize,
int recvMsgSize,int maxInLine, int timeout);
std::unordered_map<uint32_t, RdmaRepEndpoint *> *_qpRepEndpointMap;
RdmaRepEndpointGroup(int sendQueueSize, int recvQueueSize, int compQueueSize, int sendMsgSize,
int recvMsgSize, int maxInLine, int timeout, rocksdb::DB *db);
void processCmEvents();
void processCmEvent(struct rdma_cm_event *event);
void start();
struct ibv_cq *createCq(struct rdma_cm_id *id);
RdmaClientEndpoint *createEndpoint();
void startCmProcessor();
void setExecutor(Executor *ex);
RdmaClientRepEndpoint *createEndpoint();
void close();
struct ibv_cq *createCq(struct rdma_cm_id *id);
};
#endif
\ No newline at end of file
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
#include <iostream>
#include <thread>
#include <unordered_map>
#ifndef __RDMASALCQPROCESSOR__
#define __RDMASALCQPROCESSOR__
#include "Executor.hpp"
class RdmaSalCqProcessor
{
public:
struct ibv_comp_channel *_compChannel{NULL};
struct ibv_cq *_completionQueue{NULL};
std::thread *_compQueueThread{NULL};
bool _stop{false};
Executor *_executor{NULL};
RdmaSalCqProcessor(Executor *ex, ibv_context *verbs, int compQueueSize);
struct ibv_cq *getCq();
void start();
void processCQEvents();
void close();
};
#endif
#ifndef __RdmaInvEndpoint__
#define __RdmaInvEndpoint__
#ifndef __RdmaSalEndpoint__
#define __RdmaSalEndpoint__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
#include <errno.h>
#include <iostream>
#include <boost/lockfree/queue.hpp>
#include <rocksdb/db.h>
#include "RdmaEndpoint.hpp"
#include "MessageFormats.hpp"
#include <rocksdb/db.h>
#include "Logger.hpp"
class RdmaSalEndpoint : public RdmaEndpoint
class RdmaSalEndpoint
{
public:
static int CONN_STATE_INITIALIZED;
static int CONN_STATE_RESOURCES_ALLOCATED;
static int CONN_STATE_CONNECTED;
static int CONN_STATE_CLOSED;
int _state{0};
struct rdma_cm_id *_cm_id{NULL};
struct ibv_cq *_completionQueue{NULL};
struct ibv_pd *_protectionDomain{NULL};
int _sendQueueSize{0};
int _recvQueueSize{0};
int _sendMsgSize{0};
int _recvMsgSize{0};
char *_sendBuff{NULL};
char *_recvBuff{NULL};
struct ibv_mr *_sendMr{NULL};
struct ibv_mr *_recvMr{NULL};
std::queue<char*> _sendBuffers;
std::mutex _sendBuffersM;
rocksdb::DB *_db;
RdmaSalEndpoint(struct rdma_cm_id *id, struct ibv_cq *completionQueue, int sendQueueSize,
int recvQueueSize, int sendMsgSize, int recvMsgSize,rocksdb::DB *_db);
void createResources();
void processCmEvent(struct rdma_cm_event *event);
void processCqEvent(struct ibv_wc wc);
void processSendCompletion(struct ibv_wc *data);
void processRecvCompletion(struct ibv_wc *data);
void processDelete(struct SalRequestHeader *);
void processGet(struct SalRequestHeader *req);
void processPut(struct SalRequestHeader *req);
void processDelete(struct MessageHeader *req);
void processGet(struct MessageHeader *req);
void processPut(struct MessageHeader *req);
int sendMessage(const char *buffer, uint32_t size);
void close();
};
......
#ifndef __RdmaServerEndpointGroup__
#define __RdmaServerEndpointGroup__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
......@@ -9,39 +12,38 @@
#include <mutex>
#include <shared_mutex>
#ifndef __RDMASERVERENDPOINTGROUP__
#define __RDMASERVERENDPOINTGROUP__
#include "RdmaSalEndpoint.hpp"
#include "RdmaServerRepEndpoint.hpp"
#include "RdmaEndpointGroup.hpp"
#include "CqEventData.hpp"
#include "Executor.hpp"
#include "RdmaSalCqProcessor.hpp"
#include "RdmaCmProcessor.hpp"
#include "RdmaCqProcessor.hpp"
#include "Logger.hpp"
class RdmaServerEndpointGroup
{
/*
* Variables to maintain Group state
*/
* Variables to maintain Group state
*/
static int CONN_STATE_INITIALIZED;
static int CONN_STATE_BINDED;
static int CONN_STATE_CONNECTED;
static int CONN_STATE_CLOSED;
bool _stop{false};
std::thread *_cmEventThread{NULL};
struct rdma_event_channel *_eventChannel{NULL};
RdmaCqProcessor *_cqProcessor{NULL};
Executor *_executor;
struct rdma_cm_id *_cm_id{NULL};
RdmaSalCqProcessor *_salCqProcessor{NULL};
std::vector<RdmaSalEndpoint *> *_salEps{NULL};
std::unordered_map<uint32_t, RdmaSalEndpoint *> *_qpSalEndpointMap{NULL};
struct rdma_event_channel *_eventChannel{NULL};
std::thread *_cmEventThread{NULL};
bool _stopCmProcessor{false};
/*
* variables to maintain queue state
*/
* variables to maintain queue state
*/
int _sendQueueSize{0};
int _recvQueueSize{0};
int _compQueueSize{0};
......@@ -50,23 +52,24 @@ class RdmaServerEndpointGroup
rocksdb::DB *_db;
mutable std::shared_mutex _salMutex;
mutable std::shared_mutex _repMutex;
void clientClose();
public:
RdmaServerEndpointGroup(int sendQueueSize, int recvQueueSize, int compQueueSize, int sendMsgSize, int recvMsgSize);
void bind(const char *ip, const char *port, int backlog);
std::unordered_map<uint32_t, RdmaSalEndpoint *> *_qpSalEndpointMap;
std::unordered_map<uint32_t, RdmaRepEndpoint*> *_qpRepEndpointMap;
struct ibv_cq *createSalCq(struct rdma_cm_id *id);
void dispatchSalCqEvents(ibv_wc wc[], int size);
void createExecutor(int threadSize);
RdmaServerEndpointGroup(int sendQueueSize, int recvQueueSize, int compQueueSize,
int sendMsgSize, int recvMsgSize, rocksdb::DB *db);
// void setExecutor(Executor *executor);
void bind(const char *ip, const char *port, int backlog);
/*
* Sending false will run cmProcessor on calling thread used to save creation of new thread
* when main thread can also do same work
*/
void startCmProcessor(bool newThread);
struct ibv_cq *createCq(struct rdma_cm_id *id);
void dispatchCqEvents(ibv_wc wc[], int size);
void setExecutor(Executor *ex);
void processCmEvents();
void startCmProcessor(bool newThread);
void processCmEvent(struct rdma_cm_event *event);
void createEpCmEvent(struct rdma_cm_event *event);
void close();
......
#ifndef __RdmaServerRepEndpoint__
#define __RdmaServerRepEndpoint__
#include <rdma/rdma_cma.h>
#include <rdma/rdma_verbs.h>
#include <stdint.h>
#include <errno.h>
#include <iostream>
#include <rocksdb/db.h>
#include "RdmaEndpoint.hpp"
#include "MessageFormats.hpp"
#include "Logger.hpp"
#include "RdmaRepEndpoint.hpp"
class RdmaServerRepEndpoint : public RdmaRepEndpoint
{
public:
static int CONN_STATE_INITIALIZED;
static int CONN_STATE_RESOURCES_ALLOCATED;
static int CONN_STATE_CONNECTED;
static int CONN_STATE_CLOSED;
int _state{0};
struct ibv_cq *_completionQueue{NULL};
RdmaServerRepEndpoint(struct rdma_cm_id *id, struct ibv_cq *completionQueue, int sendQueueSize,
int recvQueueSize, int sendMsgSize, int recvMsgSize,rocksdb::DB *_db);
void createResources();
void processCmEvent(struct rdma_cm_event *event);
void processCqEvent(struct ibv_wc wc);
void close();
};
#endif
\ No newline at end of file
#ifndef __TaskThread__
#define __TaskThread__
#include "Runnable.hpp"
#include "CqEventData.hpp"
#include "ConcurrentQueue.hpp"
#include <pthread.h>
#include <iostream>
#include <queue>
#include <map>
#include <pthread.h>
#include "Runnable.hpp"
#include "CqEventData.hpp"
#include "MessageFormats.hpp"
#include "Logger.hpp"
#include "ConcurrentQueue.hpp"
#include "RdmaRepEndpoint.hpp"
#include "RdmaSalEndpoint.hpp"
class TaskThread
{
private:
ConcurrentQueue<struct ibv_wc *> *_taskQueue;
std::unordered_map<uint32_t, RdmaSalEndpoint *> *_qpSalEndpointMap;
ConcurrentQueue *_taskQueue;
bool _stop{false};
int _id;
pthread_t thread;
std::unordered_map<uint32_t, RdmaRepEndpoint *> *_clientRepMap;
std::unordered_map<uint32_t, RdmaRepEndpoint *> *_serverRepMap;
std::unordered_map<uint32_t, RdmaSalEndpoint *> *_salMap;
public:
TaskThread(int id, int cpu, ConcurrentQueue<struct ibv_wc *> *, std::unordered_map<uint32_t, RdmaSalEndpoint *> *_qpSalEndpointMap);
TaskThread(int id, ConcurrentQueue<struct ibv_wc *> *, std::unordered_map<uint32_t, RdmaSalEndpoint *> *_qpSalEndpointMap);
TaskThread(int id, int cpu, ConcurrentQueue *,
std::unordered_map<uint32_t, RdmaRepEndpoint *> *clientRepMap,
std::unordered_map<uint32_t, RdmaRepEndpoint *> *serverRepMap,
std::unordered_map<uint32_t, RdmaSalEndpoint *> *salMap);
void replicateSalRequest(char *salRequest, uint32_t size);
static void *run(void *object);
void stop();
void processEvent(struct ibv_wc *data);
void processEvent(RdmaSalEndpoint* ep,struct ibv_wc *data);
void processRepEvent(RdmaRepEndpoint* ep,struct ibv_wc *data);
~TaskThread();
};
......
......@@ -4,7 +4,21 @@
#All Parameters will be taken as string
# Fixed Parameters
ENABLE_LOGGING=0
sendQS=50
recvQS=50
compQS=50
sendMS=500
recvMS=500
DB_PATH=/tmp/testdb1
EXECUTOR_POOL_SIZE=4
ENABLE_LOGGING=1
SERVER_IP=192.168.200.20
SERVER_PORT=1921
EXECUTOR_POOL_SIZE=4
\ No newline at end of file
FOLLOWERS=1
FOLLOWER1_IP=192.168.200.20
FOLLOWER1_PORT=1920
FOLLOWER2_IP=192.168.200.20
FOLLOWER2_IP=1921
\ No newline at end of file
#include <TaskThread.hpp>
void ConcurrentQueue::push(struct ibv_wc *const &data)
{
std::unique_lock<std::mutex> lock(queueMutex);
std::cout << "putting data\n";
queue1.push(data);
lock.unlock();
queueCv.notify_one();
}
bool ConcurrentQueue::empty()
{
std::unique_lock<std::mutex> lock(queueMutex);
return queue1.empty() && queue2.empty();
}
struct ibv_wc *ConcurrentQueue::try_pop()
{
struct ibv_wc *value = NULL;
std::unique_lock<std::mutex> lock(queueMutex);
/* first check in queue 2 if it is empty then check in queue 1
* Since queue2 is empty and if queue1 is empty we can make thread to wait
* Currently thread will busy wait if there are contending request in queue2
* ie same request is in thread 2 and same request exist in runningRequests set
*/
if (queue2.empty())
{
queueCv.wait(lock, [&]
{ return queue1.size() > 0; });
value = queue1.front();
queue1.pop();
}
else
{
value = queue2.front();
queue2.pop();
}
/* At this point value will not be null because it either gets data
* from queue 1 or form queue2
* To make it sequential consistent we need to order only put requests
* it means ,we can return all other requests immediately
*/
MessageHeader *request = (MessageHeader *)value->wr_id;
if (value->opcode == IBV_WC_RECV && request->type == MessageType::PUT)
{
// std::cout<<"value "<<value<<std::endl;
/* Since this is the first request we need to put this in runningRequests set
* because this might contend in future and return this immediately
*/
if (runningRequests.empty())
{
runningRequests.insert(value);
return value;
}
/* If this is contending with some runningRequest we need to put it inside queue2
* and check for other request till queue1 not become empty, if there is no request in
* queue1 then we will return null and taskthread will retry(probable busy waiting scenario)
* Currently this way will make busy waiting of thread if queue1 is empty,
* and it will always check in queue2 whenever it will retry for next time
* probable solution is to add cv and wait it until we consume that runningRequest or new request come in queue1
*/
auto it = runningRequests.find(value);
if (it != runningRequests.end())
{
queue2.push(value);
// std::cout<<"found putting in 2"<<std::endl;
return NULL;
}
return value;
}
/* Returning all other request immediately*/
return value;
}
void ConcurrentQueue::removeFromSet(struct ibv_wc *data)
{
/* Remove Request from runningRequest Set after processing it so that other threads can
* process contending request from queue2
*/
std::unique_lock<std::mutex> lock(queueMutex);
// std::cout<<"removing"<<data<<std::endl;
runningRequests.erase(data);
}
/* needs correction*/
void ConcurrentQueue::wait_and_pop(struct ibv_wc *&popped_value)
{
std::unique_lock<std::mutex> lock(queueMutex);
queueCv.wait(lock, [&]
{ return queue1.size() > 0; });
popped_value = queue1.front();
queue1.pop();
}
\ No newline at end of file
#include "Executor.hpp"
Executor::Executor(int size,std::unordered_map<uint32_t, RdmaSalEndpoint *> *qpSalEndpointMap)
: _size(size), _qpSalEndpointMap(qpSalEndpointMap)
Executor::Executor(int size)
: _size(size)
{
_taskQueue = new ConcurrentQueue();
// _taskThreads = new std::vector<TaskThread *>();
_taskThreads.reserve(size);
}
void Executor::createThreads(std::unordered_map<uint32_t, RdmaRepEndpoint *> *clientRepMap,
std::unordered_map<uint32_t, RdmaRepEndpoint *> *serverRepMap,
std::unordered_map<uint32_t, RdmaSalEndpoint *> *salMap)
{
_taskQueue = new ConcurrentQueue<struct ibv_wc *>();
_taskThreads = new std::vector<TaskThread *>();
_taskThreads->reserve(size);
for (int i = 0; i < _size; i++)
{
TaskThread *thread = new TaskThread(i, _taskQueue, _qpSalEndpointMap);
_taskThreads->push_back(thread);
TaskThread *thread = new TaskThread(i, i, _taskQueue, clientRepMap,serverRepMap, salMap);
_taskThreads.push_back(thread);
}
}
void Executor::submit(struct ibv_wc *task)
void Executor::submit(struct ibv_wc *task)
{
_taskQueue->push(task);
}
void Executor::getTask()
{
}
\ No newline at end of file
}
void Executor::stop()
{
for (size_t i = 0; i < _taskThreads.size(); i++)
{
_taskThreads[i]->stop();
delete _taskThreads[i];
}
delete _taskQueue;
}
#include <string>
#include<iostream>
#include<fstream>
#include<map>
#include "Properties.hpp"
std::string Properties::ltrim(const std::string &s)
{
size_t start = s.find_first_not_of(_WHITESPACE);
return (start == std::string::npos) ? "" : s.substr(start);
}
class Properties{
private:
std::map<std::string,std::string> _props;
const std::string _WHITESPACE = " \n\r\t\f\v";
std::string ltrim(const std::string& s)
std::string Properties::rtrim(const std::string &s)
{
size_t end = s.find_last_not_of(_WHITESPACE);
return (end == std::string::npos) ? "" : s.substr(0, end + 1);
}
std::string Properties::trim(const std::string &s)
{
return rtrim(ltrim(s));
}
Properties::Properties(std::string filename)
{
// std::cout<<"Reading Properties From file named prop.config ...........\n";
std::ifstream file(filename);
if (!file.is_open())
{
size_t start = s.find_first_not_of(_WHITESPACE);
return (start == std::string::npos) ? "" : s.substr(start);
std::cout << "Confiq file opening failed\n";
exit(0);
}
std::string rtrim(const std::string& s)
std::string line;
std::string key, value;
int delimPos;
while (getline(file, line))
{
size_t end = s.find_last_not_of(_WHITESPACE);
return (end == std::string::npos) ? "" : s.substr(0, end + 1);
delimPos = line.find('#');
line = trim(line);
if (!line.empty())
{
line = line.substr(0, delimPos);
delimPos = line.find('=');
_props.insert(make_pair(trim(line.substr(0, delimPos)), trim(line.substr(delimPos + 1))));
}
}
std::string trim(const std::string& s)
}
std::string Properties::getValue(std::string key)
{
auto it = _props.find(key);
if (it == _props.end())
{
return rtrim(ltrim(s));
return "";
}
public:
Properties(std::string filename){
//std::cout<<"Reading Properties From file named prop.config ...........\n";
std::ifstream file (filename);
if(!file.is_open()){
std::cout<<"Confiq file opening failed\n";
exit(0);
}
std::string line;
std::string key,value;
int delimPos;
while(getline(file,line)){
delimPos=line.find('#');
line=trim(line);
if(!line.empty()){
line=line.substr(0,delimPos);
delimPos=line.find('=');
_props.insert(make_pair(trim(line.substr(0,delimPos)),trim(line.substr(delimPos+1))));
}
}
}
std::string getValue(std::string key){
auto it=_props.find(key);
if(it==_props.end()){
return "";
}
return it->second;
}
};
\ No newline at end of file
return it->second;
}
\ No newline at end of file
#include "RdmaCmProcessor.hpp"
#include <iostream>
RdmaCmProcessor::RdmaCmProcessor(RdmaEndpointGroup *group)
: _endpointGroup(group)
{
CPPLog::LOG_INFO("CMProcessor : Step 1 creating event channel");
_eventChannel = rdma_create_event_channel();
if (_eventChannel == NULL)
{
CPPLog::LOG_ERROR( "CMProcesor : error creating event channel");
}
}
struct rdma_cm_id *RdmaCmProcessor::createId()
{
struct rdma_cm_id *id = NULL;
int ret = rdma_create_id(_eventChannel, &id, NULL, RDMA_PS_TCP);
if (ret == -1)
CPPLog::LOG_ERROR("CMProcesor : rdma_create_id failed");
return id;
}
void RdmaCmProcessor::processCmEvent()
{
int ret;
struct rdma_cm_event *event;
CPPLog::LOG_INFO("CMProcessor : starting cm processing thread");
while (!_stop)
{
ret = rdma_get_cm_event(_eventChannel, &event);
if (ret)
{
CPPLog::LOG_ERROR("CMProcesor : rdma_get_cm_event failed");
continue;
}
_endpointGroup->processCmEvent(event);
ret = rdma_ack_cm_event(event);
if (ret)
{
CPPLog::LOG_ERROR("CMProcesor : rdma_ack_cm_event failed");
}
}
}
void RdmaCmProcessor::start(bool newThread)
{
if (newThread == true)
_cmEventThread = new std::thread(&RdmaCmProcessor::processCmEvent, this);
else
processCmEvent();
}
void RdmaCmProcessor::close()
{
CPPLog::LOG_ALWAYS("Closing CM Processor");
_stop = true;
if (_cmEventThread != NULL)
_cmEventThread->join();
delete _cmEventThread;
rdma_destroy_event_channel(_eventChannel);
}
#include "RdmaCqProcessor.hpp"
RdmaCqProcessor::RdmaCqProcessor(ibv_context *verbs, int compQueueSize)
RdmaCqProcessor::RdmaCqProcessor(Executor *ex, ibv_context *verbs, int compQueueSize)
: _executor(ex)
{
//_qpEndpointMap = new std::unordered_map<>();
_qpEndpointMap = new std::unordered_map<uint32_t, RdmaClientEndpoint *>();
_compChannel = ibv_create_comp_channel(verbs);
if (_compChannel == NULL)
{
std::cout << "CqProcessr : ibv_create_comp_channel failed\n";
CPPLog::LOG_ERROR("CqProcessr : ibv_create_comp_channel failed");
return;
}
_completionQueue = ibv_create_cq(verbs, compQueueSize, NULL, _compChannel, 0);
if (_completionQueue == NULL)
{
std::cout << "CqProcessr : ibv_create_cq failed" << std::endl;
CPPLog::LOG_INFO("CqProcessr : ibv_create_cq failed");
return;
}
int ret = ibv_req_notify_cq(_completionQueue, 0);
if (ret)
{
std::cout << "CqProcessr : ibv_req_notify_cq failed\n";
CPPLog::LOG_INFO("CqProcessr : ibv_req_notify_cq failed");
}
}
struct ibv_cq *RdmaCqProcessor::getCq()
{
return _completionQueue;
}
void RdmaCqProcessor::registerEp(uint64_t qp,RdmaClientEndpoint* ep)
{
_qpEndpointMap->emplace(qp,ep);
}
void RdmaCqProcessor::start()
{
std::cout << "CqProcessr : starting process CQ events" << std::endl;
CPPLog::LOG_ALWAYS("CqProcessr : starting process CQ events");
_compQueueThread = new std::thread(&RdmaCqProcessor::processCQEvents, this);
pthread_setname_np(_compQueueThread->native_handle(),"compQueueThread");
pthread_setname_np(_compQueueThread->native_handle(),"CQProcessor");
}
void RdmaCqProcessor::processCQEvents()
{
......@@ -45,73 +37,56 @@ void RdmaCqProcessor::processCQEvents()
void *context;
const int nevent = 10;
struct ibv_wc wc_array[nevent];
while (1)
while (!_stop)
{
/*
* get_CQ_event is a blocking call and it wait save some cpu cycles but.
* it might not be that efficient compared to polling
*/
ret = ibv_get_cq_event(_compChannel, &cq, &context);
if (ret == -1)
{
std::cout << "CqProcessr : ibv_get_cq_event failed\n";
CPPLog::LOG_ERROR("CqProcessr : ibv_get_cq_event failed");
close();
}
ibv_ack_cq_events(cq, 1);
/*
* Create a request for next completion cycle
*/
ret = ibv_req_notify_cq(_completionQueue, 0);
if (ret)
{
std::cout << "CqProcessr : ibv_req_notify_cq failed\n";
CPPLog::LOG_ERROR("CqProcessr : ibv_req_notify_cq failed");
close();
}
ret = ibv_poll_cq(cq, nevent, wc_array);
if (ret < 0)
{
std::cout << "CqProcessr : ibv_poll_cq failed\n";
CPPLog::LOG_ERROR("CqProcessr : ibv_poll_cq failed");
close();
}
if (ret == 0)
continue;
dispatchCqEvents(wc_array, ret);
}
}
inline void RdmaCqProcessor::dispatchCqEvents(ibv_wc wc[], int size)
{
for (int i = 0; i < size; i++)
{
if (wc[i].status != IBV_WC_SUCCESS)
{
std::cout << "RdmaCqProcessor : failed work completion : " << ibv_wc_status_str(wc[i].status) << " on qp " << wc[i].qp_num << std::endl;
return;
}
auto it = _qpEndpointMap->find(wc[i].qp_num);
if (it == _qpEndpointMap->end())
{
std::cout << "RdmaCqProcessor : endpoint not registered for qp num" << std::endl;
return;
}
switch (wc[i].opcode)
for (int i = 0; i < ret; i++)
{
if (wc_array[i].status != IBV_WC_SUCCESS)
{
case IBV_WC_SEND:
it->second->processSendComp(wc[i]);
break;
case IBV_WC_RECV:
it->second->processRecvComp(wc[i]);
break;
case IBV_WC_RDMA_WRITE:
std::cout << "rdma write completion\n";
break;
case IBV_WC_RDMA_READ:
std::cout << "rdma read completion\n";
break;
default:
std::cout << "RdmaCqProcessor : invalid opcode" << std::endl;
break;
std::ostringstream ss;
ss<< "RdmaCqProcessor : failed work completion : ";
ss<<ibv_wc_status_str(wc_array[i].status)<<"on qp"<<wc_array[i].qp_num;
CPPLog::LOG_ERROR(ss);
continue;
}
}
}
struct ibv_wc *data = new struct ibv_wc(wc_array[i]);
_executor->submit(data);
}
}
}
void RdmaCqProcessor::close()
{
_stop = true;
if (_compQueueThread != NULL)
_compQueueThread->join();
delete _compQueueThread;
}
\ No newline at end of file
#include "RdmaEndpoint.hpp"
#include "Logger.hpp"
int RdmaEndpoint::CONN_STATE_INITIALIZED = 1;
int RdmaEndpoint::CONN_STATE_RESOURCES_ALLOCATED = 2;
int RdmaEndpoint::CONN_STATE_CONNECTED = 3;
int RdmaEndpoint::CONN_STATE_CLOSED = 4;
RdmaEndpoint::RdmaEndpoint(struct rdma_cm_id *id, struct ibv_cq *completionQueue, int sendQueueSize,
int recvQueueSize, int sendMsgSize, int recvMsgSize)
int recvQueueSize, int sendMsgSize, int recvMsgSize)
: _cm_id(id), _completionQueue(completionQueue), _sendQueueSize(sendQueueSize),
_recvQueueSize(recvQueueSize), _sendMsgSize(sendMsgSize), _recvMsgSize(recvMsgSize)
{
_state = CONN_STATE_INITIALIZED;
_sendBuffers = new boost::lockfree::queue<void*>(_sendMsgSize);
}
void RdmaEndpoint::createResources()
{
/* These states are used to avoid errors in lifetime of rdma connection
* more erros can be tracked in future using these lifecycle states
*/
if (_state != CONN_STATE_INITIALIZED)
{
std::cout << "RdmaEndpoint : createResource invalid satte" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : createResource invalid state");
}
//Step 1 to create endpoint
_protectionDomain = ibv_alloc_pd(_cm_id->verbs);
if (_protectionDomain == NULL)
{
std::cout << "RdmaEndpoint : ibv_alloc_pd failed " << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : ibv_alloc_pd failed");
return;
}
//step 2 Creating Queue pair with completion queueu setted for send and recieve
struct ibv_qp_init_attr qp_init_attr;
memset(&qp_init_attr, 0, sizeof(qp_init_attr));
//This is used to set endpoint address with qp
/*
* Endpoint address is setted in QP context to get endpoint at run time with qp
* without using any map to map qp_num to endpoint
*/
qp_init_attr.qp_context = (void *)this;
// if not set 0, all work requests submitted to SQ will always generate a Work Completion
qp_init_attr.sq_sig_all = 1;
......@@ -49,107 +55,120 @@ void RdmaEndpoint::createResources()
int ret = rdma_create_qp(_cm_id, _protectionDomain, &qp_init_attr);
if (ret)
{
std::cout << "RdmaEndpoint : ibv_create_cq failed\n";
CPPLog::LOG_ERROR("RdmaEndpoint : ibv_create_cq failed");
}
if (_cm_id->pd == NULL)
{
std::cout << "RdmaEndpoint : pd not set" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : pd not set");
_cm_id->pd = _protectionDomain;
}
_sendBuff = (char*)malloc(_sendMsgSize * _sendQueueSize);
/*
* Step 3 register memory for send and recv queue
*/
_sendBuff = new char[(_sendMsgSize * _sendQueueSize)];
if (_sendBuff == NULL)
std::cout << "RdmaEndpoint : sendBuff malloc failed" << std::endl;
_recvBuff = (char*)malloc(_recvMsgSize * _recvQueueSize);
CPPLog::LOG_ERROR("RdmaEndpoint : sendBuff allocation failed");
_sendMr = rdma_reg_write(_cm_id, reinterpret_cast<void *>(_sendBuff), _sendMsgSize * _sendQueueSize);
if (_sendMr == NULL)
std::cout << "RdmaEndpoint : sendMr reg failed" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : sendMr reg failed");
_recvBuff = new char[(_recvMsgSize * _recvQueueSize)];
if (_recvBuff == NULL)
std::cout << "RdmaEndpoint : recvBuff malloc failed" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : recvBuff allocation failed");
_recvMr = rdma_reg_read(_cm_id, reinterpret_cast<void *>(_recvBuff), _recvMsgSize * _recvQueueSize);
if (_recvMr == NULL)
std::cout << "RdmaEndpoint : recvMr reg failed" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : recvMr reg failed");
/*
* adding buffers for recving rdma data
*/
for (int i = 0; i < _recvQueueSize; i++)
{
char *const location = _recvBuff + i * _recvMsgSize;
char* location = _recvBuff + i * _recvMsgSize;
rdma_post_recv(_cm_id, reinterpret_cast<void *>(location), reinterpret_cast<void *>(location),
_recvMsgSize, _recvMr);
}
/*
* Adding buffers to queue for receving data
*/
for (int i = 0; i < _sendQueueSize; i++)
{
void* const location = _sendBuff + i * _sendMsgSize;
_sendBuffers->push(location);
char* location = _sendBuff + i * _sendMsgSize;
std::unique_lock<std::mutex> lock(_sendBuffersM);
_sendBuffers.push((char *)location);
}
_state = CONN_STATE_RESOURCES_ALLOCATED;
}
void RdmaEndpoint::processCmEvent(struct rdma_cm_event *event)
{
std::cout << "RdmaEndpoint : Event " << rdma_event_str(event->event) << std::endl;
std::ostringstream ss;
ss<<"RdmaEndpoint : Event "<<rdma_event_str(event->event);
CPPLog::LOG_ALWAYS(ss);
if (event->event == RDMA_CM_EVENT_CONNECT_REQUEST)
{
std::cout << "RdmaEndpoint : Connect request";
CPPLog::LOG_ALWAYS("RdmaEndpoint : Connect request");
}
else if (event->event == RDMA_CM_EVENT_ESTABLISHED)
{
if (_state != CONN_STATE_RESOURCES_ALLOCATED)
{
std::cout << "RdmaEndpoint : EstablishedEvent invalid state " << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : Established_Event but resource not alloted");
}
std::cout << "RdmaEndpoint : step 6 Connected" << std::endl;
CPPLog::LOG_INFO("RdmaEndpoint : step 6 Connected");
_state = CONN_STATE_CONNECTED;
}
else if (event->event == RDMA_CM_EVENT_DISCONNECTED)
{
std::cout << "RdmaEndpoint : step 7 disconnected" << std::endl;
clientClose();
CPPLog::LOG_INFO("RdmaEndpoint : step 7 disconnected");
close();
}
}
void RdmaEndpoint::clientClose()
void RdmaEndpoint::close()
{
if (_state != CONN_STATE_CONNECTED)
if (_state < CONN_STATE_RESOURCES_ALLOCATED)
{
std::cout << "RdmaEndpoint : clientClose invalid state" << std::endl;
return;
}
std::cout<<"RdmaEndpoint : closing connection qp "<<_cm_id->qp->qp_num<< std::endl;
CPPLog::LOG_INFO("RdmaEndpoint : closing connection");
int ret;
ret = rdma_disconnect(_cm_id);
if (ret)
{
std::cout << "RdmaEndpoint : rdma_disconnect failed" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_disconnect failed");
}
ret = rdma_dereg_mr(_sendMr);
if (ret)
{
std::cout << "RdmaEndpoint : rdma_dereg_mr send failed" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_dereg_mr send failed");
}
free(_sendBuff);
delete[] _sendBuff;
ret = rdma_dereg_mr(_recvMr);
if (ret)
{
std::cout << "RdmaEndpoint : rdma_dereg_mr recv failed" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_dereg_mr recv failed");
}
free(_recvBuff);
delete[] _recvBuff;
rdma_destroy_qp(_cm_id);
std::cout<<"des qp"<<std::endl;
// ret = rdma_destroy_id(_cm_id);
CPPLog::LOG_INFO("des qp");
// rdma_destroy_id(_cm_id);
// ret = rdma_destroy_id(_cm_id);
std::cout<<"des mr"<<std::endl;
CPPLog::LOG_INFO("des mr");
if (ret)
{
std::cout << "RdmaEndpoint : rdma_destroy_id failed" << std::endl;
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_destroy_id failed");
}
_state = CONN_STATE_CLOSED;
std::cout<<"closed"<<std::endl;
CPPLog::LOG_INFO("closed");
}
RdmaEndpoint::~RdmaEndpoint()
{}
\ No newline at end of file
#include "RdmaRepEndpoint.hpp"
RdmaRepEndpoint::RdmaRepEndpoint(struct rdma_cm_id *id, int sendQueueSize, int recvQueueSize,
int sendMsgSize, int recvMsgSize, rocksdb::DB *db)
: _cm_id(id), _sendQueueSize(sendQueueSize), _recvQueueSize(recvQueueSize), _sendMsgSize(sendMsgSize),
_recvMsgSize(recvMsgSize), _db(db)
{
}
void RdmaRepEndpoint::processSendCompletion(struct ibv_wc *data)
{
/*means data has been send to other side we can use this buffer*/
std::unique_lock<std::mutex> lock(_sendBuffersM);
_sendBuffers.push((char *)data->wr_id);
}
int RdmaRepEndpoint::sendMessage(const char *buffer, uint32_t size)
{
if (size > (uint32_t)_sendMsgSize)
return -1;
char *sendBuffer = nullptr;
std::unique_lock<std::mutex> lock(_sendBuffersM);
if (_sendBuffers.size() == 0)
return -1;
sendBuffer = (char*)_sendBuffers.front();
_sendBuffers.pop();
lock.unlock();
memcpy((void *)sendBuffer, buffer, size);
return rdma_post_send(_cm_id, (void *)sendBuffer, (void *)sendBuffer, size, _sendMr, 0);
}
void RdmaRepEndpoint::processRecvCompletion(struct ibv_wc *data)
{
char *request = new char[data->byte_len];
memcpy(request, (void *)data->wr_id, data->byte_len);
struct MessageHeader *req = (struct MessageHeader *)request;
rdma_post_recv(_cm_id, (void *)data->wr_id, (void *)data->wr_id, _recvMsgSize, _recvMr);
if (req->type == MessageType::DELETE)
processDelete(req);
if (req->type == MessageType::GET)
processGet(req);
if (req->type == MessageType::PUT)
processPut(req);
delete[] request;
}
void RdmaRepEndpoint::processDelete(struct MessageHeader *req)
{
rocksdb::Status s = _db->Delete(rocksdb::WriteOptions(), {(char *)req + MessageHeaderSize, req->keySize});
void *sendBuf = nullptr;
std::unique_lock<std::mutex> lock(_sendBuffersM);
if (_sendBuffers.size() == 0)
return;
sendBuf = _sendBuffers.front();
_sendBuffers.pop();
lock.unlock();
MessageHeader *response = (MessageHeader *)sendBuf;
/*This id done to avoid else case*/
response->type = MessageType::FAILURE;
response->id = req->id;
if (s.ok())
{
response->type = MessageType::SUCCESS;
}
rdma_post_send(_cm_id, sendBuf, sendBuf, MessageHeaderSize, _sendMr, 0);
}
void RdmaRepEndpoint::processGet(struct MessageHeader *req)
{
char *sendBuf = nullptr;
std::unique_lock<std::mutex> lock(_sendBuffersM);
if (_sendBuffers.size() == 0)
{
return;
}
sendBuf = (char*)_sendBuffers.front();
_sendBuffers.pop();
lock.unlock();
std::string value;
rocksdb::Status s = _db->Get(rocksdb::ReadOptions(), {(char *)req + MessageHeaderSize, req->keySize}, &value);
MessageHeader *response = (MessageHeader *)sendBuf;
/*This id done to avoid else case*/
response->type = MessageType::FAILURE;
response->id = req->id;
if (s.ok())
{
response->type = MessageType::SUCCESS;
response->valueSize = value.size();
memcpy(response + MessageHeaderSize, value.c_str(), value.size());
}
rdma_post_send(_cm_id, sendBuf, sendBuf, MessageHeaderSize + value.size(), _sendMr, 0);
}
void RdmaRepEndpoint::processPut(struct MessageHeader *req)
{
rocksdb::Status s = _db->Put(rocksdb::WriteOptions(), {(char *)req + MessageHeaderSize, req->keySize},
{(char *)req + MessageHeaderSize + req->keySize, req->valueSize});
char *sendBuf = nullptr;
std::unique_lock<std::mutex> lock(_sendBuffersM);
if (_sendBuffers.size() == 0)
{
CPPLog::LOG_ERROR("No send Buffer");
return;
}
sendBuf = (char*)_sendBuffers.front();
_sendBuffers.pop();
lock.unlock();
MessageHeader *response = (MessageHeader *)sendBuf;
/*This id done to avoid else case*/
response->type = MessageType::FAILURE;
response->id = req->id;
if (s.ok())
response->type = MessageType::SUCCESS;
rdma_post_send(_cm_id, sendBuf, sendBuf, MessageHeaderSize, _sendMr, 0);
}
#include "RdmaClientEndpointGroup.hpp"
#include "RdmaRepEndpointGroup.hpp"
RdmaClientEndpointGroup::RdmaClientEndpointGroup(int sendQueueSize, int recvQueueSize, int compQueueSize, int sendMsgSize,
int recvMsgSize,int maxInLine, int timeout)
: _sendQueueSize(sendQueueSize), _recvQueueSize(recvQueueSize), _compQueueSize(compQueueSize),
_sendMsgSize(sendMsgSize), _recvMsgSize(recvMsgSize), _maxInLine(maxInLine),_timeoutMs(timeout)
RdmaRepEndpointGroup::RdmaRepEndpointGroup(int sendQueueSize, int recvQueueSize, int compQueueSize, int sendMsgSize,
int recvMsgSize, int maxInLine, int timeout, rocksdb::DB *db)
: _sendQueueSize(sendQueueSize), _recvQueueSize(recvQueueSize), _compQueueSize(compQueueSize), _sendMsgSize(sendMsgSize),
_recvMsgSize(recvMsgSize), _maxInLine(maxInLine), _timeoutMs(timeout), _db(db)
{
std::cout << "RdmaClientEndpointGroup : Step 1 creating event channel" << std::endl;
std::cout << "RdmaRepEndpointGroup : Step 1 creating event channel" << std::endl;
_eventChannel = rdma_create_event_channel();
_stop = false;
_stopCMThread = false;
if (_eventChannel == NULL)
{
std::cout << "RdmaClientEndpointGroup : error creating event channel";
std::cout << "RdmaRepEndpointGroup : error creating event channel";
}
_qpRepEndpointMap = new std::unordered_map<uint32_t, RdmaRepEndpoint *>();
}
void RdmaClientEndpointGroup::start()
void RdmaRepEndpointGroup::startCmProcessor()
{
_cmEventThread = new std::thread(&RdmaClientEndpointGroup::processCmEvents, this);
pthread_setname_np(_cmEventThread->native_handle(),"ClientCMProcessor");
_cmEventThread = new std::thread(&RdmaRepEndpointGroup::processCmEvents, this);
pthread_setname_np(_cmEventThread->native_handle(), "RepCMProcessor");
}
void RdmaClientEndpointGroup::processCmEvents()
struct ibv_cq *RdmaRepEndpointGroup::createCq(struct rdma_cm_id *id)
{
if (_cqProcessor == NULL)
{
CPPLog::LOG_ALWAYS("RdmaServerEndpointGroup : step 5 create repcq processor");
_cqProcessor = new RdmaCqProcessor(_executor, id->verbs, _compQueueSize);
_cqProcessor->start();
}
return _cqProcessor->getCq();
}
void RdmaRepEndpointGroup::processCmEvents()
{
int ret;
struct rdma_cm_event *event;
std::cout << "RdmaClientEndpointGroup : starting cm processing thread" << std::endl;
while (!_stop)
std::cout << "RdmaRepEndpointGroup : starting cm processing thread" << std::endl;
while (!_stopCMThread)
{
ret = rdma_get_cm_event(_eventChannel, &event);
if (ret)
{
std::cout << "RdmaClientEndpointGroup : rdma_get_cm_event failed" << std::endl;
std::cout << "RdmaRepEndpointGroup : rdma_get_cm_event failed" << std::endl;
continue;
}
processCmEvent(event);
ret = rdma_ack_cm_event(event);
if (ret)
{
std::cout << "RdmaClientEndpointGroup : rdma_ack_cm_event failed";
std::cout << "RdmaRepEndpointGroup : rdma_ack_cm_event failed";
}
}
}
void RdmaClientEndpointGroup::processCmEvent(struct rdma_cm_event *event)
void RdmaRepEndpointGroup::processCmEvent(struct rdma_cm_event *event)
{
std::cout << "RdmaClientEndpointGroup : event" << rdma_event_str(event->event) << std::endl;
std::cout << "RdmaRepEndpointGroup : event" << rdma_event_str(event->event) << std::endl;
if (event->id != NULL && event->id->context != NULL)
{
if (event->event == RDMA_CM_EVENT_ADDR_RESOLVED && event->id != NULL)
{
((RdmaClientEndpoint *)event->id->context)->createResources(createCq(event->id));
((RdmaClientRepEndpoint *)event->id->context)->createResources(createCq(event->id));
}
((RdmaClientEndpoint *)event->id->context)->processCmEvent(event);
if(event->event == RDMA_CM_EVENT_ADDR_RESOLVED)
((RdmaClientRepEndpoint *)event->id->context)->processCmEvent(event);
if (event->event == RDMA_CM_EVENT_ADDR_RESOLVED)
{
_cqProcessor->registerEp(event->id->qp->qp_num,((RdmaClientEndpoint *)event->id->context));
_qpRepEndpointMap->emplace(event->id->qp->qp_num, ((RdmaRepEndpoint *)event->id->context));
}
if (event->event == RDMA_CM_EVENT_DISCONNECTED)
{
if (_qpRepEndpointMap->find(event->id->qp->qp_num) != _qpRepEndpointMap->end())
_qpRepEndpointMap->erase(event->id->qp->qp_num);
delete ((RdmaRepEndpoint *)event->id->context);
}
}
else
{
std::cout << "RdmaClientEndpointGroup : Not able to procces CM EVent";
std::cout << "RdmaRepEndpointGroup : Not able to procces CM EVent";
std::cout << rdma_event_str(event->event) << event->id << " ";
std::cout << event->listen_id << std::endl;
}
}
RdmaClientEndpoint *RdmaClientEndpointGroup::createEndpoint()
RdmaClientRepEndpoint *RdmaRepEndpointGroup::createEndpoint()
{
struct rdma_cm_id *id = NULL;
int ret = rdma_create_id(_eventChannel, &id, NULL, RDMA_PS_TCP);
std::cout<<"id "<<id<<std::endl;
if (ret == -1)
std::cout << "CMProcesor : rdma_create_id failed" << std::endl;
RdmaClientEndpoint *endpoint = new RdmaClientEndpoint(id,
_sendQueueSize, _recvQueueSize,
_sendMsgSize, _recvMsgSize,_maxInLine, _timeoutMs);
RdmaClientRepEndpoint *endpoint = new RdmaClientRepEndpoint(id, _sendQueueSize, _recvQueueSize, _sendMsgSize,
_recvMsgSize, _maxInLine, _timeoutMs, _db);
id->context = (void *)endpoint;
return endpoint;
}
struct ibv_cq *RdmaClientEndpointGroup::createCq(struct rdma_cm_id *id)
void RdmaRepEndpointGroup::setExecutor(Executor *ex)
{
if (_cqProcessor == NULL)
{
std::cout << "RdmaClientEndpointGroup : Creating CQ processor" << std::endl;
_cqProcessor = new RdmaCqProcessor(id->verbs, _compQueueSize);
_cqProcessor->start();
}
return _cqProcessor->getCq();
_executor = ex;
}
void RdmaClientEndpointGroup::close()
void RdmaRepEndpointGroup::close()
{
_stop = true;
_stopCMThread = true;
_cmEventThread->join();
_cqProcessor->close();
delete _cmEventThread;
delete _cqProcessor;
delete _qpRepEndpointMap;
rdma_destroy_event_channel(_eventChannel);
}
\ No newline at end of file
#include "RdmaSalCqProcessor.hpp"
RdmaSalCqProcessor::RdmaSalCqProcessor(Executor *ex, ibv_context *verbs, int compQueueSize)
: _executor(ex)
{
_compChannel = ibv_create_comp_channel(verbs);
if (_compChannel == NULL)
{
std::cout << "SalCqProcessr : ibv_create_comp_channel failed\n";
return;
}
_completionQueue = ibv_create_cq(verbs, compQueueSize, NULL, _compChannel, 0);
if (_completionQueue == NULL)
{
std::cout << "SalCqProcessr : ibv_create_cq failed" << std::endl;
return;
}
int ret = ibv_req_notify_cq(_completionQueue, 0);
if (ret)
{
std::cout << "SalCqProcessr : ibv_req_notify_cq failed\n";
}
}
struct ibv_cq *RdmaSalCqProcessor::getCq()
{
return _completionQueue;
}
void RdmaSalCqProcessor::start()
{
std::cout << "SalCqProcessr : starting process CQ events" << std::endl;
_compQueueThread = new std::thread(&RdmaSalCqProcessor::processCQEvents, this);
}
void RdmaSalCqProcessor::processCQEvents()
{
int ret = 0;
struct ibv_cq *cq;
void *context;
const int nevent = 10;
struct ibv_wc wc_array[nevent];
while (!_stop)
{
/*
* get_CQ_event is a blocking call and it wait save some cpu cycles but.
* it might not be that efficient compared to polling
*/
ret = ibv_get_cq_event(_compChannel, &cq, &context);
if (ret == -1)
{
std::cout << "SalCqProcessr : ibv_get_cq_event failed\n";
close();
}
ibv_ack_cq_events(cq, 1);
/*
* Create a request for next completion cycle
*/
ret = ibv_req_notify_cq(_completionQueue, 0);
if (ret)
{
std::cout << "SalCqProcessr : ibv_req_notify_cq failed\n";
close();
}
ret = ibv_poll_cq(cq, nevent, wc_array);
if (ret < 0)
{
std::cout << "SalCqProcessr : ibv_poll_cq failed\n";
close();
}
if (ret == 0)
continue;
for (int i = 0; i < ret; i++)
{
struct ibv_wc *data = new struct ibv_wc(wc_array[i]);
/*
* vendor_err is set to check whether the request came from sal or followers
* data->vendor_err = 0;
*/
_executor->submit(data);
}
// _executor->dispatchSalCqEvents(wc_array, ret);
}
}
void RdmaSalCqProcessor::close()
{
_stop = true;
if (_compQueueThread != NULL)
_compQueueThread->join();
}
\ No newline at end of file
This diff is collapsed.
This diff is collapsed.
#include "RdmaServerRepEndpoint.hpp"
int RdmaServerRepEndpoint::CONN_STATE_INITIALIZED = 1;
int RdmaServerRepEndpoint::CONN_STATE_RESOURCES_ALLOCATED = 2;
int RdmaServerRepEndpoint::CONN_STATE_CONNECTED = 3;
int RdmaServerRepEndpoint::CONN_STATE_CLOSED = 4;
RdmaServerRepEndpoint::RdmaServerRepEndpoint(struct rdma_cm_id *id, struct ibv_cq *completionQueue, int sendQueueSize,
int recvQueueSize, int sendMsgSize, int recvMsgSize, rocksdb::DB *db)
: RdmaRepEndpoint(id,sendQueueSize, recvQueueSize, sendMsgSize ,recvMsgSize ,db),_completionQueue(completionQueue)
{
_state = CONN_STATE_INITIALIZED;
}
void RdmaServerRepEndpoint::createResources()
{
/* These states are used to avoid errors in lifetime of rdma connection
* more erros can be tracked in future using these lifecycle states
*/
if (_state != CONN_STATE_INITIALIZED)
{
CPPLog::LOG_ERROR("RdmaEndpoint : createResource invalid state");
}
// Step 1 to create endpoint
_protectionDomain = ibv_alloc_pd(_cm_id->verbs);
if (_protectionDomain == NULL)
{
CPPLog::LOG_ERROR("RdmaEndpoint : ibv_alloc_pd failed");
return;
}
// step 2 Creating Queue pair with completion queueu setted for send and recieve
struct ibv_qp_init_attr qp_init_attr;
memset(&qp_init_attr, 0, sizeof(qp_init_attr));
/*
* Endpoint address is setted in QP context to get endpoint at run time with qp
* without using any map to map qp_num to endpoint
*/
qp_init_attr.qp_context = (void *)this;
// if not set 0, all work requests submitted to SQ will always generate a Work Completion
qp_init_attr.sq_sig_all = 1;
// completion queue can be shared or you can use distinct completion queues.
qp_init_attr.send_cq = _completionQueue;
qp_init_attr.recv_cq = _completionQueue;
qp_init_attr.qp_type = IBV_QPT_RC;
// increase if you want to keep more send work requests in the SQ.
qp_init_attr.cap.max_send_wr = _sendQueueSize;
// increase if you want to keep more receive work requests in the RQ.
qp_init_attr.cap.max_recv_wr = _recvQueueSize;
// increase if you allow send work requests to have multiple scatter gather entry (SGE).
qp_init_attr.cap.max_send_sge = 1;
// increase if you allow receive work requests to have multiple scatter gather entry (SGE).
qp_init_attr.cap.max_recv_sge = 1;
int ret = rdma_create_qp(_cm_id, _protectionDomain, &qp_init_attr);
if (ret)
{
CPPLog::LOG_ERROR("RdmaEndpoint : ibv_create_cq failed");
}
if (_cm_id->pd == NULL)
{
CPPLog::LOG_ERROR("RdmaEndpoint : pd not set");
_cm_id->pd = _protectionDomain;
}
/*
* Step 3 register memory for send and recv queue
*/
_sendBuff = new char[(_sendMsgSize * _sendQueueSize)];
if (_sendBuff == NULL)
CPPLog::LOG_ERROR("RdmaEndpoint : sendBuff allocation failed");
_sendMr = rdma_reg_write(_cm_id, reinterpret_cast<void *>(_sendBuff), _sendMsgSize * _sendQueueSize);
if (_sendMr == NULL)
CPPLog::LOG_ERROR("RdmaEndpoint : sendMr reg failed");
_recvBuff = new char[(_recvMsgSize * _recvQueueSize)];
if (_recvBuff == NULL)
CPPLog::LOG_ERROR("RdmaEndpoint : recvBuff allocation failed");
_recvMr = rdma_reg_read(_cm_id, reinterpret_cast<void *>(_recvBuff), _recvMsgSize * _recvQueueSize);
if (_recvMr == NULL)
CPPLog::LOG_ERROR("RdmaEndpoint : recvMr reg failed");
/*
* adding buffers for recving rdma data
*/
for (int i = 0; i < _recvQueueSize; i++)
{
char *location = _recvBuff + i * _recvMsgSize;
rdma_post_recv(_cm_id, reinterpret_cast<void *>(location), reinterpret_cast<void *>(location),
_recvMsgSize, _recvMr);
}
/*
* Adding buffers to queue for receving data
*/
for (int i = 0; i < _sendQueueSize; i++)
{
char *location = _sendBuff + i * _sendMsgSize;
std::unique_lock<std::mutex> lock(_sendBuffersM);
_sendBuffers.push((char *)location);
}
_state = CONN_STATE_RESOURCES_ALLOCATED;
}
void RdmaServerRepEndpoint::processCmEvent(struct rdma_cm_event *event)
{
std::ostringstream ss;
ss << "RdmaEndpoint : Event " << rdma_event_str(event->event);
CPPLog::LOG_ALWAYS(ss);
if (event->event == RDMA_CM_EVENT_CONNECT_REQUEST)
{
CPPLog::LOG_ALWAYS("RdmaEndpoint : Connect request");
}
else if (event->event == RDMA_CM_EVENT_ESTABLISHED)
{
if (_state != CONN_STATE_RESOURCES_ALLOCATED)
{
CPPLog::LOG_ERROR("RdmaEndpoint : Established_Event but resource not alloted");
}
CPPLog::LOG_INFO("RdmaEndpoint : step 6 Connected");
_state = CONN_STATE_CONNECTED;
}
else if (event->event == RDMA_CM_EVENT_DISCONNECTED)
{
CPPLog::LOG_INFO("RdmaEndpoint : step 7 disconnected");
close();
}
}
void RdmaServerRepEndpoint::close()
{
if (_state < CONN_STATE_RESOURCES_ALLOCATED)
{
return;
}
CPPLog::LOG_INFO("RdmaEndpoint : closing connection");
int ret;
ret = rdma_disconnect(_cm_id);
if (ret)
{
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_disconnect failed");
}
ret = rdma_dereg_mr(_sendMr);
if (ret)
{
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_dereg_mr send failed");
}
delete[] _sendBuff;
ret = rdma_dereg_mr(_recvMr);
if (ret)
{
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_dereg_mr recv failed");
}
delete[] _recvBuff;
rdma_destroy_qp(_cm_id);
CPPLog::LOG_INFO("des qp");
// rdma_destroy_id(_cm_id);
// ret = rdma_destroy_id(_cm_id);
CPPLog::LOG_INFO("des mr");
if (ret)
{
CPPLog::LOG_ERROR("RdmaEndpoint : rdma_destroy_id failed");
}
_state = CONN_STATE_CLOSED;
CPPLog::LOG_INFO("closed");
}
#include "TaskThread.hpp"
#include "MessageFormats.hpp"
TaskThread::TaskThread(int id, int cpu, ConcurrentQueue<struct ibv_wc *> *taskqueue, std::unordered_map<uint32_t, RdmaSalEndpoint *> *qpSalEndpointMap)
: _id(id), _qpSalEndpointMap(qpSalEndpointMap)
TaskThread::TaskThread(int id, int cpu, ConcurrentQueue *taskqueue,
std::unordered_map<uint32_t, RdmaRepEndpoint *> *clientRepMap,
std::unordered_map<uint32_t, RdmaRepEndpoint *> *serverRepMap,
std::unordered_map<uint32_t, RdmaSalEndpoint *> *salMap)
: _id(id), _clientRepMap(clientRepMap), _serverRepMap(serverRepMap), _salMap(salMap)
{
_taskQueue = taskqueue;
if (pthread_create(&thread, NULL, &TaskThread::run, this))
{
std::cout << "pthread create has been failed while creating taskthread " << std::endl;
CPPLog::LOG_ERROR("pthread create has been failed while creating taskthread");
exit(0);
}
cpu_set_t cpuset;
CPU_ZERO(&cpuset);
CPU_SET(cpu, &cpuset);
std::ostringstream ss;
ss << "New Thread Setting CPU affinty " << cpu;
CPPLog::LOG_ALWAYS(ss);
if (pthread_setaffinity_np(thread, sizeof(cpu_set_t), &cpuset) != 0)
{
std::cerr << "Error calling pthread_setaffinity_np: "
<< "\n";
}
}
TaskThread::TaskThread(int id, ConcurrentQueue<struct ibv_wc *> *taskqueue, std::unordered_map<uint32_t, RdmaSalEndpoint *> *qpSalEndpointMap)
: _id(id), _qpSalEndpointMap(qpSalEndpointMap)
{
_taskQueue = taskqueue;
if (pthread_create(&thread, NULL, &TaskThread::run, this))
{
std::cout << "pthread create has been failed while creating taskthread " << std::endl;
exit(0);
CPPLog::LOG_ERROR("Error calling pthread_setaffinity_np ");
}
pthread_setname_np(thread, "TaskThread");
}
TaskThread::~TaskThread()
{
std::cout << "Task Destructed" << std::endl;
CPPLog::LOG_INFO("TaskThread Destructed");
stop();
}
void TaskThread::stop()
{
_stop = true;
if (pthread_join(thread, NULL) == 0)
{
std::cout << "pthread join failed" << std::endl;
CPPLog::LOG_ERROR("pthread join failed");
}
}
inline void *TaskThread::run(void *object)
{
/*
* This is because data from pthread_create is passed as void*
*/
TaskThread *thread = reinterpret_cast<TaskThread *>(object);
std::cout << "running task thread" << thread->_id << std::endl;
std::ostringstream ss;
ss<<"Running task thread"<<thread->_id;
CPPLog::LOG_ALWAYS(ss);
while (!thread->_stop)
{
struct ibv_wc *data = NULL;
thread->_taskQueue->wait_and_pop(data);
thread->processEvent(data);
std::cout << "Get start\n";
data = thread->_taskQueue->try_pop();
if (data != NULL)
{
std::cout << "TaskThread:: got data";
if (data == NULL || data->status != IBV_WC_SUCCESS)
{
std::ostringstream ss;
ss << "TaskThread : failed work completion : ";
ss << ibv_wc_status_str(data->status) << " on qp " << data->qp_num;
CPPLog::LOG_ERROR(ss);
continue;
}
RdmaSalEndpoint *salEp = nullptr;
RdmaRepEndpoint *repEp = nullptr;
auto it = thread->_salMap->find(data->qp_num);
if (it != thread->_salMap->end())
{
salEp = it->second;
}
auto it2 = thread->_clientRepMap->find(data->qp_num);
if (it2 != thread->_clientRepMap->end())
{
repEp = it2->second;
}
else
{
auto it3 = thread->_serverRepMap->find(data->qp_num);
if(it3 != thread->_serverRepMap->end())
repEp = it3->second;
}
if (salEp == nullptr && repEp == nullptr)
{
std::ostringstream ss;
ss<<"RdmaSal : endpoint not registered for qp"<<data->qp_num;
CPPLog::LOG_INFO(ss);
continue;
}
else if (salEp != nullptr)
thread->processEvent(salEp, data);
else if (repEp != nullptr)
thread->processRepEvent(repEp, data);
thread->_taskQueue->removeFromSet(data);
delete data;
}
}
return NULL;
}
void TaskThread::processEvent(struct ibv_wc *data)
void TaskThread::processEvent(RdmaSalEndpoint *ep, struct ibv_wc *data)
{
if (data == NULL || data->status != IBV_WC_SUCCESS)
std::cout<<"processing sal event\n";
/* sal Request*/
switch (data->opcode)
{
std::cout << "TaskThread : failed work completion : ";
std::cout << ibv_wc_status_str(data->status) << " on qp " << data->qp_num << std::endl;
return;
}
/*
* Process Request from client
*/
auto it = _qpSalEndpointMap->find(data->qp_num);
if (it == _qpSalEndpointMap->end())
case IBV_WC_SEND:
ep->processSendCompletion(data);
break;
case IBV_WC_RECV:
{
std::cout << data->qp_num << "RdmaSal : endpoint not registered for qp num\n";
return;
// it->second->processRecvCompletion(data);
char *buffer = new char[data->byte_len];
memcpy(buffer, (void *)data->wr_id, data->byte_len);
rdma_post_recv(ep->_cm_id, (void *)data->wr_id, (void *)data->wr_id,
ep->_recvMsgSize, ep->_recvMr);
struct MessageHeader *req = (struct MessageHeader *)buffer;
std::cout << "TaskThread 1\n";
switch (req->type)
{
case MessageType::GET:
ep->processGet(req);
break;
case MessageType::DELETE:
replicateSalRequest(buffer, data->byte_len);
ep->processDelete(req);
break;
case MessageType::PUT:
replicateSalRequest(buffer, data->byte_len);
ep->processPut(req);
break;
default:
CPPLog::LOG_ERROR("SalRequest invalid req type");
break;
}
delete[] buffer;
}
break;
default:
std::ostringstream ss;
ss << "TaskThread default opcode : " << data->opcode;
CPPLog::LOG_INFO(ss);
break;
}
}
void TaskThread::processRepEvent(RdmaRepEndpoint *ep, struct ibv_wc *data)
{
switch (data->opcode)
{
case IBV_WC_SEND:
it->second->processSendCompletion(data);
ep->processSendCompletion(data);
break;
case IBV_WC_RECV:
{
// it->second->processRecvCompletion(data);
char *buffer = new char[data->byte_len];
memcpy(buffer, (void *)data->wr_id, data->byte_len);
rdma_post_recv(it->second->_cm_id, (void *)data->wr_id, (void *)data->wr_id,
it->second->_recvMsgSize, it->second->_recvMr);
struct SalRequestHeader *req = (struct SalRequestHeader *)buffer;
std::cout << "recieve id : " << req->id << " " << (char *)req + SalRequestHeaderSize;
std::cout << " " << req->type << "size" << data->byte_len << "\n";
rdma_post_recv(ep->_cm_id, (void *)data->wr_id, (void *)data->wr_id,
ep->_recvMsgSize, ep->_recvMr);
struct MessageHeader *req = (struct MessageHeader *)buffer;
std::cout << "TaskThread 1\n";
switch (req->type)
{
case RequestType::GET:
it->second->processGet(req);
case MessageType::GET:
ep->processGet(req);
break;
case RequestType::DELETE:
std::cout<<"TaskThread:: incorrect delete request from client to follower"<<std::endl;
case MessageType::DELETE:
ep->processDelete(req);
break;
case RequestType::PUT:
std::cout<<"TaskThread:: incorrect put request from client to follower"<<std::endl;
case MessageType::PUT:
ep->processPut(req);
break;
default:
std::cout << "SalRequest invalid req type";
CPPLog::LOG_ERROR("SalRequest invalid req type");
break;
}
delete[] buffer;
}
break;
case IBV_WC_RDMA_WRITE:
std::cout << "rdma write completion\n";
break;
case IBV_WC_RDMA_READ:
std::cout << "rdma read completion\n";
break;
default:
std::cout << "TaskThread default opcode : " << data->opcode << std::endl;
std::ostringstream ss;
ss << "TaskThread default opcode : " << data->opcode;
CPPLog::LOG_INFO(ss);
break;
}
}
void TaskThread::replicateSalRequest(char *req, uint32_t size)
{
auto repIt = _clientRepMap->begin();
for (; repIt != _clientRepMap->end(); repIt++)
{
repIt->second->sendMessage(req, size);
}
auto serverRepIt = _serverRepMap->begin();
for (; serverRepIt != _serverRepMap->end(); serverRepIt++)
{
serverRepIt->second->sendMessage(req, size);
}
MessageHeader *salReq = (MessageHeader *)req;
char *buffer = new char[MessageHeaderSize + salReq->keySize];
MessageHeader *invRequest = (MessageHeader *)(buffer);
invRequest->type = MessageType::INVALIDATE;
invRequest->id = salReq->id;
invRequest->keySize = salReq->keySize;
memcpy(buffer + MessageHeaderSize, salReq + MessageHeaderSize, salReq->keySize);
// Send Invalidation to sal's
auto salIt = _salMap->begin();
for (;salIt != _salMap->end();salIt++)
{
salIt->second->sendMessage(buffer, MessageHeaderSize + salReq->keySize);
}
delete[] buffer;
}
\ No newline at end of file
#include <iostream>
#include "RdmaServerEndpointGroup.hpp"
#include "RdmaClientEndpointGroup.hpp"
#include "RdmaRepEndpointGroup.hpp"
#include "RdmaEndpoint.hpp"
#include "Executor.hpp"
#include "Properties.hpp"
#include "rocksdb/db.h"
#include "rocksdb/slice.h"
#include "rocksdb/options.h"
int connectToRepServer(Properties &prop, RdmaRepEndpointGroup* repGroup)
{
int followers = stoi(prop.getValue("FOLLOWERS"));
std::cout<<"followers "<<followers<<"\n";
RdmaClientRepEndpoint* clientEPs[followers];
for(int i = 0 ;i< followers ;i++)
{
clientEPs[i] = repGroup->createEndpoint();
std::string ip = prop.getValue("FOLLOWER"+std::to_string(i+1)+"_IP");
std::string port = prop.getValue("FOLLOWER"+std::to_string(i+1)+"_PORT");
std::cout<<"Connecting to follower "<<ip<<":"<<port<<"\n";
clientEPs[i]->connect(ip.c_str(), port.c_str(), "fol");
}
return 0;
}
int main()
{
RdmaClientEndpointGroup *clgroup = new RdmaClientEndpointGroup(5, 5, 5, 50, 50, 0, 1000);
clgroup->start();
RdmaClientEndpoint *clientEp = clgroup->createEndpoint();
clientEp->connect("192.168.200.20", "1921", "fol");
while (!clientEp->isConnected());
std::cout << "client : connected" << std::endl;
std::cout << "Starting Server Main Thread\n";
Properties prop("prop.config");
int sendQS = stoi(prop.getValue("sendQS"));
int recvQS = stoi(prop.getValue("recvQS"));
int compQS = stoi(prop.getValue("compQS"));
int sendMS = stoi(prop.getValue("sendMS"));
int recvMS = stoi(prop.getValue("recvMS"));
std::string dbpath = prop.getValue("DB_PATH");
std::string serverIP = prop.getValue("SERVER_IP");
std::string serverPort = prop.getValue("SERVER_PORT");
int executorPoolSize = stoi(prop.getValue("EXECUTOR_POOL_SIZE"));
int enableLogging = stoi(prop.getValue("ENABLE_LOGGING"));
if (enableLogging == 0)
CPPLog::Logger::getInstance()->updateLogLevel(CPPLog::DISABLE_LOG);
rocksdb::DB *db;
rocksdb::Options options;
options.create_if_missing = true;
// open a database with a name which corresponds to a file system directory
rocksdb::Status status = rocksdb::DB::Open(options, dbpath, &db);
if (!status.ok())
{
CPPLog::LOG_ERROR(status.ToString().c_str());
exit(1);
}
CPPLog::LOG_ALWAYS("Rocks started");
Executor *executor = new Executor(executorPoolSize);
RdmaRepEndpointGroup *repgroup = new RdmaRepEndpointGroup(sendQS, recvQS, compQS, sendMS, recvMS, 1, 100000, db);
RdmaServerEndpointGroup *sgroup = new RdmaServerEndpointGroup(sendQS, recvQS, compQS, sendMS, recvMS, db);
executor->createThreads(repgroup->_qpRepEndpointMap, sgroup->_qpRepEndpointMap, sgroup->_qpSalEndpointMap);
repgroup->setExecutor(executor);
sgroup->setExecutor(executor);
repgroup->startCmProcessor();
sgroup->bind(serverIP.c_str(), serverPort.c_str(), 2);
std::thread t(connectToRepServer, std::ref(prop), (repgroup));
RdmaServerEndpointGroup *sgroup = new RdmaServerEndpointGroup(5, 5, 5, 50, 50);
sgroup->createExecutor(4);
sgroup->bind("192.168.200.20", "1920", 2);
sgroup->startCmProcessor(false);
std::cout << "rhdhj" << std::endl;
while (1)
;
}
\ No newline at end of file
}
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment