forked from vitalif/vitastor
Begin write algorithm
parent
2f18a3d19e
commit
84c62840bd
2
Makefile
2
Makefile
|
@ -1,4 +1,4 @@
|
||||||
all: allocator.o blockstore.o blockstore_init.o blockstore_open.o blockstore_read.o crc32c.o ringloop.o test
|
all: allocator.o blockstore.o blockstore_init.o blockstore_open.o blockstore_read.o blockstore_write.o crc32c.o ringloop.o test
|
||||||
clean:
|
clean:
|
||||||
rm -f *.o
|
rm -f *.o
|
||||||
crc32c.o: crc32c.c
|
crc32c.o: crc32c.c
|
||||||
|
|
|
@ -138,14 +138,39 @@ void blockstore::loop()
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
else if (((*cur)->flags & OP_TYPE_MASK) == OP_WRITE ||
|
||||||
|
((*cur)->flags & OP_TYPE_MASK) == OP_DELETE)
|
||||||
|
{
|
||||||
|
int dequeue_op = dequeue_write(*cur);
|
||||||
|
if (dequeue_op)
|
||||||
|
{
|
||||||
|
submit_queue.erase(cur);
|
||||||
|
}
|
||||||
|
else if ((*cur)->wait_for == WAIT_SQE)
|
||||||
|
{
|
||||||
|
// ring is full, stop submission
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else if (((*cur)->flags & OP_TYPE_MASK) == OP_SYNC)
|
||||||
|
{
|
||||||
|
|
||||||
|
}
|
||||||
|
else if (((*cur)->flags & OP_TYPE_MASK) == OP_STABLE)
|
||||||
|
{
|
||||||
|
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
int blockstore::enqueue_op(blockstore_operation *op)
|
int blockstore::enqueue_op(blockstore_operation *op)
|
||||||
{
|
{
|
||||||
if (op->offset >= block_size || op->len >= block_size-op->offset)
|
if (op->offset >= block_size || op->len >= block_size-op->offset ||
|
||||||
|
(op->len % DISK_ALIGNMENT) ||
|
||||||
|
(op->flags & OP_TYPE_MASK) < OP_READ || (op->flags & OP_TYPE_MASK) > OP_DELETE)
|
||||||
{
|
{
|
||||||
|
// Basic verification not passed
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
submit_queue.push_back(op);
|
submit_queue.push_back(op);
|
||||||
|
|
12
blockstore.h
12
blockstore.h
|
@ -69,6 +69,7 @@ inline bool operator == (const object_id & a, const object_id & b)
|
||||||
}
|
}
|
||||||
|
|
||||||
// 32 bytes per "clean" entry on disk with fixed metadata tables
|
// 32 bytes per "clean" entry on disk with fixed metadata tables
|
||||||
|
// FIXME: maybe add crc32's to metadata
|
||||||
struct __attribute__((__packed__)) clean_disk_entry
|
struct __attribute__((__packed__)) clean_disk_entry
|
||||||
{
|
{
|
||||||
object_id oid;
|
object_id oid;
|
||||||
|
@ -156,7 +157,7 @@ struct blockstore_operation
|
||||||
std::map<uint64_t, struct iovec> read_vec;
|
std::map<uint64_t, struct iovec> read_vec;
|
||||||
int pending_ops;
|
int pending_ops;
|
||||||
int wait_for;
|
int wait_for;
|
||||||
uint64_t wait_version;
|
uint64_t wait_detail;
|
||||||
};
|
};
|
||||||
|
|
||||||
class blockstore;
|
class blockstore;
|
||||||
|
@ -211,11 +212,18 @@ public:
|
||||||
void handle_event(ring_data_t* data);
|
void handle_event(ring_data_t* data);
|
||||||
void loop();
|
void loop();
|
||||||
|
|
||||||
// Read
|
// Submission
|
||||||
int enqueue_op(blockstore_operation *op);
|
int enqueue_op(blockstore_operation *op);
|
||||||
|
|
||||||
|
// Read
|
||||||
int dequeue_read(blockstore_operation *read_op);
|
int dequeue_read(blockstore_operation *read_op);
|
||||||
int fulfill_read(blockstore_operation *read_op, uint32_t item_start, uint32_t item_end,
|
int fulfill_read(blockstore_operation *read_op, uint32_t item_start, uint32_t item_end,
|
||||||
uint32_t item_state, uint64_t item_version, uint64_t item_location);
|
uint32_t item_state, uint64_t item_version, uint64_t item_location);
|
||||||
int fulfill_read_push(blockstore_operation *read_op, uint32_t item_start,
|
int fulfill_read_push(blockstore_operation *read_op, uint32_t item_start,
|
||||||
uint32_t item_state, uint64_t item_version, uint64_t item_location, uint32_t cur_start, uint32_t cur_end);
|
uint32_t item_state, uint64_t item_version, uint64_t item_location, uint32_t cur_start, uint32_t cur_end);
|
||||||
|
|
||||||
|
// Write
|
||||||
|
int dequeue_write(blockstore_operation *op);
|
||||||
|
|
||||||
|
// Sync
|
||||||
};
|
};
|
||||||
|
|
|
@ -9,7 +9,7 @@ int blockstore::fulfill_read_push(blockstore_operation *read_op, uint32_t item_s
|
||||||
{
|
{
|
||||||
// Pause until it's written somewhere
|
// Pause until it's written somewhere
|
||||||
read_op->wait_for = WAIT_IN_FLIGHT;
|
read_op->wait_for = WAIT_IN_FLIGHT;
|
||||||
read_op->wait_version = item_version;
|
read_op->wait_detail = item_version;
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
else if (item_state == ST_DEL_WRITTEN || item_state == ST_DEL_SYNCED || item_state == ST_DEL_MOVED)
|
else if (item_state == ST_DEL_WRITTEN || item_state == ST_DEL_SYNCED || item_state == ST_DEL_MOVED)
|
||||||
|
|
|
@ -0,0 +1,123 @@
|
||||||
|
#include "blockstore.h"
|
||||||
|
|
||||||
|
// First step of the write algorithm: dequeue operation and submit initial write(s)
|
||||||
|
int blockstore::dequeue_write(blockstore_operation *op)
|
||||||
|
{
|
||||||
|
auto dirty_it = dirty_queue[op->oid].find(op->version); // FIXME OOPS
|
||||||
|
if (op->len == block_size)
|
||||||
|
{
|
||||||
|
// Big (redirect) write
|
||||||
|
uint64_t loc = allocator_find_free(data_alloc);
|
||||||
|
if (loc == (uint64_t)-1)
|
||||||
|
{
|
||||||
|
// no space
|
||||||
|
op->retval = -ENOSPC;
|
||||||
|
op->callback(op);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
struct io_uring_sqe *sqe = get_sqe();
|
||||||
|
if (!sqe)
|
||||||
|
{
|
||||||
|
// Pause until there are more requests available
|
||||||
|
op->wait_for = WAIT_SQE;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
struct ring_data_t *data = ((ring_data_t*)sqe->user_data);
|
||||||
|
(*dirty_it).location = loc << block_order;
|
||||||
|
//(*dirty_it).state = ST_D_SUBMITTED;
|
||||||
|
allocator_set(data_alloc, loc, true);
|
||||||
|
data->iov = (struct iovec){ op->buf, op->len };
|
||||||
|
data->op = op;
|
||||||
|
io_uring_prep_writev(
|
||||||
|
sqe, data_fd, &data->iov, 1, data_offset + (loc << block_order)
|
||||||
|
);
|
||||||
|
op->pending_ops = 1;
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
// Small (journaled) write
|
||||||
|
// First check if the journal has sufficient space
|
||||||
|
bool two_sqes = false;
|
||||||
|
uint64_t next_pos = journal_data_pos;
|
||||||
|
if (512 - journal_sector_pos < sizeof(struct journal_entry_small_write))
|
||||||
|
{
|
||||||
|
next_pos = next_pos + 512;
|
||||||
|
if (journal_len - next_pos < op->len)
|
||||||
|
two_sqes = true;
|
||||||
|
if (next_pos >= journal_len)
|
||||||
|
next_pos = 512;
|
||||||
|
}
|
||||||
|
else if (journal_sector + 512 != journal_data_pos || journal_len - journal_data_pos < op->len)
|
||||||
|
two_sqes = true;
|
||||||
|
next_pos = (journal_len - next_pos < op->len ? 512 : next_pos) + op->len;
|
||||||
|
if (next_pos >= journal_start)
|
||||||
|
{
|
||||||
|
// No space in the journal. Wait until it's available
|
||||||
|
op->wait_for = WAIT_JOURNAL;
|
||||||
|
op->wait_detail = next_pos - journal_start;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
// There is sufficient space. Get SQE(s)
|
||||||
|
struct io_uring_sqe *sqe1 = get_sqe(), *sqe2 = two_sqes ? get_sqe() : NULL;
|
||||||
|
if (!sqe1 || two_sqes && !sqe2)
|
||||||
|
{
|
||||||
|
// Pause until there are more requests available
|
||||||
|
op->wait_for = WAIT_SQE;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
struct ring_data_t *data1 = ((ring_data_t*)sqe1->user_data);
|
||||||
|
struct ring_data_t *data2 = two_sqes ? ((ring_data_t*)sqe2->user_data) : NULL;
|
||||||
|
// Got SQEs. Prepare journal sector write
|
||||||
|
if (512 - journal_sector_pos < sizeof(struct journal_entry_small_write))
|
||||||
|
{
|
||||||
|
// Move to the next journal sector
|
||||||
|
next_pos = journal_data_pos + 512;
|
||||||
|
if (next_pos >= journal_len)
|
||||||
|
next_pos = 512;
|
||||||
|
journal_sector = journal_data_pos;
|
||||||
|
journal_sector_pos = 0;
|
||||||
|
journal_data_pos = next_pos;
|
||||||
|
memset(journal_sector_buf, 0, 512);
|
||||||
|
}
|
||||||
|
journal_entry_small_write *je = (struct journal_entry_small_write*)(journal_sector_buf + journal_sector_pos);
|
||||||
|
*je = {
|
||||||
|
.crc32 = 0,
|
||||||
|
.magic = JOURNAL_MAGIC,
|
||||||
|
.type = JE_SMALL_WRITE,
|
||||||
|
.size = sizeof(struct journal_entry_small_write),
|
||||||
|
.crc32_prev = journal_crc32_last,
|
||||||
|
.oid = op->oid,
|
||||||
|
.version = op->version,
|
||||||
|
.offset = op->offset,
|
||||||
|
.len = op->len,
|
||||||
|
};
|
||||||
|
je.crc32 = je_crc32((journal_entry*)je);
|
||||||
|
data1->iov = (struct iovec){ journal_sector_buf, 512 };
|
||||||
|
data1->op = op;
|
||||||
|
io_uring_prep_writev(
|
||||||
|
sqe1, journal_fd, &data1->iov, 1, journal_offset + journal_sector
|
||||||
|
);
|
||||||
|
// Prepare journal data write
|
||||||
|
if (journal_len - journal_data_pos < op->len)
|
||||||
|
journal_data_pos = 512;
|
||||||
|
data2->iov = (struct iovec){ op->buf, op->len };
|
||||||
|
data2->op = op;
|
||||||
|
io_uring_prep_writev(
|
||||||
|
sqe2, journal_fd, &data2->iov, 1, journal_offset + journal_data_pos
|
||||||
|
);
|
||||||
|
(*dirty_it).location = journal_data_pos;
|
||||||
|
//(*dirty_it).state = ST_J_SUBMITTED;
|
||||||
|
// Move journal_data_pos
|
||||||
|
journal_data_pos += op->len;
|
||||||
|
if (journal_data_pos >= journal_len)
|
||||||
|
journal_data_pos = 512;
|
||||||
|
op->pending_ops = 2;
|
||||||
|
}
|
||||||
|
in_process_ops.insert(op);
|
||||||
|
int ret = ringloop->submit();
|
||||||
|
if (ret < 0)
|
||||||
|
{
|
||||||
|
throw new std::runtime_error(std::string("io_uring_submit: ") + strerror(-ret));
|
||||||
|
}
|
||||||
|
return 1;
|
||||||
|
}
|
Loading…
Reference in New Issue