forked from vitalif/vitastor
Rename object_db to clean_db
parent
69581e6d0c
commit
ff57dd420e
|
@ -229,7 +229,7 @@ class blockstore
|
|||
struct ring_consumer_t ring_consumer;
|
||||
|
||||
// Another option is https://github.com/algorithm-ninja/cpp-btree
|
||||
spp::sparse_hash_map<object_id, clean_entry, oid_hash> object_db;
|
||||
spp::sparse_hash_map<object_id, clean_entry, oid_hash> clean_db;
|
||||
std::map<obj_ver_id, dirty_entry> dirty_db;
|
||||
std::list<blockstore_operation*> submit_queue;
|
||||
std::deque<obj_ver_id> unsynced_big_writes, unsynced_small_writes;
|
||||
|
|
|
@ -75,7 +75,7 @@ void blockstore_init_meta::handle_entries(struct clean_disk_entry* entries, int
|
|||
if (entries[i].oid.inode > 0)
|
||||
{
|
||||
allocator_set(bs->data_alloc, done_cnt+i, true);
|
||||
bs->object_db[entries[i].oid] = (struct clean_entry){
|
||||
bs->clean_db[entries[i].oid] = (struct clean_entry){
|
||||
entries[i].version,
|
||||
(uint32_t)(entries[i].flags ? ST_CURRENT : ST_D_META_SYNCED),
|
||||
done_cnt+i
|
||||
|
|
|
@ -71,13 +71,13 @@ int blockstore::fulfill_read(blockstore_operation *read_op, uint32_t item_start,
|
|||
|
||||
int blockstore::dequeue_read(blockstore_operation *read_op)
|
||||
{
|
||||
auto clean_it = object_db.find(read_op->oid);
|
||||
auto clean_it = clean_db.find(read_op->oid);
|
||||
auto dirty_it = dirty_db.upper_bound((obj_ver_id){
|
||||
.oid = read_op->oid,
|
||||
.version = UINT64_MAX,
|
||||
});
|
||||
dirty_it--;
|
||||
bool clean_found = clean_it != object_db.end();
|
||||
bool clean_found = clean_it != clean_db.end();
|
||||
bool dirty_found = (dirty_it != dirty_db.end() && dirty_it->first.oid == read_op->oid);
|
||||
if (!clean_found && !dirty_found)
|
||||
{
|
||||
|
@ -107,7 +107,7 @@ int blockstore::dequeue_read(blockstore_operation *read_op)
|
|||
dirty_it--;
|
||||
}
|
||||
}
|
||||
if (clean_it != object_db.end())
|
||||
if (clean_it != clean_db.end())
|
||||
{
|
||||
if (!fulfill_read(read_op, 0, block_size, ST_CURRENT, 0, clean_it->second.location))
|
||||
{
|
||||
|
|
|
@ -8,8 +8,8 @@ int blockstore::dequeue_stable(blockstore_operation *op)
|
|||
});
|
||||
if (dirty_it == dirty_db.end())
|
||||
{
|
||||
auto clean_it = object_db.find(op->oid);
|
||||
if (clean_it == object_db.end() || clean_it->second.version < op->version)
|
||||
auto clean_it = clean_db.find(op->oid);
|
||||
if (clean_it == clean_db.end() || clean_it->second.version < op->version)
|
||||
{
|
||||
// No such object version
|
||||
op->retval = EINVAL;
|
||||
|
@ -116,7 +116,7 @@ void blockstore::handle_stable_event(ring_data_t *data, blockstore_operation *op
|
|||
dirty_it->second.state = ST_J_STABLE;
|
||||
// Copy data from the journal to the data device
|
||||
// -> increase version on the metadata device
|
||||
// -> advance object_db entry's version and clear previous journal entries
|
||||
// -> advance clean_db entry's version and clear previous journal entries
|
||||
// This makes 1 4K small write look like:
|
||||
// 512b+4K (journal) + sync + 512b (journal) + sync + 512b (metadata) + 4K (data) + sync.
|
||||
// WA = 2.375. It's not the best, SSD FTL-like redirect-write with defragmentation
|
||||
|
@ -127,7 +127,7 @@ void blockstore::handle_stable_event(ring_data_t *data, blockstore_operation *op
|
|||
{
|
||||
dirty_it->second.state = ST_D_STABLE;
|
||||
// Copy metadata from the journal to the metadata device
|
||||
// -> move dirty_db entry to object_db and clear previous journal entries
|
||||
// -> move dirty_db entry to clean_db and clear previous journal entries
|
||||
// This makes 1 128K big write look like:
|
||||
// 128K (data) + sync + 512b (journal) + sync + 512b (journal) + sync + 512b (metadata) + sync.
|
||||
// WA = 1.012. Very good :)
|
||||
|
|
|
@ -14,8 +14,8 @@ void blockstore::enqueue_write(blockstore_operation *op)
|
|||
}
|
||||
else
|
||||
{
|
||||
auto clean_it = object_db.find(op->oid);
|
||||
if (clean_it != object_db.end())
|
||||
auto clean_it = clean_db.find(op->oid);
|
||||
if (clean_it != clean_db.end())
|
||||
{
|
||||
op->version = clean_it->second.version + 1;
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue