mbed-os/features/storage/kvstore/tdbstore/TDBStore.cpp

1514 lines
44 KiB
C++

/*
* Copyright (c) 2018 ARM Limited. All rights reserved.
* SPDX-License-Identifier: Apache-2.0
* Licensed under the Apache License, Version 2.0 (the License); you may
* not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an AS IS BASIS, WITHOUT
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
// ----------------------------------------------------------- Includes -----------------------------------------------------------
#include "TDBStore.h"
#include <algorithm>
#include <string.h>
#include <stdio.h>
#include "mbed_error.h"
#include "mbed_wait_api.h"
#include "MbedCRC.h"
#include "SystemStorage.h"
using namespace mbed;
// --------------------------------------------------------- Definitions ----------------------------------------------------------
static const uint32_t delete_flag = (1UL << 31);
static const uint32_t internal_flags = delete_flag;
static const uint32_t supported_flags = KVStore::WRITE_ONCE_FLAG;
namespace {
typedef struct {
uint32_t magic;
uint16_t header_size;
uint16_t revision;
uint32_t flags;
uint16_t key_size;
uint16_t reserved;
uint32_t data_size;
uint32_t crc;
} record_header_t;
typedef struct {
uint32_t hash;
bd_size_t bd_offset;
} ram_table_entry_t;
static const char *master_rec_key = "TDBS";
static const uint32_t tdbstore_magic = 0x54686683; // "TDBS" in ASCII
static const uint32_t tdbstore_revision = 1;
typedef struct {
uint16_t version;
uint16_t tdbstore_revision;
uint32_t reserved;
} master_record_data_t;
typedef enum {
TDBSTORE_AREA_STATE_NONE = 0,
TDBSTORE_AREA_STATE_EMPTY,
TDBSTORE_AREA_STATE_VALID,
} area_state_e;
typedef struct {
uint16_t trailer_size;
uint16_t data_size;
uint32_t crc;
} reserved_trailer_t;
static const uint32_t work_buf_size = 64;
static const uint32_t initial_crc = 0xFFFFFFFF;
static const uint32_t initial_max_keys = 16;
// incremental set handle
typedef struct {
record_header_t header;
bd_size_t bd_base_offset;
bd_size_t bd_curr_offset;
uint32_t offset_in_data;
uint32_t ram_table_ind;
uint32_t hash;
bool new_key;
} inc_set_handle_t;
// iterator handle
typedef struct {
int iterator_num;
uint32_t ram_table_ind;
char *prefix;
} key_iterator_handle_t;
} // anonymous namespace
// -------------------------------------------------- Local Functions Declaration ----------------------------------------------------
// -------------------------------------------------- Functions Implementation ----------------------------------------------------
static inline uint32_t align_up(uint32_t val, uint32_t size)
{
return (((val - 1) / size) + 1) * size;
}
static uint32_t calc_crc(uint32_t init_crc, uint32_t data_size, const void *data_buf)
{
uint32_t crc;
MbedCRC<POLY_32BIT_ANSI, 32> ct(init_crc, 0x0, true, false);
ct.compute(const_cast<void *>(data_buf), data_size, &crc);
return crc;
}
// Class member functions
TDBStore::TDBStore(BlockDevice *bd) : _ram_table(0), _max_keys(0),
_num_keys(0), _bd(bd), _buff_bd(0), _free_space_offset(0), _master_record_offset(0),
_master_record_size(0), _is_initialized(false), _active_area(0), _active_area_version(0), _size(0),
_prog_size(0), _work_buf(0), _key_buf(0), _variant_bd_erase_unit_size(false), _inc_set_handle(0)
{
}
TDBStore::~TDBStore()
{
deinit();
}
int TDBStore::read_area(uint8_t area, uint32_t offset, uint32_t size, void *buf)
{
int os_ret = _buff_bd->read(buf, _area_params[area].address + offset, size);
if (os_ret) {
return MBED_ERROR_READ_FAILED;
}
return MBED_SUCCESS;
}
int TDBStore::write_area(uint8_t area, uint32_t offset, uint32_t size, const void *buf)
{
int os_ret = _buff_bd->program(buf, _area_params[area].address + offset, size);
if (os_ret) {
return MBED_ERROR_WRITE_FAILED;
}
return MBED_SUCCESS;
}
int TDBStore::erase_erase_unit(uint8_t area, uint32_t offset)
{
uint32_t bd_offset = _area_params[area].address + offset;
uint32_t eu_size = _buff_bd->get_erase_size(bd_offset);
int os_ret = _buff_bd->erase(bd_offset, eu_size);
if (os_ret) {
return MBED_ERROR_WRITE_FAILED;
}
return MBED_SUCCESS;
}
void TDBStore::calc_area_params()
{
// TDBStore can't exceed 32 bits
bd_size_t bd_size = std::min(_bd->size(), (bd_size_t) 0x80000000L);
memset(_area_params, 0, sizeof(_area_params));
size_t area_0_size = 0;
bd_size_t prev_erase_unit_size = _bd->get_erase_size(area_0_size);
_variant_bd_erase_unit_size = 0;
while (area_0_size < bd_size / 2) {
bd_size_t erase_unit_size = _bd->get_erase_size(area_0_size);
_variant_bd_erase_unit_size |= (erase_unit_size != prev_erase_unit_size);
area_0_size += erase_unit_size;
}
_area_params[0].address = 0;
_area_params[0].size = area_0_size;
_area_params[1].address = area_0_size;
_area_params[1].size = bd_size - area_0_size;
}
// This function, reading a record from the BD, is used for multiple purposes:
// - Init (scan all records, no need to return file name and data)
// - Get (return file data)
// - Get first/next file (check whether name matches, return name if so)
int TDBStore::read_record(uint8_t area, uint32_t offset, char *key,
void *data_buf, uint32_t data_buf_size,
uint32_t &actual_data_size, size_t data_offset, bool copy_key,
bool copy_data, bool check_expected_key, bool calc_hash,
uint32_t &hash, uint32_t &flags, uint32_t &next_offset)
{
int ret;
record_header_t header;
uint32_t total_size, key_size, data_size;
uint32_t curr_data_offset;
char *user_key_ptr;
uint32_t crc = initial_crc;
// Upper layers typically use non zero offsets for reading the records chunk by chunk,
// so only validate entire record at first chunk (otherwise we'll have a serious performance penalty).
bool validate = (data_offset == 0);
ret = MBED_SUCCESS;
// next offset should only be updated to the end of record if successful
next_offset = offset;
ret = read_area(area, offset, sizeof(header), &header);
if (ret) {
return ret;
}
if (header.magic != tdbstore_magic) {
return MBED_ERROR_INVALID_DATA_DETECTED;
}
offset += align_up(sizeof(header), _prog_size);
key_size = header.key_size;
data_size = header.data_size;
flags = header.flags;
if ((!key_size) || (key_size >= MAX_KEY_SIZE)) {
return MBED_ERROR_INVALID_DATA_DETECTED;
}
total_size = key_size + data_size;
// Make sure our read sizes didn't cause any wraparounds
if ((total_size < key_size) || (total_size < data_size)) {
return MBED_ERROR_INVALID_DATA_DETECTED;
}
if (offset + total_size >= _size) {
return MBED_ERROR_INVALID_DATA_DETECTED;
}
if (data_offset > data_size) {
return MBED_ERROR_INVALID_SIZE;
}
actual_data_size = std::min(data_buf_size, data_size - data_offset);
if (copy_data && actual_data_size && !data_buf) {
return MBED_ERROR_INVALID_ARGUMENT;
}
if (validate) {
// Calculate CRC on header (excluding CRC itself)
crc = calc_crc(crc, sizeof(record_header_t) - sizeof(crc), &header);
curr_data_offset = 0;
} else {
// Non validation case: No need to read the key, nor the parts before data_offset
// or after the actual part requested by the user.
total_size = actual_data_size;
curr_data_offset = data_offset;
offset += data_offset + key_size;
// Mark code that key handling is finished
key_size = 0;
}
user_key_ptr = key;
hash = initial_crc;
while (total_size) {
uint8_t *dest_buf;
uint32_t chunk_size;
if (key_size) {
// This means that we're on the key part
if (copy_key) {
dest_buf = reinterpret_cast<uint8_t *>(user_key_ptr);
chunk_size = key_size;
user_key_ptr[key_size] = '\0';
} else {
dest_buf = _work_buf;
chunk_size = std::min(key_size, work_buf_size);
}
} else {
// This means that we're on the data part
// We have four cases that need different handling:
// 1. Before data_offset - read to work buffer
// 2. After data_offset, but before actual part is finished - read to user buffer
// 3. After actual part is finished - read to work buffer
// 4. Copy data flag not set - read to work buffer
if (curr_data_offset < data_offset) {
chunk_size = std::min(work_buf_size, data_offset - curr_data_offset);
dest_buf = _work_buf;
} else if (copy_data && (curr_data_offset < data_offset + actual_data_size)) {
chunk_size = actual_data_size;
dest_buf = static_cast<uint8_t *>(data_buf);
} else {
chunk_size = std::min(work_buf_size, total_size);
dest_buf = _work_buf;
}
}
ret = read_area(area, offset, chunk_size, dest_buf);
if (ret) {
goto end;
}
if (validate) {
// calculate CRC on current read chunk
crc = calc_crc(crc, chunk_size, dest_buf);
}
if (key_size) {
// We're on key part. May need to calculate hash or check whether key is the expected one
if (check_expected_key) {
if (memcmp(user_key_ptr, dest_buf, chunk_size)) {
ret = MBED_ERROR_ITEM_NOT_FOUND;
}
}
if (calc_hash) {
hash = calc_crc(hash, chunk_size, dest_buf);
}
user_key_ptr += chunk_size;
key_size -= chunk_size;
if (!key_size) {
offset += data_offset;
}
} else {
curr_data_offset += chunk_size;
}
total_size -= chunk_size;
offset += chunk_size;
}
if (validate && (crc != header.crc)) {
ret = MBED_ERROR_INVALID_DATA_DETECTED;
goto end;
}
next_offset = align_up(offset, _prog_size);
end:
return ret;
}
int TDBStore::find_record(uint8_t area, const char *key, uint32_t &offset,
uint32_t &ram_table_ind, uint32_t &hash)
{
ram_table_entry_t *ram_table = (ram_table_entry_t *) _ram_table;
ram_table_entry_t *entry;
int ret = MBED_ERROR_ITEM_NOT_FOUND;
uint32_t actual_data_size;
uint32_t flags, dummy_hash, next_offset;
hash = calc_crc(initial_crc, strlen(key), key);
for (ram_table_ind = 0; ram_table_ind < _num_keys; ram_table_ind++) {
entry = &ram_table[ram_table_ind];
offset = entry->bd_offset;
if (hash < entry->hash) {
continue;
}
if (hash > entry->hash) {
return MBED_ERROR_ITEM_NOT_FOUND;
}
ret = read_record(_active_area, offset, const_cast<char *>(key), 0, 0, actual_data_size, 0,
false, false, true, false, dummy_hash, flags, next_offset);
// not found return code here means that hash doesn't belong to name. Continue searching.
if (ret != MBED_ERROR_ITEM_NOT_FOUND) {
break;
}
}
return ret;
}
uint32_t TDBStore::record_size(const char *key, uint32_t data_size)
{
return align_up(sizeof(record_header_t), _prog_size) +
align_up(strlen(key) + data_size, _prog_size);
}
int TDBStore::set_start(set_handle_t *handle, const char *key, size_t final_data_size,
uint32_t create_flags)
{
int ret;
uint32_t offset;
uint32_t hash, ram_table_ind;
inc_set_handle_t *ih;
bool need_gc = false;
if (!is_valid_key(key)) {
return MBED_ERROR_INVALID_ARGUMENT;
}
if (create_flags & ~(supported_flags | internal_flags)) {
return MBED_ERROR_INVALID_ARGUMENT;
}
*handle = reinterpret_cast<set_handle_t>(_inc_set_handle);
ih = reinterpret_cast<inc_set_handle_t *>(*handle);
if (!strcmp(key, master_rec_key)) {
// Master record - special case (no need to protect by the mutex, as it is already covered
// in the upper layers).
ih->bd_base_offset = _master_record_offset;
ih->new_key = false;
} else {
_mutex.lock();
// A valid magic in the header means that this function has been called after an aborted
// incremental set process. This means that our media may be in a bad state - call GC.
if (ih->header.magic == tdbstore_magic) {
ret = garbage_collection();
if (ret) {
goto fail;
}
}
// If we have no room for the record, perform garbage collection
uint32_t rec_size = record_size(key, final_data_size);
if (_free_space_offset + rec_size > _size) {
ret = garbage_collection();
if (ret) {
goto fail;
}
}
// If even after GC we have no room for the record, return error
if (_free_space_offset + rec_size > _size) {
ret = MBED_ERROR_MEDIA_FULL;
goto fail;
}
ret = find_record(_active_area, key, offset, ram_table_ind, hash);
if (ret == MBED_SUCCESS) {
ret = read_area(_active_area, offset, sizeof(ih->header), &ih->header);
if (ret) {
goto fail;
}
if (ih->header.flags & WRITE_ONCE_FLAG) {
ret = MBED_ERROR_WRITE_PROTECTED;
goto fail;
}
ih->new_key = false;
} else if (ret == MBED_ERROR_ITEM_NOT_FOUND) {
if (create_flags & delete_flag) {
goto fail;
}
if (_num_keys >= _max_keys) {
increment_max_keys();
}
ih->new_key = true;
} else {
goto fail;
}
ih->bd_base_offset = _free_space_offset;
check_erase_before_write(_active_area, ih->bd_base_offset, rec_size);
}
ret = MBED_SUCCESS;
// Fill handle and header fields
// Jump to offset after header (header will be written at finalize phase)
ih->bd_curr_offset = ih->bd_base_offset + align_up(sizeof(record_header_t), _prog_size);
ih->offset_in_data = 0;
ih->hash = hash;
ih->ram_table_ind = ram_table_ind;
ih->header.magic = tdbstore_magic;
ih->header.header_size = sizeof(record_header_t);
ih->header.revision = tdbstore_revision;
ih->header.flags = create_flags;
ih->header.key_size = strlen(key);
ih->header.reserved = 0;
ih->header.data_size = final_data_size;
// Calculate CRC on header and key
ih->header.crc = calc_crc(initial_crc, sizeof(record_header_t) - sizeof(ih->header.crc), &ih->header);
ih->header.crc = calc_crc(ih->header.crc, ih->header.key_size, key);
// Write key now
ret = write_area(_active_area, ih->bd_curr_offset, ih->header.key_size, key);
if (ret) {
need_gc = true;
goto fail;
}
ih->bd_curr_offset += ih->header.key_size;
goto end;
fail:
if ((need_gc) && (ih->bd_base_offset != _master_record_offset)) {
garbage_collection();
}
// mark handle as invalid by clearing magic field in header
ih->header.magic = 0;
_mutex.unlock();
end:
return ret;
}
int TDBStore::set_add_data(set_handle_t handle, const void *value_data, size_t data_size)
{
int ret = MBED_SUCCESS;
inc_set_handle_t *ih;
bool need_gc = false;
if (handle != _inc_set_handle) {
return MBED_ERROR_INVALID_ARGUMENT;
}
if (!value_data && data_size) {
return MBED_ERROR_INVALID_ARGUMENT;
}
_inc_set_mutex.lock();
ih = reinterpret_cast<inc_set_handle_t *>(handle);
if (!ih->header.magic) {
ret = MBED_ERROR_INVALID_ARGUMENT;
goto end;
}
if (ih->offset_in_data + data_size > ih->header.data_size) {
ret = MBED_ERROR_INVALID_SIZE;
goto end;
}
// Update CRC with data chunk
ih->header.crc = calc_crc(ih->header.crc, data_size, value_data);
// Write the data chunk
ret = write_area(_active_area, ih->bd_curr_offset, data_size, value_data);
if (ret) {
need_gc = true;
goto end;
}
ih->bd_curr_offset += data_size;
ih->offset_in_data += data_size;
end:
if ((need_gc) && (ih->bd_base_offset != _master_record_offset)) {
garbage_collection();
}
_inc_set_mutex.unlock();
return ret;
}
int TDBStore::set_finalize(set_handle_t handle)
{
int os_ret, ret = MBED_SUCCESS;
inc_set_handle_t *ih;
ram_table_entry_t *ram_table = (ram_table_entry_t *) _ram_table;
ram_table_entry_t *entry;
bool need_gc = false;
uint32_t actual_data_size, hash, flags, next_offset;
if (handle != _inc_set_handle) {
return MBED_ERROR_INVALID_ARGUMENT;
}
ih = reinterpret_cast<inc_set_handle_t *>(handle);
if (!ih->header.magic) {
return MBED_ERROR_INVALID_ARGUMENT;
}
_inc_set_mutex.lock();
if (ih->offset_in_data != ih->header.data_size) {
ret = MBED_ERROR_INVALID_SIZE;
need_gc = true;
goto end;
}
// Write header
ret = write_area(_active_area, ih->bd_base_offset, sizeof(record_header_t), &ih->header);
if (ret) {
need_gc = true;
goto end;
}
// Need to flush buffered BD as our record is totally written now
os_ret = _buff_bd->sync();
if (os_ret) {
ret = MBED_ERROR_WRITE_FAILED;
need_gc = true;
goto end;
}
// In master record case we don't update RAM table
if (ih->bd_base_offset == _master_record_offset) {
goto end;
}
// Writes may fail without returning a failure (specially in flash components). Reread the record
// to ensure write success (this won't read the data anywhere - just use the CRC calculation).
ret = read_record(_active_area, ih->bd_base_offset, 0, 0, (uint32_t) -1,
actual_data_size, 0, false, false, false, false,
hash, flags, next_offset);
if (ret) {
need_gc = true;
goto end;
}
// Update RAM table
if (ih->header.flags & delete_flag) {
_num_keys--;
if (ih->ram_table_ind < _num_keys) {
memmove(&ram_table[ih->ram_table_ind], &ram_table[ih->ram_table_ind + 1],
sizeof(ram_table_entry_t) * (_num_keys - ih->ram_table_ind));
}
update_all_iterators(false, ih->ram_table_ind);
} else {
if (ih->new_key) {
if (ih->ram_table_ind < _num_keys) {
memmove(&ram_table[ih->ram_table_ind + 1], &ram_table[ih->ram_table_ind],
sizeof(ram_table_entry_t) * (_num_keys - ih->ram_table_ind));
}
_num_keys++;
update_all_iterators(true, ih->ram_table_ind);
}
entry = &ram_table[ih->ram_table_ind];
entry->hash = ih->hash;
entry->bd_offset = ih->bd_base_offset;
}
_free_space_offset = align_up(ih->bd_curr_offset, _prog_size);
end:
if ((need_gc) && (ih->bd_base_offset != _master_record_offset)) {
garbage_collection();
}
// mark handle as invalid by clearing magic field in header
ih->header.magic = 0;
_inc_set_mutex.unlock();
if (ih->bd_base_offset != _master_record_offset) {
_mutex.unlock();
}
return ret;
}
int TDBStore::set(const char *key, const void *buffer, size_t size, uint32_t create_flags)
{
int ret;
set_handle_t handle;
// Don't wait till we get to set_add_data to catch this
if (!buffer && size) {
return MBED_ERROR_INVALID_ARGUMENT;
}
ret = set_start(&handle, key, size, create_flags);
if (ret) {
return ret;
}
ret = set_add_data(handle, buffer, size);
if (ret) {
return ret;
}
ret = set_finalize(handle);
return ret;
}
int TDBStore::remove(const char *key)
{
return set(key, 0, 0, delete_flag);
}
int TDBStore::get(const char *key, void *buffer, size_t buffer_size, size_t *actual_size, size_t offset)
{
int ret;
uint32_t actual_data_size;
uint32_t bd_offset, next_bd_offset;
uint32_t flags, hash, ram_table_ind;
if (!is_valid_key(key)) {
return MBED_ERROR_INVALID_ARGUMENT;
}
_mutex.lock();
ret = find_record(_active_area, key, bd_offset, ram_table_ind, hash);
if (ret != MBED_SUCCESS) {
goto end;
}
ret = read_record(_active_area, bd_offset, const_cast<char *>(key), buffer, buffer_size,
actual_data_size, offset, false, true, false, false, hash, flags, next_bd_offset);
if (actual_size) {
*actual_size = actual_data_size;
}
end:
_mutex.unlock();
return ret;
}
int TDBStore::get_info(const char *key, info_t *info)
{
int ret;
uint32_t bd_offset, next_bd_offset;
uint32_t flags, hash, ram_table_ind;
uint32_t actual_data_size;
if (!is_valid_key(key)) {
return MBED_ERROR_INVALID_ARGUMENT;
}
_mutex.lock();
ret = find_record(_active_area, key, bd_offset, ram_table_ind, hash);
if (ret) {
goto end;
}
// Give a large dummy buffer size in order to achieve actual data size
// (as copy_data flag is not set, data won't be copied anywhere)
ret = read_record(_active_area, bd_offset, const_cast<char *>(key), 0, (uint32_t) -1,
actual_data_size, 0, false, false, false, false, hash, flags,
next_bd_offset);
if (ret) {
goto end;
}
if (info) {
info->flags = flags;
info->size = actual_data_size;
}
end:
_mutex.unlock();
return ret;
}
int TDBStore::write_master_record(uint8_t area, uint16_t version, uint32_t &next_offset)
{
master_record_data_t master_rec;
master_rec.version = version;
master_rec.tdbstore_revision = tdbstore_revision;
master_rec.reserved = 0;
next_offset = _master_record_offset + _master_record_size;
return set(master_rec_key, &master_rec, sizeof(master_rec), 0);
}
int TDBStore::copy_record(uint8_t from_area, uint32_t from_offset, uint32_t to_offset,
uint32_t &to_next_offset)
{
int ret;
record_header_t header;
uint32_t total_size;
uint16_t chunk_size;
ret = read_area(from_area, from_offset, sizeof(header), &header);
if (ret) {
return ret;
}
total_size = align_up(sizeof(record_header_t), _prog_size) +
align_up(header.key_size + header.data_size, _prog_size);;
ret = check_erase_before_write(1 - from_area, to_offset, total_size);
if (ret) {
return ret;
}
chunk_size = align_up(sizeof(record_header_t), _prog_size);
ret = write_area(1 - from_area, to_offset, chunk_size, &header);
if (ret) {
return ret;
}
from_offset += chunk_size;
to_offset += chunk_size;
total_size -= chunk_size;
while (total_size) {
chunk_size = std::min(total_size, work_buf_size);
ret = read_area(from_area, from_offset, chunk_size, _work_buf);
if (ret) {
return ret;
}
ret = write_area(1 - from_area, to_offset, chunk_size, _work_buf);
if (ret) {
return ret;
}
from_offset += chunk_size;
to_offset += chunk_size;
total_size -= chunk_size;
}
to_next_offset = align_up(to_offset, _prog_size);
return MBED_SUCCESS;
}
int TDBStore::garbage_collection()
{
ram_table_entry_t *ram_table = (ram_table_entry_t *) _ram_table;
uint32_t to_offset, to_next_offset;
uint32_t chunk_size, reserved_size;
int ret;
size_t ind;
ret = check_erase_before_write(1 - _active_area, 0, _master_record_offset + _master_record_size);
if (ret) {
return ret;
}
ret = do_reserved_data_get(0, RESERVED_AREA_SIZE);
if (!ret) {
// Copy reserved data
to_offset = 0;
reserved_size = _master_record_offset;
while (reserved_size) {
chunk_size = std::min(work_buf_size, reserved_size);
ret = read_area(_active_area, to_offset, chunk_size, _work_buf);
if (ret) {
return ret;
}
ret = write_area(1 - _active_area, to_offset, chunk_size, _work_buf);
if (ret) {
return ret;
}
to_offset += chunk_size;
reserved_size -= chunk_size;
}
}
to_offset = _master_record_offset + _master_record_size;
// Initialize in case table is empty
to_next_offset = to_offset;
// Go over ram table and copy all entries to opposite area
for (ind = 0; ind < _num_keys; ind++) {
uint32_t from_offset = ram_table[ind].bd_offset;
ret = copy_record(_active_area, from_offset, to_offset, to_next_offset);
if (ret) {
return ret;
}
// Update RAM table
ram_table[ind].bd_offset = to_offset;
to_offset = to_next_offset;
}
to_offset = to_next_offset;
_free_space_offset = to_next_offset;
// Now we can switch to the new active area
_active_area = 1 - _active_area;
// Now write master record, with version incremented by 1.
_active_area_version++;
ret = write_master_record(_active_area, _active_area_version, to_offset);
if (ret) {
return ret;
}
// Now reset standby area
ret = reset_area(1 - _active_area);
if (ret) {
return ret;
}
return MBED_SUCCESS;
}
int TDBStore::build_ram_table()
{
ram_table_entry_t *ram_table = (ram_table_entry_t *) _ram_table;
uint32_t offset, next_offset = 0, dummy;
int ret = MBED_SUCCESS;
uint32_t hash;
uint32_t flags;
uint32_t actual_data_size;
uint32_t ram_table_ind;
_num_keys = 0;
offset = _master_record_offset;
while (offset < _free_space_offset) {
ret = read_record(_active_area, offset, _key_buf, 0, 0, actual_data_size, 0,
true, false, false, true, hash, flags, next_offset);
if (ret) {
goto end;
}
ret = find_record(_active_area, _key_buf, dummy, ram_table_ind, hash);
if ((ret != MBED_SUCCESS) && (ret != MBED_ERROR_ITEM_NOT_FOUND)) {
goto end;
}
uint32_t save_offset = offset;
offset = next_offset;
if (ret == MBED_ERROR_ITEM_NOT_FOUND) {
// Key doesn't exist, need to add it to RAM table
ret = MBED_SUCCESS;
if (flags & delete_flag) {
continue;
}
if (_num_keys >= _max_keys) {
// In order to avoid numerous reallocations of ram table,
// Add a chunk of entries now
increment_max_keys(reinterpret_cast<void **>(&ram_table));
}
memmove(&ram_table[ram_table_ind + 1], &ram_table[ram_table_ind],
sizeof(ram_table_entry_t) * (_num_keys - ram_table_ind));
_num_keys++;
} else if (flags & delete_flag) {
_num_keys--;
memmove(&ram_table[ram_table_ind], &ram_table[ram_table_ind + 1],
sizeof(ram_table_entry_t) * (_num_keys - ram_table_ind));
continue;
}
// update record parameters
ram_table[ram_table_ind].hash = hash;
ram_table[ram_table_ind].bd_offset = save_offset;
}
end:
_free_space_offset = next_offset;
return ret;
}
int TDBStore::increment_max_keys(void **ram_table)
{
// Reallocate ram table with new size
ram_table_entry_t *old_ram_table = (ram_table_entry_t *) _ram_table;
ram_table_entry_t *new_ram_table = new ram_table_entry_t[_max_keys + 1];
// Copy old content to new table
memcpy(new_ram_table, old_ram_table, sizeof(ram_table_entry_t) * _max_keys);
_max_keys++;
_ram_table = new_ram_table;
delete[] old_ram_table;
if (ram_table) {
*ram_table = _ram_table;
}
return MBED_SUCCESS;
}
int TDBStore::init()
{
ram_table_entry_t *ram_table;
area_state_e area_state[_num_areas];
uint32_t next_offset;
uint32_t flags, hash;
uint32_t actual_data_size;
int os_ret, ret = MBED_SUCCESS, reserved_ret;
uint16_t versions[_num_areas];
_mutex.lock();
if (_is_initialized) {
goto end;
}
//Check if we are on internal memory && try to set the internal memory for TDBStore use.
if (strcmp(_bd->get_type(), "FLASHIAP") == 0 &&
avoid_conflict_nvstore_tdbstore(TDBSTORE) == MBED_ERROR_ALREADY_INITIALIZED) {
MBED_ERROR(MBED_ERROR_ALREADY_INITIALIZED, "TDBStore in internal memory can not be initialize when NVStore is in use");
}
_max_keys = initial_max_keys;
ram_table = new ram_table_entry_t[_max_keys];
_ram_table = ram_table;
_num_keys = 0;
_size = (size_t) -1;
_buff_bd = new BufferedBlockDevice(_bd);
_buff_bd->init();
// Underlying BD must have flash attributes, i.e. have an erase value
if (_bd->get_erase_value() == -1) {
MBED_ERROR(MBED_ERROR_INVALID_ARGUMENT, "Underlying BD must have flash attributes");
}
_prog_size = _bd->get_program_size();
_work_buf = new uint8_t[work_buf_size];
_key_buf = new char[MAX_KEY_SIZE];
_inc_set_handle = new inc_set_handle_t;
memset(_inc_set_handle, 0, sizeof(inc_set_handle_t));
memset(_iterator_table, 0, sizeof(_iterator_table));
_master_record_offset = align_up(RESERVED_AREA_SIZE + sizeof(reserved_trailer_t), _prog_size);
_master_record_size = record_size(master_rec_key, sizeof(master_record_data_t));
calc_area_params();
for (uint8_t area = 0; area < _num_areas; area++) {
area_state[area] = TDBSTORE_AREA_STATE_NONE;
versions[area] = 0;
_size = std::min(_size, _area_params[area].size);
// Check validity of master record
master_record_data_t master_rec;
ret = read_record(area, _master_record_offset, const_cast<char *>(master_rec_key),
&master_rec, sizeof(master_rec), actual_data_size, 0, false, true, true, false,
hash, flags, next_offset);
if ((ret != MBED_SUCCESS) && (ret != MBED_ERROR_INVALID_DATA_DETECTED)) {
MBED_ERROR(ret, "TDBSTORE: Unable to read record at init");
}
// Master record may be either corrupt or erased - either way erase it
// (this will do nothing if already erased)
if (ret == MBED_ERROR_INVALID_DATA_DETECTED) {
if (check_erase_before_write(area, _master_record_offset, _master_record_size, true)) {
MBED_ERROR(MBED_ERROR_READ_FAILED, "TDBSTORE: Unable reset area at init");
}
area_state[area] = TDBSTORE_AREA_STATE_EMPTY;
continue;
}
versions[area] = master_rec.version;
area_state[area] = TDBSTORE_AREA_STATE_VALID;
// Unless both areas are valid (a case handled later), getting here means
// that we found our active area.
_active_area = area;
_active_area_version = versions[area];
}
// In case we have two empty areas, arbitrarily use area 0 as the active one.
if ((area_state[0] == TDBSTORE_AREA_STATE_EMPTY) && (area_state[1] == TDBSTORE_AREA_STATE_EMPTY)) {
_active_area = 0;
_active_area_version = 1;
ret = write_master_record(_active_area, _active_area_version, _free_space_offset);
if (ret) {
MBED_ERROR(ret, "TDBSTORE: Unable to write master record at init");
}
// Nothing more to do here if active area is empty
goto end;
}
// In case we have two valid areas, choose the one having the higher version (or 0
// in case of wrap around). Reset the other one.
if ((area_state[0] == TDBSTORE_AREA_STATE_VALID) && (area_state[1] == TDBSTORE_AREA_STATE_VALID)) {
if ((versions[0] > versions[1]) || (!versions[0])) {
_active_area = 0;
} else {
_active_area = 1;
}
_active_area_version = versions[_active_area];
ret = reset_area(1 - _active_area);
if (ret) {
MBED_ERROR(ret, "TDBSTORE: Unable to reset area at init");
}
}
// Currently set free space offset pointer to the end of free space.
// Ram table build process needs it, but will update it.
_free_space_offset = _size;
ret = build_ram_table();
if ((ret != MBED_SUCCESS) && (ret != MBED_ERROR_INVALID_DATA_DETECTED)) {
MBED_ERROR(ret, "TDBSTORE: Unable to build RAM table at init");
}
if ((ret == MBED_ERROR_INVALID_DATA_DETECTED) && (_free_space_offset < _size)) {
// Space after last valid record may be erased, hence "corrupt". Now check if it really is erased.
bool erased;
if (is_erase_unit_erased(_active_area, _free_space_offset, erased)) {
MBED_ERROR(MBED_ERROR_READ_FAILED, "TDBSTORE: Unable to check whether erase unit is erased at init");
}
if (erased) {
// Erased - all good
ret = MBED_SUCCESS;
}
}
reserved_ret = do_reserved_data_get(0, RESERVED_AREA_SIZE);
// If we either have a corrupt record somewhere, or the reserved area is corrupt,
// perform garbage collection to salvage all preceding records and/or clean reserved area.
if ((ret == MBED_ERROR_INVALID_DATA_DETECTED) || (reserved_ret == MBED_ERROR_INVALID_DATA_DETECTED)) {
ret = garbage_collection();
if (ret) {
MBED_ERROR(ret, "TDBSTORE: Unable to perform GC at init");
}
os_ret = _buff_bd->sync();
if (os_ret) {
MBED_ERROR(MBED_ERROR_WRITE_FAILED, "TDBSTORE: Unable to sync BD at init");
}
}
end:
_is_initialized = true;
_mutex.unlock();
return ret;
}
int TDBStore::deinit()
{
_mutex.lock();
if (_is_initialized) {
_buff_bd->deinit();
delete _buff_bd;
ram_table_entry_t *ram_table = (ram_table_entry_t *) _ram_table;
delete[] ram_table;
delete[] _work_buf;
delete[] _key_buf;
}
_is_initialized = false;
_mutex.unlock();
return MBED_SUCCESS;
}
int TDBStore::reset_area(uint8_t area)
{
// Erase reserved area and master record
return check_erase_before_write(area, 0, _master_record_offset + _master_record_size, true);
}
int TDBStore::reset()
{
uint8_t area;
int ret;
if (!_is_initialized) {
return MBED_ERROR_NOT_READY;
}
_mutex.lock();
// Reset both areas
for (area = 0; area < _num_areas; area++) {
ret = reset_area(area);
if (ret) {
goto end;
}
}
_active_area = 0;
_num_keys = 0;
_free_space_offset = _master_record_offset;
_active_area_version = 1;
// Write an initial master record on active area
ret = write_master_record(_active_area, _active_area_version, _free_space_offset);
end:
_mutex.unlock();
return ret;
}
int TDBStore::iterator_open(iterator_t *it, const char *prefix)
{
key_iterator_handle_t *handle;
int ret = MBED_SUCCESS;
if (!_is_initialized) {
return MBED_ERROR_NOT_READY;
}
if (!it) {
return MBED_ERROR_INVALID_ARGUMENT;
}
_mutex.lock();
int it_num;
for (it_num = 0; it_num < _max_open_iterators; it_num++) {
if (!_iterator_table[it_num]) {
break;
}
}
if (it_num == _max_open_iterators) {
ret = MBED_ERROR_OUT_OF_RESOURCES;
goto end;
}
handle = new key_iterator_handle_t;
*it = reinterpret_cast<iterator_t>(handle);
if (prefix && strcmp(prefix, "")) {
handle->prefix = new char[strlen(prefix) + 1];
strcpy(handle->prefix, prefix);
} else {
handle->prefix = 0;
}
handle->ram_table_ind = 0;
handle->iterator_num = it_num;
_iterator_table[it_num] = handle;
end:
_mutex.unlock();
return ret;
}
int TDBStore::iterator_next(iterator_t it, char *key, size_t key_size)
{
ram_table_entry_t *ram_table = (ram_table_entry_t *) _ram_table;
key_iterator_handle_t *handle;
int ret;
uint32_t actual_data_size, hash, flags, next_offset;
if (!_is_initialized) {
return MBED_ERROR_NOT_READY;
}
_mutex.lock();
handle = reinterpret_cast<key_iterator_handle_t *>(it);
ret = MBED_ERROR_ITEM_NOT_FOUND;
while (ret && (handle->ram_table_ind < _num_keys)) {
ret = read_record(_active_area, ram_table[handle->ram_table_ind].bd_offset, _key_buf,
0, 0, actual_data_size, 0, true, false, false, false, hash, flags, next_offset);
if (ret) {
goto end;
}
if (!handle->prefix || (strstr(_key_buf, handle->prefix) == _key_buf)) {
if (strlen(_key_buf) >= key_size) {
ret = MBED_ERROR_INVALID_SIZE;
goto end;
}
strcpy(key, _key_buf);
} else {
ret = MBED_ERROR_ITEM_NOT_FOUND;
}
handle->ram_table_ind++;
}
end:
_mutex.unlock();
return ret;
}
int TDBStore::iterator_close(iterator_t it)
{
key_iterator_handle_t *handle;
if (!_is_initialized) {
return MBED_ERROR_NOT_READY;
}
_mutex.lock();
handle = reinterpret_cast<key_iterator_handle_t *>(it);
delete[] handle->prefix;
_iterator_table[handle->iterator_num] = 0;
delete handle;
_mutex.unlock();
return MBED_SUCCESS;
}
void TDBStore::update_all_iterators(bool added, uint32_t ram_table_ind)
{
for (int it_num = 0; it_num < _max_open_iterators; it_num++) {
key_iterator_handle_t *handle = static_cast <key_iterator_handle_t *>(_iterator_table[it_num]);
if (!handle) {
continue;
}
if (ram_table_ind >= handle->ram_table_ind) {
continue;
}
if (added) {
handle->ram_table_ind++;
} else {
handle->ram_table_ind--;
}
}
}
int TDBStore::reserved_data_set(const void *reserved_data, size_t reserved_data_buf_size)
{
reserved_trailer_t trailer;
int os_ret, ret = MBED_SUCCESS;
if (reserved_data_buf_size > RESERVED_AREA_SIZE) {
return MBED_ERROR_INVALID_SIZE;
}
_mutex.lock();
ret = do_reserved_data_get(0, RESERVED_AREA_SIZE);
if ((ret == MBED_SUCCESS) || (ret == MBED_ERROR_INVALID_DATA_DETECTED)) {
ret = MBED_ERROR_WRITE_FAILED;
goto end;
} else if (ret != MBED_ERROR_ITEM_NOT_FOUND) {
goto end;
}
ret = write_area(_active_area, 0, reserved_data_buf_size, reserved_data);
if (ret) {
goto end;
}
trailer.trailer_size = sizeof(trailer);
trailer.data_size = reserved_data_buf_size;
trailer.crc = calc_crc(initial_crc, reserved_data_buf_size, reserved_data);
ret = write_area(_active_area, RESERVED_AREA_SIZE, sizeof(trailer), &trailer);
if (ret) {
goto end;
}
os_ret = _buff_bd->sync();
if (os_ret) {
ret = MBED_ERROR_WRITE_FAILED;
goto end;
}
end:
_mutex.unlock();
return ret;
}
int TDBStore::do_reserved_data_get(void *reserved_data, size_t reserved_data_buf_size, size_t *actual_data_size)
{
reserved_trailer_t trailer;
uint8_t *buf;
int ret;
bool erased = true;
size_t actual_size;
uint32_t crc = initial_crc;
uint32_t offset;
uint8_t blank = _buff_bd->get_erase_value();
ret = read_area(_active_area, RESERVED_AREA_SIZE, sizeof(trailer), &trailer);
if (ret) {
return ret;
}
buf = reinterpret_cast <uint8_t *>(&trailer);
for (uint32_t i = 0; i < sizeof(trailer); i++) {
if (buf[i] != blank) {
erased = false;
break;
}
}
if (!erased) {
actual_size = trailer.data_size;
if (actual_data_size) {
*actual_data_size = actual_size;
}
if (reserved_data_buf_size < actual_size) {
return MBED_ERROR_INVALID_SIZE;
}
} else {
actual_size = std::min((size_t) RESERVED_AREA_SIZE, reserved_data_buf_size);
}
if (reserved_data) {
buf = reinterpret_cast <uint8_t *>(reserved_data);
} else {
buf = _work_buf;
}
offset = 0;
while (actual_size) {
uint32_t chunk = std::min(work_buf_size, (uint32_t) actual_size);
ret = read_area(_active_area, offset, chunk, buf);
if (ret) {
return ret;
}
for (uint32_t i = 0; i < chunk; i++) {
if (buf[i] != blank) {
erased = false;
break;
}
}
crc = calc_crc(crc, chunk, buf + offset);
offset += chunk;
actual_size -= chunk;
}
if (erased) {
return MBED_ERROR_ITEM_NOT_FOUND;
} else if (crc != trailer.crc) {
return MBED_ERROR_INVALID_DATA_DETECTED;
}
return MBED_SUCCESS;
}
int TDBStore::reserved_data_get(void *reserved_data, size_t reserved_data_buf_size, size_t *actual_data_size)
{
_mutex.lock();
int ret = do_reserved_data_get(reserved_data, reserved_data_buf_size, actual_data_size);
_mutex.unlock();
return ret;
}
void TDBStore::offset_in_erase_unit(uint8_t area, uint32_t offset,
uint32_t &offset_from_start, uint32_t &dist_to_end)
{
uint32_t bd_offset = _area_params[area].address + offset;
if (!_variant_bd_erase_unit_size) {
uint32_t eu_size = _buff_bd->get_erase_size();
offset_from_start = bd_offset % eu_size;
dist_to_end = eu_size - offset_from_start;
return;
}
uint32_t agg_offset = 0;
while (bd_offset >= agg_offset + _buff_bd->get_erase_size(agg_offset)) {
agg_offset += _buff_bd->get_erase_size(agg_offset);
}
offset_from_start = bd_offset - agg_offset;
dist_to_end = _buff_bd->get_erase_size(agg_offset) - offset_from_start;
}
int TDBStore::is_erase_unit_erased(uint8_t area, uint32_t offset, bool &erased)
{
uint32_t offset_from_start, dist;
offset_in_erase_unit(area, offset, offset_from_start, dist);
uint8_t buf[sizeof(record_header_t)], blanks[sizeof(record_header_t)];
memset(blanks, _buff_bd->get_erase_value(), sizeof(blanks));
while (dist) {
uint32_t chunk = std::min(dist, (uint32_t) sizeof(buf));
int ret = read_area(area, offset, chunk, buf);
if (ret) {
return MBED_ERROR_READ_FAILED;
}
if (memcmp(buf, blanks, chunk)) {
erased = false;
return MBED_SUCCESS;
}
offset += chunk;
dist -= chunk;
}
erased = true;
return MBED_SUCCESS;
}
int TDBStore::check_erase_before_write(uint8_t area, uint32_t offset, uint32_t size, bool force_check)
{
// In order to save init time, we don't check that the entire area is erased.
// Instead, whenever reaching an erase unit start, check that it's erased, and if not -
// erase it.
while (size) {
uint32_t dist, offset_from_start;
int ret;
offset_in_erase_unit(area, offset, offset_from_start, dist);
uint32_t chunk = std::min(size, dist);
if (!offset_from_start || force_check) {
// We're at the start of an erase unit. Here (and only here, if not forced),
// check if it's erased.
bool erased;
ret = is_erase_unit_erased(area, offset, erased);
if (ret) {
return MBED_ERROR_WRITE_FAILED;
}
if (!erased) {
ret = erase_erase_unit(area, offset - offset_from_start);
if (ret) {
return MBED_ERROR_WRITE_FAILED;
}
}
}
offset += chunk;
size -= chunk;
}
return MBED_SUCCESS;
}