2019-01-18 18:45:36 -04:00
|
|
|
/*
|
|
|
|
block based logging, for boards with flash logging
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "AP_Logger_Block.h"
|
|
|
|
|
2021-01-06 22:04:09 -04:00
|
|
|
#if HAL_LOGGING_BLOCK_ENABLED
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
#include <AP_HAL/AP_HAL.h>
|
|
|
|
#include <stdio.h>
|
2020-05-05 14:00:54 -03:00
|
|
|
#include <AP_RTC/AP_RTC.h>
|
2019-12-15 18:08:10 -04:00
|
|
|
#include <GCS_MAVLink/GCS.h>
|
2019-01-18 18:45:36 -04:00
|
|
|
|
2019-10-11 11:57:17 -03:00
|
|
|
const extern AP_HAL::HAL& hal;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
// the last page holds the log format in first 4 bytes. Please change
|
|
|
|
// this if (and only if!) the low level format changes
|
2020-05-05 14:00:54 -03:00
|
|
|
#define DF_LOGGING_FORMAT 0x1901201B
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
AP_Logger_Block::AP_Logger_Block(AP_Logger &front, LoggerMessageWriter_DFLogStart *writer) :
|
|
|
|
writebuf(0),
|
|
|
|
AP_Logger_Backend(front, writer)
|
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
// buffer is used for both reads and writes so access must always be within the semaphore
|
2019-01-18 18:45:36 -04:00
|
|
|
buffer = (uint8_t *)hal.util->malloc_type(page_size_max, AP_HAL::Util::MEM_DMA_SAFE);
|
|
|
|
if (buffer == nullptr) {
|
|
|
|
AP_HAL::panic("Out of DMA memory for logging");
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
df_stats_clear();
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// init is called after backend init
|
|
|
|
void AP_Logger_Block::Init(void)
|
|
|
|
{
|
|
|
|
if (CardInserted()) {
|
|
|
|
// reserve space for version in last sector
|
2019-12-15 18:08:10 -04:00
|
|
|
df_NumPages -= df_PagePerBlock;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
// determine and limit file backend buffersize
|
|
|
|
uint32_t bufsize = _front._params.file_bufsize;
|
|
|
|
if (bufsize > 64) {
|
|
|
|
bufsize = 64;
|
|
|
|
}
|
|
|
|
bufsize *= 1024;
|
|
|
|
|
|
|
|
// If we can't allocate the full size, try to reduce it until we can allocate it
|
2019-12-15 18:08:10 -04:00
|
|
|
while (!writebuf.set_size(bufsize) && bufsize >= df_PageSize * df_PagePerBlock) {
|
2019-01-18 18:45:36 -04:00
|
|
|
hal.console->printf("AP_Logger_Block: Couldn't set buffer size to=%u\n", (unsigned)bufsize);
|
|
|
|
bufsize >>= 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!writebuf.get_size()) {
|
|
|
|
hal.console->printf("Out of memory for logging\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
hal.console->printf("AP_Logger_Block: buffer size=%u\n", (unsigned)bufsize);
|
|
|
|
_initialised = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
WITH_SEMAPHORE(sem);
|
2021-01-06 00:32:16 -04:00
|
|
|
|
|
|
|
if (NeedErase()) {
|
|
|
|
EraseAll();
|
|
|
|
} else {
|
|
|
|
validate_log_structure();
|
|
|
|
}
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t AP_Logger_Block::bufferspace_available()
|
|
|
|
{
|
|
|
|
// because AP_Logger_Block devices are ring buffers, we *always*
|
|
|
|
// have room...
|
|
|
|
return df_NumPages * df_PageSize;
|
|
|
|
}
|
|
|
|
|
2019-02-11 04:38:01 -04:00
|
|
|
// *** LOGGER PUBLIC FUNCTIONS ***
|
2019-01-18 18:45:36 -04:00
|
|
|
void AP_Logger_Block::StartWrite(uint32_t PageAdr)
|
|
|
|
{
|
|
|
|
df_PageAdr = PageAdr;
|
|
|
|
}
|
|
|
|
|
|
|
|
void AP_Logger_Block::FinishWrite(void)
|
|
|
|
{
|
|
|
|
// Write Buffer to flash
|
|
|
|
BufferToPage(df_PageAdr);
|
|
|
|
df_PageAdr++;
|
|
|
|
|
|
|
|
// If we reach the end of the memory, start from the beginning
|
|
|
|
if (df_PageAdr > df_NumPages) {
|
|
|
|
df_PageAdr = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
// when starting a new sector, erase it
|
2019-12-15 18:08:10 -04:00
|
|
|
if ((df_PageAdr-1) % df_PagePerBlock == 0) {
|
2020-05-05 14:00:54 -03:00
|
|
|
// if we have wrapped over an existing log, force the oldest to be recalculated
|
|
|
|
if (_cached_oldest_log > 0) {
|
|
|
|
uint16_t log_num = StartRead(df_PageAdr);
|
|
|
|
if (log_num != 0xFFFF && log_num >= _cached_oldest_log) {
|
|
|
|
_cached_oldest_log = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// are we about to erase a sector with our own headers in it?
|
|
|
|
if (df_Write_FilePage > df_NumPages - df_PagePerBlock) {
|
|
|
|
chip_full = true;
|
|
|
|
return;
|
|
|
|
}
|
2019-12-15 18:08:10 -04:00
|
|
|
SectorErase(df_PageAdr / df_PagePerBlock);
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AP_Logger_Block::WritesOK() const
|
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
if (!CardInserted() || erase_started) {
|
2019-01-18 18:45:36 -04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AP_Logger_Block::_WritePrioritisedBlock(const void *pBuffer, uint16_t size, bool is_critical)
|
|
|
|
{
|
|
|
|
// is_critical is ignored - we're a ring buffer and never run out
|
|
|
|
// of space. possibly if we do more complicated bandwidth
|
|
|
|
// limiting we can reserve bandwidth based on is_critical
|
|
|
|
if (!WritesOK()) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (!WriteBlockCheckStartupMessages()) {
|
|
|
|
_dropped++;
|
2019-01-18 18:45:36 -04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (!write_sem.take(1)) {
|
2020-09-05 04:36:12 -03:00
|
|
|
_dropped++;
|
2020-05-05 14:00:54 -03:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2020-09-05 04:36:12 -03:00
|
|
|
const uint32_t space = writebuf.space();
|
2020-05-05 14:00:54 -03:00
|
|
|
|
|
|
|
if (_writing_startup_messages &&
|
|
|
|
_startup_messagewriter->fmt_done()) {
|
|
|
|
// the state machine has called us, and it has finished
|
|
|
|
// writing format messages out. It can always get back to us
|
|
|
|
// with more messages later, so let's leave room for other
|
|
|
|
// things:
|
|
|
|
const uint32_t now = AP_HAL::millis();
|
|
|
|
const bool must_dribble = (now - last_messagewrite_message_sent) > 100;
|
|
|
|
if (!must_dribble &&
|
|
|
|
space < non_messagewriter_message_reserved_space(writebuf.get_size())) {
|
|
|
|
// this message isn't dropped, it will be sent again...
|
|
|
|
write_sem.give();
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
last_messagewrite_message_sent = now;
|
|
|
|
} else {
|
|
|
|
// we reserve some amount of space for critical messages:
|
|
|
|
if (!is_critical && space < critical_message_reserved_space(writebuf.get_size())) {
|
|
|
|
_dropped++;
|
|
|
|
write_sem.give();
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// if no room for entire message - drop it:
|
|
|
|
if (space < size) {
|
|
|
|
_dropped++;
|
|
|
|
write_sem.give();
|
2019-01-18 18:45:36 -04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
writebuf.write((uint8_t*)pBuffer, size);
|
2020-05-05 14:00:54 -03:00
|
|
|
df_stats_gather(size, writebuf.space());
|
|
|
|
write_sem.give();
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// read from the page address and return the file number at that location
|
|
|
|
uint16_t AP_Logger_Block::StartRead(uint32_t PageAdr)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
|
|
|
df_Read_PageAdr = PageAdr;
|
|
|
|
|
|
|
|
// copy flash page to buffer
|
|
|
|
if (erase_started) {
|
|
|
|
memset(buffer, 0xff, df_PageSize);
|
|
|
|
} else {
|
|
|
|
PageToBuffer(df_Read_PageAdr);
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
return ReadHeaders();
|
|
|
|
}
|
2019-01-18 18:45:36 -04:00
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// read the headers at the current read point returning the file number
|
|
|
|
uint16_t AP_Logger_Block::ReadHeaders()
|
|
|
|
{
|
2019-01-18 18:45:36 -04:00
|
|
|
// We are starting a new page - read FileNumber and FilePage
|
|
|
|
struct PageHeader ph;
|
|
|
|
BlockRead(0, &ph, sizeof(ph));
|
|
|
|
df_FileNumber = ph.FileNumber;
|
|
|
|
df_FilePage = ph.FilePage;
|
2020-05-05 14:00:54 -03:00
|
|
|
#if BLOCK_LOG_VALIDATE
|
|
|
|
if (ph.crc != DF_LOGGING_FORMAT + df_FilePage && df_FileNumber != 0xFFFF) {
|
|
|
|
printf("ReadHeaders: invalid block read at %d\n", df_Read_PageAdr);
|
|
|
|
}
|
|
|
|
#endif
|
2019-01-18 18:45:36 -04:00
|
|
|
df_Read_BufferIdx = sizeof(ph);
|
2020-05-05 14:00:54 -03:00
|
|
|
// we are at the start of a file, read the file header
|
|
|
|
if (df_FilePage == 1) {
|
|
|
|
struct FileHeader fh;
|
|
|
|
BlockRead(0, &fh, sizeof(fh));
|
|
|
|
df_FileTime = fh.utc_secs;
|
|
|
|
df_Read_BufferIdx += sizeof(fh);
|
|
|
|
}
|
|
|
|
|
|
|
|
return df_FileNumber;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
bool AP_Logger_Block::ReadBlock(void *pBuffer, uint16_t size)
|
|
|
|
{
|
|
|
|
if (erase_started) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
while (size > 0) {
|
|
|
|
uint16_t n = df_PageSize - df_Read_BufferIdx;
|
|
|
|
if (n > size) {
|
|
|
|
n = size;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!BlockRead(df_Read_BufferIdx, pBuffer, n)) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
size -= n;
|
|
|
|
pBuffer = (void *)(n + (uintptr_t)pBuffer);
|
|
|
|
|
|
|
|
df_Read_BufferIdx += n;
|
|
|
|
|
|
|
|
if (df_Read_BufferIdx == df_PageSize) {
|
|
|
|
df_Read_PageAdr++;
|
|
|
|
if (df_Read_PageAdr > df_NumPages) {
|
|
|
|
df_Read_PageAdr = 1;
|
|
|
|
}
|
|
|
|
if (erase_started) {
|
|
|
|
memset(buffer, 0xff, df_PageSize);
|
|
|
|
} else {
|
|
|
|
PageToBuffer(df_Read_PageAdr);
|
|
|
|
}
|
|
|
|
|
|
|
|
// We are starting a new page - read FileNumber and FilePage
|
2020-05-05 14:00:54 -03:00
|
|
|
ReadHeaders();
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// initialize the log data for the given file number
|
|
|
|
void AP_Logger_Block::StartLogFile(uint16_t FileNumber)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
|
|
|
df_FileNumber = FileNumber;
|
2020-05-05 14:00:54 -03:00
|
|
|
df_Write_FileNumber = FileNumber;
|
2019-01-18 18:45:36 -04:00
|
|
|
df_FilePage = 1;
|
2020-05-05 14:00:54 -03:00
|
|
|
df_Write_FilePage = 1;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
2021-02-01 12:26:29 -04:00
|
|
|
uint16_t AP_Logger_Block::GetFileNumber() const
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
|
|
|
return df_FileNumber;
|
|
|
|
}
|
|
|
|
|
|
|
|
void AP_Logger_Block::EraseAll()
|
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
if (hal.util->get_soft_armed()) {
|
|
|
|
// do not want to do any filesystem operations while we are e.g. flying
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// push out the message before stopping logging
|
|
|
|
if (!erase_started) {
|
|
|
|
GCS_SEND_TEXT(MAV_SEVERITY_INFO, "Chip erase started");
|
|
|
|
}
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
WITH_SEMAPHORE(sem);
|
2020-05-05 14:00:54 -03:00
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
if (erase_started) {
|
|
|
|
// already erasing
|
|
|
|
return;
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
erase_started = true;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// remember what we were doing
|
|
|
|
new_log_pending = log_write_started;
|
2020-01-05 18:01:32 -04:00
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// throw away everything
|
2019-01-18 18:45:36 -04:00
|
|
|
log_write_started = false;
|
2020-05-05 14:00:54 -03:00
|
|
|
writebuf.clear();
|
|
|
|
|
|
|
|
// reset the format version and wrapped status so that any incomplete erase will be caught
|
|
|
|
Sector4kErase(get_sector(df_NumPages));
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
StartErase();
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
void AP_Logger_Block::periodic_1Hz()
|
|
|
|
{
|
|
|
|
AP_Logger_Backend::periodic_1Hz();
|
|
|
|
|
2021-07-28 03:52:35 -03:00
|
|
|
if (rate_limiter == nullptr && _front._params.blk_ratemax > 0) {
|
|
|
|
// setup rate limiting
|
|
|
|
rate_limiter = new AP_Logger_RateLimiter(_front, _front._params.blk_ratemax);
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (!io_thread_alive()) {
|
|
|
|
if (warning_decimation_counter == 0 && _initialised) {
|
|
|
|
// we don't print this error unless we did initialise. When _initialised is set to true
|
|
|
|
// we register the IO timer callback
|
|
|
|
GCS_SEND_TEXT(MAV_SEVERITY_CRITICAL, "AP_Logger: IO thread died");
|
|
|
|
}
|
|
|
|
if (warning_decimation_counter++ > 57) {
|
|
|
|
warning_decimation_counter = 0;
|
|
|
|
}
|
|
|
|
_initialised = false;
|
|
|
|
} else if (chip_full) {
|
|
|
|
if (warning_decimation_counter == 0) {
|
|
|
|
GCS_SEND_TEXT(MAV_SEVERITY_WARNING, "Chip full, logging stopped");
|
|
|
|
}
|
|
|
|
if (warning_decimation_counter++ > 57) {
|
|
|
|
warning_decimation_counter = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// EraseAll is asynchronous, but we must not start a new
|
|
|
|
// log in a child thread so this task picks up the hint from the io timer
|
|
|
|
// keeping locking to a minimum
|
|
|
|
void AP_Logger_Block::periodic_10Hz(const uint32_t now)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
if (erase_started || InErase()) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// don't print status messages in io thread, do it here
|
|
|
|
switch (status_msg) {
|
|
|
|
case StatusMessage::ERASE_COMPLETE:
|
|
|
|
GCS_SEND_TEXT(MAV_SEVERITY_INFO, "Chip erase complete");
|
|
|
|
status_msg = StatusMessage::NONE;
|
|
|
|
break;
|
|
|
|
case StatusMessage::RECOVERY_COMPLETE:
|
|
|
|
GCS_SEND_TEXT(MAV_SEVERITY_WARNING, "Log recovery complete");
|
|
|
|
status_msg = StatusMessage::NONE;
|
|
|
|
break;
|
|
|
|
case StatusMessage::NONE:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// EraseAll should only set this in the main thread
|
|
|
|
if (new_log_pending) {
|
|
|
|
start_new_log();
|
|
|
|
}
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* we need to erase if the logging format has changed
|
|
|
|
*/
|
|
|
|
bool AP_Logger_Block::NeedErase(void)
|
|
|
|
{
|
|
|
|
uint32_t version = 0;
|
2020-05-05 14:00:54 -03:00
|
|
|
PageToBuffer(df_NumPages+1); // last page
|
2019-01-18 18:45:36 -04:00
|
|
|
BlockRead(0, &version, sizeof(version));
|
|
|
|
if (version == DF_LOGGING_FORMAT) {
|
2020-05-05 14:00:54 -03:00
|
|
|
// only leave the read point in a sane place if we are not about to destroy everything
|
|
|
|
StartRead(1);
|
2019-01-18 18:45:36 -04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-12-15 18:08:10 -04:00
|
|
|
/*
|
|
|
|
* iterate through all of the logs files looking for ones that are corrupted and correct.
|
|
|
|
*/
|
|
|
|
void AP_Logger_Block::validate_log_structure()
|
|
|
|
{
|
|
|
|
WITH_SEMAPHORE(sem);
|
2020-05-05 14:00:54 -03:00
|
|
|
bool wrapped = is_wrapped();
|
2019-12-15 18:08:10 -04:00
|
|
|
uint32_t page = 1;
|
|
|
|
uint32_t page_start = 1;
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
uint16_t file = StartRead(page);
|
2019-12-15 18:08:10 -04:00
|
|
|
uint16_t first_file = file;
|
|
|
|
uint16_t next_file = file;
|
2020-01-04 07:29:22 -04:00
|
|
|
uint16_t last_file = 0;
|
2019-12-15 18:08:10 -04:00
|
|
|
|
|
|
|
while (file != 0xFFFF && page <= df_NumPages && (file == next_file || (wrapped && file < next_file))) {
|
|
|
|
uint32_t end_page = find_last_page_of_log(file);
|
|
|
|
if (end_page == 0 || end_page < page) { // this can happen and may be responsible for corruption that we have seen
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
page = end_page + 1;
|
2020-05-05 14:00:54 -03:00
|
|
|
file = StartRead(page);
|
2019-12-15 18:08:10 -04:00
|
|
|
next_file++;
|
2020-05-05 14:00:54 -03:00
|
|
|
// skip over the rest of an erased block
|
2020-01-05 18:01:32 -04:00
|
|
|
if (wrapped && file == 0xFFFF) {
|
2020-05-05 14:00:54 -03:00
|
|
|
file = StartRead((get_block(page) + 1) * df_PagePerBlock + 1);
|
2020-01-05 18:01:32 -04:00
|
|
|
}
|
2019-12-15 18:08:10 -04:00
|
|
|
if (wrapped && file < next_file) {
|
|
|
|
page_start = page;
|
|
|
|
next_file = file;
|
|
|
|
first_file = file;
|
|
|
|
} else if (last_file < next_file) {
|
|
|
|
last_file = file;
|
|
|
|
}
|
|
|
|
if (file == next_file) {
|
|
|
|
hal.console->printf("Found complete log %d at %X-%X\n", int(file), unsigned(page), unsigned(find_last_page_of_log(file)));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (file != 0xFFFF && file != next_file && page <= df_NumPages && page > 0) {
|
2020-05-05 14:00:54 -03:00
|
|
|
hal.console->printf("Found corrupt log %d at 0x%04X, erasing", int(file), unsigned(page));
|
2019-12-15 18:08:10 -04:00
|
|
|
df_EraseFrom = page;
|
|
|
|
} else if (next_file != 0xFFFF && page > 0 && next_file > 1) { // chip is empty
|
2020-05-05 14:00:54 -03:00
|
|
|
hal.console->printf("Found %d complete logs at 0x%04X-0x%04X", int(next_file - first_file), unsigned(page_start), unsigned(page - 1));
|
2019-12-15 18:08:10 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
/**
|
2020-05-05 14:00:54 -03:00
|
|
|
* get raw data from a log - page is the start page of the log, offset is the offset within the log starting at that page
|
2019-01-18 18:45:36 -04:00
|
|
|
*/
|
|
|
|
int16_t AP_Logger_Block::get_log_data_raw(uint16_t log_num, uint32_t page, uint32_t offset, uint16_t len, uint8_t *data)
|
|
|
|
{
|
|
|
|
WITH_SEMAPHORE(sem);
|
2020-05-05 14:00:54 -03:00
|
|
|
const uint16_t data_page_size = df_PageSize - sizeof(struct PageHeader);
|
|
|
|
const uint16_t first_page_size = data_page_size - sizeof(struct FileHeader);
|
|
|
|
|
|
|
|
// offset is the true offset in the file, so we have to calculate the offset accounting for page headers
|
|
|
|
if (offset >= first_page_size) {
|
|
|
|
offset -= first_page_size;
|
|
|
|
page = page + offset / data_page_size + 1;
|
|
|
|
offset %= data_page_size;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
if (page > df_NumPages) {
|
2020-05-05 14:00:54 -03:00
|
|
|
page = page % df_NumPages;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
|
|
|
|
// Sanity check we haven't been asked for an offset beyond the end of the log
|
|
|
|
if (StartRead(page) != log_num) {
|
|
|
|
return -1;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
df_Read_BufferIdx += offset;
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
if (!ReadBlock(data, len)) {
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return (int16_t)len;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
get data from a log, accounting for adding FMT headers
|
|
|
|
*/
|
2020-05-05 14:00:54 -03:00
|
|
|
int16_t AP_Logger_Block::get_log_data(uint16_t list_entry, uint16_t page, uint32_t offset, uint16_t len, uint8_t *data)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
const uint16_t log_num = log_num_from_list_entry(list_entry);
|
2019-01-18 18:45:36 -04:00
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (log_num == 0) {
|
|
|
|
// that failed - probably no logs
|
|
|
|
return -1;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
//printf("get_log_data(%d, %d, %d, %d)\n", log_num, page, offset, len);
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
|
|
|
uint16_t ret = 0;
|
2019-01-18 18:45:36 -04:00
|
|
|
if (len > 0) {
|
2019-04-18 18:43:24 -03:00
|
|
|
const int16_t bytes = get_log_data_raw(log_num, page, offset, len, data);
|
|
|
|
if (bytes == -1) {
|
2020-05-05 14:00:54 -03:00
|
|
|
return -1;
|
2019-04-18 18:43:24 -03:00
|
|
|
}
|
|
|
|
ret += bytes;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// This function determines the number of whole log files in the AP_Logger
|
|
|
|
// partial logs are rejected as without the headers they are relatively useless
|
2019-01-18 18:45:36 -04:00
|
|
|
uint16_t AP_Logger_Block::get_num_logs(void)
|
|
|
|
{
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
uint32_t lastpage;
|
|
|
|
uint32_t last;
|
|
|
|
|
|
|
|
if (!CardInserted() || find_last_page() == 1) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
uint32_t first = StartRead(1);
|
2019-12-15 18:08:10 -04:00
|
|
|
|
|
|
|
if (first == 0xFFFF) {
|
2019-01-18 18:45:36 -04:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
lastpage = find_last_page();
|
2020-05-05 14:00:54 -03:00
|
|
|
last = StartRead(lastpage);
|
|
|
|
|
|
|
|
if (is_wrapped()) {
|
2020-01-05 18:01:32 -04:00
|
|
|
// if we wrapped then the rest of the block will be filled with 0xFFFF because we always erase
|
|
|
|
// a block before writing to it, in order to find the first page we therefore have to read after the
|
|
|
|
// next block boundary
|
2020-05-05 14:00:54 -03:00
|
|
|
first = StartRead((get_block(lastpage) + 1) * df_PagePerBlock + 1);
|
|
|
|
// unless we happen to land on the first page of the file that is being overwritten we skip to the next file
|
|
|
|
if (df_FilePage > 1) {
|
|
|
|
first++;
|
|
|
|
}
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
if (last == first) {
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return (last - first + 1);
|
|
|
|
}
|
|
|
|
|
2020-09-10 16:57:40 -03:00
|
|
|
// stop logging immediately
|
2020-05-05 14:00:54 -03:00
|
|
|
void AP_Logger_Block::stop_logging(void)
|
|
|
|
{
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
|
|
|
log_write_started = false;
|
|
|
|
|
2020-09-10 16:57:40 -03:00
|
|
|
// nuke writing any previous log
|
2020-05-05 14:00:54 -03:00
|
|
|
writebuf.clear();
|
|
|
|
}
|
2019-01-18 18:45:36 -04:00
|
|
|
|
2020-09-10 16:57:40 -03:00
|
|
|
// stop logging and flush any remaining data
|
|
|
|
void AP_Logger_Block::stop_logging_async(void)
|
|
|
|
{
|
|
|
|
stop_log_pending = true;
|
|
|
|
}
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
// This function starts a new log file in the AP_Logger
|
2020-05-05 14:00:54 -03:00
|
|
|
// no actual data should be written to the storage here
|
|
|
|
// that should all be handled by the IO thread
|
2020-01-20 13:27:12 -04:00
|
|
|
void AP_Logger_Block::start_new_log(void)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
if (erase_started) {
|
|
|
|
// already erasing
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
WITH_SEMAPHORE(sem);
|
2020-05-05 14:00:54 -03:00
|
|
|
|
|
|
|
if (logging_started()) {
|
|
|
|
stop_logging();
|
|
|
|
}
|
|
|
|
|
|
|
|
// no need to schedule this anymore
|
|
|
|
new_log_pending = false;
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
uint32_t last_page = find_last_page();
|
|
|
|
|
|
|
|
StartRead(last_page);
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
log_write_started = true;
|
|
|
|
uint16_t new_log_num = 1;
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
if (find_last_log() == 0 || GetFileNumber() == 0xFFFF) {
|
2020-05-05 14:00:54 -03:00
|
|
|
StartLogFile(new_log_num);
|
2019-01-18 18:45:36 -04:00
|
|
|
StartWrite(1);
|
|
|
|
// Check for log of length 1 page and suppress
|
2020-05-05 14:00:54 -03:00
|
|
|
} else if (df_FilePage <= 1) {
|
2019-01-18 18:45:36 -04:00
|
|
|
new_log_num = GetFileNumber();
|
|
|
|
// Last log too short, reuse its number
|
|
|
|
// and overwrite it
|
2020-05-05 14:00:54 -03:00
|
|
|
StartLogFile(new_log_num);
|
2019-01-18 18:45:36 -04:00
|
|
|
StartWrite(last_page);
|
|
|
|
} else {
|
|
|
|
new_log_num = GetFileNumber()+1;
|
|
|
|
if (last_page == 0xFFFF) {
|
|
|
|
last_page=0;
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
StartLogFile(new_log_num);
|
2019-01-18 18:45:36 -04:00
|
|
|
StartWrite(last_page + 1);
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
|
|
|
|
// save UTC time in the first 4 bytes so that we can retrieve it later
|
|
|
|
uint64_t utc_usec;
|
|
|
|
FileHeader hdr {};
|
|
|
|
if (AP::rtc().get_utc_usec(utc_usec)) {
|
|
|
|
hdr.utc_secs = utc_usec / 1000000U;
|
|
|
|
}
|
|
|
|
writebuf.write((uint8_t*)&hdr, sizeof(FileHeader));
|
|
|
|
|
|
|
|
start_new_log_reset_variables();
|
|
|
|
|
2020-01-20 13:27:12 -04:00
|
|
|
return;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// This function finds the first and last pages of a log file
|
|
|
|
// The first page may be greater than the last page if the AP_Logger has been filled and partially overwritten.
|
2020-05-05 14:00:54 -03:00
|
|
|
void AP_Logger_Block::get_log_boundaries(uint16_t list_entry, uint32_t & start_page, uint32_t & end_page)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
const uint16_t log_num = log_num_from_list_entry(list_entry);
|
|
|
|
if (log_num == 0) {
|
|
|
|
// that failed - probably no logs
|
|
|
|
start_page = 0;
|
|
|
|
end_page = 0;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
uint16_t num = get_num_logs();
|
|
|
|
uint32_t look;
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
end_page = find_last_page_of_log(log_num);
|
|
|
|
|
|
|
|
if (num == 1 || log_num == 1) {
|
|
|
|
if (!is_wrapped()) {
|
2019-01-18 18:45:36 -04:00
|
|
|
start_page = 1;
|
|
|
|
} else {
|
2020-05-05 14:00:54 -03:00
|
|
|
StartRead(end_page);
|
|
|
|
start_page = (end_page + df_NumPages - df_FilePage) % df_NumPages + 1;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
} else {
|
2020-05-05 14:00:54 -03:00
|
|
|
// looking for the first log which might have a gap in front of it
|
|
|
|
if (list_entry == 1) {
|
|
|
|
StartRead(end_page);
|
|
|
|
if (end_page > df_FilePage) { // log is not wrapped
|
|
|
|
start_page = end_page - df_FilePage + 1;
|
|
|
|
} else { // log is wrapped
|
|
|
|
start_page = (end_page + df_NumPages - df_FilePage) % df_NumPages + 1;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
} else {
|
2020-05-05 14:00:54 -03:00
|
|
|
look = log_num-1;
|
|
|
|
do {
|
|
|
|
start_page = find_last_page_of_log(look) + 1;
|
|
|
|
look--;
|
|
|
|
} while (start_page <= 0 && look >=1);
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
|
|
|
|
if (start_page == df_NumPages + 1 || start_page == 0) {
|
2019-01-18 18:45:36 -04:00
|
|
|
start_page = 1;
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
if (end_page == 0) {
|
|
|
|
end_page = start_page;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// return true if logging has wrapped around to the beginning of the chip
|
|
|
|
bool AP_Logger_Block::is_wrapped(void)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
return StartRead(df_NumPages) != 0xFFFF;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2019-12-15 18:08:10 -04:00
|
|
|
// This function finds the last log number
|
2019-01-18 18:45:36 -04:00
|
|
|
uint16_t AP_Logger_Block::find_last_log(void)
|
|
|
|
{
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
uint32_t last_page = find_last_page();
|
2020-05-05 14:00:54 -03:00
|
|
|
return StartRead(last_page);
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// This function finds the last page of the last file
|
|
|
|
uint32_t AP_Logger_Block::find_last_page(void)
|
|
|
|
{
|
|
|
|
uint32_t look;
|
|
|
|
uint32_t bottom = 1;
|
|
|
|
uint32_t top = df_NumPages;
|
2019-10-03 11:58:13 -03:00
|
|
|
uint64_t look_hash;
|
|
|
|
uint64_t bottom_hash;
|
|
|
|
uint64_t top_hash;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
|
|
|
StartRead(bottom);
|
2019-10-03 11:58:13 -03:00
|
|
|
bottom_hash = ((int64_t)GetFileNumber()<<32) | df_FilePage;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
while (top-bottom > 1) {
|
|
|
|
look = (top+bottom)/2;
|
|
|
|
StartRead(look);
|
2019-10-03 11:58:13 -03:00
|
|
|
look_hash = (int64_t)GetFileNumber()<<32 | df_FilePage;
|
2020-01-05 18:01:32 -04:00
|
|
|
// erased sector so can discount everything above
|
2019-10-03 11:58:13 -03:00
|
|
|
if (look_hash >= 0xFFFF00000000) {
|
2019-01-18 18:45:36 -04:00
|
|
|
look_hash = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (look_hash < bottom_hash) {
|
|
|
|
// move down
|
|
|
|
top = look;
|
|
|
|
} else {
|
|
|
|
// move up
|
|
|
|
bottom = look;
|
|
|
|
bottom_hash = look_hash;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
StartRead(top);
|
2019-10-03 11:58:13 -03:00
|
|
|
top_hash = ((int64_t)GetFileNumber()<<32) | df_FilePage;
|
|
|
|
if (top_hash >= 0xFFFF00000000) {
|
2019-01-18 18:45:36 -04:00
|
|
|
top_hash = 0;
|
|
|
|
}
|
|
|
|
if (top_hash > bottom_hash) {
|
|
|
|
return top;
|
|
|
|
}
|
|
|
|
|
|
|
|
return bottom;
|
|
|
|
}
|
|
|
|
|
|
|
|
// This function finds the last page of a particular log file
|
|
|
|
uint32_t AP_Logger_Block::find_last_page_of_log(uint16_t log_number)
|
|
|
|
{
|
|
|
|
uint32_t look;
|
|
|
|
uint32_t bottom;
|
|
|
|
uint32_t top;
|
2019-10-03 11:58:13 -03:00
|
|
|
uint64_t look_hash;
|
|
|
|
uint64_t check_hash;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (is_wrapped()) {
|
|
|
|
bottom = StartRead(1);
|
2019-01-18 18:45:36 -04:00
|
|
|
if (bottom > log_number) {
|
|
|
|
bottom = find_last_page();
|
|
|
|
top = df_NumPages;
|
|
|
|
} else {
|
|
|
|
bottom = 1;
|
|
|
|
top = find_last_page();
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
bottom = 1;
|
|
|
|
top = find_last_page();
|
|
|
|
}
|
|
|
|
|
2019-10-03 11:58:13 -03:00
|
|
|
check_hash = (int64_t)log_number<<32 | 0xFFFFFFFF;
|
2019-01-18 18:45:36 -04:00
|
|
|
|
|
|
|
while (top-bottom > 1) {
|
|
|
|
look = (top+bottom)/2;
|
|
|
|
StartRead(look);
|
2019-10-03 11:58:13 -03:00
|
|
|
look_hash = (int64_t)GetFileNumber()<<32 | df_FilePage;
|
|
|
|
if (look_hash >= 0xFFFF00000000) {
|
2019-01-18 18:45:36 -04:00
|
|
|
look_hash = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (look_hash > check_hash) {
|
|
|
|
// move down
|
|
|
|
top = look;
|
|
|
|
} else {
|
|
|
|
// move up
|
|
|
|
bottom = look;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (StartRead(top) == log_number) {
|
2019-01-18 18:45:36 -04:00
|
|
|
return top;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (StartRead(bottom) == log_number) {
|
2019-01-18 18:45:36 -04:00
|
|
|
return bottom;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
GCS_SEND_TEXT(MAV_SEVERITY_ERROR, "No last page of log %d at top=%X or bot=%X", int(log_number), unsigned(top), unsigned(bottom));
|
2019-01-18 18:45:36 -04:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
void AP_Logger_Block::get_log_info(uint16_t list_entry, uint32_t &size, uint32_t &time_utc)
|
2019-01-18 18:45:36 -04:00
|
|
|
{
|
|
|
|
uint32_t start, end;
|
|
|
|
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
get_log_boundaries(list_entry, start, end);
|
2019-01-18 18:45:36 -04:00
|
|
|
if (end >= start) {
|
2020-05-05 14:00:54 -03:00
|
|
|
size = (end + 1 - start) * (uint32_t)(df_PageSize - sizeof(PageHeader));
|
2019-01-18 18:45:36 -04:00
|
|
|
} else {
|
2020-05-05 14:00:54 -03:00
|
|
|
size = (df_NumPages + end + 1 - start) * (uint32_t)(df_PageSize - sizeof(PageHeader));
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
size -= sizeof(FileHeader);
|
2019-01-18 18:45:36 -04:00
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
//printf("LOG %d(%d), %d-%d, size %d\n", log_num_from_list_entry(list_entry), list_entry, start, end, size);
|
|
|
|
|
|
|
|
StartRead(start);
|
|
|
|
|
|
|
|
// the log we are currently writing
|
|
|
|
if (df_FileTime == 0 && df_FileNumber == df_Write_FileNumber) {
|
|
|
|
uint64_t utc_usec;
|
|
|
|
if (AP::rtc().get_utc_usec(utc_usec)) {
|
|
|
|
df_FileTime = utc_usec / 1000000U;
|
|
|
|
}
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
time_utc = df_FileTime;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// read size bytes of data from the buffer
|
|
|
|
bool AP_Logger_Block::BlockRead(uint16_t IntPageAdr, void *pBuffer, uint16_t size)
|
|
|
|
{
|
|
|
|
memcpy(pBuffer, &buffer[IntPageAdr], size);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
bool AP_Logger_Block::logging_failed() const
|
|
|
|
{
|
|
|
|
if (!_initialised) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
if (!io_thread_alive()) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
if (chip_full) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2020-11-21 16:07:37 -04:00
|
|
|
// detect whether the IO thread is running, since this is considered a catastrophic failure for the logging system
|
|
|
|
// better be really, really sure
|
2020-05-05 14:00:54 -03:00
|
|
|
bool AP_Logger_Block::io_thread_alive() const
|
|
|
|
{
|
2020-11-21 16:07:37 -04:00
|
|
|
// if the io thread hasn't had a heartbeat in 3s it is dead
|
2020-12-18 16:39:42 -04:00
|
|
|
return (AP_HAL::millis() - io_timer_heartbeat) < 3000U || !hal.scheduler->is_system_initialized();
|
2020-05-05 14:00:54 -03:00
|
|
|
}
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
/*
|
|
|
|
IO timer running on IO thread
|
2020-05-05 14:00:54 -03:00
|
|
|
The IO timer runs every 1ms or at 1Khz. The standard flash chip can write rougly 130Kb/s
|
|
|
|
so there is little point in trying to write more than 130 bytes - or 1 page (256 bytes).
|
|
|
|
The W25Q128FV datasheet gives tpp as typically 0.7ms yielding an absolute maximum rate of
|
|
|
|
365Kb/s or just over a page per cycle.
|
2019-01-18 18:45:36 -04:00
|
|
|
*/
|
|
|
|
void AP_Logger_Block::io_timer(void)
|
|
|
|
{
|
2020-05-05 14:00:54 -03:00
|
|
|
uint32_t tnow = AP_HAL::millis();
|
|
|
|
io_timer_heartbeat = tnow;
|
|
|
|
|
|
|
|
// don't write anything for the first 2s to give the dataflash chip a chance to be ready
|
|
|
|
if (!_initialised || tnow < 2000) {
|
2019-01-18 18:45:36 -04:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (erase_started) {
|
2020-05-05 14:00:54 -03:00
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
2019-01-18 18:45:36 -04:00
|
|
|
if (InErase()) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
// write the logging format in the last page
|
|
|
|
StartWrite(df_NumPages+1);
|
|
|
|
uint32_t version = DF_LOGGING_FORMAT;
|
|
|
|
memset(buffer, 0, df_PageSize);
|
|
|
|
memcpy(buffer, &version, sizeof(version));
|
|
|
|
FinishWrite();
|
|
|
|
erase_started = false;
|
2020-05-05 14:00:54 -03:00
|
|
|
chip_full = false;
|
|
|
|
status_msg = StatusMessage::ERASE_COMPLETE;
|
2020-01-05 18:01:32 -04:00
|
|
|
return;
|
2019-12-15 18:08:10 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
if (df_EraseFrom > 0) {
|
2020-05-05 14:00:54 -03:00
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
2019-12-15 18:08:10 -04:00
|
|
|
const uint32_t sectors = df_NumPages / df_PagePerSector;
|
|
|
|
const uint32_t sectors_in_64k = 0x10000 / (df_PagePerSector * df_PageSize);
|
|
|
|
uint32_t next_sector = get_sector(df_EraseFrom);
|
|
|
|
const uint32_t aligned_sector = sectors - (((df_NumPages - df_EraseFrom + 1) / df_PagePerSector) / sectors_in_64k) * sectors_in_64k;
|
|
|
|
while (next_sector < aligned_sector) {
|
|
|
|
Sector4kErase(next_sector);
|
2020-11-21 16:07:37 -04:00
|
|
|
io_timer_heartbeat = AP_HAL::millis();
|
2019-12-15 18:08:10 -04:00
|
|
|
next_sector++;
|
|
|
|
}
|
|
|
|
uint16_t blocks_erased = 0;
|
|
|
|
while (next_sector < sectors) {
|
|
|
|
blocks_erased++;
|
|
|
|
SectorErase(next_sector / sectors_in_64k);
|
2020-11-21 16:07:37 -04:00
|
|
|
io_timer_heartbeat = AP_HAL::millis();
|
2019-12-15 18:08:10 -04:00
|
|
|
next_sector += sectors_in_64k;
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
status_msg = StatusMessage::RECOVERY_COMPLETE;
|
2019-12-15 18:08:10 -04:00
|
|
|
df_EraseFrom = 0;
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
if (!CardInserted() || new_log_pending || chip_full) {
|
2019-01-18 18:45:36 -04:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2020-09-10 16:57:40 -03:00
|
|
|
// we have been asked to stop logging, flush everything
|
|
|
|
if (stop_log_pending) {
|
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
|
|
|
log_write_started = false;
|
|
|
|
|
2020-09-17 18:03:28 -03:00
|
|
|
// complete writing any previous log, a page at a time to avoid holding the lock for too long
|
|
|
|
if (writebuf.available()) {
|
2020-09-10 16:57:40 -03:00
|
|
|
write_log_page();
|
2020-09-17 18:03:28 -03:00
|
|
|
} else {
|
|
|
|
writebuf.clear();
|
|
|
|
stop_log_pending = false;
|
2020-09-10 16:57:40 -03:00
|
|
|
}
|
|
|
|
|
2020-05-05 14:00:54 -03:00
|
|
|
// write at most one page
|
2020-09-10 16:57:40 -03:00
|
|
|
} else if (writebuf.available() >= df_PageSize - sizeof(struct PageHeader)) {
|
2020-05-05 14:00:54 -03:00
|
|
|
WITH_SEMAPHORE(sem);
|
|
|
|
|
|
|
|
write_log_page();
|
2019-01-18 18:45:36 -04:00
|
|
|
}
|
|
|
|
}
|
2020-05-05 14:00:54 -03:00
|
|
|
|
|
|
|
// write out a page of log data
|
|
|
|
void AP_Logger_Block::write_log_page()
|
|
|
|
{
|
|
|
|
struct PageHeader ph;
|
|
|
|
ph.FileNumber = df_Write_FileNumber;
|
|
|
|
ph.FilePage = df_Write_FilePage;
|
|
|
|
#if BLOCK_LOG_VALIDATE
|
|
|
|
ph.crc = DF_LOGGING_FORMAT + df_Write_FilePage;
|
|
|
|
#endif
|
|
|
|
memcpy(buffer, &ph, sizeof(ph));
|
|
|
|
const uint32_t pagesize = df_PageSize - sizeof(ph);
|
|
|
|
uint32_t nbytes = writebuf.read(&buffer[sizeof(ph)], pagesize);
|
|
|
|
if (nbytes < pagesize) {
|
|
|
|
memset(&buffer[sizeof(ph) + nbytes], 0, pagesize - nbytes);
|
|
|
|
}
|
|
|
|
FinishWrite();
|
|
|
|
df_Write_FilePage++;
|
|
|
|
}
|
|
|
|
|
2021-01-06 22:04:09 -04:00
|
|
|
#endif // HAL_LOGGING_BLOCK_ENABLED
|