You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

161 lines
5.2 KiB

/*
* Copyright (C) 2020 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "SerializedFlushToState.h"
#include <limits>
#include <android-base/logging.h>
SerializedFlushToState::SerializedFlushToState(uint64_t start, LogMask log_mask,
std::list<SerializedLogChunk>* logs)
: FlushToState(start, log_mask), logs_(logs) {
log_id_for_each(i) {
if (((1 << i) & log_mask) == 0) {
continue;
}
logs_needed_from_next_position_[i] = true;
}
}
SerializedFlushToState::~SerializedFlushToState() {
log_id_for_each(i) {
if (log_positions_[i]) {
log_positions_[i]->buffer_it->DetachReader(this);
}
}
}
void SerializedFlushToState::CreateLogPosition(log_id_t log_id) {
CHECK(!logs_[log_id].empty());
LogPosition log_position;
auto it = logs_[log_id].begin();
while (it != logs_[log_id].end() && start() > it->highest_sequence_number()) {
++it;
}
if (it == logs_[log_id].end()) {
--it;
}
it->AttachReader(this);
log_position.buffer_it = it;
// Find the offset of the first log with sequence number >= start().
int read_offset = 0;
while (read_offset < it->write_offset()) {
const auto* entry = it->log_entry(read_offset);
if (entry->sequence() >= start()) {
break;
}
read_offset += entry->total_len();
}
log_position.read_offset = read_offset;
log_positions_[log_id].emplace(log_position);
}
void SerializedFlushToState::UpdateLogsNeeded(log_id_t log_id) {
auto& buffer_it = log_positions_[log_id]->buffer_it;
auto read_offset = log_positions_[log_id]->read_offset;
// If there is another log to read in this buffer, let it be read.
if (read_offset < buffer_it->write_offset()) {
logs_needed_from_next_position_[log_id] = false;
} else if (read_offset == buffer_it->write_offset()) {
// If there are no more logs to read in this buffer and it's the last buffer, then
// set logs_needed_from_next_position_ to wait until more logs get logged.
if (buffer_it == std::prev(logs_[log_id].end())) {
logs_needed_from_next_position_[log_id] = true;
} else {
// Otherwise, if there is another buffer piece, move to that and do the same check.
buffer_it->DetachReader(this);
++buffer_it;
buffer_it->AttachReader(this);
log_positions_[log_id]->read_offset = 0;
if (buffer_it->write_offset() == 0) {
logs_needed_from_next_position_[log_id] = true;
} else {
logs_needed_from_next_position_[log_id] = false;
}
}
} else {
// read_offset > buffer_it->write_offset() should never happen.
LOG(FATAL) << "read_offset (" << read_offset << ") > buffer_it->write_offset() ("
<< buffer_it->write_offset() << ")";
}
}
void SerializedFlushToState::CheckForNewLogs() {
log_id_for_each(i) {
if (!logs_needed_from_next_position_[i]) {
continue;
}
if (!log_positions_[i]) {
if (logs_[i].empty()) {
continue;
}
CreateLogPosition(i);
}
UpdateLogsNeeded(i);
}
}
bool SerializedFlushToState::HasUnreadLogs() {
CheckForNewLogs();
log_id_for_each(i) {
if (log_positions_[i] && !logs_needed_from_next_position_[i]) {
return true;
}
}
return false;
}
LogWithId SerializedFlushToState::PopNextUnreadLog() {
uint64_t min_sequence = std::numeric_limits<uint64_t>::max();
log_id_t log_id;
const SerializedLogEntry* entry = nullptr;
log_id_for_each(i) {
if (!log_positions_[i] || logs_needed_from_next_position_[i]) {
continue;
}
if (log_positions_[i]->log_entry()->sequence() < min_sequence) {
log_id = i;
entry = log_positions_[i]->log_entry();
min_sequence = entry->sequence();
}
}
CHECK_NE(nullptr, entry);
log_positions_[log_id]->read_offset += entry->total_len();
logs_needed_from_next_position_[log_id] = true;
return {log_id, entry};
}
void SerializedFlushToState::Prune(log_id_t log_id) {
CHECK(log_positions_[log_id].has_value());
// Decrease the ref count since we're deleting our reference.
log_positions_[log_id]->buffer_it->DetachReader(this);
// Delete in the reference.
log_positions_[log_id].reset();
// Finally set logs_needed_from_next_position_, so CheckForNewLogs() will re-create the
// log_position_ object during the next read.
logs_needed_from_next_position_[log_id] = true;
}