mirror of
https://github.com/scottlamb/moonfire-nvr.git
synced 2024-12-26 07:05:56 -05:00
b9db9c11cc
...in particular, pass the error as a tracing param rather than directly in the main error message. The new behavior is nicer with e.g. OpenTelemetry.
489 lines
16 KiB
Rust
489 lines
16 KiB
Rust
// This file is part of Moonfire NVR, a security camera network video recorder.
|
|
// Copyright (C) 2020 The Moonfire NVR Authors; see AUTHORS and LICENSE.txt.
|
|
// SPDX-License-Identifier: GPL-v3.0-or-later WITH GPL-3.0-linking-exception.
|
|
|
|
//! Subcommand to check the database and sample file dir for errors.
|
|
|
|
use crate::compare;
|
|
use crate::db::{self, CompositeId, SqlUuid};
|
|
use crate::dir;
|
|
use crate::json::SampleFileDirConfig;
|
|
use crate::raw;
|
|
use crate::recording;
|
|
use crate::schema;
|
|
use base::{err, Error};
|
|
use fnv::{FnvHashMap, FnvHashSet};
|
|
use nix::fcntl::AtFlags;
|
|
use rusqlite::params;
|
|
use std::os::unix::io::AsRawFd;
|
|
use tracing::{error, info, warn};
|
|
|
|
pub struct Options {
|
|
pub compare_lens: bool,
|
|
pub trash_orphan_sample_files: bool,
|
|
pub delete_orphan_rows: bool,
|
|
pub trash_corrupt_rows: bool,
|
|
}
|
|
|
|
#[derive(Default)]
|
|
pub struct Context {
|
|
rows_to_delete: FnvHashSet<CompositeId>,
|
|
files_to_trash: FnvHashSet<(i32, CompositeId)>, // (dir_id, composite_id)
|
|
}
|
|
|
|
pub fn run(conn: &mut rusqlite::Connection, opts: &Options) -> Result<i32, Error> {
|
|
let mut printed_error = false;
|
|
|
|
info!("Checking SQLite database integrity...");
|
|
{
|
|
let mut stmt = conn.prepare("pragma integrity_check")?;
|
|
let mut rows = stmt.query(params![])?;
|
|
while let Some(row) = rows.next()? {
|
|
let e: String = row.get(0)?;
|
|
if e == "ok" {
|
|
continue;
|
|
}
|
|
error!(err = %e, "sqlite integrity error");
|
|
printed_error = true;
|
|
}
|
|
}
|
|
info!("...done");
|
|
|
|
// Compare stated schema version.
|
|
if let Err(e) = db::check_schema_version(conn) {
|
|
error!("Schema version is not as expected:\n{}", e);
|
|
printed_error = true;
|
|
} else {
|
|
info!("Schema at expected version {}.", db::EXPECTED_VERSION);
|
|
}
|
|
|
|
// Compare schemas.
|
|
{
|
|
let mut expected = rusqlite::Connection::open_in_memory()?;
|
|
db::init(&mut expected)?;
|
|
if let Some(diffs) = compare::get_diffs("actual", conn, "expected", &expected)? {
|
|
error!("Schema is not as expected:\n{}", &diffs);
|
|
printed_error = true;
|
|
} else {
|
|
info!("Schema is as expected.");
|
|
}
|
|
}
|
|
|
|
if printed_error {
|
|
warn!("The following analysis may be incorrect or encounter errors due to schema differences.");
|
|
}
|
|
|
|
let (db_uuid, _config) = raw::read_meta(conn)?;
|
|
|
|
// Scan directories.
|
|
let mut dirs_by_id: FnvHashMap<i32, Dir> = FnvHashMap::default();
|
|
{
|
|
let mut dir_stmt = conn.prepare(
|
|
r#"
|
|
select d.id, d.config, d.uuid, d.last_complete_open_id, o.uuid
|
|
from sample_file_dir d left join open o on (d.last_complete_open_id = o.id)
|
|
"#,
|
|
)?;
|
|
let mut garbage_stmt =
|
|
conn.prepare_cached("select composite_id from garbage where sample_file_dir_id = ?")?;
|
|
let mut rows = dir_stmt.query(params![])?;
|
|
while let Some(row) = rows.next()? {
|
|
let mut meta = schema::DirMeta::default();
|
|
let dir_id: i32 = row.get(0)?;
|
|
let config: SampleFileDirConfig = row.get(1)?;
|
|
let dir_uuid: SqlUuid = row.get(2)?;
|
|
let open_id = row.get(3)?;
|
|
let open_uuid: SqlUuid = row.get(4)?;
|
|
meta.db_uuid.extend_from_slice(&db_uuid.as_bytes()[..]);
|
|
meta.dir_uuid.extend_from_slice(&dir_uuid.0.as_bytes()[..]);
|
|
{
|
|
let o = meta.last_complete_open.mut_or_insert_default();
|
|
o.id = open_id;
|
|
o.uuid.extend_from_slice(&open_uuid.0.as_bytes()[..]);
|
|
}
|
|
|
|
// Open the directory (checking its metadata) and hold it open (for the lock).
|
|
let dir = dir::SampleFileDir::open(&config.path, &meta)
|
|
.map_err(|e| err!(e, msg("unable to open dir {}", config.path.display())))?;
|
|
let mut streams = read_dir(&dir, opts)?;
|
|
let mut rows = garbage_stmt.query(params![dir_id])?;
|
|
while let Some(row) = rows.next()? {
|
|
let id = CompositeId(row.get(0)?);
|
|
let s = streams.entry(id.stream()).or_insert_with(Stream::default);
|
|
s.recordings
|
|
.entry(id.recording())
|
|
.or_insert_with(Recording::default)
|
|
.garbage_row = true;
|
|
}
|
|
dirs_by_id.insert(dir_id, streams);
|
|
}
|
|
}
|
|
|
|
// Scan known streams.
|
|
let mut ctx = Context::default();
|
|
{
|
|
let mut stmt = conn.prepare(
|
|
r#"
|
|
select
|
|
id,
|
|
sample_file_dir_id,
|
|
cum_recordings
|
|
from
|
|
stream
|
|
where
|
|
sample_file_dir_id is not null
|
|
"#,
|
|
)?;
|
|
let mut rows = stmt.query(params![])?;
|
|
while let Some(row) = rows.next()? {
|
|
let stream_id = row.get(0)?;
|
|
let dir_id = row.get(1)?;
|
|
let cum_recordings = row.get(2)?;
|
|
let mut stream = match dirs_by_id.get_mut(&dir_id) {
|
|
None => Stream::default(),
|
|
Some(d) => d.remove(&stream_id).unwrap_or_default(),
|
|
};
|
|
stream.cum_recordings = Some(cum_recordings);
|
|
printed_error |= compare_stream(conn, dir_id, stream_id, opts, stream, &mut ctx)?;
|
|
}
|
|
}
|
|
|
|
// Expect the rest to have only garbage.
|
|
for (&dir_id, streams) in &dirs_by_id {
|
|
for (&stream_id, stream) in streams {
|
|
for (&recording_id, r) in &stream.recordings {
|
|
let id = CompositeId::new(stream_id, recording_id);
|
|
if r.recording_row.is_some()
|
|
|| r.playback_row.is_some()
|
|
|| r.integrity_row
|
|
|| !r.garbage_row
|
|
{
|
|
error!(
|
|
"dir {} recording {} for unknown stream: {:#?}",
|
|
dir_id, id, r
|
|
);
|
|
printed_error = true;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if !ctx.rows_to_delete.is_empty() || !ctx.files_to_trash.is_empty() {
|
|
let tx = conn.transaction()?;
|
|
if !ctx.rows_to_delete.is_empty() {
|
|
info!("Deleting {} recording rows", ctx.rows_to_delete.len());
|
|
let mut d1 = tx.prepare("delete from recording_playback where composite_id = ?")?;
|
|
let mut d2 = tx.prepare("delete from recording_integrity where composite_id = ?")?;
|
|
let mut d3 = tx.prepare("delete from recording where composite_id = ?")?;
|
|
for &id in &ctx.rows_to_delete {
|
|
d1.execute(params![id.0])?;
|
|
d2.execute(params![id.0])?;
|
|
d3.execute(params![id.0])?;
|
|
}
|
|
}
|
|
if !ctx.files_to_trash.is_empty() {
|
|
info!("Trashing {} recording files", ctx.files_to_trash.len());
|
|
let mut g = tx.prepare(
|
|
"insert or ignore into garbage (sample_file_dir_id, composite_id) values (?, ?)",
|
|
)?;
|
|
for (dir_id, composite_id) in &ctx.files_to_trash {
|
|
g.execute(params![dir_id, composite_id.0])?;
|
|
}
|
|
}
|
|
tx.commit()?;
|
|
}
|
|
|
|
Ok(if printed_error { 1 } else { 0 })
|
|
}
|
|
|
|
#[derive(Debug, Eq, PartialEq)]
|
|
struct RecordingSummary {
|
|
bytes: u64,
|
|
video_samples: i32,
|
|
video_sync_samples: i32,
|
|
media_duration: i32,
|
|
flags: i32,
|
|
}
|
|
|
|
#[derive(Debug, Default)]
|
|
struct Recording {
|
|
/// Present iff there is a file. When `args.compare_lens` is true, the length; otherwise 0.
|
|
file: Option<u64>,
|
|
|
|
/// Iff a `recording` row is present, a `RecordingSummary` from those fields.
|
|
recording_row: Option<RecordingSummary>,
|
|
|
|
/// Iff a `recording_playback` row is present, a `RecordingSummary` computed from the index.
|
|
/// This should match the recording row.
|
|
playback_row: Option<RecordingSummary>,
|
|
|
|
/// True iff a `recording_integrity` row is present.
|
|
integrity_row: bool,
|
|
|
|
/// True iff a `garbage` row is present.
|
|
garbage_row: bool,
|
|
}
|
|
|
|
#[derive(Default)]
|
|
struct Stream {
|
|
recordings: FnvHashMap<i32, Recording>,
|
|
cum_recordings: Option<i32>,
|
|
}
|
|
|
|
type Dir = FnvHashMap<i32, Stream>;
|
|
|
|
fn summarize_index(video_index: &[u8]) -> Result<RecordingSummary, Error> {
|
|
let mut it = recording::SampleIndexIterator::default();
|
|
let mut media_duration = 0;
|
|
let mut video_samples = 0;
|
|
let mut video_sync_samples = 0;
|
|
let mut bytes = 0;
|
|
while it.next(video_index)? {
|
|
bytes += it.bytes as u64;
|
|
media_duration += it.duration_90k;
|
|
video_samples += 1;
|
|
video_sync_samples += it.is_key() as i32;
|
|
}
|
|
Ok(RecordingSummary {
|
|
bytes,
|
|
video_samples,
|
|
video_sync_samples,
|
|
media_duration,
|
|
flags: if it.duration_90k == 0 {
|
|
db::RecordingFlags::TrailingZero as i32
|
|
} else {
|
|
0
|
|
},
|
|
})
|
|
}
|
|
|
|
/// Reads through the given sample file directory.
|
|
/// Logs unexpected files and creates a hash map of the files found there.
|
|
/// If `opts.compare_lens` is set, the values are lengths; otherwise they're insignificant.
|
|
fn read_dir(d: &dir::SampleFileDir, opts: &Options) -> Result<Dir, Error> {
|
|
let mut dir = Dir::default();
|
|
let mut d = d.opendir()?;
|
|
let fd = d.as_raw_fd();
|
|
for e in d.iter() {
|
|
let e = e?;
|
|
let f = e.file_name();
|
|
match f.to_bytes() {
|
|
b"." | b".." | b"meta" => continue,
|
|
_ => {}
|
|
};
|
|
let id = match dir::parse_id(f.to_bytes()) {
|
|
Ok(id) => id,
|
|
Err(_) => {
|
|
error!(
|
|
"sample file directory contains file {:?} which isn't an id",
|
|
f
|
|
);
|
|
continue;
|
|
}
|
|
};
|
|
let len = if opts.compare_lens {
|
|
nix::sys::stat::fstatat(fd, f, AtFlags::empty())?.st_size as u64
|
|
} else {
|
|
0
|
|
};
|
|
let stream = dir.entry(id.stream()).or_insert_with(Stream::default);
|
|
stream
|
|
.recordings
|
|
.entry(id.recording())
|
|
.or_insert_with(Recording::default)
|
|
.file = Some(len);
|
|
}
|
|
Ok(dir)
|
|
}
|
|
|
|
/// Looks through a known stream for errors.
|
|
fn compare_stream(
|
|
conn: &rusqlite::Connection,
|
|
dir_id: i32,
|
|
stream_id: i32,
|
|
opts: &Options,
|
|
mut stream: Stream,
|
|
ctx: &mut Context,
|
|
) -> Result<bool, Error> {
|
|
let start = CompositeId::new(stream_id, 0);
|
|
let end = CompositeId::new(stream_id, i32::max_value());
|
|
let mut printed_error = false;
|
|
let cum_recordings = stream
|
|
.cum_recordings
|
|
.expect("cum_recordings must be set on known stream");
|
|
|
|
// recording row.
|
|
{
|
|
let mut stmt = conn.prepare_cached(
|
|
r#"
|
|
select
|
|
composite_id,
|
|
flags,
|
|
sample_file_bytes,
|
|
wall_duration_90k + media_duration_delta_90k,
|
|
video_samples,
|
|
video_sync_samples
|
|
from
|
|
recording
|
|
where
|
|
composite_id between ? and ?
|
|
"#,
|
|
)?;
|
|
let mut rows = stmt.query(params![start.0, end.0])?;
|
|
while let Some(row) = rows.next()? {
|
|
let id = CompositeId(row.get(0)?);
|
|
let s = RecordingSummary {
|
|
flags: row.get(1)?,
|
|
bytes: row.get::<_, i64>(2)? as u64,
|
|
media_duration: row.get(3)?,
|
|
video_samples: row.get(4)?,
|
|
video_sync_samples: row.get(5)?,
|
|
};
|
|
stream
|
|
.recordings
|
|
.entry(id.recording())
|
|
.or_insert_with(Recording::default)
|
|
.recording_row = Some(s);
|
|
}
|
|
}
|
|
|
|
// recording_playback row.
|
|
{
|
|
let mut stmt = conn.prepare_cached(
|
|
r#"
|
|
select
|
|
composite_id,
|
|
video_index
|
|
from
|
|
recording_playback
|
|
where
|
|
composite_id between ? and ?
|
|
"#,
|
|
)?;
|
|
let mut rows = stmt.query(params![start.0, end.0])?;
|
|
while let Some(row) = rows.next()? {
|
|
let id = CompositeId(row.get(0)?);
|
|
let video_index: Vec<u8> = row.get(1)?;
|
|
let s = match summarize_index(&video_index) {
|
|
Ok(s) => s,
|
|
Err(e) => {
|
|
error!("id {} has bad video_index: {}", id, e);
|
|
printed_error = true;
|
|
if opts.trash_corrupt_rows {
|
|
ctx.rows_to_delete.insert(id);
|
|
ctx.files_to_trash.insert((dir_id, id));
|
|
}
|
|
continue;
|
|
}
|
|
};
|
|
stream
|
|
.recordings
|
|
.entry(id.recording())
|
|
.or_insert_with(Recording::default)
|
|
.playback_row = Some(s);
|
|
}
|
|
}
|
|
|
|
// recording_integrity row.
|
|
{
|
|
let mut stmt = conn.prepare_cached(
|
|
r#"
|
|
select
|
|
composite_id
|
|
from
|
|
recording_integrity
|
|
where
|
|
composite_id between ? and ?
|
|
"#,
|
|
)?;
|
|
let mut rows = stmt.query(params![start.0, end.0])?;
|
|
while let Some(row) = rows.next()? {
|
|
let id = CompositeId(row.get(0)?);
|
|
stream
|
|
.recordings
|
|
.entry(id.recording())
|
|
.or_insert_with(Recording::default)
|
|
.integrity_row = true;
|
|
}
|
|
}
|
|
|
|
for (&id, recording) in &stream.recordings {
|
|
let id = CompositeId::new(stream_id, id);
|
|
|
|
// Files should have recording and playback rows if they aren't marked
|
|
// as garbage (deletion in progress) and aren't newer than
|
|
// cum_recordings (were being written when the process died).
|
|
let db_rows_expected = !recording.garbage_row && id.recording() < cum_recordings;
|
|
|
|
let r = match recording.recording_row {
|
|
Some(ref r) => {
|
|
if !db_rows_expected {
|
|
error!("Unexpected recording row for {}: {:#?}", id, recording);
|
|
printed_error = true;
|
|
continue;
|
|
}
|
|
r
|
|
}
|
|
None => {
|
|
if db_rows_expected {
|
|
error!("Missing recording row for {}: {:#?}", id, recording);
|
|
if opts.trash_orphan_sample_files {
|
|
ctx.files_to_trash.insert((dir_id, id));
|
|
}
|
|
if opts.delete_orphan_rows {
|
|
// also delete playback/integrity rows, if any.
|
|
ctx.rows_to_delete.insert(id);
|
|
}
|
|
printed_error = true;
|
|
} else if recording.playback_row.is_some() {
|
|
error!("Unexpected playback row for {}: {:#?}", id, recording);
|
|
if opts.delete_orphan_rows {
|
|
ctx.rows_to_delete.insert(id);
|
|
}
|
|
printed_error = true;
|
|
}
|
|
continue;
|
|
}
|
|
};
|
|
match recording.playback_row {
|
|
Some(ref p) => {
|
|
if r != p {
|
|
error!(
|
|
"Recording {} summary doesn't match video_index: {:#?}",
|
|
id, recording
|
|
);
|
|
printed_error = true;
|
|
}
|
|
}
|
|
None => {
|
|
error!("Recording {} missing playback row: {:#?}", id, recording);
|
|
printed_error = true;
|
|
if opts.trash_orphan_sample_files {
|
|
ctx.files_to_trash.insert((dir_id, id));
|
|
}
|
|
if opts.delete_orphan_rows {
|
|
// also delete recording/integrity rows, if any.
|
|
ctx.rows_to_delete.insert(id);
|
|
}
|
|
}
|
|
}
|
|
match recording.file {
|
|
Some(len) => {
|
|
if opts.compare_lens && r.bytes != len {
|
|
error!("Recording {} length mismatch: {:#?}", id, recording);
|
|
printed_error = true;
|
|
}
|
|
}
|
|
None => {
|
|
error!("Recording {} missing file: {:#?}", id, recording);
|
|
if opts.delete_orphan_rows {
|
|
ctx.rows_to_delete.insert(id);
|
|
}
|
|
printed_error = true;
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(printed_error)
|
|
}
|