// This file is part of Moonfire NVR, a security camera network video recorder. // Copyright (C) 2020 The Moonfire NVR Authors; see AUTHORS and LICENSE.txt. // SPDX-License-Identifier: GPL-v3.0-or-later WITH GPL-3.0-linking-exception. //! Subcommand to check the database and sample file dir for errors. use crate::compare; use crate::db::{self, CompositeId, FromSqlUuid}; use crate::dir; use crate::raw; use crate::recording; use crate::schema; use failure::Error; use fnv::{FnvHashMap, FnvHashSet}; use log::{error, info, warn}; use nix::fcntl::AtFlags; use rusqlite::params; use std::os::unix::io::AsRawFd; pub struct Options { pub compare_lens: bool, pub trash_orphan_sample_files: bool, pub delete_orphan_rows: bool, pub trash_corrupt_rows: bool, } #[derive(Default)] pub struct Context { rows_to_delete: FnvHashSet, files_to_trash: FnvHashSet<(i32, CompositeId)>, // (dir_id, composite_id) } pub fn run(conn: &mut rusqlite::Connection, opts: &Options) -> Result { let mut printed_error = false; info!("Checking SQLite database integrity..."); { let mut stmt = conn.prepare("pragma integrity_check")?; let mut rows = stmt.query(params![])?; while let Some(row) = rows.next()? { let e: String = row.get(0)?; if e == "ok" { continue; } error!("{}", e); printed_error = true; } } info!("...done"); // Compare stated schema version. if let Err(e) = db::check_schema_version(conn) { error!("Schema version is not as expected:\n{}", e); printed_error = true; } else { info!("Schema at expected version {}.", db::EXPECTED_VERSION); } // Compare schemas. { let mut expected = rusqlite::Connection::open_in_memory()?; db::init(&mut expected)?; if let Some(diffs) = compare::get_diffs("actual", conn, "expected", &expected)? { error!("Schema is not as expected:\n{}", &diffs); printed_error = true; } else { info!("Schema is as expected."); } } if printed_error { warn!("The following analysis may be incorrect or encounter errors due to schema differences."); } let db_uuid = raw::get_db_uuid(&conn)?; // Scan directories. let mut dirs_by_id: FnvHashMap = FnvHashMap::default(); { let mut dir_stmt = conn.prepare( r#" select d.id, d.path, d.uuid, d.last_complete_open_id, o.uuid from sample_file_dir d left join open o on (d.last_complete_open_id = o.id) "#, )?; let mut garbage_stmt = conn.prepare_cached("select composite_id from garbage where sample_file_dir_id = ?")?; let mut rows = dir_stmt.query(params![])?; while let Some(row) = rows.next()? { let mut meta = schema::DirMeta::default(); let dir_id: i32 = row.get(0)?; let dir_path: String = row.get(1)?; let dir_uuid: FromSqlUuid = row.get(2)?; let open_id = row.get(3)?; let open_uuid: FromSqlUuid = row.get(4)?; meta.db_uuid.extend_from_slice(&db_uuid.as_bytes()[..]); meta.dir_uuid.extend_from_slice(&dir_uuid.0.as_bytes()[..]); { let o = meta.last_complete_open.set_default(); o.id = open_id; o.uuid.extend_from_slice(&open_uuid.0.as_bytes()[..]); } // Open the directory (checking its metadata) and hold it open (for the lock). let dir = dir::SampleFileDir::open(&dir_path, &meta) .map_err(|e| e.context(format!("unable to open dir {}", dir_path)))?; let mut streams = read_dir(&dir, opts)?; let mut rows = garbage_stmt.query(params![dir_id])?; while let Some(row) = rows.next()? { let id = CompositeId(row.get(0)?); let s = streams.entry(id.stream()).or_insert_with(Stream::default); s.recordings .entry(id.recording()) .or_insert_with(Recording::default) .garbage_row = true; } dirs_by_id.insert(dir_id, streams); } } // Scan known streams. let mut ctx = Context::default(); { let mut stmt = conn.prepare( r#" select id, sample_file_dir_id, cum_recordings from stream where sample_file_dir_id is not null "#, )?; let mut rows = stmt.query(params![])?; while let Some(row) = rows.next()? { let stream_id = row.get(0)?; let dir_id = row.get(1)?; let cum_recordings = row.get(2)?; let mut stream = match dirs_by_id.get_mut(&dir_id) { None => Stream::default(), Some(d) => d.remove(&stream_id).unwrap_or_else(Stream::default), }; stream.cum_recordings = Some(cum_recordings); printed_error |= compare_stream(conn, dir_id, stream_id, opts, stream, &mut ctx)?; } } // Expect the rest to have only garbage. for (&dir_id, streams) in &dirs_by_id { for (&stream_id, stream) in streams { for (&recording_id, r) in &stream.recordings { let id = CompositeId::new(stream_id, recording_id); if r.recording_row.is_some() || r.playback_row.is_some() || r.integrity_row || !r.garbage_row { error!( "dir {} recording {} for unknown stream: {:#?}", dir_id, id, r ); printed_error = true; } } } } if !ctx.rows_to_delete.is_empty() || !ctx.files_to_trash.is_empty() { let tx = conn.transaction()?; if !ctx.rows_to_delete.is_empty() { info!("Deleting {} recording rows", ctx.rows_to_delete.len()); let mut d1 = tx.prepare("delete from recording_playback where composite_id = ?")?; let mut d2 = tx.prepare("delete from recording_integrity where composite_id = ?")?; let mut d3 = tx.prepare("delete from recording where composite_id = ?")?; for &id in &ctx.rows_to_delete { d1.execute(params![id.0])?; d2.execute(params![id.0])?; d3.execute(params![id.0])?; } } if !ctx.files_to_trash.is_empty() { info!("Trashing {} recording files", ctx.files_to_trash.len()); let mut g = tx.prepare( "insert or ignore into garbage (sample_file_dir_id, composite_id) values (?, ?)", )?; for (dir_id, composite_id) in &ctx.files_to_trash { g.execute(params![dir_id, composite_id.0])?; } } tx.commit()?; } Ok(if printed_error { 1 } else { 0 }) } #[derive(Debug, Eq, PartialEq)] struct RecordingSummary { bytes: u64, video_samples: i32, video_sync_samples: i32, media_duration: i32, flags: i32, } #[derive(Debug, Default)] struct Recording { /// Present iff there is a file. When `args.compare_lens` is true, the length; otherwise 0. file: Option, /// Iff a `recording` row is present, a `RecordingSummary` from those fields. recording_row: Option, /// Iff a `recording_playback` row is present, a `RecordingSummary` computed from the index. /// This should match the recording row. playback_row: Option, /// True iff a `recording_integrity` row is present. integrity_row: bool, /// True iff a `garbage` row is present. garbage_row: bool, } #[derive(Default)] struct Stream { recordings: FnvHashMap, cum_recordings: Option, } type Dir = FnvHashMap; fn summarize_index(video_index: &[u8]) -> Result { let mut it = recording::SampleIndexIterator::default(); let mut media_duration = 0; let mut video_samples = 0; let mut video_sync_samples = 0; let mut bytes = 0; while it.next(video_index)? { bytes += it.bytes as u64; media_duration += it.duration_90k; video_samples += 1; video_sync_samples += it.is_key() as i32; } Ok(RecordingSummary { bytes, video_samples, video_sync_samples, media_duration, flags: if it.duration_90k == 0 { db::RecordingFlags::TrailingZero as i32 } else { 0 }, }) } /// Reads through the given sample file directory. /// Logs unexpected files and creates a hash map of the files found there. /// If `opts.compare_lens` is set, the values are lengths; otherwise they're insignificant. fn read_dir(d: &dir::SampleFileDir, opts: &Options) -> Result { let mut dir = Dir::default(); let mut d = d.opendir()?; let fd = d.as_raw_fd(); for e in d.iter() { let e = e?; let f = e.file_name(); match f.to_bytes() { b"." | b".." | b"meta" => continue, _ => {} }; let id = match dir::parse_id(f.to_bytes()) { Ok(id) => id, Err(_) => { error!( "sample file directory contains file {:?} which isn't an id", f ); continue; } }; let len = if opts.compare_lens { nix::sys::stat::fstatat(fd, f, AtFlags::empty())?.st_size as u64 } else { 0 }; let stream = dir.entry(id.stream()).or_insert_with(Stream::default); stream .recordings .entry(id.recording()) .or_insert_with(Recording::default) .file = Some(len); } Ok(dir) } /// Looks through a known stream for errors. fn compare_stream( conn: &rusqlite::Connection, dir_id: i32, stream_id: i32, opts: &Options, mut stream: Stream, ctx: &mut Context, ) -> Result { let start = CompositeId::new(stream_id, 0); let end = CompositeId::new(stream_id, i32::max_value()); let mut printed_error = false; let cum_recordings = stream .cum_recordings .expect("cum_recordings must be set on known stream"); // recording row. { let mut stmt = conn.prepare_cached( r#" select composite_id, flags, sample_file_bytes, wall_duration_90k + media_duration_delta_90k, video_samples, video_sync_samples from recording where composite_id between ? and ? "#, )?; let mut rows = stmt.query(params![start.0, end.0])?; while let Some(row) = rows.next()? { let id = CompositeId(row.get(0)?); let s = RecordingSummary { flags: row.get(1)?, bytes: row.get::<_, i64>(2)? as u64, media_duration: row.get(3)?, video_samples: row.get(4)?, video_sync_samples: row.get(5)?, }; stream .recordings .entry(id.recording()) .or_insert_with(Recording::default) .recording_row = Some(s); } } // recording_playback row. { let mut stmt = conn.prepare_cached( r#" select composite_id, video_index from recording_playback where composite_id between ? and ? "#, )?; let mut rows = stmt.query(params![start.0, end.0])?; while let Some(row) = rows.next()? { let id = CompositeId(row.get(0)?); let video_index: Vec = row.get(1)?; let s = match summarize_index(&video_index) { Ok(s) => s, Err(e) => { error!("id {} has bad video_index: {}", id, e); printed_error = true; if opts.trash_corrupt_rows { ctx.rows_to_delete.insert(id); ctx.files_to_trash.insert((dir_id, id)); } continue; } }; stream .recordings .entry(id.recording()) .or_insert_with(Recording::default) .playback_row = Some(s); } } // recording_integrity row. { let mut stmt = conn.prepare_cached( r#" select composite_id from recording_integrity where composite_id between ? and ? "#, )?; let mut rows = stmt.query(params![start.0, end.0])?; while let Some(row) = rows.next()? { let id = CompositeId(row.get(0)?); stream .recordings .entry(id.recording()) .or_insert_with(Recording::default) .integrity_row = true; } } for (&id, recording) in &stream.recordings { let id = CompositeId::new(stream_id, id); // Files should have recording and playback rows if they aren't marked // as garbage (deletion in progress) and aren't newer than // cum_recordings (were being written when the process died). let db_rows_expected = !recording.garbage_row && id.recording() < cum_recordings; let r = match recording.recording_row { Some(ref r) => { if !db_rows_expected { error!("Unexpected recording row for {}: {:#?}", id, recording); printed_error = true; continue; } r } None => { if db_rows_expected { error!("Missing recording row for {}: {:#?}", id, recording); if opts.trash_orphan_sample_files { ctx.files_to_trash.insert((dir_id, id)); } if opts.delete_orphan_rows { // also delete playback/integrity rows, if any. ctx.rows_to_delete.insert(id); } printed_error = true; } else if recording.playback_row.is_some() { error!("Unexpected playback row for {}: {:#?}", id, recording); if opts.delete_orphan_rows { ctx.rows_to_delete.insert(id); } printed_error = true; } continue; } }; match recording.playback_row { Some(ref p) => { if r != p { error!( "Recording {} summary doesn't match video_index: {:#?}", id, recording ); printed_error = true; } } None => { error!("Recording {} missing playback row: {:#?}", id, recording); printed_error = true; if opts.trash_orphan_sample_files { ctx.files_to_trash.insert((dir_id, id)); } if opts.delete_orphan_rows { // also delete recording/integrity rows, if any. ctx.rows_to_delete.insert(id); } } } match recording.file { Some(len) => { if opts.compare_lens && r.bytes != len { error!("Recording {} length mismatch: {:#?}", id, recording); printed_error = true; } } None => { error!("Recording {} missing file: {:#?}", id, recording); if opts.delete_orphan_rows { ctx.rows_to_delete.insert(id); } printed_error = true; } } } Ok(printed_error) }