use crate::{ concurrent_processor::ProcessMap, config::Configuration, error::{Error, UploadError}, formats::InputProcessableFormat, future::LocalBoxFuture, repo::{Alias, DeleteToken, FullRepo, Hash, JobId, UploadId}, serde_str::Serde, store::Store, }; use reqwest_middleware::ClientWithMiddleware; use std::{ path::PathBuf, sync::Arc, time::{Duration, Instant}, }; use tracing::Instrument; pub(crate) mod cleanup; mod process; const CLEANUP_QUEUE: &str = "cleanup"; const PROCESS_QUEUE: &str = "process"; #[derive(Debug, serde::Deserialize, serde::Serialize)] enum Cleanup { Hash { hash: Hash, }, Identifier { identifier: String, }, Alias { alias: Serde, token: Serde, }, Variant { hash: Hash, #[serde(skip_serializing_if = "Option::is_none")] variant: Option, }, AllVariants, OutdatedVariants, OutdatedProxies, } #[derive(Debug, serde::Deserialize, serde::Serialize)] enum Process { Ingest { identifier: String, upload_id: Serde, declared_alias: Option>, }, Generate { target_format: InputProcessableFormat, source: Serde, process_path: PathBuf, process_args: Vec, }, } pub(crate) async fn cleanup_alias( repo: &Arc, alias: Alias, token: DeleteToken, ) -> Result<(), Error> { let job = serde_json::to_value(Cleanup::Alias { alias: Serde::new(alias), token: Serde::new(token), }) .map_err(UploadError::PushJob)?; repo.push(CLEANUP_QUEUE, job).await?; Ok(()) } pub(crate) async fn cleanup_hash(repo: &Arc, hash: Hash) -> Result<(), Error> { let job = serde_json::to_value(Cleanup::Hash { hash }).map_err(UploadError::PushJob)?; repo.push(CLEANUP_QUEUE, job).await?; Ok(()) } pub(crate) async fn cleanup_identifier( repo: &Arc, identifier: &Arc, ) -> Result<(), Error> { let job = serde_json::to_value(Cleanup::Identifier { identifier: identifier.to_string(), }) .map_err(UploadError::PushJob)?; repo.push(CLEANUP_QUEUE, job).await?; Ok(()) } async fn cleanup_variants( repo: &Arc, hash: Hash, variant: Option, ) -> Result<(), Error> { let job = serde_json::to_value(Cleanup::Variant { hash, variant }).map_err(UploadError::PushJob)?; repo.push(CLEANUP_QUEUE, job).await?; Ok(()) } pub(crate) async fn cleanup_outdated_proxies(repo: &Arc) -> Result<(), Error> { let job = serde_json::to_value(Cleanup::OutdatedProxies).map_err(UploadError::PushJob)?; repo.push(CLEANUP_QUEUE, job).await?; Ok(()) } pub(crate) async fn cleanup_outdated_variants(repo: &Arc) -> Result<(), Error> { let job = serde_json::to_value(Cleanup::OutdatedVariants).map_err(UploadError::PushJob)?; repo.push(CLEANUP_QUEUE, job).await?; Ok(()) } pub(crate) async fn cleanup_all_variants(repo: &Arc) -> Result<(), Error> { let job = serde_json::to_value(Cleanup::AllVariants).map_err(UploadError::PushJob)?; repo.push(CLEANUP_QUEUE, job).await?; Ok(()) } pub(crate) async fn queue_ingest( repo: &Arc, identifier: &Arc, upload_id: UploadId, declared_alias: Option, ) -> Result<(), Error> { let job = serde_json::to_value(Process::Ingest { identifier: identifier.to_string(), declared_alias: declared_alias.map(Serde::new), upload_id: Serde::new(upload_id), }) .map_err(UploadError::PushJob)?; repo.push(PROCESS_QUEUE, job).await?; Ok(()) } pub(crate) async fn queue_generate( repo: &Arc, target_format: InputProcessableFormat, source: Alias, process_path: PathBuf, process_args: Vec, ) -> Result<(), Error> { let job = serde_json::to_value(Process::Generate { target_format, source: Serde::new(source), process_path, process_args, }) .map_err(UploadError::PushJob)?; repo.push(PROCESS_QUEUE, job).await?; Ok(()) } pub(crate) async fn process_cleanup( repo: Arc, store: S, config: Configuration, ) { process_jobs(&repo, &store, &config, CLEANUP_QUEUE, cleanup::perform).await } pub(crate) async fn process_images( repo: Arc, store: S, client: ClientWithMiddleware, process_map: ProcessMap, config: Configuration, ) { process_image_jobs( &repo, &store, &client, &process_map, &config, PROCESS_QUEUE, process::perform, ) .await } async fn process_jobs( repo: &Arc, store: &S, config: &Configuration, queue: &'static str, callback: F, ) where S: Store, for<'a> F: Fn( &'a Arc, &'a S, &'a Configuration, serde_json::Value, ) -> LocalBoxFuture<'a, Result<(), Error>> + Copy, { let worker_id = uuid::Uuid::new_v4(); loop { let res = job_loop(repo, store, config, worker_id, queue, callback).await; if let Err(e) = res { tracing::warn!("Error processing jobs: {}", format!("{e}")); tracing::warn!("{}", format!("{e:?}")); if e.is_disconnected() { actix_web::rt::time::sleep(Duration::from_secs(10)).await; } continue; } break; } } struct MetricsGuard { worker_id: uuid::Uuid, queue: &'static str, start: Instant, armed: bool, } impl MetricsGuard { fn guard(worker_id: uuid::Uuid, queue: &'static str) -> Self { metrics::increment_counter!("pict-rs.job.start", "queue" => queue, "worker-id" => worker_id.to_string()); Self { worker_id, queue, start: Instant::now(), armed: true, } } fn disarm(mut self) { self.armed = false; } } impl Drop for MetricsGuard { fn drop(&mut self) { metrics::histogram!("pict-rs.job.duration", self.start.elapsed().as_secs_f64(), "queue" => self.queue, "worker-id" => self.worker_id.to_string(), "completed" => (!self.armed).to_string()); metrics::increment_counter!("pict-rs.job.end", "queue" => self.queue, "worker-id" => self.worker_id.to_string(), "completed" => (!self.armed).to_string()); } } async fn job_loop( repo: &Arc, store: &S, config: &Configuration, worker_id: uuid::Uuid, queue: &'static str, callback: F, ) -> Result<(), Error> where S: Store, for<'a> F: Fn( &'a Arc, &'a S, &'a Configuration, serde_json::Value, ) -> LocalBoxFuture<'a, Result<(), Error>> + Copy, { loop { let fut = async { let (job_id, job) = repo.pop(queue, worker_id).await?; let span = tracing::info_span!("Running Job"); let guard = MetricsGuard::guard(worker_id, queue); let res = span .in_scope(|| { heartbeat( repo, queue, worker_id, job_id, (callback)(repo, store, config, job), ) }) .instrument(span) .await; repo.complete_job(queue, worker_id, job_id).await?; res?; guard.disarm(); Ok(()) as Result<(), Error> }; fut.instrument(tracing::info_span!("tick", worker_id = %worker_id)) .await?; } } async fn process_image_jobs( repo: &Arc, store: &S, client: &ClientWithMiddleware, process_map: &ProcessMap, config: &Configuration, queue: &'static str, callback: F, ) where S: Store, for<'a> F: Fn( &'a Arc, &'a S, &'a ClientWithMiddleware, &'a ProcessMap, &'a Configuration, serde_json::Value, ) -> LocalBoxFuture<'a, Result<(), Error>> + Copy, { let worker_id = uuid::Uuid::new_v4(); loop { let res = image_job_loop( repo, store, client, process_map, config, worker_id, queue, callback, ) .await; if let Err(e) = res { tracing::warn!("Error processing jobs: {}", format!("{e}")); tracing::warn!("{}", format!("{e:?}")); if e.is_disconnected() { actix_web::rt::time::sleep(Duration::from_secs(3)).await; } continue; } break; } } #[allow(clippy::too_many_arguments)] async fn image_job_loop( repo: &Arc, store: &S, client: &ClientWithMiddleware, process_map: &ProcessMap, config: &Configuration, worker_id: uuid::Uuid, queue: &'static str, callback: F, ) -> Result<(), Error> where S: Store, for<'a> F: Fn( &'a Arc, &'a S, &'a ClientWithMiddleware, &'a ProcessMap, &'a Configuration, serde_json::Value, ) -> LocalBoxFuture<'a, Result<(), Error>> + Copy, { loop { let fut = async { let (job_id, job) = repo.pop(queue, worker_id).await?; let span = tracing::info_span!("Running Job"); let guard = MetricsGuard::guard(worker_id, queue); let res = span .in_scope(|| { heartbeat( repo, queue, worker_id, job_id, (callback)(repo, store, client, process_map, config, job), ) }) .instrument(span) .await; repo.complete_job(queue, worker_id, job_id).await?; res?; guard.disarm(); Ok(()) as Result<(), Error> }; fut.instrument(tracing::info_span!("tick", worker_id = %worker_id)) .await?; } } async fn heartbeat( repo: &Arc, queue: &'static str, worker_id: uuid::Uuid, job_id: JobId, fut: Fut, ) -> Fut::Output where Fut: std::future::Future, { let mut fut = std::pin::pin!(fut.instrument(tracing::info_span!("job-future", job_id = ?job_id))); let mut interval = actix_web::rt::time::interval(Duration::from_secs(5)); let mut hb = None; loop { tokio::select! { output = &mut fut => { return output; } _ = interval.tick() => { if hb.is_none() { hb = Some(repo.heartbeat(queue, worker_id, job_id)); } } opt = poll_opt(hb.as_mut()), if hb.is_some() => { hb.take(); if let Some(Err(e)) = opt { tracing::warn!("Failed heartbeat\n{}", format!("{e:?}")); } } } } } async fn poll_opt(opt: Option<&mut Fut>) -> Option where Fut: std::future::Future + Unpin, { match opt { None => None, Some(fut) => Some(fut.await), } }