use actix_form_data::{Field, Form, Value}; use actix_web::{ guard, http::header::{CacheControl, CacheDirective, LastModified, ACCEPT_RANGES}, web, App, HttpResponse, HttpResponseBuilder, HttpServer, }; use awc::Client; use dashmap::{mapref::entry::Entry, DashMap}; use futures_util::{ stream::{once, LocalBoxStream}, Stream, }; use once_cell::sync::Lazy; use std::{ collections::HashSet, future::{ready, Future}, path::PathBuf, pin::Pin, task::{Context, Poll}, time::SystemTime, }; use structopt::StructOpt; use tokio::{ io::{AsyncReadExt, AsyncWriteExt}, sync::{ oneshot::{Receiver, Sender}, Semaphore, }, }; use tracing::{debug, error, info, instrument, subscriber::set_global_default, Span}; use tracing_actix_web::TracingLogger; use tracing_bunyan_formatter::{BunyanFormattingLayer, JsonStorageLayer}; use tracing_error::ErrorLayer; use tracing_log::LogTracer; use tracing_subscriber::{layer::SubscriberExt, EnvFilter, Registry}; mod config; mod error; mod exiftool; mod ffmpeg; mod magick; mod middleware; mod migrate; mod processor; mod range; mod root_span_builder; mod stream; mod upload_manager; mod validate; use self::{ config::{Config, Format}, error::{Error, UploadError}, middleware::{Deadline, Internal}, root_span_builder::RootSpanBuilder, upload_manager::{Details, UploadManager, UploadManagerSession}, validate::{image_webp, video_mp4}, }; const MEGABYTES: usize = 1024 * 1024; const MINUTES: u32 = 60; const HOURS: u32 = 60 * MINUTES; const DAYS: u32 = 24 * HOURS; static TMP_DIR: Lazy = Lazy::new(|| { use rand::{ distributions::{Alphanumeric, Distribution}, thread_rng, }; let mut rng = thread_rng(); let tmp_nonce = Alphanumeric .sample_iter(&mut rng) .take(7) .map(char::from) .collect::(); let mut path = std::env::temp_dir(); path.push(format!("pict-rs-{}", tmp_nonce)); path }); static CONFIG: Lazy = Lazy::new(Config::from_args); static PROCESS_SEMAPHORE: Lazy = Lazy::new(|| Semaphore::new(num_cpus::get().saturating_sub(1).max(1))); static PROCESS_MAP: Lazy = Lazy::new(DashMap::new); type OutcomeSender = Sender<(Details, web::Bytes)>; type ProcessMap = DashMap>; struct CancelSafeProcessor { path: PathBuf, receiver: Option>, fut: F, } impl CancelSafeProcessor where F: Future> + Unpin, { pub(crate) fn new(path: PathBuf, fut: F) -> Self { let entry = PROCESS_MAP.entry(path.clone()); let receiver = match entry { Entry::Vacant(vacant) => { vacant.insert(Vec::new()); None } Entry::Occupied(mut occupied) => { let (tx, rx) = tokio::sync::oneshot::channel(); occupied.get_mut().push(tx); Some(rx) } }; CancelSafeProcessor { path, receiver, fut, } } } impl Future for CancelSafeProcessor where F: Future> + Unpin, { type Output = Result<(Details, web::Bytes), Error>; fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll { if let Some(ref mut rx) = self.receiver { Pin::new(rx) .poll(cx) .map(|res| res.map_err(|_| UploadError::Canceled.into())) } else { Pin::new(&mut self.fut).poll(cx).map(|res| { let opt = PROCESS_MAP.remove(&self.path); res.map(|tup| { if let Some((_, vec)) = opt { for sender in vec { let _ = sender.send(tup.clone()); } } tup }) }) } } } impl Drop for CancelSafeProcessor { fn drop(&mut self) { if self.receiver.is_none() { PROCESS_MAP.remove(&self.path); } } } // try moving a file #[instrument(name = "Moving file")] async fn safe_move_file(from: PathBuf, to: PathBuf) -> Result<(), Error> { if let Some(path) = to.parent() { debug!("Creating directory {:?}", path); tokio::fs::create_dir_all(path).await?; } debug!("Checking if {:?} already exists", to); if let Err(e) = tokio::fs::metadata(&to).await { if e.kind() != std::io::ErrorKind::NotFound { return Err(e.into()); } } else { return Err(UploadError::FileExists.into()); } debug!("Moving {:?} to {:?}", from, to); tokio::fs::copy(&from, to).await?; tokio::fs::remove_file(from).await?; Ok(()) } async fn safe_create_parent

(path: P) -> Result<(), Error> where P: AsRef, { if let Some(path) = path.as_ref().parent() { debug!("Creating directory {:?}", path); tokio::fs::create_dir_all(path).await?; } Ok(()) } // Try writing to a file #[instrument(name = "Saving file", skip(bytes))] async fn safe_save_file(path: PathBuf, mut bytes: web::Bytes) -> Result<(), Error> { if let Some(path) = path.parent() { // create the directory for the file debug!("Creating directory {:?}", path); tokio::fs::create_dir_all(path).await?; } // Only write the file if it doesn't already exist debug!("Checking if {:?} already exists", path); if let Err(e) = tokio::fs::metadata(&path).await { if e.kind() != std::io::ErrorKind::NotFound { return Err(e.into()); } } else { return Ok(()); } // Open the file for writing debug!("Creating {:?}", path); let mut file = tokio::fs::File::create(&path).await?; // try writing debug!("Writing to {:?}", path); if let Err(e) = file.write_all_buf(&mut bytes).await { error!("Error writing {:?}, {}", path, e); // remove file if writing failed before completion tokio::fs::remove_file(path).await?; return Err(e.into()); } debug!("{:?} written", path); Ok(()) } pub(crate) fn tmp_file() -> PathBuf { use rand::distributions::{Alphanumeric, Distribution}; let limit: usize = 10; let rng = rand::thread_rng(); let s: String = Alphanumeric .sample_iter(rng) .take(limit) .map(char::from) .collect(); let name = format!("{}.tmp", s); let mut path = TMP_DIR.clone(); path.push(&name); path } fn to_ext(mime: mime::Mime) -> Result<&'static str, Error> { if mime == mime::IMAGE_PNG { Ok(".png") } else if mime == mime::IMAGE_JPEG { Ok(".jpg") } else if mime == video_mp4() { Ok(".mp4") } else if mime == image_webp() { Ok(".webp") } else { Err(UploadError::UnsupportedFormat.into()) } } /// Handle responding to succesful uploads #[instrument(name = "Uploaded files", skip(value, manager))] async fn upload( value: Value, manager: web::Data, ) -> Result { let images = value .map() .and_then(|mut m| m.remove("images")) .and_then(|images| images.array()) .ok_or(UploadError::NoFiles)?; let mut files = Vec::new(); let images = images .into_iter() .filter_map(|i| i.file()) .collect::>(); for image in &images { if let Some(alias) = image.result.alias() { info!("Uploaded {} as {:?}", image.filename, alias); let delete_token = image.result.delete_token().await?; let name = manager.from_alias(alias.to_owned()).await?; let mut path = manager.image_dir(); path.push(name.clone()); let details = manager.variant_details(path.clone(), name.clone()).await?; let details = if let Some(details) = details { debug!("details exist"); details } else { debug!("generating new details from {:?}", path); let new_details = Details::from_path(path.clone()).await?; debug!("storing details for {:?} {}", path, name); manager .store_variant_details(path, name, &new_details) .await?; debug!("stored"); new_details }; files.push(serde_json::json!({ "file": alias, "delete_token": delete_token, "details": details, })); } } for image in images { image.result.succeed(); } Ok(HttpResponse::Created().json(&serde_json::json!({ "msg": "ok", "files": files }))) } #[derive(Debug, serde::Deserialize)] struct UrlQuery { url: String, } /// download an image from a URL #[instrument(name = "Downloading file", skip(client, manager))] async fn download( client: web::Data, manager: web::Data, query: web::Query, ) -> Result { let mut res = client.get(&query.url).send().await?; if !res.status().is_success() { return Err(UploadError::Download(res.status()).into()); } let fut = res.body().limit(CONFIG.max_file_size() * MEGABYTES); let stream = Box::pin(once(fut)); let permit = PROCESS_SEMAPHORE.acquire().await?; let session = manager.session().upload(stream).await?; let alias = session.alias().unwrap().to_owned(); drop(permit); let delete_token = session.delete_token().await?; let name = manager.from_alias(alias.to_owned()).await?; let mut path = manager.image_dir(); path.push(name.clone()); let details = manager.variant_details(path.clone(), name.clone()).await?; let details = if let Some(details) = details { details } else { let new_details = Details::from_path(path.clone()).await?; manager .store_variant_details(path, name, &new_details) .await?; new_details }; session.succeed(); Ok(HttpResponse::Created().json(&serde_json::json!({ "msg": "ok", "files": [{ "file": alias, "delete_token": delete_token, "details": details, }] }))) } /// Delete aliases and files #[instrument(name = "Deleting file", skip(manager))] async fn delete( manager: web::Data, path_entries: web::Path<(String, String)>, ) -> Result { let (alias, token) = path_entries.into_inner(); manager.delete(token, alias).await?; Ok(HttpResponse::NoContent().finish()) } type ProcessQuery = Vec<(String, String)>; async fn prepare_process( query: web::Query, ext: &str, manager: &UploadManager, whitelist: &Option>, ) -> Result<(Format, String, PathBuf, Vec), Error> { let (alias, operations) = query .into_inner() .into_iter() .fold((String::new(), Vec::new()), |(s, mut acc), (k, v)| { if k == "src" { (v, acc) } else { acc.push((k, v)); (s, acc) } }); if alias.is_empty() { return Err(UploadError::MissingFilename.into()); } let name = manager.from_alias(alias).await?; let operations = if let Some(whitelist) = whitelist.as_ref() { operations .into_iter() .filter(|(k, _)| whitelist.contains(&k.to_lowercase())) .collect() } else { operations }; let chain = self::processor::build_chain(&operations); let format = ext .parse::() .map_err(|_| UploadError::UnsupportedFormat)?; let processed_name = format!("{}.{}", name, ext); let base = manager.image_dir(); let thumbnail_path = self::processor::build_path(base, &chain, processed_name); let thumbnail_args = self::processor::build_args(&chain); Ok((format, name, thumbnail_path, thumbnail_args)) } #[instrument(name = "Fetching derived details", skip(manager, whitelist))] async fn process_details( query: web::Query, ext: web::Path, manager: web::Data, whitelist: web::Data>>, ) -> Result { let (_, name, thumbnail_path, _) = prepare_process(query, ext.as_str(), &manager, &whitelist).await?; let details = manager.variant_details(thumbnail_path, name).await?; let details = details.ok_or(UploadError::NoFiles)?; Ok(HttpResponse::Ok().json(&details)) } /// Process files #[instrument(name = "Processing image", skip(manager, whitelist))] async fn process( range: Option, query: web::Query, ext: web::Path, manager: web::Data, whitelist: web::Data>>, ) -> Result { let (format, name, thumbnail_path, thumbnail_args) = prepare_process(query, ext.as_str(), &manager, &whitelist).await?; // If the thumbnail doesn't exist, we need to create it let thumbnail_exists = if let Err(e) = tokio::fs::metadata(&thumbnail_path).await { if e.kind() != std::io::ErrorKind::NotFound { error!("Error looking up processed image, {}", e); return Err(e.into()); } false } else { true }; let details = manager .variant_details(thumbnail_path.clone(), name.clone()) .await?; if !thumbnail_exists || details.is_none() { let mut original_path = manager.image_dir(); original_path.push(name.clone()); let thumbnail_path2 = thumbnail_path.clone(); let process_fut = async { let thumbnail_path = thumbnail_path2; // Create and save a JPG for motion images (gif, mp4) if let Some((updated_path, exists)) = self::processor::prepare_image(original_path.clone()).await? { original_path = updated_path.clone(); if exists.is_new() { // Save the transcoded file in another task debug!("Spawning storage task"); let span = Span::current(); let manager2 = manager.clone(); let name = name.clone(); actix_rt::spawn(async move { let entered = span.enter(); if let Err(e) = manager2.store_variant(updated_path, name).await { error!("Error storing variant, {}", e); return; } drop(entered); }); } } let permit = PROCESS_SEMAPHORE.acquire().await?; let file = tokio::fs::File::open(original_path.clone()).await?; let mut processed_reader = crate::magick::process_image_write_read(file, thumbnail_args, format)?; let mut vec = Vec::new(); processed_reader.read_to_end(&mut vec).await?; let bytes = web::Bytes::from(vec); drop(permit); let details = if let Some(details) = details { details } else { Details::from_bytes(bytes.clone()).await? }; let span = tracing::Span::current(); let details2 = details.clone(); let bytes2 = bytes.clone(); actix_rt::spawn(async move { let entered = span.enter(); if let Err(e) = safe_save_file(thumbnail_path.clone(), bytes2).await { tracing::warn!("Error saving thumbnail: {}", e); return; } if let Err(e) = manager .store_variant_details(thumbnail_path.clone(), name.clone(), &details2) .await { tracing::warn!("Error saving variant details: {}", e); return; } if let Err(e) = manager.store_variant(thumbnail_path, name.clone()).await { tracing::warn!("Error saving variant info: {}", e); } drop(entered); }); Ok((details, bytes)) as Result<(Details, web::Bytes), Error> }; let (details, bytes) = CancelSafeProcessor::new(thumbnail_path.clone(), Box::pin(process_fut)).await?; return match range { Some(range_header) => { if !range_header.is_bytes() { return Err(UploadError::Range.into()); } if range_header.is_empty() { Err(UploadError::Range.into()) } else if range_header.len() == 1 { let range = range_header.ranges().next().unwrap(); let content_range = range.to_content_range(bytes.len() as u64); let stream = range.chop_bytes(bytes); let mut builder = HttpResponse::PartialContent(); builder.insert_header(content_range); Ok(srv_response( builder, stream, details.content_type(), 7 * DAYS, details.system_time(), )) } else { Err(UploadError::Range.into()) } } None => Ok(srv_response( HttpResponse::Ok(), once(ready(Ok(bytes) as Result<_, Error>)), details.content_type(), 7 * DAYS, details.system_time(), )), }; } let details = if let Some(details) = details { details } else { let details = Details::from_path(thumbnail_path.clone()).await?; manager .store_variant_details(thumbnail_path.clone(), name, &details) .await?; details }; ranged_file_resp(thumbnail_path, range, details).await } /// Fetch file details #[instrument(name = "Fetching details", skip(manager))] async fn details( alias: web::Path, manager: web::Data, ) -> Result { let name = manager.from_alias(alias.into_inner()).await?; let mut path = manager.image_dir(); path.push(name.clone()); let details = manager.variant_details(path.clone(), name.clone()).await?; let details = if let Some(details) = details { details } else { let new_details = Details::from_path(path.clone()).await?; manager .store_variant_details(path.clone(), name, &new_details) .await?; new_details }; Ok(HttpResponse::Ok().json(&details)) } /// Serve files #[instrument(name = "Serving file", skip(manager))] async fn serve( range: Option, alias: web::Path, manager: web::Data, ) -> Result { let name = manager.from_alias(alias.into_inner()).await?; let mut path = manager.image_dir(); path.push(name.clone()); let details = manager.variant_details(path.clone(), name.clone()).await?; let details = if let Some(details) = details { details } else { let details = Details::from_path(path.clone()).await?; manager .store_variant_details(path.clone(), name, &details) .await?; details }; ranged_file_resp(path, range, details).await } async fn ranged_file_resp( path: PathBuf, range: Option, details: Details, ) -> Result { let (builder, stream) = match range { //Range header exists - return as ranged Some(range_header) => { if !range_header.is_bytes() { return Err(UploadError::Range.into()); } if range_header.is_empty() { return Err(UploadError::Range.into()); } else if range_header.len() == 1 { let file = tokio::fs::File::open(path).await?; let meta = file.metadata().await?; let range = range_header.ranges().next().unwrap(); let mut builder = HttpResponse::PartialContent(); builder.insert_header(range.to_content_range(meta.len())); (builder, range.chop_file(file).await?) } else { return Err(UploadError::Range.into()); } } //No Range header in the request - return the entire document None => { let file = tokio::fs::File::open(path).await?; let stream = Box::pin(crate::stream::bytes_stream(file)) as LocalBoxStream<'_, _>; (HttpResponse::Ok(), stream) } }; Ok(srv_response( builder, stream, details.content_type(), 7 * DAYS, details.system_time(), )) } // A helper method to produce responses with proper cache headers fn srv_response( mut builder: HttpResponseBuilder, stream: S, ext: mime::Mime, expires: u32, modified: SystemTime, ) -> HttpResponse where S: Stream> + Unpin + 'static, E: std::error::Error + 'static, actix_web::Error: From, { builder .insert_header(LastModified(modified.into())) .insert_header(CacheControl(vec![ CacheDirective::Public, CacheDirective::MaxAge(expires), CacheDirective::Extension("immutable".to_owned(), None), ])) .insert_header((ACCEPT_RANGES, "bytes")) .content_type(ext.to_string()) .streaming(stream) } #[derive(Debug, serde::Deserialize)] #[serde(untagged)] enum FileOrAlias { File { file: String }, Alias { alias: String }, } #[instrument(name = "Purging file", skip(upload_manager))] async fn purge( query: web::Query, upload_manager: web::Data, ) -> Result { let aliases = match query.into_inner() { FileOrAlias::File { file } => upload_manager.aliases_by_filename(file).await?, FileOrAlias::Alias { alias } => upload_manager.aliases_by_alias(alias).await?, }; for alias in aliases.iter() { upload_manager .delete_without_token(alias.to_owned()) .await?; } Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", "aliases": aliases }))) } #[instrument(name = "Fetching aliases", skip(upload_manager))] async fn aliases( query: web::Query, upload_manager: web::Data, ) -> Result { let aliases = match query.into_inner() { FileOrAlias::File { file } => upload_manager.aliases_by_filename(file).await?, FileOrAlias::Alias { alias } => upload_manager.aliases_by_alias(alias).await?, }; Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", "aliases": aliases, }))) } #[derive(Debug, serde::Deserialize)] struct ByAlias { alias: String, } #[instrument(name = "Fetching filename", skip(upload_manager))] async fn filename_by_alias( query: web::Query, upload_manager: web::Data, ) -> Result { let filename = upload_manager.from_alias(query.into_inner().alias).await?; Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", "filename": filename, }))) } #[actix_rt::main] async fn main() -> Result<(), anyhow::Error> { let env_filter = EnvFilter::try_from_default_env().unwrap_or_else(|_| EnvFilter::new("info")); LogTracer::init()?; let subscriber = Registry::default() .with(env_filter) .with(ErrorLayer::default()); if CONFIG.json_logging() { let formatting_layer = BunyanFormattingLayer::new("pict-rs".into(), std::io::stdout); let subscriber = subscriber.with(JsonStorageLayer).with(formatting_layer); set_global_default(subscriber)?; } else { let subscriber = subscriber.with(tracing_subscriber::fmt::layer()); set_global_default(subscriber)?; }; let manager = UploadManager::new(CONFIG.data_dir(), CONFIG.format()).await?; // Create a new Multipart Form validator // // This form is expecting a single array field, 'images' with at most 10 files in it let manager2 = manager.clone(); let form = Form::new() .max_files(10) .max_file_size(CONFIG.max_file_size() * MEGABYTES) .transform_error(|e| Error::from(e).into()) .field( "images", Field::array(Field::file(move |filename, _, stream| { let manager = manager2.clone(); async move { let span = tracing::info_span!("file-upload", ?filename); let entered = span.enter(); let permit = PROCESS_SEMAPHORE.acquire().await?; let res = manager.session().upload(stream).await; drop(permit); drop(entered); res } })), ); // Create a new Multipart Form validator for internal imports // // This form is expecting a single array field, 'images' with at most 10 files in it let validate_imports = CONFIG.validate_imports(); let manager2 = manager.clone(); let import_form = Form::new() .max_files(10) .max_file_size(CONFIG.max_file_size() * MEGABYTES) .transform_error(|e| Error::from(e).into()) .field( "images", Field::array(Field::file(move |filename, content_type, stream| { let manager = manager2.clone(); async move { let span = tracing::info_span!("file-import", ?filename); let entered = span.enter(); let permit = PROCESS_SEMAPHORE.acquire().await?; let res = manager .session() .import(filename, content_type, validate_imports, stream) .await; drop(permit); drop(entered); res } })), ); HttpServer::new(move || { let client = Client::builder() .header("User-Agent", "pict-rs v0.1.0-master") .finish(); App::new() .wrap(TracingLogger::::new()) .wrap(Deadline) .app_data(web::Data::new(manager.clone())) .app_data(web::Data::new(client)) .app_data(web::Data::new(CONFIG.filter_whitelist())) .service( web::scope("/image") .service( web::resource("") .guard(guard::Post()) .wrap(form.clone()) .route(web::post().to(upload)), ) .service(web::resource("/download").route(web::get().to(download))) .service( web::resource("/delete/{delete_token}/{filename}") .route(web::delete().to(delete)) .route(web::get().to(delete)), ) .service(web::resource("/original/{filename}").route(web::get().to(serve))) .service(web::resource("/process.{ext}").route(web::get().to(process))) .service( web::scope("/details") .service( web::resource("/original/{filename}").route(web::get().to(details)), ) .service( web::resource("/process.{ext}") .route(web::get().to(process_details)), ), ), ) .service( web::scope("/internal") .wrap(Internal(CONFIG.api_key().map(|s| s.to_owned()))) .service( web::resource("/import") .wrap(import_form.clone()) .route(web::post().to(upload)), ) .service(web::resource("/purge").route(web::post().to(purge))) .service(web::resource("/aliases").route(web::get().to(aliases))) .service(web::resource("/filename").route(web::get().to(filename_by_alias))), ) }) .bind(CONFIG.bind_address())? .run() .await?; if tokio::fs::metadata(&*TMP_DIR).await.is_ok() { tokio::fs::remove_dir_all(&*TMP_DIR).await?; } Ok(()) }