use super::ids::*; use super::DatabaseError; use std::collections::HashMap; pub struct VersionBuilder { pub version_id: VersionId, pub project_id: ProjectId, pub author_id: UserId, pub name: String, pub version_number: String, pub changelog: String, pub files: Vec, pub dependencies: Vec<(VersionId, String)>, pub game_versions: Vec, pub loaders: Vec, pub release_channel: ChannelId, pub featured: bool, } pub struct VersionFileBuilder { pub url: String, pub filename: String, pub hashes: Vec, pub primary: bool, } impl VersionFileBuilder { pub async fn insert( self, version_id: VersionId, transaction: &mut sqlx::Transaction<'_, sqlx::Postgres>, ) -> Result { let file_id = generate_file_id(&mut *transaction).await?; sqlx::query!( " INSERT INTO files (id, version_id, url, filename) VALUES ($1, $2, $3, $4) ", file_id as FileId, version_id as VersionId, self.url, self.filename, ) .execute(&mut *transaction) .await?; for hash in self.hashes { sqlx::query!( " INSERT INTO hashes (file_id, algorithm, hash) VALUES ($1, $2, $3) ", file_id as FileId, hash.algorithm, hash.hash, ) .execute(&mut *transaction) .await?; } Ok(file_id) } } pub struct HashBuilder { pub algorithm: String, pub hash: Vec, } impl VersionBuilder { pub async fn insert( self, transaction: &mut sqlx::Transaction<'_, sqlx::Postgres>, ) -> Result { let version = Version { id: self.version_id, project_id: self.project_id, author_id: self.author_id, name: self.name, version_number: self.version_number, changelog: self.changelog, changelog_url: None, date_published: chrono::Utc::now(), downloads: 0, release_channel: self.release_channel, featured: self.featured, }; version.insert(&mut *transaction).await?; sqlx::query!( " UPDATE mods SET updated = NOW() WHERE id = $1 ", self.project_id as ProjectId, ) .execute(&mut *transaction) .await?; for file in self.files { file.insert(self.version_id, transaction).await?; } for dependency in self.dependencies { sqlx::query!( " INSERT INTO dependencies (dependent_id, dependency_id, dependency_type) VALUES ($1, $2, $3) ", self.version_id as VersionId, dependency.0 as VersionId, dependency.1, ) .execute(&mut *transaction) .await?; } for loader in self.loaders { sqlx::query!( " INSERT INTO loaders_versions (loader_id, version_id) VALUES ($1, $2) ", loader as LoaderId, self.version_id as VersionId, ) .execute(&mut *transaction) .await?; } for game_version in self.game_versions { sqlx::query!( " INSERT INTO game_versions_versions (game_version_id, joining_version_id) VALUES ($1, $2) ", game_version as GameVersionId, self.version_id as VersionId, ) .execute(&mut *transaction) .await?; } Ok(self.version_id) } } pub struct Version { pub id: VersionId, pub project_id: ProjectId, pub author_id: UserId, pub name: String, pub version_number: String, pub changelog: String, pub changelog_url: Option, pub date_published: chrono::DateTime, pub downloads: i32, pub release_channel: ChannelId, pub featured: bool, } impl Version { pub async fn insert( &self, transaction: &mut sqlx::Transaction<'_, sqlx::Postgres>, ) -> Result<(), sqlx::error::Error> { sqlx::query!( " INSERT INTO versions ( id, mod_id, author_id, name, version_number, changelog, changelog_url, date_published, downloads, release_channel, featured ) VALUES ( $1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11 ) ", self.id as VersionId, self.project_id as ProjectId, self.author_id as UserId, &self.name, &self.version_number, self.changelog, self.changelog_url.as_ref(), self.date_published, self.downloads, self.release_channel as ChannelId, self.featured ) .execute(&mut *transaction) .await?; Ok(()) } // TODO: someone verify this pub async fn remove_full<'a, E>(id: VersionId, exec: E) -> Result, sqlx::Error> where E: sqlx::Executor<'a, Database = sqlx::Postgres> + Copy, { let result = sqlx::query!( " SELECT EXISTS(SELECT 1 FROM versions WHERE id = $1) ", id as VersionId, ) .fetch_one(exec) .await?; if !result.exists.unwrap_or(false) { return Ok(None); } sqlx::query!( " DELETE FROM reports WHERE version_id = $1 ", id as VersionId, ) .execute(exec) .await?; sqlx::query!( " DELETE FROM game_versions_versions gvv WHERE gvv.joining_version_id = $1 ", id as VersionId, ) .execute(exec) .await?; sqlx::query!( " DELETE FROM loaders_versions WHERE loaders_versions.version_id = $1 ", id as VersionId, ) .execute(exec) .await?; sqlx::query!( " DELETE FROM downloads WHERE downloads.version_id = $1 ", id as VersionId, ) .execute(exec) .await?; use futures::TryStreamExt; let files = sqlx::query!( " SELECT files.id, files.url, files.filename, files.is_primary FROM files WHERE files.version_id = $1 ", id as VersionId, ) .fetch_many(exec) .try_filter_map(|e| async { Ok(e.right().map(|c| VersionFile { id: FileId(c.id), version_id: id, url: c.url, filename: c.filename, primary: c.is_primary, })) }) .try_collect::>() .await?; for file in files { // TODO: store backblaze id in database so that we can delete the files here // For now, we can't delete the files since we don't have the backblaze id log::warn!( "Can't delete version file id: {} (url: {}, name: {})", file.id.0, file.url, file.filename ) } sqlx::query!( " DELETE FROM hashes WHERE EXISTS( SELECT 1 FROM files WHERE (files.version_id = $1) AND (hashes.file_id = files.id) ) ", id as VersionId ) .execute(exec) .await?; sqlx::query!( " DELETE FROM files WHERE files.version_id = $1 ", id as VersionId, ) .execute(exec) .await?; sqlx::query!( " DELETE FROM versions WHERE id = $1 ", id as VersionId, ) .execute(exec) .await?; sqlx::query!( " DELETE FROM dependencies WHERE dependent_id = $1 ", id as VersionId, ) .execute(exec) .await?; Ok(Some(())) } pub async fn get_dependencies<'a, E>( id: VersionId, exec: E, ) -> Result, sqlx::Error> where E: sqlx::Executor<'a, Database = sqlx::Postgres>, { use futures::stream::TryStreamExt; let vec = sqlx::query!( " SELECT dependency_id id FROM dependencies WHERE dependent_id = $1 ", id as VersionId, ) .fetch_many(exec) .try_filter_map(|e| async { Ok(e.right().map(|v| VersionId(v.id))) }) .try_collect::>() .await?; Ok(vec) } pub async fn get_project_versions<'a, E>( project_id: ProjectId, game_versions: Option>, loaders: Option>, exec: E, ) -> Result, sqlx::Error> where E: sqlx::Executor<'a, Database = sqlx::Postgres>, { use futures::stream::TryStreamExt; let vec = sqlx::query!( " SELECT version.id FROM ( SELECT DISTINCT ON(v.id) v.id, v.date_published FROM versions v INNER JOIN game_versions_versions gvv ON gvv.joining_version_id = v.id INNER JOIN game_versions gv on gvv.game_version_id = gv.id AND (cardinality($2::varchar[]) = 0 OR gv.version = ANY($2::varchar[])) INNER JOIN loaders_versions lv ON lv.version_id = v.id INNER JOIN loaders l on lv.loader_id = l.id AND (cardinality($3::varchar[]) = 0 OR l.loader = ANY($3::varchar[])) WHERE v.mod_id = $1 ) AS version ORDER BY version.date_published ASC ", project_id as ProjectId, &game_versions.unwrap_or_default(), &loaders.unwrap_or_default(), ) .fetch_many(exec) .try_filter_map(|e| async { Ok(e.right().map(|v| VersionId(v.id))) }) .try_collect::>() .await?; Ok(vec) } pub async fn get<'a, 'b, E>( id: VersionId, executor: E, ) -> Result, sqlx::error::Error> where E: sqlx::Executor<'a, Database = sqlx::Postgres>, { let result = sqlx::query!( " SELECT v.mod_id, v.author_id, v.name, v.version_number, v.changelog, v.changelog_url, v.date_published, v.downloads, v.release_channel, v.featured FROM versions v WHERE v.id = $1 ", id as VersionId, ) .fetch_optional(executor) .await?; if let Some(row) = result { Ok(Some(Version { id, project_id: ProjectId(row.mod_id), author_id: UserId(row.author_id), name: row.name, version_number: row.version_number, changelog: row.changelog, changelog_url: row.changelog_url, date_published: row.date_published, downloads: row.downloads, release_channel: ChannelId(row.release_channel), featured: row.featured, })) } else { Ok(None) } } pub async fn get_many<'a, E>( version_ids: Vec, exec: E, ) -> Result, sqlx::Error> where E: sqlx::Executor<'a, Database = sqlx::Postgres> + Copy, { use futures::stream::TryStreamExt; let version_ids_parsed: Vec = version_ids.into_iter().map(|x| x.0).collect(); let versions = sqlx::query!( " SELECT v.id, v.mod_id, v.author_id, v.name, v.version_number, v.changelog, v.changelog_url, v.date_published, v.downloads, v.release_channel, v.featured FROM versions v WHERE v.id IN (SELECT * FROM UNNEST($1::bigint[])) ORDER BY v.date_published ASC ", &version_ids_parsed ) .fetch_many(exec) .try_filter_map(|e| async { Ok(e.right().map(|v| Version { id: VersionId(v.id), project_id: ProjectId(v.mod_id), author_id: UserId(v.author_id), name: v.name, version_number: v.version_number, changelog: v.changelog, changelog_url: v.changelog_url, date_published: v.date_published, downloads: v.downloads, release_channel: ChannelId(v.release_channel), featured: v.featured, })) }) .try_collect::>() .await?; Ok(versions) } pub async fn get_full<'a, 'b, E>( id: VersionId, executor: E, ) -> Result, sqlx::error::Error> where E: sqlx::Executor<'a, Database = sqlx::Postgres>, { let result = sqlx::query!( " SELECT v.id id, v.mod_id mod_id, v.author_id author_id, v.name version_name, v.version_number version_number, v.changelog changelog, v.changelog_url changelog_url, v.date_published date_published, v.downloads downloads, rc.channel release_channel, v.featured featured, STRING_AGG(DISTINCT gv.version, ',') game_versions, STRING_AGG(DISTINCT l.loader, ',') loaders, STRING_AGG(DISTINCT f.id || ', ' || f.filename || ', ' || f.is_primary || ', ' || f.url, ' ,') files, STRING_AGG(DISTINCT h.algorithm || ', ' || encode(h.hash, 'escape') || ', ' || h.file_id, ' ,') hashes, STRING_AGG(DISTINCT d.dependency_id || ', ' || d.dependency_type, ' ,') dependencies FROM versions v INNER JOIN release_channels rc on v.release_channel = rc.id LEFT OUTER JOIN game_versions_versions gvv on v.id = gvv.joining_version_id LEFT OUTER JOIN game_versions gv on gvv.game_version_id = gv.id LEFT OUTER JOIN loaders_versions lv on v.id = lv.version_id LEFT OUTER JOIN loaders l on lv.loader_id = l.id LEFT OUTER JOIN files f on v.id = f.version_id LEFT OUTER JOIN hashes h on f.id = h.file_id LEFT OUTER JOIN dependencies d on v.id = d.dependent_id WHERE v.id = $1 GROUP BY v.id, rc.id; ", id as VersionId, ) .fetch_optional(executor) .await?; if let Some(v) = result { let mut hashes: Vec<(FileId, String, Vec)> = Vec::new(); v.hashes.unwrap_or_default().split(" ,").for_each(|f| { let hash: Vec<&str> = f.split(", ").collect(); if hash.len() >= 3 { hashes.push(( FileId(hash[2].parse().unwrap_or(0)), hash[0].to_string(), hash[1].to_string().into_bytes(), )); } }); let mut files = Vec::new(); v.files.unwrap_or_default().split(" ,").for_each(|f| { let file: Vec<&str> = f.split(", ").collect(); if file.len() >= 4 { let file_id = FileId(file[0].parse().unwrap_or(0)); let mut file_hashes = HashMap::new(); for hash in &hashes { if (hash.0).0 == file_id.0 { file_hashes.insert(hash.1.clone(), hash.2.clone()); } } files.push(QueryFile { id: file_id, url: file[3].to_string(), filename: file[1].to_string(), hashes: file_hashes, primary: file[2].parse().unwrap_or(false), }) } }); let mut dependencies = Vec::new(); v.dependencies .unwrap_or_default() .split(" ,") .for_each(|f| { let dependency: Vec<&str> = f.split(", ").collect(); if dependency.len() >= 2 { dependencies.push(( VersionId(dependency[0].parse().unwrap_or(0)), dependency[1].to_string(), )) } }); Ok(Some(QueryVersion { id: VersionId(v.id), project_id: ProjectId(v.mod_id), author_id: UserId(v.author_id), name: v.version_name, version_number: v.version_number, changelog: v.changelog, changelog_url: v.changelog_url, date_published: v.date_published, downloads: v.downloads, release_channel: v.release_channel, files, game_versions: v .game_versions .unwrap_or_default() .split(',') .map(|x| x.to_string()) .collect(), loaders: v .loaders .unwrap_or_default() .split(',') .map(|x| x.to_string()) .collect(), featured: v.featured, dependencies, })) } else { Ok(None) } } pub async fn get_many_full<'a, E>( version_ids: Vec, exec: E, ) -> Result, sqlx::Error> where E: sqlx::Executor<'a, Database = sqlx::Postgres> + Copy, { use futures::stream::TryStreamExt; let version_ids_parsed: Vec = version_ids.into_iter().map(|x| x.0).collect(); sqlx::query!( " SELECT v.id id, v.mod_id mod_id, v.author_id author_id, v.name version_name, v.version_number version_number, v.changelog changelog, v.changelog_url changelog_url, v.date_published date_published, v.downloads downloads, rc.channel release_channel, v.featured featured, STRING_AGG(DISTINCT gv.version, ',') game_versions, STRING_AGG(DISTINCT l.loader, ',') loaders, STRING_AGG(DISTINCT f.id || ', ' || f.filename || ', ' || f.is_primary || ', ' || f.url, ' ,') files, STRING_AGG(DISTINCT h.algorithm || ', ' || encode(h.hash, 'escape') || ', ' || h.file_id, ' ,') hashes, STRING_AGG(DISTINCT d.dependency_id || ', ' || d.dependency_type, ' ,') dependencies FROM versions v INNER JOIN release_channels rc on v.release_channel = rc.id LEFT OUTER JOIN game_versions_versions gvv on v.id = gvv.joining_version_id LEFT OUTER JOIN game_versions gv on gvv.game_version_id = gv.id LEFT OUTER JOIN loaders_versions lv on v.id = lv.version_id LEFT OUTER JOIN loaders l on lv.loader_id = l.id LEFT OUTER JOIN files f on v.id = f.version_id LEFT OUTER JOIN hashes h on f.id = h.file_id LEFT OUTER JOIN dependencies d on v.id = d.dependent_id WHERE v.id IN (SELECT * FROM UNNEST($1::bigint[])) GROUP BY v.id, rc.id ORDER BY v.date_published ASC; ", &version_ids_parsed ) .fetch_many(exec) .try_filter_map(|e| async { Ok(e.right().map(|v| { let mut hashes : Vec<(FileId, String, Vec)> = Vec::new(); v.hashes.unwrap_or_default().split(" ,").for_each(|f| { let hash: Vec<&str> = f.split(", ").collect(); if hash.len() >= 3 { hashes.push(( FileId(hash[2].parse().unwrap_or(0)), hash[0].to_string(), hash[1].to_string().into_bytes(), )); } }); let mut files = Vec::new(); v.files.unwrap_or_default().split(" ,").for_each(|f| { let file : Vec<&str> = f.split(", ").collect(); if file.len() >= 4 { let file_id = FileId(file[0].parse().unwrap_or(0)); let mut file_hashes = HashMap::new(); for hash in &hashes { if (hash.0).0 == file_id.0 { file_hashes.insert(hash.1.clone(), hash.2.clone()); } } files.push(QueryFile { id: file_id, url: file[3].to_string(), filename: file[1].to_string(), hashes: file_hashes, primary: file[2].parse().unwrap_or(false) }) } }); let mut dependencies = Vec::new(); v.dependencies.unwrap_or_default().split(" ,").for_each(|f| { let dependency: Vec<&str> = f.split(", ").collect(); if dependency.len() >= 2 { dependencies.push((VersionId(dependency[0].parse().unwrap_or(0)), dependency[1].to_string())) } }); QueryVersion { id: VersionId(v.id), project_id: ProjectId(v.mod_id), author_id: UserId(v.author_id), name: v.version_name, version_number: v.version_number, changelog: v.changelog, changelog_url: v.changelog_url, date_published: v.date_published, downloads: v.downloads, release_channel: v.release_channel, files, game_versions: v.game_versions.unwrap_or_default().split(',').map(|x| x.to_string()).collect(), loaders: v.loaders.unwrap_or_default().split(',').map(|x| x.to_string()).collect(), featured: v.featured, dependencies, } })) }) .try_collect::>() .await } } pub struct ReleaseChannel { pub id: ChannelId, pub channel: String, } pub struct VersionFile { pub id: FileId, pub version_id: VersionId, pub url: String, pub filename: String, pub primary: bool, } pub struct FileHash { pub file_id: FileId, pub algorithm: String, pub hash: Vec, } #[derive(Clone)] pub struct QueryVersion { pub id: VersionId, pub project_id: ProjectId, pub author_id: UserId, pub name: String, pub version_number: String, pub changelog: String, pub changelog_url: Option, pub date_published: chrono::DateTime, pub downloads: i32, pub release_channel: String, pub files: Vec, pub game_versions: Vec, pub loaders: Vec, pub featured: bool, pub dependencies: Vec<(VersionId, String)>, } #[derive(Clone)] pub struct QueryFile { pub id: FileId, pub url: String, pub filename: String, pub hashes: HashMap>, pub primary: bool, }