You've already forked AstralRinth
forked from didirus/AstralRinth
* v3_reroute 404 error * hash change * fixed issue with error conversion * added new model confirmation tests + title name change * renaming, fields * owner; test changes * clippy prepare * fmt * merge fixes * clippy * working merge * revs * merge fixes
365 lines
12 KiB
Rust
365 lines
12 KiB
Rust
use super::ApiError;
|
|
use crate::database::redis::RedisPool;
|
|
use crate::models::projects::{Project, Version, VersionType};
|
|
use crate::models::v2::projects::{LegacyProject, LegacyVersion};
|
|
use crate::queue::session::AuthQueue;
|
|
use crate::routes::v3::version_file::HashQuery;
|
|
use crate::routes::{v2_reroute, v3};
|
|
use actix_web::{delete, get, post, web, HttpRequest, HttpResponse};
|
|
use serde::{Deserialize, Serialize};
|
|
use sqlx::PgPool;
|
|
use std::collections::HashMap;
|
|
|
|
pub fn config(cfg: &mut web::ServiceConfig) {
|
|
cfg.service(
|
|
web::scope("version_file")
|
|
.service(delete_file)
|
|
.service(get_version_from_hash)
|
|
.service(download_version)
|
|
.service(get_update_from_hash)
|
|
.service(get_projects_from_hashes),
|
|
);
|
|
|
|
cfg.service(
|
|
web::scope("version_files")
|
|
.service(get_versions_from_hashes)
|
|
.service(update_files)
|
|
.service(update_individual_files),
|
|
);
|
|
}
|
|
|
|
// under /api/v1/version_file/{hash}
|
|
#[get("{version_id}")]
|
|
pub async fn get_version_from_hash(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let response =
|
|
v3::version_file::get_version_from_hash(req, info, pool, redis, hash_query, session_queue)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)?;
|
|
|
|
// Convert response to V2 format
|
|
match v2_reroute::extract_ok_json::<Version>(response).await {
|
|
Ok(version) => {
|
|
let v2_version = LegacyVersion::from(version);
|
|
Ok(HttpResponse::Ok().json(v2_version))
|
|
}
|
|
Err(response) => Ok(response),
|
|
}
|
|
}
|
|
|
|
// under /api/v1/version_file/{hash}/download
|
|
#[get("{version_id}/download")]
|
|
pub async fn download_version(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
v3::version_file::download_version(req, info, pool, redis, hash_query, session_queue)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)
|
|
}
|
|
|
|
// under /api/v1/version_file/{hash}
|
|
#[delete("{version_id}")]
|
|
pub async fn delete_file(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
v3::version_file::delete_file(req, info, pool, redis, hash_query, session_queue)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub struct UpdateData {
|
|
pub loaders: Option<Vec<String>>,
|
|
pub game_versions: Option<Vec<String>>,
|
|
pub version_types: Option<Vec<VersionType>>,
|
|
}
|
|
|
|
#[post("{version_id}/update")]
|
|
pub async fn get_update_from_hash(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
update_data: web::Json<UpdateData>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let update_data = update_data.into_inner();
|
|
let mut loader_fields = HashMap::new();
|
|
let mut game_versions = vec![];
|
|
for gv in update_data.game_versions.into_iter().flatten() {
|
|
game_versions.push(serde_json::json!(gv.clone()));
|
|
}
|
|
if !game_versions.is_empty() {
|
|
loader_fields.insert("game_versions".to_string(), game_versions);
|
|
}
|
|
let update_data = v3::version_file::UpdateData {
|
|
loaders: update_data.loaders.clone(),
|
|
version_types: update_data.version_types.clone(),
|
|
loader_fields: Some(loader_fields),
|
|
};
|
|
|
|
let response = v3::version_file::get_update_from_hash(
|
|
req,
|
|
info,
|
|
pool,
|
|
redis,
|
|
hash_query,
|
|
web::Json(update_data),
|
|
session_queue,
|
|
)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)?;
|
|
|
|
// Convert response to V2 format
|
|
match v2_reroute::extract_ok_json::<Version>(response).await {
|
|
Ok(version) => {
|
|
let v2_version = LegacyVersion::from(version);
|
|
Ok(HttpResponse::Ok().json(v2_version))
|
|
}
|
|
Err(response) => Ok(response),
|
|
}
|
|
}
|
|
|
|
// Requests above with multiple versions below
|
|
#[derive(Deserialize)]
|
|
pub struct FileHashes {
|
|
pub algorithm: Option<String>,
|
|
pub hashes: Vec<String>,
|
|
}
|
|
|
|
// under /api/v2/version_files
|
|
#[post("")]
|
|
pub async fn get_versions_from_hashes(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
file_data: web::Json<FileHashes>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let file_data = file_data.into_inner();
|
|
let file_data = v3::version_file::FileHashes {
|
|
algorithm: file_data.algorithm,
|
|
hashes: file_data.hashes,
|
|
};
|
|
let response = v3::version_file::get_versions_from_hashes(
|
|
req,
|
|
pool,
|
|
redis,
|
|
web::Json(file_data),
|
|
session_queue,
|
|
)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)?;
|
|
|
|
// Convert to V2
|
|
match v2_reroute::extract_ok_json::<HashMap<String, Version>>(response).await {
|
|
Ok(versions) => {
|
|
let v2_versions = versions
|
|
.into_iter()
|
|
.map(|(hash, version)| {
|
|
let v2_version = LegacyVersion::from(version);
|
|
(hash, v2_version)
|
|
})
|
|
.collect::<HashMap<_, _>>();
|
|
Ok(HttpResponse::Ok().json(v2_versions))
|
|
}
|
|
Err(response) => Ok(response),
|
|
}
|
|
}
|
|
|
|
#[post("project")]
|
|
pub async fn get_projects_from_hashes(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
file_data: web::Json<FileHashes>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let file_data = file_data.into_inner();
|
|
let file_data = v3::version_file::FileHashes {
|
|
algorithm: file_data.algorithm,
|
|
hashes: file_data.hashes,
|
|
};
|
|
let response = v3::version_file::get_projects_from_hashes(
|
|
req,
|
|
pool.clone(),
|
|
redis.clone(),
|
|
web::Json(file_data),
|
|
session_queue,
|
|
)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)?;
|
|
|
|
// Convert to V2
|
|
match v2_reroute::extract_ok_json::<HashMap<String, Project>>(response).await {
|
|
Ok(projects_hashes) => {
|
|
let hash_to_project_id = projects_hashes
|
|
.iter()
|
|
.map(|(hash, project)| {
|
|
let project_id = project.id;
|
|
(hash.clone(), project_id)
|
|
})
|
|
.collect::<HashMap<_, _>>();
|
|
let legacy_projects =
|
|
LegacyProject::from_many(projects_hashes.into_values().collect(), &**pool, &redis)
|
|
.await?;
|
|
let legacy_projects_hashes = hash_to_project_id
|
|
.into_iter()
|
|
.filter_map(|(hash, project_id)| {
|
|
let legacy_project =
|
|
legacy_projects.iter().find(|x| x.id == project_id)?.clone();
|
|
Some((hash, legacy_project))
|
|
})
|
|
.collect::<HashMap<_, _>>();
|
|
|
|
Ok(HttpResponse::Ok().json(legacy_projects_hashes))
|
|
}
|
|
Err(response) => Ok(response),
|
|
}
|
|
}
|
|
|
|
#[derive(Deserialize)]
|
|
pub struct ManyUpdateData {
|
|
pub algorithm: Option<String>, // Defaults to calculation based on size of hash
|
|
pub hashes: Vec<String>,
|
|
pub loaders: Option<Vec<String>>,
|
|
pub game_versions: Option<Vec<String>>,
|
|
pub version_types: Option<Vec<VersionType>>,
|
|
}
|
|
|
|
#[post("update")]
|
|
pub async fn update_files(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
update_data: web::Json<ManyUpdateData>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let update_data = update_data.into_inner();
|
|
let mut loader_fields = HashMap::new();
|
|
let mut game_versions = vec![];
|
|
for gv in update_data.game_versions.into_iter().flatten() {
|
|
game_versions.push(serde_json::json!(gv.clone()));
|
|
}
|
|
if !game_versions.is_empty() {
|
|
loader_fields.insert("game_versions".to_string(), game_versions);
|
|
}
|
|
let update_data = v3::version_file::ManyUpdateData {
|
|
loaders: update_data.loaders.clone(),
|
|
version_types: update_data.version_types.clone(),
|
|
loader_fields: Some(loader_fields),
|
|
algorithm: update_data.algorithm,
|
|
hashes: update_data.hashes,
|
|
};
|
|
|
|
let response =
|
|
v3::version_file::update_files(req, pool, redis, web::Json(update_data), session_queue)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)?;
|
|
|
|
// Convert response to V2 format
|
|
match v2_reroute::extract_ok_json::<HashMap<String, Version>>(response).await {
|
|
Ok(returned_versions) => {
|
|
let v3_versions = returned_versions
|
|
.into_iter()
|
|
.map(|(hash, version)| {
|
|
let v2_version = LegacyVersion::from(version);
|
|
(hash, v2_version)
|
|
})
|
|
.collect::<HashMap<_, _>>();
|
|
Ok(HttpResponse::Ok().json(v3_versions))
|
|
}
|
|
Err(response) => Ok(response),
|
|
}
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub struct FileUpdateData {
|
|
pub hash: String,
|
|
pub loaders: Option<Vec<String>>,
|
|
pub game_versions: Option<Vec<String>>,
|
|
pub version_types: Option<Vec<VersionType>>,
|
|
}
|
|
|
|
#[derive(Deserialize)]
|
|
pub struct ManyFileUpdateData {
|
|
pub algorithm: Option<String>, // Defaults to calculation based on size of hash
|
|
pub hashes: Vec<FileUpdateData>,
|
|
}
|
|
|
|
#[post("update_individual")]
|
|
pub async fn update_individual_files(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
update_data: web::Json<ManyFileUpdateData>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let update_data = update_data.into_inner();
|
|
let update_data = v3::version_file::ManyFileUpdateData {
|
|
algorithm: update_data.algorithm,
|
|
hashes: update_data
|
|
.hashes
|
|
.into_iter()
|
|
.map(|x| {
|
|
let mut loader_fields = HashMap::new();
|
|
let mut game_versions = vec![];
|
|
for gv in x.game_versions.into_iter().flatten() {
|
|
game_versions.push(serde_json::json!(gv.clone()));
|
|
}
|
|
if !game_versions.is_empty() {
|
|
loader_fields.insert("game_versions".to_string(), game_versions);
|
|
}
|
|
v3::version_file::FileUpdateData {
|
|
hash: x.hash.clone(),
|
|
loaders: x.loaders.clone(),
|
|
loader_fields: Some(loader_fields),
|
|
version_types: x.version_types,
|
|
}
|
|
})
|
|
.collect(),
|
|
};
|
|
|
|
let response = v3::version_file::update_individual_files(
|
|
req,
|
|
pool,
|
|
redis,
|
|
web::Json(update_data),
|
|
session_queue,
|
|
)
|
|
.await
|
|
.or_else(v2_reroute::flatten_404_error)?;
|
|
|
|
// Convert response to V2 format
|
|
match v2_reroute::extract_ok_json::<HashMap<String, Version>>(response).await {
|
|
Ok(returned_versions) => {
|
|
let v3_versions = returned_versions
|
|
.into_iter()
|
|
.map(|(hash, version)| {
|
|
let v2_version = LegacyVersion::from(version);
|
|
(hash, v2_version)
|
|
})
|
|
.collect::<HashMap<_, _>>();
|
|
Ok(HttpResponse::Ok().json(v3_versions))
|
|
}
|
|
Err(response) => Ok(response),
|
|
}
|
|
}
|