You've already forked AstralRinth
forked from didirus/AstralRinth
* v3_reroute 404 error * hash change * fixed issue with error conversion * added new model confirmation tests + title name change * renaming, fields * owner; test changes * clippy prepare * fmt * merge fixes * clippy * working merge * revs * merge fixes
708 lines
22 KiB
Rust
708 lines
22 KiB
Rust
use super::ApiError;
|
|
use crate::auth::{
|
|
filter_authorized_projects, filter_authorized_versions, get_user_from_headers,
|
|
is_authorized_version,
|
|
};
|
|
use crate::database::redis::RedisPool;
|
|
use crate::models::ids::VersionId;
|
|
use crate::models::pats::Scopes;
|
|
use crate::models::projects::VersionType;
|
|
use crate::models::teams::ProjectPermissions;
|
|
use crate::queue::session::AuthQueue;
|
|
use crate::{database, models};
|
|
use actix_web::{web, HttpRequest, HttpResponse};
|
|
use itertools::Itertools;
|
|
use serde::{Deserialize, Serialize};
|
|
use sqlx::PgPool;
|
|
use std::collections::HashMap;
|
|
|
|
pub fn config(cfg: &mut web::ServiceConfig) {
|
|
cfg.service(
|
|
web::scope("version_file")
|
|
.route("{version_id}", web::get().to(get_version_from_hash))
|
|
.route("{version_id}/update", web::post().to(get_update_from_hash))
|
|
.route("project", web::post().to(get_projects_from_hashes))
|
|
.route("{version_id}", web::delete().to(delete_file))
|
|
.route("{version_id}/download", web::get().to(download_version)),
|
|
);
|
|
cfg.service(
|
|
web::scope("version_files")
|
|
.route("update", web::post().to(update_files))
|
|
.route("update_individual", web::post().to(update_individual_files))
|
|
.route("", web::post().to(get_versions_from_hashes)),
|
|
);
|
|
}
|
|
|
|
pub async fn get_version_from_hash(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user_option = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::VERSION_READ]),
|
|
)
|
|
.await
|
|
.map(|x| x.1)
|
|
.ok();
|
|
let hash = info.into_inner().0.to_lowercase();
|
|
let algorithm = hash_query
|
|
.algorithm
|
|
.clone()
|
|
.unwrap_or_else(|| default_algorithm_from_hashes(&[hash.clone()]));
|
|
let file = database::models::Version::get_file_from_hash(
|
|
algorithm,
|
|
hash,
|
|
hash_query.version_id.map(|x| x.into()),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
if let Some(file) = file {
|
|
let version = database::models::Version::get(file.version_id, &**pool, &redis).await?;
|
|
if let Some(version) = version {
|
|
if !is_authorized_version(&version.inner, &user_option, &pool).await? {
|
|
return Err(ApiError::NotFound);
|
|
}
|
|
|
|
Ok(HttpResponse::Ok().json(models::projects::Version::from(version)))
|
|
} else {
|
|
Err(ApiError::NotFound)
|
|
}
|
|
} else {
|
|
Err(ApiError::NotFound)
|
|
}
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub struct HashQuery {
|
|
pub algorithm: Option<String>, // Defaults to calculation based on size of hash
|
|
pub version_id: Option<VersionId>,
|
|
}
|
|
|
|
// Calculates whether or not to use sha1 or sha512 based on the size of the hash
|
|
pub fn default_algorithm_from_hashes(hashes: &[String]) -> String {
|
|
// Gets first hash, optionally
|
|
let empty_string = "".into();
|
|
let hash = hashes.first().unwrap_or(&empty_string);
|
|
let hash_len = hash.len();
|
|
// Sha1 = 40 characters
|
|
// Sha512 = 128 characters
|
|
// Favour sha1 as default, unless the hash is longer or equal to 128 characters
|
|
if hash_len >= 128 {
|
|
return "sha512".into();
|
|
}
|
|
"sha1".into()
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub struct UpdateData {
|
|
pub loaders: Option<Vec<String>>,
|
|
pub version_types: Option<Vec<VersionType>>,
|
|
/*
|
|
Loader fields to filter with:
|
|
"game_versions": ["1.16.5", "1.17"]
|
|
|
|
Returns if it matches any of the values
|
|
*/
|
|
pub loader_fields: Option<HashMap<String, Vec<serde_json::Value>>>,
|
|
}
|
|
|
|
pub async fn get_update_from_hash(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
update_data: web::Json<UpdateData>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user_option = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::VERSION_READ]),
|
|
)
|
|
.await
|
|
.map(|x| x.1)
|
|
.ok();
|
|
let hash = info.into_inner().0.to_lowercase();
|
|
|
|
if let Some(file) = database::models::Version::get_file_from_hash(
|
|
hash_query
|
|
.algorithm
|
|
.clone()
|
|
.unwrap_or_else(|| default_algorithm_from_hashes(&[hash.clone()])),
|
|
hash,
|
|
hash_query.version_id.map(|x| x.into()),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?
|
|
{
|
|
if let Some(project) =
|
|
database::models::Project::get_id(file.project_id, &**pool, &redis).await?
|
|
{
|
|
let versions = database::models::Version::get_many(&project.versions, &**pool, &redis)
|
|
.await?
|
|
.into_iter()
|
|
.filter(|x| {
|
|
let mut bool = true;
|
|
if let Some(version_types) = &update_data.version_types {
|
|
bool &= version_types
|
|
.iter()
|
|
.any(|y| y.as_str() == x.inner.version_type);
|
|
}
|
|
if let Some(loaders) = &update_data.loaders {
|
|
bool &= x.loaders.iter().any(|y| loaders.contains(y));
|
|
}
|
|
if let Some(loader_fields) = &update_data.loader_fields {
|
|
for (key, values) in loader_fields {
|
|
bool &= if let Some(x_vf) =
|
|
x.version_fields.iter().find(|y| y.field_name == *key)
|
|
{
|
|
values.iter().any(|v| x_vf.value.contains_json_value(v))
|
|
} else {
|
|
true
|
|
};
|
|
}
|
|
}
|
|
bool
|
|
})
|
|
.sorted();
|
|
|
|
if let Some(first) = versions.last() {
|
|
if !is_authorized_version(&first.inner, &user_option, &pool).await? {
|
|
return Err(ApiError::NotFound);
|
|
}
|
|
|
|
return Ok(HttpResponse::Ok().json(models::projects::Version::from(first)));
|
|
}
|
|
}
|
|
}
|
|
|
|
Err(ApiError::NotFound)
|
|
}
|
|
|
|
// Requests above with multiple versions below
|
|
#[derive(Deserialize)]
|
|
pub struct FileHashes {
|
|
pub algorithm: Option<String>, // Defaults to calculation based on size of hash
|
|
pub hashes: Vec<String>,
|
|
}
|
|
|
|
pub async fn get_versions_from_hashes(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
file_data: web::Json<FileHashes>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user_option = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::VERSION_READ]),
|
|
)
|
|
.await
|
|
.map(|x| x.1)
|
|
.ok();
|
|
|
|
let algorithm = file_data
|
|
.algorithm
|
|
.clone()
|
|
.unwrap_or_else(|| default_algorithm_from_hashes(&file_data.hashes));
|
|
|
|
let files = database::models::Version::get_files_from_hash(
|
|
algorithm.clone(),
|
|
&file_data.hashes,
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
let version_ids = files.iter().map(|x| x.version_id).collect::<Vec<_>>();
|
|
let versions_data = filter_authorized_versions(
|
|
database::models::Version::get_many(&version_ids, &**pool, &redis).await?,
|
|
&user_option,
|
|
&pool,
|
|
)
|
|
.await?;
|
|
|
|
let mut response = HashMap::new();
|
|
|
|
for version in versions_data {
|
|
for file in files.iter().filter(|x| x.version_id == version.id.into()) {
|
|
if let Some(hash) = file.hashes.get(&algorithm) {
|
|
response.insert(hash.clone(), version.clone());
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(HttpResponse::Ok().json(response))
|
|
}
|
|
|
|
pub async fn get_projects_from_hashes(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
file_data: web::Json<FileHashes>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user_option = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::PROJECT_READ, Scopes::VERSION_READ]),
|
|
)
|
|
.await
|
|
.map(|x| x.1)
|
|
.ok();
|
|
|
|
let algorithm = file_data
|
|
.algorithm
|
|
.clone()
|
|
.unwrap_or_else(|| default_algorithm_from_hashes(&file_data.hashes));
|
|
let files = database::models::Version::get_files_from_hash(
|
|
algorithm.clone(),
|
|
&file_data.hashes,
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
let project_ids = files.iter().map(|x| x.project_id).collect::<Vec<_>>();
|
|
|
|
let projects_data = filter_authorized_projects(
|
|
database::models::Project::get_many_ids(&project_ids, &**pool, &redis).await?,
|
|
&user_option,
|
|
&pool,
|
|
)
|
|
.await?;
|
|
|
|
let mut response = HashMap::new();
|
|
|
|
for project in projects_data {
|
|
for file in files.iter().filter(|x| x.project_id == project.id.into()) {
|
|
if let Some(hash) = file.hashes.get(&algorithm) {
|
|
response.insert(hash.clone(), project.clone());
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(HttpResponse::Ok().json(response))
|
|
}
|
|
|
|
#[derive(Deserialize)]
|
|
pub struct ManyUpdateData {
|
|
pub algorithm: Option<String>, // Defaults to calculation based on size of hash
|
|
pub hashes: Vec<String>,
|
|
pub loaders: Option<Vec<String>>,
|
|
pub loader_fields: Option<HashMap<String, Vec<serde_json::Value>>>,
|
|
pub version_types: Option<Vec<VersionType>>,
|
|
}
|
|
pub async fn update_files(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
update_data: web::Json<ManyUpdateData>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user_option = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::VERSION_READ]),
|
|
)
|
|
.await
|
|
.map(|x| x.1)
|
|
.ok();
|
|
|
|
let algorithm = update_data
|
|
.algorithm
|
|
.clone()
|
|
.unwrap_or_else(|| default_algorithm_from_hashes(&update_data.hashes));
|
|
let files = database::models::Version::get_files_from_hash(
|
|
algorithm.clone(),
|
|
&update_data.hashes,
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
let projects = database::models::Project::get_many_ids(
|
|
&files.iter().map(|x| x.project_id).collect::<Vec<_>>(),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
let all_versions = database::models::Version::get_many(
|
|
&projects
|
|
.iter()
|
|
.flat_map(|x| x.versions.clone())
|
|
.collect::<Vec<_>>(),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
let mut response = HashMap::new();
|
|
|
|
for project in projects {
|
|
for file in files.iter().filter(|x| x.project_id == project.inner.id) {
|
|
let version = all_versions
|
|
.iter()
|
|
.filter(|x| x.inner.project_id == file.project_id)
|
|
.filter(|x| {
|
|
// TODO: Behaviour here is repeated in a few other filtering places, should be abstracted
|
|
let mut bool = true;
|
|
|
|
if let Some(version_types) = &update_data.version_types {
|
|
bool &= version_types
|
|
.iter()
|
|
.any(|y| y.as_str() == x.inner.version_type);
|
|
}
|
|
if let Some(loaders) = &update_data.loaders {
|
|
bool &= x.loaders.iter().any(|y| loaders.contains(y));
|
|
}
|
|
if let Some(loader_fields) = &update_data.loader_fields {
|
|
for (key, values) in loader_fields {
|
|
bool &= if let Some(x_vf) =
|
|
x.version_fields.iter().find(|y| y.field_name == *key)
|
|
{
|
|
values.iter().any(|v| x_vf.value.contains_json_value(v))
|
|
} else {
|
|
true
|
|
};
|
|
}
|
|
}
|
|
|
|
bool
|
|
})
|
|
.sorted()
|
|
.last();
|
|
|
|
if let Some(version) = version {
|
|
if is_authorized_version(&version.inner, &user_option, &pool).await? {
|
|
if let Some(hash) = file.hashes.get(&algorithm) {
|
|
response.insert(
|
|
hash.clone(),
|
|
models::projects::Version::from(version.clone()),
|
|
);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(HttpResponse::Ok().json(response))
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub struct FileUpdateData {
|
|
pub hash: String,
|
|
pub loaders: Option<Vec<String>>,
|
|
pub loader_fields: Option<HashMap<String, Vec<serde_json::Value>>>,
|
|
pub version_types: Option<Vec<VersionType>>,
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub struct ManyFileUpdateData {
|
|
pub algorithm: Option<String>, // Defaults to calculation based on size of hash
|
|
pub hashes: Vec<FileUpdateData>,
|
|
}
|
|
|
|
pub async fn update_individual_files(
|
|
req: HttpRequest,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
update_data: web::Json<ManyFileUpdateData>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user_option = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::VERSION_READ]),
|
|
)
|
|
.await
|
|
.map(|x| x.1)
|
|
.ok();
|
|
|
|
let algorithm = update_data.algorithm.clone().unwrap_or_else(|| {
|
|
default_algorithm_from_hashes(
|
|
&update_data
|
|
.hashes
|
|
.iter()
|
|
.map(|x| x.hash.clone())
|
|
.collect::<Vec<_>>(),
|
|
)
|
|
});
|
|
let files = database::models::Version::get_files_from_hash(
|
|
algorithm.clone(),
|
|
&update_data
|
|
.hashes
|
|
.iter()
|
|
.map(|x| x.hash.clone())
|
|
.collect::<Vec<_>>(),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
let projects = database::models::Project::get_many_ids(
|
|
&files.iter().map(|x| x.project_id).collect::<Vec<_>>(),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
let all_versions = database::models::Version::get_many(
|
|
&projects
|
|
.iter()
|
|
.flat_map(|x| x.versions.clone())
|
|
.collect::<Vec<_>>(),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
let mut response = HashMap::new();
|
|
|
|
for project in projects {
|
|
for file in files.iter().filter(|x| x.project_id == project.inner.id) {
|
|
if let Some(hash) = file.hashes.get(&algorithm) {
|
|
if let Some(query_file) = update_data.hashes.iter().find(|x| &x.hash == hash) {
|
|
let version = all_versions
|
|
.iter()
|
|
.filter(|x| x.inner.project_id == file.project_id)
|
|
.filter(|x| {
|
|
let mut bool = true;
|
|
|
|
if let Some(version_types) = &query_file.version_types {
|
|
bool &= version_types
|
|
.iter()
|
|
.any(|y| y.as_str() == x.inner.version_type);
|
|
}
|
|
if let Some(loaders) = &query_file.loaders {
|
|
bool &= x.loaders.iter().any(|y| loaders.contains(y));
|
|
}
|
|
|
|
if let Some(loader_fields) = &query_file.loader_fields {
|
|
for (key, values) in loader_fields {
|
|
bool &= if let Some(x_vf) =
|
|
x.version_fields.iter().find(|y| y.field_name == *key)
|
|
{
|
|
values.iter().any(|v| x_vf.value.contains_json_value(v))
|
|
} else {
|
|
true
|
|
};
|
|
}
|
|
}
|
|
bool
|
|
})
|
|
.sorted()
|
|
.last();
|
|
|
|
if let Some(version) = version {
|
|
if is_authorized_version(&version.inner, &user_option, &pool).await? {
|
|
response.insert(
|
|
hash.clone(),
|
|
models::projects::Version::from(version.clone()),
|
|
);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(HttpResponse::Ok().json(response))
|
|
}
|
|
|
|
// under /api/v1/version_file/{hash}
|
|
pub async fn delete_file(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::VERSION_WRITE]),
|
|
)
|
|
.await?
|
|
.1;
|
|
|
|
let hash = info.into_inner().0.to_lowercase();
|
|
let algorithm = hash_query
|
|
.algorithm
|
|
.clone()
|
|
.unwrap_or_else(|| default_algorithm_from_hashes(&[hash.clone()]));
|
|
let file = database::models::Version::get_file_from_hash(
|
|
algorithm.clone(),
|
|
hash,
|
|
hash_query.version_id.map(|x| x.into()),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
if let Some(row) = file {
|
|
if !user.role.is_admin() {
|
|
let team_member = database::models::TeamMember::get_from_user_id_version(
|
|
row.version_id,
|
|
user.id.into(),
|
|
&**pool,
|
|
)
|
|
.await
|
|
.map_err(ApiError::Database)?;
|
|
|
|
let organization =
|
|
database::models::Organization::get_associated_organization_project_id(
|
|
row.project_id,
|
|
&**pool,
|
|
)
|
|
.await
|
|
.map_err(ApiError::Database)?;
|
|
|
|
let organization_team_member = if let Some(organization) = &organization {
|
|
database::models::TeamMember::get_from_user_id_organization(
|
|
organization.id,
|
|
user.id.into(),
|
|
&**pool,
|
|
)
|
|
.await
|
|
.map_err(ApiError::Database)?
|
|
} else {
|
|
None
|
|
};
|
|
|
|
let permissions = ProjectPermissions::get_permissions_by_role(
|
|
&user.role,
|
|
&team_member,
|
|
&organization_team_member,
|
|
)
|
|
.unwrap_or_default();
|
|
|
|
if !permissions.contains(ProjectPermissions::DELETE_VERSION) {
|
|
return Err(ApiError::CustomAuthentication(
|
|
"You don't have permission to delete this file!".to_string(),
|
|
));
|
|
}
|
|
}
|
|
|
|
let version = database::models::Version::get(row.version_id, &**pool, &redis).await?;
|
|
if let Some(version) = version {
|
|
if version.files.len() < 2 {
|
|
return Err(ApiError::InvalidInput(
|
|
"Versions must have at least one file uploaded to them".to_string(),
|
|
));
|
|
}
|
|
|
|
database::models::Version::clear_cache(&version, &redis).await?;
|
|
}
|
|
|
|
let mut transaction = pool.begin().await?;
|
|
|
|
sqlx::query!(
|
|
"
|
|
DELETE FROM hashes
|
|
WHERE file_id = $1
|
|
",
|
|
row.id.0
|
|
)
|
|
.execute(&mut *transaction)
|
|
.await?;
|
|
|
|
sqlx::query!(
|
|
"
|
|
DELETE FROM files
|
|
WHERE files.id = $1
|
|
",
|
|
row.id.0,
|
|
)
|
|
.execute(&mut *transaction)
|
|
.await?;
|
|
|
|
transaction.commit().await?;
|
|
|
|
Ok(HttpResponse::NoContent().body(""))
|
|
} else {
|
|
Err(ApiError::NotFound)
|
|
}
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub struct DownloadRedirect {
|
|
pub url: String,
|
|
}
|
|
|
|
// under /api/v1/version_file/{hash}/download
|
|
pub async fn download_version(
|
|
req: HttpRequest,
|
|
info: web::Path<(String,)>,
|
|
pool: web::Data<PgPool>,
|
|
redis: web::Data<RedisPool>,
|
|
hash_query: web::Query<HashQuery>,
|
|
session_queue: web::Data<AuthQueue>,
|
|
) -> Result<HttpResponse, ApiError> {
|
|
let user_option = get_user_from_headers(
|
|
&req,
|
|
&**pool,
|
|
&redis,
|
|
&session_queue,
|
|
Some(&[Scopes::VERSION_READ]),
|
|
)
|
|
.await
|
|
.map(|x| x.1)
|
|
.ok();
|
|
|
|
let hash = info.into_inner().0.to_lowercase();
|
|
let algorithm = hash_query
|
|
.algorithm
|
|
.clone()
|
|
.unwrap_or_else(|| default_algorithm_from_hashes(&[hash.clone()]));
|
|
let file = database::models::Version::get_file_from_hash(
|
|
algorithm.clone(),
|
|
hash,
|
|
hash_query.version_id.map(|x| x.into()),
|
|
&**pool,
|
|
&redis,
|
|
)
|
|
.await?;
|
|
|
|
if let Some(file) = file {
|
|
let version = database::models::Version::get(file.version_id, &**pool, &redis).await?;
|
|
|
|
if let Some(version) = version {
|
|
if !is_authorized_version(&version.inner, &user_option, &pool).await? {
|
|
return Err(ApiError::NotFound);
|
|
}
|
|
|
|
Ok(HttpResponse::TemporaryRedirect()
|
|
.append_header(("Location", &*file.url))
|
|
.json(DownloadRedirect { url: file.url }))
|
|
} else {
|
|
Err(ApiError::NotFound)
|
|
}
|
|
} else {
|
|
Err(ApiError::NotFound)
|
|
}
|
|
}
|