You've already forked AstralRinth
forked from didirus/AstralRinth
* chore: fix typo in status message * feat(labrinth): overhaul malware scanner report storage and routes * chore: address some review comments * feat: add Delphi to Docker Compose `with-delphi` profile * chore: fix unused import Clippy lint * feat(labrinth/delphi): use PAT token authorization with project read scopes * chore: expose file IDs in version queries * fix: accept null decompiled source payloads from Delphi * tweak(labrinth): expose base62 file IDs more consistently for Delphi * feat(labrinth/delphi): support new Delphi report severity field * chore(labrinth): run `cargo sqlx prepare` to fix Docker build errors * tweak: add route for fetching Delphi issue type schema, abstract Labrinth away from issue types * chore: run `cargo sqlx prepare` * chore: fix typo on frontend generated state file message * feat: update to use new Delphi issue schema * wip: tech review endpoints * wip: add ToSchema for dependent types * wip: report issues return * wip * wip: returning more data * wip * Fix up db query * Delphi configuration to talk to Labrinth * Get Delphi working with Labrinth * Add Delphi dummy fixture * Better Delphi logging * Improve utoipa for tech review routes * Add more sorting options for tech review queue * Oops join * New routes for fetching issues and reports * Fix which kind of ID is returned in tech review endpoints * Deduplicate tech review report rows * Reduce info sent for projects * Fetch more thread info * Address PR comments * fix ci * fix postgres version mismatch * fix version creation * Implement routes * fix up tech review * Allow adding a moderation comment to Delphi rejections * fix up rebase * exclude rejected projects from tech review * add status change msg to tech review thread * cargo sqlx prepare * also ignore withheld projects * More filtering on issue search * wip: report routes * Fix up for build * cargo sqlx prepare * fix thread message privacy * New tech review search route * submit route * details have statuses now * add default to drid status * dedup issue details * fix sqlx query on empty files * fixes * Dedupe issue detail statuses and message on entering tech rev * Fix qa issues * Fix qa issues * fix review comments * typos * fix ci * feat: tech review frontend (#4781) * chore: fix typo in status message * feat(labrinth): overhaul malware scanner report storage and routes * chore: address some review comments * feat: add Delphi to Docker Compose `with-delphi` profile * chore: fix unused import Clippy lint * feat(labrinth/delphi): use PAT token authorization with project read scopes * chore: expose file IDs in version queries * fix: accept null decompiled source payloads from Delphi * tweak(labrinth): expose base62 file IDs more consistently for Delphi * feat(labrinth/delphi): support new Delphi report severity field * chore(labrinth): run `cargo sqlx prepare` to fix Docker build errors * tweak: add route for fetching Delphi issue type schema, abstract Labrinth away from issue types * chore: run `cargo sqlx prepare` * chore: fix typo on frontend generated state file message * feat: update to use new Delphi issue schema * wip: tech review endpoints * wip: add ToSchema for dependent types * wip: report issues return * wip * wip: returning more data * wip * Fix up db query * Delphi configuration to talk to Labrinth * Get Delphi working with Labrinth * Add Delphi dummy fixture * Better Delphi logging * Improve utoipa for tech review routes * Add more sorting options for tech review queue * Oops join * New routes for fetching issues and reports * Fix which kind of ID is returned in tech review endpoints * Deduplicate tech review report rows * Reduce info sent for projects * Fetch more thread info * Address PR comments * fix ci * fix ci * fix postgres version mismatch * fix version creation * Implement routes * feat: batch scan alert * feat: layout * feat: introduce surface variables * fix: theme selector * feat: rough draft of tech review card * feat: tab switcher * feat: batch scan btn * feat: api-client module for tech review * draft: impl * feat: auto icons * fix: layout issues * feat: fixes to code blocks + flag labels * feat: temp remove mock data * fix: search sort types * fix: intl & lint * chore: re-enable mock data * fix: flag badges + auto open first issue in file tab * feat: update for new routes * fix: more qa issues * feat: lazy load sources * fix: re-enable auth middleware * feat: impl threads * fix: lint & severity * feat: download btn + switch to using NavTabs with new local mode option * feat: re-add toplevel btns * feat: reports page consistency * fix: consistency on project queue * fix: icons + sizing * fix: colors and gaps * fix: impl endpoints * feat: load all flags on file tab * feat: thread generics changes * feat: more qa * feat: fix collapse * fix: qa * feat: msg modal * fix: ISO import * feat: qa fixes * fix: empty state basic * fix: collapsible region * fix: collapse thread by default * feat: rough draft of new process/flow * fix labrinth build * fix thread message privacy * New tech review search route * feat: qa fixes * feat: QA changes * fix: verdict on detail not whole issue * fix: lint + intl * fix: lint * fix: thread message for tech rev verdict * feat: use anim frames * fix: exports + typecheck * polish: qa changes * feat: qa * feat: qa polish * feat: fix malic modal * fix: lint * fix: qa + lint * fix: pagination * fix: lint * fix: qa * intl extract * fix ci --------- Signed-off-by: Calum H. <contact@cal.engineer> Co-authored-by: Alejandro González <me@alegon.dev> Co-authored-by: aecsocket <aecsocket@tutanota.com> --------- Signed-off-by: Calum H. <contact@cal.engineer> Co-authored-by: Alejandro González <me@alegon.dev> Co-authored-by: Calum H. <contact@cal.engineer>
294 lines
9.8 KiB
Rust
294 lines
9.8 KiB
Rust
use actix_web::dev::Service;
|
|
use actix_web::middleware::from_fn;
|
|
use actix_web::{App, HttpServer};
|
|
use actix_web_prom::PrometheusMetricsBuilder;
|
|
use clap::Parser;
|
|
|
|
use labrinth::app_config;
|
|
use labrinth::background_task::BackgroundTask;
|
|
use labrinth::database::redis::RedisPool;
|
|
use labrinth::file_hosting::{S3BucketConfig, S3Host};
|
|
use labrinth::queue::email::EmailQueue;
|
|
use labrinth::search;
|
|
use labrinth::util::anrok;
|
|
use labrinth::util::env::parse_var;
|
|
use labrinth::util::gotenberg::GotenbergClient;
|
|
use labrinth::util::ratelimit::rate_limit_middleware;
|
|
use labrinth::utoipa_app_config;
|
|
use labrinth::{check_env_vars, clickhouse, database, file_hosting};
|
|
use std::ffi::CStr;
|
|
use std::sync::Arc;
|
|
use tracing::{Instrument, error, info, info_span};
|
|
use tracing_actix_web::TracingLogger;
|
|
use utoipa::OpenApi;
|
|
use utoipa::openapi::security::{ApiKey, ApiKeyValue, SecurityScheme};
|
|
use utoipa_actix_web::AppExt;
|
|
use utoipa_swagger_ui::SwaggerUi;
|
|
|
|
#[cfg(target_os = "linux")]
|
|
#[global_allocator]
|
|
static ALLOC: tikv_jemallocator::Jemalloc = tikv_jemallocator::Jemalloc;
|
|
|
|
#[unsafe(export_name = "malloc_conf")]
|
|
pub static MALLOC_CONF: &CStr = c"prof:true,prof_active:true,lg_prof_sample:19";
|
|
|
|
#[derive(Clone)]
|
|
pub struct Pepper {
|
|
pub pepper: String,
|
|
}
|
|
|
|
#[derive(Parser)]
|
|
#[command(version)]
|
|
struct Args {
|
|
/// Don't run regularly scheduled background tasks. This means the tasks should be run
|
|
/// manually with --run-background-task.
|
|
#[arg(long)]
|
|
no_background_tasks: bool,
|
|
|
|
/// Don't automatically run migrations. This means the migrations should be run via --run-background-task.
|
|
#[arg(long)]
|
|
no_migrations: bool,
|
|
|
|
/// Run a single background task and then exit. Perfect for cron jobs.
|
|
#[arg(long, value_enum, id = "task")]
|
|
run_background_task: Option<BackgroundTask>,
|
|
}
|
|
|
|
#[actix_rt::main]
|
|
async fn main() -> std::io::Result<()> {
|
|
let args = Args::parse();
|
|
|
|
color_eyre::install().expect("failed to install `color-eyre`");
|
|
dotenvy::dotenv().ok();
|
|
modrinth_util::log::init().expect("failed to initialize logging");
|
|
|
|
if check_env_vars() {
|
|
error!("Some environment variables are missing!");
|
|
std::process::exit(1);
|
|
}
|
|
|
|
rustls::crypto::aws_lc_rs::default_provider()
|
|
.install_default()
|
|
.unwrap();
|
|
|
|
// DSN is from SENTRY_DSN env variable.
|
|
// Has no effect if not set.
|
|
let sentry = sentry::init(sentry::ClientOptions {
|
|
release: sentry::release_name!(),
|
|
traces_sample_rate: 0.1,
|
|
..Default::default()
|
|
});
|
|
if sentry.is_enabled() {
|
|
info!("Enabled Sentry integration");
|
|
unsafe {
|
|
std::env::set_var("RUST_BACKTRACE", "1");
|
|
}
|
|
}
|
|
|
|
if args.run_background_task.is_none() {
|
|
info!(
|
|
"Starting labrinth on {}",
|
|
dotenvy::var("BIND_ADDR").unwrap()
|
|
);
|
|
|
|
if !args.no_migrations {
|
|
database::check_for_migrations()
|
|
.await
|
|
.expect("An error occurred while running migrations.");
|
|
}
|
|
}
|
|
|
|
// Database Connector
|
|
let (pool, ro_pool) = database::connect_all()
|
|
.await
|
|
.expect("Database connection failed");
|
|
|
|
// Redis connector
|
|
let redis_pool = RedisPool::new(None);
|
|
|
|
let storage_backend =
|
|
dotenvy::var("STORAGE_BACKEND").unwrap_or_else(|_| "local".to_string());
|
|
|
|
let file_host: Arc<dyn file_hosting::FileHost + Send + Sync> =
|
|
match storage_backend.as_str() {
|
|
"s3" => {
|
|
let config_from_env = |bucket_type| S3BucketConfig {
|
|
name: parse_var(&format!("S3_{bucket_type}_BUCKET_NAME"))
|
|
.unwrap(),
|
|
uses_path_style: parse_var(&format!(
|
|
"S3_{bucket_type}_USES_PATH_STYLE_BUCKET"
|
|
))
|
|
.unwrap(),
|
|
region: parse_var(&format!("S3_{bucket_type}_REGION"))
|
|
.unwrap(),
|
|
url: parse_var(&format!("S3_{bucket_type}_URL")).unwrap(),
|
|
access_token: parse_var(&format!(
|
|
"S3_{bucket_type}_ACCESS_TOKEN"
|
|
))
|
|
.unwrap(),
|
|
secret: parse_var(&format!("S3_{bucket_type}_SECRET"))
|
|
.unwrap(),
|
|
};
|
|
|
|
Arc::new(
|
|
S3Host::new(
|
|
config_from_env("PUBLIC"),
|
|
config_from_env("PRIVATE"),
|
|
)
|
|
.unwrap(),
|
|
)
|
|
}
|
|
"local" => Arc::new(file_hosting::MockHost::new()),
|
|
_ => panic!("Invalid storage backend specified. Aborting startup!"),
|
|
};
|
|
|
|
info!("Initializing clickhouse connection");
|
|
let mut clickhouse = clickhouse::init_client().await.unwrap();
|
|
|
|
let search_config = search::SearchConfig::new(None);
|
|
|
|
let stripe_client =
|
|
stripe::Client::new(dotenvy::var("STRIPE_API_KEY").unwrap());
|
|
|
|
let anrok_client = anrok::Client::from_env().unwrap();
|
|
let email_queue =
|
|
EmailQueue::init(pool.clone(), redis_pool.clone()).unwrap();
|
|
|
|
let gotenberg_client = GotenbergClient::from_env(redis_pool.clone())
|
|
.expect("Failed to create Gotenberg client");
|
|
let muralpay = labrinth::queue::payouts::create_muralpay_client()
|
|
.expect("Failed to create MuralPay client");
|
|
|
|
if let Some(task) = args.run_background_task {
|
|
info!("Running task {task:?} and exiting");
|
|
task.run(
|
|
pool,
|
|
redis_pool,
|
|
search_config,
|
|
clickhouse,
|
|
stripe_client,
|
|
anrok_client.clone(),
|
|
email_queue,
|
|
muralpay,
|
|
)
|
|
.await;
|
|
return Ok(());
|
|
}
|
|
|
|
let maxmind_reader = modrinth_maxmind::MaxMind::new().await;
|
|
|
|
let prometheus = PrometheusMetricsBuilder::new("labrinth")
|
|
.endpoint("/metrics")
|
|
.exclude_regex(r"^/api/v1/.*$")
|
|
.exclude_regex(r"^/maven/.*$")
|
|
.exclude("/_internal/launcher_socket")
|
|
.mask_unmatched_patterns("UNKNOWN")
|
|
.build()
|
|
.expect("Failed to create prometheus metrics middleware");
|
|
|
|
database::register_and_set_metrics(&pool, &prometheus.registry)
|
|
.await
|
|
.expect("Failed to register database metrics");
|
|
redis_pool
|
|
.register_and_set_metrics(&prometheus.registry)
|
|
.await
|
|
.expect("Failed to register redis metrics");
|
|
|
|
#[cfg(target_os = "linux")]
|
|
labrinth::routes::debug::jemalloc_memory_stats(&prometheus.registry)
|
|
.expect("Failed to register jemalloc metrics");
|
|
|
|
let labrinth_config = labrinth::app_setup(
|
|
pool.clone(),
|
|
ro_pool.clone(),
|
|
redis_pool.clone(),
|
|
search_config.clone(),
|
|
&mut clickhouse,
|
|
file_host.clone(),
|
|
maxmind_reader.clone(),
|
|
stripe_client,
|
|
anrok_client.clone(),
|
|
email_queue,
|
|
gotenberg_client,
|
|
!args.no_background_tasks,
|
|
);
|
|
|
|
info!("Starting Actix HTTP server!");
|
|
|
|
HttpServer::new(move || {
|
|
App::new()
|
|
.wrap(TracingLogger::default())
|
|
.wrap_fn(|req, srv| {
|
|
// We capture the same fields as `tracing-actix-web`'s `RootSpanBuilder`.
|
|
// See `root_span!` macro.
|
|
let span = info_span!(
|
|
"HTTP request",
|
|
http.method = %req.method(),
|
|
http.client_ip = %req.connection_info().realip_remote_addr().unwrap_or(""),
|
|
http.user_agent = %req.headers().get("User-Agent").map_or("", |h| h.to_str().unwrap_or("")),
|
|
http.target = %req.uri().path_and_query().map_or("", |p| p.as_str()),
|
|
http.authenticated = %req.headers().get("Authorization").is_some()
|
|
);
|
|
|
|
let fut = srv.call(req);
|
|
async move {
|
|
fut.await.inspect(|resp| {
|
|
let _span = info_span!(
|
|
"HTTP response",
|
|
http.status = %resp.response().status().as_u16(),
|
|
).entered();
|
|
|
|
resp.response()
|
|
.error()
|
|
.inspect(|err| log_error(err));
|
|
})
|
|
}
|
|
.instrument(span)
|
|
})
|
|
.wrap(prometheus.clone())
|
|
.wrap(from_fn(rate_limit_middleware))
|
|
.wrap(actix_web::middleware::Compress::default())
|
|
.wrap(sentry_actix::Sentry::new())
|
|
.into_utoipa_app()
|
|
.configure(|cfg| utoipa_app_config(cfg, labrinth_config.clone()))
|
|
.openapi_service(|api| SwaggerUi::new("/docs/swagger-ui/{_:.*}")
|
|
.config(utoipa_swagger_ui::Config::default().try_it_out_enabled(true))
|
|
.url("/docs/openapi.json", ApiDoc::openapi().merge_from(api)))
|
|
.into_app()
|
|
.configure(|cfg| app_config(cfg, labrinth_config.clone()))
|
|
})
|
|
.bind(dotenvy::var("BIND_ADDR").unwrap())?
|
|
.run()
|
|
.await
|
|
}
|
|
|
|
#[derive(utoipa::OpenApi)]
|
|
#[openapi(info(title = "Labrinth"), modifiers(&SecurityAddon))]
|
|
struct ApiDoc;
|
|
|
|
struct SecurityAddon;
|
|
|
|
impl utoipa::Modify for SecurityAddon {
|
|
fn modify(&self, openapi: &mut utoipa::openapi::OpenApi) {
|
|
let components = openapi.components.as_mut().unwrap();
|
|
components.add_security_scheme(
|
|
"bearer_auth",
|
|
SecurityScheme::ApiKey(ApiKey::Header(ApiKeyValue::new(
|
|
"authorization",
|
|
))),
|
|
);
|
|
}
|
|
}
|
|
|
|
fn log_error(err: &actix_web::Error) {
|
|
if err.as_response_error().status_code().is_client_error() {
|
|
tracing::debug!(
|
|
"Error encountered while processing the incoming HTTP request: {err:#?}"
|
|
);
|
|
} else {
|
|
tracing::error!(
|
|
"Error encountered while processing the incoming HTTP request: {err:#?}"
|
|
);
|
|
}
|
|
}
|