fix: local_files hot-reload via RwLock state fields + rebuild_registry local_files case
This commit is contained in:
@@ -113,6 +113,65 @@ async fn rebuild_registry(state: &AppState) -> DomainResult<()> {
|
||||
);
|
||||
}
|
||||
}
|
||||
#[cfg(feature = "local-files")]
|
||||
"local_files" => {
|
||||
let config: std::collections::HashMap<String, String> =
|
||||
match serde_json::from_str(&row.config_json) {
|
||||
Ok(c) => c,
|
||||
Err(_) => continue,
|
||||
};
|
||||
|
||||
let files_dir = match config.get("files_dir") {
|
||||
Some(d) => std::path::PathBuf::from(d),
|
||||
None => continue,
|
||||
};
|
||||
|
||||
let transcode_dir = config
|
||||
.get("transcode_dir")
|
||||
.filter(|s| !s.is_empty())
|
||||
.map(std::path::PathBuf::from);
|
||||
|
||||
let cleanup_ttl_hours: u32 = config
|
||||
.get("cleanup_ttl_hours")
|
||||
.and_then(|s| s.parse().ok())
|
||||
.unwrap_or(24);
|
||||
|
||||
let base_url = state.config.base_url.clone();
|
||||
|
||||
let sqlite_pool = match &state.raw_sqlite_pool {
|
||||
Some(p) => p.clone(),
|
||||
None => {
|
||||
tracing::warn!("local_files provider requires SQLite; skipping");
|
||||
continue;
|
||||
}
|
||||
};
|
||||
|
||||
let lf_cfg = infra::LocalFilesConfig {
|
||||
root_dir: files_dir,
|
||||
base_url,
|
||||
transcode_dir: transcode_dir.clone(),
|
||||
cleanup_ttl_hours,
|
||||
};
|
||||
|
||||
let idx = Arc::new(infra::LocalIndex::new(&lf_cfg, sqlite_pool.clone()).await);
|
||||
|
||||
let scan_idx = Arc::clone(&idx);
|
||||
tokio::spawn(async move { scan_idx.rescan().await; });
|
||||
|
||||
let tm = transcode_dir.as_ref().map(|td| {
|
||||
std::fs::create_dir_all(td).ok();
|
||||
infra::TranscodeManager::new(td.clone(), cleanup_ttl_hours)
|
||||
});
|
||||
|
||||
new_registry.register(
|
||||
"local",
|
||||
Arc::new(infra::LocalFilesProvider::new(Arc::clone(&idx), lf_cfg, tm.clone())),
|
||||
);
|
||||
|
||||
*state.local_index.write().await = Some(idx);
|
||||
*state.transcode_manager.write().await = tm;
|
||||
*state.sqlite_pool.write().await = Some(sqlite_pool);
|
||||
}
|
||||
_ => {}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -147,9 +147,7 @@ async fn trigger_rescan(
|
||||
State(state): State<AppState>,
|
||||
CurrentUser(_user): CurrentUser,
|
||||
) -> Result<Json<serde_json::Value>, ApiError> {
|
||||
let index = state
|
||||
.local_index
|
||||
.as_ref()
|
||||
let index = state.local_index.read().await.clone()
|
||||
.ok_or_else(|| ApiError::not_implemented("no local files provider active"))?;
|
||||
let count = index.rescan().await;
|
||||
Ok(Json(serde_json::json!({ "items_found": count })))
|
||||
@@ -164,9 +162,7 @@ async fn transcode_playlist(
|
||||
State(state): State<AppState>,
|
||||
Path(id): Path<String>,
|
||||
) -> Result<Response, ApiError> {
|
||||
let tm = state
|
||||
.transcode_manager
|
||||
.as_ref()
|
||||
let tm = state.transcode_manager.read().await.clone()
|
||||
.ok_or_else(|| ApiError::not_implemented("TRANSCODE_DIR not configured"))?;
|
||||
|
||||
let root = state.config.local_files_dir.as_ref().ok_or_else(|| {
|
||||
@@ -219,9 +215,7 @@ async fn transcode_segment(
|
||||
return Err(ApiError::Forbidden("invalid segment path".into()));
|
||||
}
|
||||
|
||||
let tm = state
|
||||
.transcode_manager
|
||||
.as_ref()
|
||||
let tm = state.transcode_manager.read().await.clone()
|
||||
.ok_or_else(|| ApiError::not_implemented("TRANSCODE_DIR not configured"))?;
|
||||
|
||||
let file_path = tm.transcode_dir.join(&id).join(&segment);
|
||||
@@ -262,14 +256,12 @@ async fn get_transcode_settings(
|
||||
State(state): State<AppState>,
|
||||
CurrentUser(_user): CurrentUser,
|
||||
) -> Result<Json<TranscodeSettingsResponse>, ApiError> {
|
||||
let pool = state
|
||||
.sqlite_pool
|
||||
.as_ref()
|
||||
let pool = state.sqlite_pool.read().await.clone()
|
||||
.ok_or_else(|| ApiError::not_implemented("sqlite not available"))?;
|
||||
|
||||
let (ttl,): (i64,) =
|
||||
sqlx::query_as("SELECT cleanup_ttl_hours FROM transcode_settings WHERE id = 1")
|
||||
.fetch_one(pool)
|
||||
.fetch_one(&pool)
|
||||
.await
|
||||
.map_err(|e| ApiError::internal(e.to_string()))?;
|
||||
|
||||
@@ -284,19 +276,18 @@ async fn update_transcode_settings(
|
||||
CurrentUser(_user): CurrentUser,
|
||||
Json(req): Json<UpdateTranscodeSettingsRequest>,
|
||||
) -> Result<Json<TranscodeSettingsResponse>, ApiError> {
|
||||
let pool = state
|
||||
.sqlite_pool
|
||||
.as_ref()
|
||||
let pool = state.sqlite_pool.read().await.clone()
|
||||
.ok_or_else(|| ApiError::not_implemented("sqlite not available"))?;
|
||||
|
||||
let ttl = req.cleanup_ttl_hours as i64;
|
||||
sqlx::query("UPDATE transcode_settings SET cleanup_ttl_hours = ? WHERE id = 1")
|
||||
.bind(ttl)
|
||||
.execute(pool)
|
||||
.execute(&pool)
|
||||
.await
|
||||
.map_err(|e| ApiError::internal(e.to_string()))?;
|
||||
|
||||
if let Some(tm) = &state.transcode_manager {
|
||||
let tm_opt = state.transcode_manager.read().await.clone();
|
||||
if let Some(tm) = tm_opt {
|
||||
tm.set_cleanup_ttl(req.cleanup_ttl_hours);
|
||||
}
|
||||
|
||||
@@ -310,9 +301,7 @@ async fn get_transcode_stats(
|
||||
State(state): State<AppState>,
|
||||
CurrentUser(_user): CurrentUser,
|
||||
) -> Result<Json<TranscodeStatsResponse>, ApiError> {
|
||||
let tm = state
|
||||
.transcode_manager
|
||||
.as_ref()
|
||||
let tm = state.transcode_manager.read().await.clone()
|
||||
.ok_or_else(|| ApiError::not_implemented("TRANSCODE_DIR not configured"))?;
|
||||
let (cache_size_bytes, item_count) = tm.cache_stats().await;
|
||||
Ok(Json(TranscodeStatsResponse {
|
||||
@@ -326,9 +315,7 @@ async fn clear_transcode_cache(
|
||||
State(state): State<AppState>,
|
||||
CurrentUser(_user): CurrentUser,
|
||||
) -> Result<StatusCode, ApiError> {
|
||||
let tm = state
|
||||
.transcode_manager
|
||||
.as_ref()
|
||||
let tm = state.transcode_manager.read().await.clone()
|
||||
.ok_or_else(|| ApiError::not_implemented("TRANSCODE_DIR not configured"))?;
|
||||
tm.clear_cache()
|
||||
.await
|
||||
|
||||
Reference in New Issue
Block a user