- workflows.yaml: declare `shared_volume: { mount_path: /workspace,
size: 30Gi }` on the ci workflow so all sub-workflows share a PVC;
set `shell: /bin/bash` on ci_config/ci_long_config anchors.
- log_search.rs: fix opensearch_url() TCP probe to resolve hostnames
(not just IPs); make ensure_index handle resource_already_exists
races gracefully.
- webhook.rs: 14 new handler-level tests covering generic event auth
(accept/reject/missing), GitHub/Gitea HMAC verification, bad JSON
400s, trigger matching, trigger ref-mismatch skip, and real
workflow-start side effect verification.
575 lines
18 KiB
Rust
575 lines
18 KiB
Rust
use chrono::{DateTime, Utc};
|
|
use opensearch::http::transport::Transport;
|
|
use opensearch::{IndexParts, OpenSearch, SearchParts};
|
|
use serde::{Deserialize, Serialize};
|
|
use serde_json::json;
|
|
use wfe_core::traits::{LogChunk, LogStreamType};
|
|
|
|
const LOG_INDEX: &str = "wfe-build-logs";
|
|
|
|
/// Document structure for a log line stored in OpenSearch.
|
|
#[derive(Debug, Serialize, Deserialize)]
|
|
struct LogDocument {
|
|
workflow_id: String,
|
|
definition_id: String,
|
|
step_id: usize,
|
|
step_name: String,
|
|
stream: String,
|
|
line: String,
|
|
timestamp: String,
|
|
}
|
|
|
|
impl LogDocument {
|
|
fn from_chunk(chunk: &LogChunk) -> Self {
|
|
Self {
|
|
workflow_id: chunk.workflow_id.clone(),
|
|
definition_id: chunk.definition_id.clone(),
|
|
step_id: chunk.step_id,
|
|
step_name: chunk.step_name.clone(),
|
|
stream: match chunk.stream {
|
|
LogStreamType::Stdout => "stdout".to_string(),
|
|
LogStreamType::Stderr => "stderr".to_string(),
|
|
},
|
|
line: String::from_utf8_lossy(&chunk.data).trim_end().to_string(),
|
|
timestamp: chunk.timestamp.to_rfc3339(),
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Result from a log search query.
|
|
#[derive(Debug, Clone)]
|
|
pub struct LogSearchHit {
|
|
pub workflow_id: String,
|
|
pub definition_id: String,
|
|
pub step_name: String,
|
|
pub line: String,
|
|
pub stream: String,
|
|
pub timestamp: DateTime<Utc>,
|
|
}
|
|
|
|
/// OpenSearch-backed log search index.
|
|
pub struct LogSearchIndex {
|
|
client: OpenSearch,
|
|
}
|
|
|
|
impl LogSearchIndex {
|
|
pub fn new(url: &str) -> wfe_core::Result<Self> {
|
|
let transport = Transport::single_node(url)
|
|
.map_err(|e| Box::new(e) as Box<dyn std::error::Error + Send + Sync>)?;
|
|
Ok(Self {
|
|
client: OpenSearch::new(transport),
|
|
})
|
|
}
|
|
|
|
/// Create the log index if it doesn't exist.
|
|
pub async fn ensure_index(&self) -> wfe_core::Result<()> {
|
|
let exists = self
|
|
.client
|
|
.indices()
|
|
.exists(opensearch::indices::IndicesExistsParts::Index(&[LOG_INDEX]))
|
|
.send()
|
|
.await
|
|
.map_err(|e| Box::new(e) as Box<dyn std::error::Error + Send + Sync>)?;
|
|
|
|
if exists.status_code().is_success() {
|
|
return Ok(());
|
|
}
|
|
|
|
let body = json!({
|
|
"mappings": {
|
|
"properties": {
|
|
"workflow_id": { "type": "keyword" },
|
|
"definition_id": { "type": "keyword" },
|
|
"step_id": { "type": "integer" },
|
|
"step_name": { "type": "keyword" },
|
|
"stream": { "type": "keyword" },
|
|
"line": { "type": "text", "analyzer": "standard" },
|
|
"timestamp": { "type": "date" }
|
|
}
|
|
}
|
|
});
|
|
|
|
let response = self
|
|
.client
|
|
.indices()
|
|
.create(opensearch::indices::IndicesCreateParts::Index(LOG_INDEX))
|
|
.body(body)
|
|
.send()
|
|
.await
|
|
.map_err(|e| Box::new(e) as Box<dyn std::error::Error + Send + Sync>)?;
|
|
|
|
if !response.status_code().is_success() {
|
|
let text = response.text().await.unwrap_or_default();
|
|
// Race: another caller created the index between our
|
|
// `exists` probe and the `create` call. OpenSearch returns
|
|
// a 400 with `resource_already_exists_exception`; treat that
|
|
// as a successful no-op rather than failing the call.
|
|
if text.contains("resource_already_exists_exception") {
|
|
return Ok(());
|
|
}
|
|
return Err(wfe_core::WfeError::Persistence(format!(
|
|
"Failed to create log index: {text}"
|
|
)));
|
|
}
|
|
|
|
tracing::info!(index = LOG_INDEX, "log search index created");
|
|
Ok(())
|
|
}
|
|
|
|
/// Index a single log chunk.
|
|
pub async fn index_chunk(&self, chunk: &LogChunk) -> wfe_core::Result<()> {
|
|
let doc = LogDocument::from_chunk(chunk);
|
|
let body = serde_json::to_value(&doc)?;
|
|
|
|
let response = self
|
|
.client
|
|
.index(IndexParts::Index(LOG_INDEX))
|
|
.body(body)
|
|
.send()
|
|
.await
|
|
.map_err(|e| Box::new(e) as Box<dyn std::error::Error + Send + Sync>)?;
|
|
|
|
if !response.status_code().is_success() {
|
|
let text = response.text().await.unwrap_or_default();
|
|
return Err(wfe_core::WfeError::Persistence(format!(
|
|
"failed to index log chunk: {text}"
|
|
)));
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Search log lines.
|
|
pub async fn search(
|
|
&self,
|
|
query: &str,
|
|
workflow_id: Option<&str>,
|
|
step_name: Option<&str>,
|
|
stream_filter: Option<&str>,
|
|
skip: u64,
|
|
take: u64,
|
|
) -> wfe_core::Result<(Vec<LogSearchHit>, u64)> {
|
|
let mut must_clauses = Vec::new();
|
|
let mut filter_clauses = Vec::new();
|
|
|
|
if !query.is_empty() {
|
|
must_clauses.push(json!({
|
|
"match": { "line": query }
|
|
}));
|
|
}
|
|
|
|
if let Some(wf_id) = workflow_id {
|
|
filter_clauses.push(json!({ "term": { "workflow_id": wf_id } }));
|
|
}
|
|
if let Some(sn) = step_name {
|
|
filter_clauses.push(json!({ "term": { "step_name": sn } }));
|
|
}
|
|
if let Some(stream) = stream_filter {
|
|
filter_clauses.push(json!({ "term": { "stream": stream } }));
|
|
}
|
|
|
|
let query_body = if must_clauses.is_empty() && filter_clauses.is_empty() {
|
|
json!({ "match_all": {} })
|
|
} else {
|
|
let mut bool_q = serde_json::Map::new();
|
|
if !must_clauses.is_empty() {
|
|
bool_q.insert("must".to_string(), json!(must_clauses));
|
|
}
|
|
if !filter_clauses.is_empty() {
|
|
bool_q.insert("filter".to_string(), json!(filter_clauses));
|
|
}
|
|
json!({ "bool": bool_q })
|
|
};
|
|
|
|
let body = json!({
|
|
"query": query_body,
|
|
"from": skip,
|
|
"size": take,
|
|
"sort": [{ "timestamp": "asc" }]
|
|
});
|
|
|
|
let response = self
|
|
.client
|
|
.search(SearchParts::Index(&[LOG_INDEX]))
|
|
.body(body)
|
|
.send()
|
|
.await
|
|
.map_err(|e| Box::new(e) as Box<dyn std::error::Error + Send + Sync>)?;
|
|
|
|
if !response.status_code().is_success() {
|
|
let text = response.text().await.unwrap_or_default();
|
|
return Err(wfe_core::WfeError::Persistence(format!(
|
|
"Log search failed: {text}"
|
|
)));
|
|
}
|
|
|
|
let resp_body: serde_json::Value = response
|
|
.json()
|
|
.await
|
|
.map_err(|e| Box::new(e) as Box<dyn std::error::Error + Send + Sync>)?;
|
|
|
|
let total = resp_body["hits"]["total"]["value"].as_u64().unwrap_or(0);
|
|
let hits = resp_body["hits"]["hits"]
|
|
.as_array()
|
|
.cloned()
|
|
.unwrap_or_default();
|
|
|
|
let results = hits
|
|
.iter()
|
|
.filter_map(|hit| {
|
|
let src = &hit["_source"];
|
|
Some(LogSearchHit {
|
|
workflow_id: src["workflow_id"].as_str()?.to_string(),
|
|
definition_id: src["definition_id"].as_str()?.to_string(),
|
|
step_name: src["step_name"].as_str()?.to_string(),
|
|
line: src["line"].as_str()?.to_string(),
|
|
stream: src["stream"].as_str()?.to_string(),
|
|
timestamp: src["timestamp"]
|
|
.as_str()
|
|
.and_then(|s| DateTime::parse_from_rfc3339(s).ok())
|
|
.map(|dt| dt.with_timezone(&Utc))
|
|
.unwrap_or_else(Utc::now),
|
|
})
|
|
})
|
|
.collect();
|
|
|
|
Ok((results, total))
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
|
|
#[test]
|
|
fn log_document_from_chunk_stdout() {
|
|
let chunk = LogChunk {
|
|
workflow_id: "wf-1".to_string(),
|
|
definition_id: "ci".to_string(),
|
|
step_id: 0,
|
|
step_name: "build".to_string(),
|
|
stream: LogStreamType::Stdout,
|
|
data: b"compiling wfe-core\n".to_vec(),
|
|
timestamp: Utc::now(),
|
|
};
|
|
let doc = LogDocument::from_chunk(&chunk);
|
|
assert_eq!(doc.workflow_id, "wf-1");
|
|
assert_eq!(doc.stream, "stdout");
|
|
assert_eq!(doc.line, "compiling wfe-core");
|
|
assert_eq!(doc.step_name, "build");
|
|
}
|
|
|
|
#[test]
|
|
fn log_document_from_chunk_stderr() {
|
|
let chunk = LogChunk {
|
|
workflow_id: "wf-2".to_string(),
|
|
definition_id: "deploy".to_string(),
|
|
step_id: 1,
|
|
step_name: "test".to_string(),
|
|
stream: LogStreamType::Stderr,
|
|
data: b"warning: unused variable\n".to_vec(),
|
|
timestamp: Utc::now(),
|
|
};
|
|
let doc = LogDocument::from_chunk(&chunk);
|
|
assert_eq!(doc.stream, "stderr");
|
|
assert_eq!(doc.line, "warning: unused variable");
|
|
}
|
|
|
|
#[test]
|
|
fn log_document_trims_trailing_newline() {
|
|
let chunk = LogChunk {
|
|
workflow_id: "wf-1".to_string(),
|
|
definition_id: "ci".to_string(),
|
|
step_id: 0,
|
|
step_name: "build".to_string(),
|
|
stream: LogStreamType::Stdout,
|
|
data: b"line with newline\n".to_vec(),
|
|
timestamp: Utc::now(),
|
|
};
|
|
let doc = LogDocument::from_chunk(&chunk);
|
|
assert_eq!(doc.line, "line with newline");
|
|
}
|
|
|
|
#[test]
|
|
fn log_document_serializes_to_json() {
|
|
let chunk = LogChunk {
|
|
workflow_id: "wf-1".to_string(),
|
|
definition_id: "ci".to_string(),
|
|
step_id: 2,
|
|
step_name: "clippy".to_string(),
|
|
stream: LogStreamType::Stdout,
|
|
data: b"all good\n".to_vec(),
|
|
timestamp: Utc::now(),
|
|
};
|
|
let doc = LogDocument::from_chunk(&chunk);
|
|
let json = serde_json::to_value(&doc).unwrap();
|
|
assert_eq!(json["step_name"], "clippy");
|
|
assert_eq!(json["step_id"], 2);
|
|
assert!(json["timestamp"].is_string());
|
|
}
|
|
|
|
// ── OpenSearch integration tests ────────────────────────────────
|
|
|
|
fn opensearch_url() -> Option<String> {
|
|
let url =
|
|
std::env::var("WFE_SEARCH_URL").unwrap_or_else(|_| "http://localhost:9200".to_string());
|
|
// Quick TCP probe to check if OpenSearch is reachable. Use
|
|
// `to_socket_addrs` so hostnames resolve — the previous
|
|
// implementation parsed `"localhost:9200"` as a SocketAddr, which
|
|
// fails (hostnames aren't valid SocketAddrs), silently skipping
|
|
// every OpenSearch test even when the daemon was available.
|
|
use std::net::ToSocketAddrs;
|
|
let host_port = url
|
|
.strip_prefix("http://")
|
|
.or_else(|| url.strip_prefix("https://"))
|
|
.unwrap_or("localhost:9200");
|
|
let mut addrs = host_port.to_socket_addrs().ok()?;
|
|
let addr = addrs.next()?;
|
|
match std::net::TcpStream::connect_timeout(&addr, std::time::Duration::from_secs(1)) {
|
|
Ok(_) => Some(url),
|
|
Err(_) => None,
|
|
}
|
|
}
|
|
|
|
fn make_test_chunk(
|
|
workflow_id: &str,
|
|
step_name: &str,
|
|
stream: LogStreamType,
|
|
line: &str,
|
|
) -> LogChunk {
|
|
LogChunk {
|
|
workflow_id: workflow_id.to_string(),
|
|
definition_id: "test-def".to_string(),
|
|
step_id: 0,
|
|
step_name: step_name.to_string(),
|
|
stream,
|
|
data: format!("{line}\n").into_bytes(),
|
|
timestamp: Utc::now(),
|
|
}
|
|
}
|
|
|
|
/// Delete the test index to start clean.
|
|
async fn cleanup_index(url: &str) {
|
|
let client = reqwest::Client::new();
|
|
let _ = client.delete(format!("{url}/{LOG_INDEX}")).send().await;
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn opensearch_ensure_index_creates_index() {
|
|
let Some(url) = opensearch_url() else {
|
|
eprintln!("SKIP: OpenSearch not available");
|
|
return;
|
|
};
|
|
cleanup_index(&url).await;
|
|
|
|
let index = LogSearchIndex::new(&url).unwrap();
|
|
index.ensure_index().await.unwrap();
|
|
|
|
// Calling again should be idempotent.
|
|
index.ensure_index().await.unwrap();
|
|
|
|
cleanup_index(&url).await;
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn opensearch_index_and_search_chunk() {
|
|
let Some(url) = opensearch_url() else {
|
|
eprintln!("SKIP: OpenSearch not available");
|
|
return;
|
|
};
|
|
cleanup_index(&url).await;
|
|
|
|
let index = LogSearchIndex::new(&url).unwrap();
|
|
index.ensure_index().await.unwrap();
|
|
|
|
// Index some log chunks.
|
|
let chunk = make_test_chunk(
|
|
"wf-search-1",
|
|
"build",
|
|
LogStreamType::Stdout,
|
|
"compiling wfe-core v1.5.0",
|
|
);
|
|
index.index_chunk(&chunk).await.unwrap();
|
|
|
|
let chunk = make_test_chunk(
|
|
"wf-search-1",
|
|
"build",
|
|
LogStreamType::Stderr,
|
|
"warning: unused variable",
|
|
);
|
|
index.index_chunk(&chunk).await.unwrap();
|
|
|
|
let chunk = make_test_chunk(
|
|
"wf-search-1",
|
|
"test",
|
|
LogStreamType::Stdout,
|
|
"test result: ok. 79 passed",
|
|
);
|
|
index.index_chunk(&chunk).await.unwrap();
|
|
|
|
// OpenSearch needs a refresh to make docs searchable.
|
|
let client = reqwest::Client::new();
|
|
client
|
|
.post(format!("{url}/{LOG_INDEX}/_refresh"))
|
|
.send()
|
|
.await
|
|
.unwrap();
|
|
|
|
// Search by text.
|
|
let (results, total) = index
|
|
.search("wfe-core", None, None, None, 0, 10)
|
|
.await
|
|
.unwrap();
|
|
assert!(total >= 1, "expected at least 1 hit, got {total}");
|
|
assert!(results.iter().any(|r| r.line.contains("wfe-core")));
|
|
|
|
// Search by workflow_id filter.
|
|
let (results, _) = index
|
|
.search("", Some("wf-search-1"), None, None, 0, 10)
|
|
.await
|
|
.unwrap();
|
|
assert_eq!(results.len(), 3);
|
|
|
|
// Search by step_name filter.
|
|
let (results, _) = index
|
|
.search("", Some("wf-search-1"), Some("test"), None, 0, 10)
|
|
.await
|
|
.unwrap();
|
|
assert_eq!(results.len(), 1);
|
|
assert!(results[0].line.contains("79 passed"));
|
|
|
|
// Search by stream filter.
|
|
let (results, _) = index
|
|
.search("", Some("wf-search-1"), None, Some("stderr"), 0, 10)
|
|
.await
|
|
.unwrap();
|
|
assert_eq!(results.len(), 1);
|
|
assert!(results[0].line.contains("unused variable"));
|
|
|
|
cleanup_index(&url).await;
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn opensearch_search_empty_index() {
|
|
let Some(url) = opensearch_url() else {
|
|
eprintln!("SKIP: OpenSearch not available");
|
|
return;
|
|
};
|
|
cleanup_index(&url).await;
|
|
|
|
let index = LogSearchIndex::new(&url).unwrap();
|
|
index.ensure_index().await.unwrap();
|
|
|
|
let (results, total) = index
|
|
.search("nonexistent", None, None, None, 0, 10)
|
|
.await
|
|
.unwrap();
|
|
assert_eq!(total, 0);
|
|
assert!(results.is_empty());
|
|
|
|
cleanup_index(&url).await;
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn opensearch_search_pagination() {
|
|
let Some(url) = opensearch_url() else {
|
|
eprintln!("SKIP: OpenSearch not available");
|
|
return;
|
|
};
|
|
cleanup_index(&url).await;
|
|
|
|
let index = LogSearchIndex::new(&url).unwrap();
|
|
index.ensure_index().await.unwrap();
|
|
|
|
// Index 5 chunks.
|
|
for i in 0..5 {
|
|
let chunk = make_test_chunk(
|
|
"wf-page",
|
|
"build",
|
|
LogStreamType::Stdout,
|
|
&format!("line {i}"),
|
|
);
|
|
index.index_chunk(&chunk).await.unwrap();
|
|
}
|
|
|
|
let client = reqwest::Client::new();
|
|
client
|
|
.post(format!("{url}/{LOG_INDEX}/_refresh"))
|
|
.send()
|
|
.await
|
|
.unwrap();
|
|
|
|
// Get first 2.
|
|
let (results, total) = index
|
|
.search("", Some("wf-page"), None, None, 0, 2)
|
|
.await
|
|
.unwrap();
|
|
assert_eq!(total, 5);
|
|
assert_eq!(results.len(), 2);
|
|
|
|
// Get next 2.
|
|
let (results, _) = index
|
|
.search("", Some("wf-page"), None, None, 2, 2)
|
|
.await
|
|
.unwrap();
|
|
assert_eq!(results.len(), 2);
|
|
|
|
// Get last 1.
|
|
let (results, _) = index
|
|
.search("", Some("wf-page"), None, None, 4, 2)
|
|
.await
|
|
.unwrap();
|
|
assert_eq!(results.len(), 1);
|
|
|
|
cleanup_index(&url).await;
|
|
}
|
|
|
|
#[test]
|
|
fn log_search_index_new_constructs_ok() {
|
|
// Construction should succeed even for unreachable URLs (fails on first use).
|
|
let result = LogSearchIndex::new("http://localhost:19876");
|
|
assert!(result.is_ok());
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn opensearch_index_chunk_result_fields() {
|
|
let Some(url) = opensearch_url() else {
|
|
eprintln!("SKIP: OpenSearch not available");
|
|
return;
|
|
};
|
|
cleanup_index(&url).await;
|
|
|
|
let index = LogSearchIndex::new(&url).unwrap();
|
|
index.ensure_index().await.unwrap();
|
|
|
|
let chunk = make_test_chunk(
|
|
"wf-fields",
|
|
"clippy",
|
|
LogStreamType::Stderr,
|
|
"error: type mismatch",
|
|
);
|
|
index.index_chunk(&chunk).await.unwrap();
|
|
|
|
let client = reqwest::Client::new();
|
|
client
|
|
.post(format!("{url}/{LOG_INDEX}/_refresh"))
|
|
.send()
|
|
.await
|
|
.unwrap();
|
|
|
|
let (results, _) = index
|
|
.search("type mismatch", None, None, None, 0, 10)
|
|
.await
|
|
.unwrap();
|
|
assert!(!results.is_empty());
|
|
let hit = &results[0];
|
|
assert_eq!(hit.workflow_id, "wf-fields");
|
|
assert_eq!(hit.definition_id, "test-def");
|
|
assert_eq!(hit.step_name, "clippy");
|
|
assert_eq!(hit.stream, "stderr");
|
|
assert!(hit.line.contains("type mismatch"));
|
|
|
|
cleanup_index(&url).await;
|
|
}
|
|
}
|