Compare commits
33 Commits
ccc3f9fd92
...
8c2215ba58
| Author | SHA1 | Date | |
|---|---|---|---|
| 8c2215ba58 | |||
|
|
c316e80d7f | ||
|
|
796be06a09 | ||
|
|
26f5a2ce3b | ||
|
|
d23b899f00 | ||
|
|
288c7a86b5 | ||
|
|
9f74e47ecf | ||
|
|
39f9925e71 | ||
|
|
af03ecdf42 | ||
|
|
0723f48e5b | ||
|
|
2634f60e59 | ||
|
|
3c0e615fb7 | ||
|
|
589da1b6d6 | ||
|
|
41e2038190 | ||
|
|
ea687a342c | ||
|
|
cea16b38ed | ||
|
|
a344be3a49 | ||
|
|
565b548ac1 | ||
|
|
7dbfaf0932 | ||
|
|
f7d7302694 | ||
|
|
c48e0bdfde | ||
|
|
5932b5d969 | ||
|
|
527e2cf017 | ||
|
|
9fd02c4edb | ||
|
|
85ebe76111 | ||
|
|
adf4b44d78 | ||
|
|
84b93d74c7 | ||
|
|
a55fc3c10d | ||
|
|
55ed3be71a | ||
|
|
41a9a975fd | ||
|
|
0eeae70369 | ||
|
|
8a637df65c | ||
|
|
5bed108d3b |
@@ -161,10 +161,17 @@ jobs:
|
|||||||
--from-literal=CAMELEER_JWT_SECRET="${CAMELEER_JWT_SECRET}" \
|
--from-literal=CAMELEER_JWT_SECRET="${CAMELEER_JWT_SECRET}" \
|
||||||
--dry-run=client -o yaml | kubectl apply -f -
|
--dry-run=client -o yaml | kubectl apply -f -
|
||||||
|
|
||||||
kubectl create secret generic clickhouse-credentials \
|
kubectl create secret generic postgres-credentials \
|
||||||
--namespace=cameleer \
|
--namespace=cameleer \
|
||||||
--from-literal=CLICKHOUSE_USER="$CLICKHOUSE_USER" \
|
--from-literal=POSTGRES_USER="$POSTGRES_USER" \
|
||||||
--from-literal=CLICKHOUSE_PASSWORD="$CLICKHOUSE_PASSWORD" \
|
--from-literal=POSTGRES_PASSWORD="$POSTGRES_PASSWORD" \
|
||||||
|
--from-literal=POSTGRES_DB="${POSTGRES_DB:-cameleer}" \
|
||||||
|
--dry-run=client -o yaml | kubectl apply -f -
|
||||||
|
|
||||||
|
kubectl create secret generic opensearch-credentials \
|
||||||
|
--namespace=cameleer \
|
||||||
|
--from-literal=OPENSEARCH_USER="${OPENSEARCH_USER:-admin}" \
|
||||||
|
--from-literal=OPENSEARCH_PASSWORD="$OPENSEARCH_PASSWORD" \
|
||||||
--dry-run=client -o yaml | kubectl apply -f -
|
--dry-run=client -o yaml | kubectl apply -f -
|
||||||
|
|
||||||
kubectl create secret generic authentik-credentials \
|
kubectl create secret generic authentik-credentials \
|
||||||
@@ -182,8 +189,11 @@ jobs:
|
|||||||
--from-literal=CAMELEER_OIDC_CLIENT_SECRET="${CAMELEER_OIDC_CLIENT_SECRET}" \
|
--from-literal=CAMELEER_OIDC_CLIENT_SECRET="${CAMELEER_OIDC_CLIENT_SECRET}" \
|
||||||
--dry-run=client -o yaml | kubectl apply -f -
|
--dry-run=client -o yaml | kubectl apply -f -
|
||||||
|
|
||||||
kubectl apply -f deploy/clickhouse.yaml
|
kubectl apply -f deploy/postgres.yaml
|
||||||
kubectl -n cameleer rollout status statefulset/clickhouse --timeout=120s
|
kubectl -n cameleer rollout status statefulset/postgres --timeout=120s
|
||||||
|
|
||||||
|
kubectl apply -f deploy/opensearch.yaml
|
||||||
|
kubectl -n cameleer rollout status statefulset/opensearch --timeout=180s
|
||||||
|
|
||||||
kubectl apply -f deploy/authentik.yaml
|
kubectl apply -f deploy/authentik.yaml
|
||||||
kubectl -n cameleer rollout status deployment/authentik-server --timeout=180s
|
kubectl -n cameleer rollout status deployment/authentik-server --timeout=180s
|
||||||
@@ -203,8 +213,11 @@ jobs:
|
|||||||
CAMELEER_JWT_SECRET: ${{ secrets.CAMELEER_JWT_SECRET }}
|
CAMELEER_JWT_SECRET: ${{ secrets.CAMELEER_JWT_SECRET }}
|
||||||
CAMELEER_UI_USER: ${{ secrets.CAMELEER_UI_USER }}
|
CAMELEER_UI_USER: ${{ secrets.CAMELEER_UI_USER }}
|
||||||
CAMELEER_UI_PASSWORD: ${{ secrets.CAMELEER_UI_PASSWORD }}
|
CAMELEER_UI_PASSWORD: ${{ secrets.CAMELEER_UI_PASSWORD }}
|
||||||
CLICKHOUSE_USER: ${{ secrets.CLICKHOUSE_USER }}
|
POSTGRES_USER: ${{ secrets.POSTGRES_USER }}
|
||||||
CLICKHOUSE_PASSWORD: ${{ secrets.CLICKHOUSE_PASSWORD }}
|
POSTGRES_PASSWORD: ${{ secrets.POSTGRES_PASSWORD }}
|
||||||
|
POSTGRES_DB: ${{ secrets.POSTGRES_DB }}
|
||||||
|
OPENSEARCH_USER: ${{ secrets.OPENSEARCH_USER }}
|
||||||
|
OPENSEARCH_PASSWORD: ${{ secrets.OPENSEARCH_PASSWORD }}
|
||||||
AUTHENTIK_PG_USER: ${{ secrets.AUTHENTIK_PG_USER }}
|
AUTHENTIK_PG_USER: ${{ secrets.AUTHENTIK_PG_USER }}
|
||||||
AUTHENTIK_PG_PASSWORD: ${{ secrets.AUTHENTIK_PG_PASSWORD }}
|
AUTHENTIK_PG_PASSWORD: ${{ secrets.AUTHENTIK_PG_PASSWORD }}
|
||||||
AUTHENTIK_SECRET_KEY: ${{ secrets.AUTHENTIK_SECRET_KEY }}
|
AUTHENTIK_SECRET_KEY: ${{ secrets.AUTHENTIK_SECRET_KEY }}
|
||||||
|
|||||||
10
CLAUDE.md
10
CLAUDE.md
@@ -38,10 +38,10 @@ java -jar cameleer3-server-app/target/cameleer3-server-app-1.0-SNAPSHOT.jar
|
|||||||
- Jackson `JavaTimeModule` for `Instant` deserialization
|
- Jackson `JavaTimeModule` for `Instant` deserialization
|
||||||
- Communication: receives HTTP POST data from agents, serves SSE event streams for config push/commands
|
- Communication: receives HTTP POST data from agents, serves SSE event streams for config push/commands
|
||||||
- Maintains agent instance registry with states: LIVE → STALE → DEAD
|
- Maintains agent instance registry with states: LIVE → STALE → DEAD
|
||||||
- Storage: ClickHouse for structured data, text index for full-text search
|
- Storage: PostgreSQL (TimescaleDB) for structured data, OpenSearch for full-text search
|
||||||
- Security: JWT auth with RBAC (AGENT/VIEWER/OPERATOR/ADMIN roles), Ed25519 config signing, bootstrap token for registration
|
- Security: JWT auth with RBAC (AGENT/VIEWER/OPERATOR/ADMIN roles), Ed25519 config signing, bootstrap token for registration
|
||||||
- OIDC: Optional external identity provider support (token exchange pattern). Configured via `CAMELEER_OIDC_*` env vars
|
- OIDC: Optional external identity provider support (token exchange pattern). Configured via `CAMELEER_OIDC_*` env vars
|
||||||
- User persistence: ClickHouse `users` table, admin CRUD at `/api/v1/admin/users`
|
- User persistence: PostgreSQL `users` table, admin CRUD at `/api/v1/admin/users`
|
||||||
|
|
||||||
## CI/CD & Deployment
|
## CI/CD & Deployment
|
||||||
|
|
||||||
@@ -50,8 +50,8 @@ java -jar cameleer3-server-app/target/cameleer3-server-app-1.0-SNAPSHOT.jar
|
|||||||
- Docker: multi-stage build (`Dockerfile`), `$BUILDPLATFORM` for native Maven on ARM64 runner, amd64 runtime
|
- Docker: multi-stage build (`Dockerfile`), `$BUILDPLATFORM` for native Maven on ARM64 runner, amd64 runtime
|
||||||
- `REGISTRY_TOKEN` build arg required for `cameleer3-common` dependency resolution
|
- `REGISTRY_TOKEN` build arg required for `cameleer3-common` dependency resolution
|
||||||
- Registry: `gitea.siegeln.net/cameleer/cameleer3-server` (container images)
|
- Registry: `gitea.siegeln.net/cameleer/cameleer3-server` (container images)
|
||||||
- K8s manifests in `deploy/` — ClickHouse StatefulSet + server Deployment + NodePort Service (30081)
|
- K8s manifests in `deploy/` — PostgreSQL + OpenSearch StatefulSets, server Deployment + NodePort Service (30081)
|
||||||
- Deployment target: k3s at 192.168.50.86, namespace `cameleer`
|
- Deployment target: k3s at 192.168.50.86, namespace `cameleer`
|
||||||
- Secrets managed in CI deploy step (idempotent `--dry-run=client | kubectl apply`): `cameleer-auth`, `clickhouse-credentials`, `CAMELEER_JWT_SECRET`
|
- Secrets managed in CI deploy step (idempotent `--dry-run=client | kubectl apply`): `cameleer-auth`, `postgres-credentials`, `opensearch-credentials`, `CAMELEER_JWT_SECRET`
|
||||||
- K8s probes: server uses `/api/v1/health`, ClickHouse uses `/ping`
|
- K8s probes: server uses `/api/v1/health`, PostgreSQL uses `pg_isready`, OpenSearch uses `/_cluster/health`
|
||||||
- Docker build uses buildx registry cache + `--provenance=false` for Gitea compatibility
|
- Docker build uses buildx registry cache + `--provenance=false` for Gitea compatibility
|
||||||
|
|||||||
@@ -18,9 +18,10 @@ FROM eclipse-temurin:17-jre
|
|||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
COPY --from=build /build/cameleer3-server-app/target/cameleer3-server-app-*.jar /app/server.jar
|
COPY --from=build /build/cameleer3-server-app/target/cameleer3-server-app-*.jar /app/server.jar
|
||||||
|
|
||||||
ENV SPRING_DATASOURCE_URL=jdbc:ch://clickhouse:8123/cameleer3
|
ENV SPRING_DATASOURCE_URL=jdbc:postgresql://postgres:5432/cameleer3
|
||||||
ENV SPRING_DATASOURCE_USERNAME=cameleer
|
ENV SPRING_DATASOURCE_USERNAME=cameleer
|
||||||
ENV SPRING_DATASOURCE_PASSWORD=cameleer_dev
|
ENV SPRING_DATASOURCE_PASSWORD=cameleer_dev
|
||||||
|
ENV OPENSEARCH_URL=http://opensearch:9200
|
||||||
|
|
||||||
EXPOSE 8081
|
EXPOSE 8081
|
||||||
ENTRYPOINT exec java -jar /app/server.jar
|
ENTRYPOINT exec java -jar /app/server.jar
|
||||||
|
|||||||
40
HOWTO.md
40
HOWTO.md
@@ -21,20 +21,20 @@ mvn clean verify # compile + run all tests (needs Docker for integrati
|
|||||||
|
|
||||||
## Infrastructure Setup
|
## Infrastructure Setup
|
||||||
|
|
||||||
Start ClickHouse:
|
Start PostgreSQL and OpenSearch:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker compose up -d
|
docker compose up -d
|
||||||
```
|
```
|
||||||
|
|
||||||
This starts ClickHouse 25.3 and automatically runs the schema init scripts (`clickhouse/init/01-schema.sql`, `clickhouse/init/02-search-columns.sql`, `clickhouse/init/03-users.sql`).
|
This starts TimescaleDB (PostgreSQL 16) and OpenSearch 2.19. The database schema is applied automatically via Flyway migrations on server startup.
|
||||||
|
|
||||||
| Service | Port | Purpose |
|
| Service | Port | Purpose |
|
||||||
|------------|------|------------------|
|
|------------|------|----------------------|
|
||||||
| ClickHouse | 8123 | HTTP API (JDBC) |
|
| PostgreSQL | 5432 | JDBC (Spring JDBC) |
|
||||||
| ClickHouse | 9000 | Native protocol |
|
| OpenSearch | 9200 | REST API (full-text) |
|
||||||
|
|
||||||
ClickHouse credentials: `cameleer` / `cameleer_dev`, database `cameleer3`.
|
PostgreSQL credentials: `cameleer` / `cameleer_dev`, database `cameleer3`.
|
||||||
|
|
||||||
## Run the Server
|
## Run the Server
|
||||||
|
|
||||||
@@ -109,7 +109,7 @@ The env-var local user gets `ADMIN` role. Agents get `AGENT` role at registratio
|
|||||||
|
|
||||||
### OIDC Login (Optional)
|
### OIDC Login (Optional)
|
||||||
|
|
||||||
OIDC configuration is stored in ClickHouse and managed via the admin API or UI. The SPA checks if OIDC is available:
|
OIDC configuration is stored in PostgreSQL and managed via the admin API or UI. The SPA checks if OIDC is available:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# 1. SPA checks if OIDC is available (returns 404 if not configured)
|
# 1. SPA checks if OIDC is available (returns 404 if not configured)
|
||||||
@@ -340,9 +340,8 @@ Key settings in `cameleer3-server-app/src/main/resources/application.yml`:
|
|||||||
|---------|---------|-------------|
|
|---------|---------|-------------|
|
||||||
| `server.port` | 8081 | Server port |
|
| `server.port` | 8081 | Server port |
|
||||||
| `ingestion.buffer-capacity` | 50000 | Max items in write buffer |
|
| `ingestion.buffer-capacity` | 50000 | Max items in write buffer |
|
||||||
| `ingestion.batch-size` | 5000 | Items per ClickHouse batch insert |
|
| `ingestion.batch-size` | 5000 | Items per batch insert |
|
||||||
| `ingestion.flush-interval-ms` | 1000 | Buffer flush interval (ms) |
|
| `ingestion.flush-interval-ms` | 1000 | Buffer flush interval (ms) |
|
||||||
| `ingestion.data-ttl-days` | 30 | ClickHouse TTL for auto-deletion |
|
|
||||||
| `agent-registry.heartbeat-interval-seconds` | 30 | Expected heartbeat interval |
|
| `agent-registry.heartbeat-interval-seconds` | 30 | Expected heartbeat interval |
|
||||||
| `agent-registry.stale-threshold-seconds` | 90 | Time before agent marked STALE |
|
| `agent-registry.stale-threshold-seconds` | 90 | Time before agent marked STALE |
|
||||||
| `agent-registry.dead-threshold-seconds` | 300 | Time after STALE before DEAD |
|
| `agent-registry.dead-threshold-seconds` | 300 | Time after STALE before DEAD |
|
||||||
@@ -386,7 +385,7 @@ npm run generate-api # Requires backend running on :8081
|
|||||||
|
|
||||||
## Running Tests
|
## Running Tests
|
||||||
|
|
||||||
Integration tests use Testcontainers (starts ClickHouse automatically — requires Docker):
|
Integration tests use Testcontainers (starts PostgreSQL and OpenSearch automatically — requires Docker):
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# All tests
|
# All tests
|
||||||
@@ -399,14 +398,13 @@ mvn test -pl cameleer3-server-core
|
|||||||
mvn test -pl cameleer3-server-app -Dtest=ExecutionControllerIT
|
mvn test -pl cameleer3-server-app -Dtest=ExecutionControllerIT
|
||||||
```
|
```
|
||||||
|
|
||||||
## Verify ClickHouse Data
|
## Verify Database Data
|
||||||
|
|
||||||
After posting data and waiting for the flush interval (1s default):
|
After posting data and waiting for the flush interval (1s default):
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker exec -it cameleer3-server-clickhouse-1 clickhouse-client \
|
docker exec -it cameleer3-server-postgres-1 psql -U cameleer -d cameleer3 \
|
||||||
--user cameleer --password cameleer_dev -d cameleer3 \
|
-c "SELECT count(*) FROM route_executions"
|
||||||
-q "SELECT count() FROM route_executions"
|
|
||||||
```
|
```
|
||||||
|
|
||||||
## Kubernetes Deployment
|
## Kubernetes Deployment
|
||||||
@@ -417,7 +415,8 @@ The full stack is deployed to k3s via CI/CD on push to `main`. K8s manifests are
|
|||||||
|
|
||||||
```
|
```
|
||||||
cameleer namespace:
|
cameleer namespace:
|
||||||
ClickHouse (StatefulSet, 2Gi PVC) ← clickhouse:8123 (ClusterIP)
|
PostgreSQL (StatefulSet, 10Gi PVC) ← postgres:5432 (ClusterIP)
|
||||||
|
OpenSearch (StatefulSet, 10Gi PVC) ← opensearch:9200 (ClusterIP)
|
||||||
cameleer3-server (Deployment) ← NodePort 30081
|
cameleer3-server (Deployment) ← NodePort 30081
|
||||||
cameleer3-ui (Deployment, Nginx) ← NodePort 30090
|
cameleer3-ui (Deployment, Nginx) ← NodePort 30090
|
||||||
Authentik Server (Deployment) ← NodePort 30950
|
Authentik Server (Deployment) ← NodePort 30950
|
||||||
@@ -439,7 +438,7 @@ cameleer namespace:
|
|||||||
|
|
||||||
Push to `main` triggers: **build** (UI npm + Maven, unit tests) → **docker** (buildx amd64 for server + UI, push to Gitea registry) → **deploy** (kubectl apply + rolling update).
|
Push to `main` triggers: **build** (UI npm + Maven, unit tests) → **docker** (buildx amd64 for server + UI, push to Gitea registry) → **deploy** (kubectl apply + rolling update).
|
||||||
|
|
||||||
Required Gitea org secrets: `REGISTRY_TOKEN`, `KUBECONFIG_BASE64`, `CAMELEER_AUTH_TOKEN`, `CAMELEER_JWT_SECRET`, `CLICKHOUSE_USER`, `CLICKHOUSE_PASSWORD`, `CAMELEER_UI_USER` (optional), `CAMELEER_UI_PASSWORD` (optional), `AUTHENTIK_PG_PASSWORD`, `AUTHENTIK_SECRET_KEY`, `CAMELEER_OIDC_ENABLED`, `CAMELEER_OIDC_ISSUER`, `CAMELEER_OIDC_CLIENT_ID`, `CAMELEER_OIDC_CLIENT_SECRET`.
|
Required Gitea org secrets: `REGISTRY_TOKEN`, `KUBECONFIG_BASE64`, `CAMELEER_AUTH_TOKEN`, `CAMELEER_JWT_SECRET`, `POSTGRES_USER`, `POSTGRES_PASSWORD`, `POSTGRES_DB`, `OPENSEARCH_USER`, `OPENSEARCH_PASSWORD`, `CAMELEER_UI_USER` (optional), `CAMELEER_UI_PASSWORD` (optional), `AUTHENTIK_PG_USER`, `AUTHENTIK_PG_PASSWORD`, `AUTHENTIK_SECRET_KEY`, `CAMELEER_OIDC_ENABLED`, `CAMELEER_OIDC_ISSUER`, `CAMELEER_OIDC_CLIENT_ID`, `CAMELEER_OIDC_CLIENT_SECRET`.
|
||||||
|
|
||||||
### Manual K8s Commands
|
### Manual K8s Commands
|
||||||
|
|
||||||
@@ -450,8 +449,11 @@ kubectl -n cameleer get pods
|
|||||||
# View server logs
|
# View server logs
|
||||||
kubectl -n cameleer logs -f deploy/cameleer3-server
|
kubectl -n cameleer logs -f deploy/cameleer3-server
|
||||||
|
|
||||||
# View ClickHouse logs
|
# View PostgreSQL logs
|
||||||
kubectl -n cameleer logs -f statefulset/clickhouse
|
kubectl -n cameleer logs -f statefulset/postgres
|
||||||
|
|
||||||
|
# View OpenSearch logs
|
||||||
|
kubectl -n cameleer logs -f statefulset/opensearch
|
||||||
|
|
||||||
# Restart server
|
# Restart server
|
||||||
kubectl -n cameleer rollout restart deployment/cameleer3-server
|
kubectl -n cameleer rollout restart deployment/cameleer3-server
|
||||||
|
|||||||
@@ -36,10 +36,26 @@
|
|||||||
<artifactId>spring-boot-starter-jdbc</artifactId>
|
<artifactId>spring-boot-starter-jdbc</artifactId>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>com.clickhouse</groupId>
|
<groupId>org.postgresql</groupId>
|
||||||
<artifactId>clickhouse-jdbc</artifactId>
|
<artifactId>postgresql</artifactId>
|
||||||
<version>0.9.7</version>
|
</dependency>
|
||||||
<classifier>all</classifier>
|
<dependency>
|
||||||
|
<groupId>org.flywaydb</groupId>
|
||||||
|
<artifactId>flyway-core</artifactId>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.flywaydb</groupId>
|
||||||
|
<artifactId>flyway-database-postgresql</artifactId>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.opensearch.client</groupId>
|
||||||
|
<artifactId>opensearch-java</artifactId>
|
||||||
|
<version>2.19.0</version>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.opensearch.client</groupId>
|
||||||
|
<artifactId>opensearch-rest-client</artifactId>
|
||||||
|
<version>2.19.0</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.springdoc</groupId>
|
<groupId>org.springdoc</groupId>
|
||||||
@@ -96,8 +112,18 @@
|
|||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.testcontainers</groupId>
|
<groupId>org.testcontainers</groupId>
|
||||||
<artifactId>testcontainers-clickhouse</artifactId>
|
<artifactId>testcontainers-postgresql</artifactId>
|
||||||
<version>2.0.3</version>
|
<scope>test</scope>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.testcontainers</groupId>
|
||||||
|
<artifactId>testcontainers-junit-jupiter</artifactId>
|
||||||
|
<scope>test</scope>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.opensearch</groupId>
|
||||||
|
<artifactId>opensearch-testcontainers</artifactId>
|
||||||
|
<version>2.1.1</version>
|
||||||
<scope>test</scope>
|
<scope>test</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
@@ -148,7 +174,7 @@
|
|||||||
<artifactId>maven-failsafe-plugin</artifactId>
|
<artifactId>maven-failsafe-plugin</artifactId>
|
||||||
<configuration>
|
<configuration>
|
||||||
<forkCount>1</forkCount>
|
<forkCount>1</forkCount>
|
||||||
<reuseForks>false</reuseForks>
|
<reuseForks>true</reuseForks>
|
||||||
</configuration>
|
</configuration>
|
||||||
<executions>
|
<executions>
|
||||||
<execution>
|
<execution>
|
||||||
|
|||||||
@@ -1,80 +0,0 @@
|
|||||||
package com.cameleer3.server.app.config;
|
|
||||||
|
|
||||||
import org.slf4j.Logger;
|
|
||||||
import org.slf4j.LoggerFactory;
|
|
||||||
import org.springframework.context.annotation.Bean;
|
|
||||||
import org.springframework.context.annotation.Configuration;
|
|
||||||
import org.springframework.core.io.Resource;
|
|
||||||
import org.springframework.core.io.support.PathMatchingResourcePatternResolver;
|
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
|
|
||||||
import jakarta.annotation.PostConstruct;
|
|
||||||
import javax.sql.DataSource;
|
|
||||||
import java.nio.charset.StandardCharsets;
|
|
||||||
import java.util.Arrays;
|
|
||||||
import java.util.Comparator;
|
|
||||||
import java.util.stream.Collectors;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* ClickHouse configuration.
|
|
||||||
* <p>
|
|
||||||
* Spring Boot auto-configures the DataSource from {@code spring.datasource.*} properties.
|
|
||||||
* This class exposes a JdbcTemplate bean and initializes the schema on startup.
|
|
||||||
* <p>
|
|
||||||
* The ClickHouse container's {@code CLICKHOUSE_DB} env var creates the database;
|
|
||||||
* this class creates the tables within it.
|
|
||||||
* <p>
|
|
||||||
* Migration files are discovered automatically from {@code classpath:clickhouse/*.sql}
|
|
||||||
* and executed in filename order (numeric prefix sort).
|
|
||||||
*/
|
|
||||||
@Configuration
|
|
||||||
public class ClickHouseConfig {
|
|
||||||
|
|
||||||
private static final Logger log = LoggerFactory.getLogger(ClickHouseConfig.class);
|
|
||||||
private static final String MIGRATION_PATTERN = "classpath:clickhouse/*.sql";
|
|
||||||
|
|
||||||
private final DataSource dataSource;
|
|
||||||
|
|
||||||
public ClickHouseConfig(DataSource dataSource) {
|
|
||||||
this.dataSource = dataSource;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Bean
|
|
||||||
public JdbcTemplate jdbcTemplate() {
|
|
||||||
return new JdbcTemplate(dataSource);
|
|
||||||
}
|
|
||||||
|
|
||||||
@PostConstruct
|
|
||||||
void initSchema() {
|
|
||||||
var jdbc = new JdbcTemplate(dataSource);
|
|
||||||
try {
|
|
||||||
Resource[] resources = new PathMatchingResourcePatternResolver()
|
|
||||||
.getResources(MIGRATION_PATTERN);
|
|
||||||
Arrays.sort(resources, Comparator.comparing(Resource::getFilename));
|
|
||||||
|
|
||||||
for (Resource resource : resources) {
|
|
||||||
String filename = resource.getFilename();
|
|
||||||
try {
|
|
||||||
String sql = resource.getContentAsString(StandardCharsets.UTF_8);
|
|
||||||
String stripped = sql.lines()
|
|
||||||
.filter(line -> !line.trim().startsWith("--"))
|
|
||||||
.collect(Collectors.joining("\n"));
|
|
||||||
for (String statement : stripped.split(";")) {
|
|
||||||
String trimmed = statement.trim();
|
|
||||||
if (!trimmed.isEmpty()) {
|
|
||||||
jdbc.execute(trimmed);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
log.info("Applied schema: {}", filename);
|
|
||||||
} catch (Exception e) {
|
|
||||||
log.error("Failed to apply schema: {}", filename, e);
|
|
||||||
throw new RuntimeException("Schema initialization failed: " + filename, e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} catch (RuntimeException e) {
|
|
||||||
throw e;
|
|
||||||
} catch (Exception e) {
|
|
||||||
throw new RuntimeException("Failed to discover migration files", e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,41 +1,22 @@
|
|||||||
package com.cameleer3.server.app.config;
|
package com.cameleer3.server.app.config;
|
||||||
|
|
||||||
import com.cameleer3.server.core.ingestion.IngestionService;
|
|
||||||
import com.cameleer3.server.core.ingestion.TaggedDiagram;
|
|
||||||
import com.cameleer3.server.core.ingestion.TaggedExecution;
|
|
||||||
import com.cameleer3.server.core.ingestion.WriteBuffer;
|
import com.cameleer3.server.core.ingestion.WriteBuffer;
|
||||||
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
||||||
import org.springframework.context.annotation.Bean;
|
import org.springframework.context.annotation.Bean;
|
||||||
import org.springframework.context.annotation.Configuration;
|
import org.springframework.context.annotation.Configuration;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Creates the write buffer and ingestion service beans.
|
* Creates the write buffer bean for metrics.
|
||||||
* <p>
|
* <p>
|
||||||
* The {@link WriteBuffer} instances are shared between the
|
* The {@link WriteBuffer} instance is shared between the
|
||||||
* {@link IngestionService} (producer side) and the flush scheduler (consumer side).
|
* {@link com.cameleer3.server.core.ingestion.IngestionService} (producer side)
|
||||||
|
* and the flush scheduler (consumer side).
|
||||||
*/
|
*/
|
||||||
@Configuration
|
@Configuration
|
||||||
public class IngestionBeanConfig {
|
public class IngestionBeanConfig {
|
||||||
|
|
||||||
@Bean
|
|
||||||
public WriteBuffer<TaggedExecution> executionBuffer(IngestionConfig config) {
|
|
||||||
return new WriteBuffer<>(config.getBufferCapacity());
|
|
||||||
}
|
|
||||||
|
|
||||||
@Bean
|
|
||||||
public WriteBuffer<TaggedDiagram> diagramBuffer(IngestionConfig config) {
|
|
||||||
return new WriteBuffer<>(config.getBufferCapacity());
|
|
||||||
}
|
|
||||||
|
|
||||||
@Bean
|
@Bean
|
||||||
public WriteBuffer<MetricsSnapshot> metricsBuffer(IngestionConfig config) {
|
public WriteBuffer<MetricsSnapshot> metricsBuffer(IngestionConfig config) {
|
||||||
return new WriteBuffer<>(config.getBufferCapacity());
|
return new WriteBuffer<>(config.getBufferCapacity());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Bean
|
|
||||||
public IngestionService ingestionService(WriteBuffer<TaggedExecution> executionBuffer,
|
|
||||||
WriteBuffer<TaggedDiagram> diagramBuffer,
|
|
||||||
WriteBuffer<MetricsSnapshot> metricsBuffer) {
|
|
||||||
return new IngestionService(executionBuffer, diagramBuffer, metricsBuffer);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,28 @@
|
|||||||
|
package com.cameleer3.server.app.config;
|
||||||
|
|
||||||
|
import org.apache.http.HttpHost;
|
||||||
|
import org.opensearch.client.RestClient;
|
||||||
|
import org.opensearch.client.json.jackson.JacksonJsonpMapper;
|
||||||
|
import org.opensearch.client.opensearch.OpenSearchClient;
|
||||||
|
import org.opensearch.client.transport.rest_client.RestClientTransport;
|
||||||
|
import org.springframework.beans.factory.annotation.Value;
|
||||||
|
import org.springframework.context.annotation.Bean;
|
||||||
|
import org.springframework.context.annotation.Configuration;
|
||||||
|
|
||||||
|
@Configuration
|
||||||
|
public class OpenSearchConfig {
|
||||||
|
|
||||||
|
@Value("${opensearch.url:http://localhost:9200}")
|
||||||
|
private String opensearchUrl;
|
||||||
|
|
||||||
|
@Bean(destroyMethod = "close")
|
||||||
|
public RestClient opensearchRestClient() {
|
||||||
|
return RestClient.builder(HttpHost.create(opensearchUrl)).build();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Bean
|
||||||
|
public OpenSearchClient openSearchClient(RestClient restClient) {
|
||||||
|
var transport = new RestClientTransport(restClient, new JacksonJsonpMapper());
|
||||||
|
return new OpenSearchClient(transport);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,32 +1,19 @@
|
|||||||
package com.cameleer3.server.app.config;
|
package com.cameleer3.server.app.config;
|
||||||
|
|
||||||
import com.cameleer3.server.app.search.ClickHouseSearchEngine;
|
|
||||||
import com.cameleer3.server.core.detail.DetailService;
|
|
||||||
import com.cameleer3.server.core.search.SearchEngine;
|
|
||||||
import com.cameleer3.server.core.search.SearchService;
|
import com.cameleer3.server.core.search.SearchService;
|
||||||
import com.cameleer3.server.core.storage.ExecutionRepository;
|
import com.cameleer3.server.core.storage.SearchIndex;
|
||||||
|
import com.cameleer3.server.core.storage.StatsStore;
|
||||||
import org.springframework.context.annotation.Bean;
|
import org.springframework.context.annotation.Bean;
|
||||||
import org.springframework.context.annotation.Configuration;
|
import org.springframework.context.annotation.Configuration;
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Creates beans for the search and detail layers.
|
* Creates beans for the search layer.
|
||||||
*/
|
*/
|
||||||
@Configuration
|
@Configuration
|
||||||
public class SearchBeanConfig {
|
public class SearchBeanConfig {
|
||||||
|
|
||||||
@Bean
|
@Bean
|
||||||
public SearchEngine searchEngine(JdbcTemplate jdbcTemplate) {
|
public SearchService searchService(SearchIndex searchIndex, StatsStore statsStore) {
|
||||||
return new ClickHouseSearchEngine(jdbcTemplate);
|
return new SearchService(searchIndex, statsStore);
|
||||||
}
|
|
||||||
|
|
||||||
@Bean
|
|
||||||
public SearchService searchService(SearchEngine searchEngine) {
|
|
||||||
return new SearchService(searchEngine);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Bean
|
|
||||||
public DetailService detailService(ExecutionRepository executionRepository) {
|
|
||||||
return new DetailService(executionRepository);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,37 @@
|
|||||||
|
package com.cameleer3.server.app.config;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.detail.DetailService;
|
||||||
|
import com.cameleer3.server.core.indexing.SearchIndexer;
|
||||||
|
import com.cameleer3.server.core.ingestion.IngestionService;
|
||||||
|
import com.cameleer3.server.core.ingestion.WriteBuffer;
|
||||||
|
import com.cameleer3.server.core.storage.*;
|
||||||
|
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
||||||
|
import org.springframework.beans.factory.annotation.Value;
|
||||||
|
import org.springframework.context.annotation.Bean;
|
||||||
|
import org.springframework.context.annotation.Configuration;
|
||||||
|
|
||||||
|
@Configuration
|
||||||
|
public class StorageBeanConfig {
|
||||||
|
|
||||||
|
@Bean
|
||||||
|
public DetailService detailService(ExecutionStore executionStore) {
|
||||||
|
return new DetailService(executionStore);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Bean(destroyMethod = "shutdown")
|
||||||
|
public SearchIndexer searchIndexer(ExecutionStore executionStore, SearchIndex searchIndex,
|
||||||
|
@Value("${opensearch.debounce-ms:2000}") long debounceMs,
|
||||||
|
@Value("${opensearch.queue-size:10000}") int queueSize) {
|
||||||
|
return new SearchIndexer(executionStore, searchIndex, debounceMs, queueSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Bean
|
||||||
|
public IngestionService ingestionService(ExecutionStore executionStore,
|
||||||
|
DiagramStore diagramStore,
|
||||||
|
WriteBuffer<MetricsSnapshot> metricsBuffer,
|
||||||
|
SearchIndexer searchIndexer,
|
||||||
|
@Value("${cameleer.body-size-limit:16384}") int bodySizeLimit) {
|
||||||
|
return new IngestionService(executionStore, diagramStore, metricsBuffer,
|
||||||
|
searchIndexer::onExecutionUpdated, bodySizeLimit);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,8 +1,9 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.storage.ClickHouseExecutionRepository;
|
|
||||||
import com.cameleer3.server.core.detail.DetailService;
|
import com.cameleer3.server.core.detail.DetailService;
|
||||||
import com.cameleer3.server.core.detail.ExecutionDetail;
|
import com.cameleer3.server.core.detail.ExecutionDetail;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ProcessorRecord;
|
||||||
import io.swagger.v3.oas.annotations.Operation;
|
import io.swagger.v3.oas.annotations.Operation;
|
||||||
import io.swagger.v3.oas.annotations.responses.ApiResponse;
|
import io.swagger.v3.oas.annotations.responses.ApiResponse;
|
||||||
import io.swagger.v3.oas.annotations.tags.Tag;
|
import io.swagger.v3.oas.annotations.tags.Tag;
|
||||||
@@ -12,14 +13,16 @@ import org.springframework.web.bind.annotation.PathVariable;
|
|||||||
import org.springframework.web.bind.annotation.RequestMapping;
|
import org.springframework.web.bind.annotation.RequestMapping;
|
||||||
import org.springframework.web.bind.annotation.RestController;
|
import org.springframework.web.bind.annotation.RestController;
|
||||||
|
|
||||||
|
import java.util.LinkedHashMap;
|
||||||
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Endpoints for retrieving execution details and processor snapshots.
|
* Endpoints for retrieving execution details and processor snapshots.
|
||||||
* <p>
|
* <p>
|
||||||
* The detail endpoint returns a nested processor tree reconstructed from
|
* The detail endpoint returns a nested processor tree reconstructed from
|
||||||
* flat parallel arrays stored in ClickHouse. The snapshot endpoint returns
|
* individual processor records stored in PostgreSQL. The snapshot endpoint
|
||||||
* per-processor exchange data (bodies and headers).
|
* returns per-processor exchange data (bodies and headers).
|
||||||
*/
|
*/
|
||||||
@RestController
|
@RestController
|
||||||
@RequestMapping("/api/v1/executions")
|
@RequestMapping("/api/v1/executions")
|
||||||
@@ -27,12 +30,12 @@ import java.util.Map;
|
|||||||
public class DetailController {
|
public class DetailController {
|
||||||
|
|
||||||
private final DetailService detailService;
|
private final DetailService detailService;
|
||||||
private final ClickHouseExecutionRepository executionRepository;
|
private final ExecutionStore executionStore;
|
||||||
|
|
||||||
public DetailController(DetailService detailService,
|
public DetailController(DetailService detailService,
|
||||||
ClickHouseExecutionRepository executionRepository) {
|
ExecutionStore executionStore) {
|
||||||
this.detailService = detailService;
|
this.detailService = detailService;
|
||||||
this.executionRepository = executionRepository;
|
this.executionStore = executionStore;
|
||||||
}
|
}
|
||||||
|
|
||||||
@GetMapping("/{executionId}")
|
@GetMapping("/{executionId}")
|
||||||
@@ -52,8 +55,18 @@ public class DetailController {
|
|||||||
public ResponseEntity<Map<String, String>> getProcessorSnapshot(
|
public ResponseEntity<Map<String, String>> getProcessorSnapshot(
|
||||||
@PathVariable String executionId,
|
@PathVariable String executionId,
|
||||||
@PathVariable int index) {
|
@PathVariable int index) {
|
||||||
return executionRepository.findProcessorSnapshot(executionId, index)
|
List<ProcessorRecord> processors = executionStore.findProcessors(executionId);
|
||||||
.map(ResponseEntity::ok)
|
if (index < 0 || index >= processors.size()) {
|
||||||
.orElse(ResponseEntity.notFound().build());
|
return ResponseEntity.notFound().build();
|
||||||
|
}
|
||||||
|
|
||||||
|
ProcessorRecord p = processors.get(index);
|
||||||
|
Map<String, String> snapshot = new LinkedHashMap<>();
|
||||||
|
if (p.inputBody() != null) snapshot.put("inputBody", p.inputBody());
|
||||||
|
if (p.outputBody() != null) snapshot.put("outputBody", p.outputBody());
|
||||||
|
if (p.inputHeaders() != null) snapshot.put("inputHeaders", p.inputHeaders());
|
||||||
|
if (p.outputHeaders() != null) snapshot.put("outputHeaders", p.outputHeaders());
|
||||||
|
|
||||||
|
return ResponseEntity.ok(snapshot);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -11,7 +11,6 @@ import io.swagger.v3.oas.annotations.responses.ApiResponse;
|
|||||||
import io.swagger.v3.oas.annotations.tags.Tag;
|
import io.swagger.v3.oas.annotations.tags.Tag;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
import org.springframework.http.HttpStatus;
|
|
||||||
import org.springframework.http.ResponseEntity;
|
import org.springframework.http.ResponseEntity;
|
||||||
import org.springframework.security.core.Authentication;
|
import org.springframework.security.core.Authentication;
|
||||||
import org.springframework.security.core.context.SecurityContextHolder;
|
import org.springframework.security.core.context.SecurityContextHolder;
|
||||||
@@ -25,8 +24,8 @@ import java.util.List;
|
|||||||
/**
|
/**
|
||||||
* Ingestion endpoint for route diagrams.
|
* Ingestion endpoint for route diagrams.
|
||||||
* <p>
|
* <p>
|
||||||
* Accepts both single {@link RouteGraph} and arrays. Data is buffered
|
* Accepts both single {@link RouteGraph} and arrays. Data is written
|
||||||
* and flushed to ClickHouse by the flush scheduler.
|
* synchronously to PostgreSQL via {@link IngestionService}.
|
||||||
*/
|
*/
|
||||||
@RestController
|
@RestController
|
||||||
@RequestMapping("/api/v1/data")
|
@RequestMapping("/api/v1/data")
|
||||||
@@ -47,26 +46,12 @@ public class DiagramController {
|
|||||||
@Operation(summary = "Ingest route diagram data",
|
@Operation(summary = "Ingest route diagram data",
|
||||||
description = "Accepts a single RouteGraph or an array of RouteGraphs")
|
description = "Accepts a single RouteGraph or an array of RouteGraphs")
|
||||||
@ApiResponse(responseCode = "202", description = "Data accepted for processing")
|
@ApiResponse(responseCode = "202", description = "Data accepted for processing")
|
||||||
@ApiResponse(responseCode = "503", description = "Buffer full, retry later")
|
|
||||||
public ResponseEntity<Void> ingestDiagrams(@RequestBody String body) throws JsonProcessingException {
|
public ResponseEntity<Void> ingestDiagrams(@RequestBody String body) throws JsonProcessingException {
|
||||||
String agentId = extractAgentId();
|
String agentId = extractAgentId();
|
||||||
List<RouteGraph> graphs = parsePayload(body);
|
List<RouteGraph> graphs = parsePayload(body);
|
||||||
List<TaggedDiagram> tagged = graphs.stream()
|
|
||||||
.map(graph -> new TaggedDiagram(agentId, graph))
|
|
||||||
.toList();
|
|
||||||
|
|
||||||
boolean accepted;
|
for (RouteGraph graph : graphs) {
|
||||||
if (tagged.size() == 1) {
|
ingestionService.ingestDiagram(new TaggedDiagram(agentId, graph));
|
||||||
accepted = ingestionService.acceptDiagram(tagged.get(0));
|
|
||||||
} else {
|
|
||||||
accepted = ingestionService.acceptDiagrams(tagged);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!accepted) {
|
|
||||||
log.warn("Diagram buffer full, returning 503");
|
|
||||||
return ResponseEntity.status(HttpStatus.SERVICE_UNAVAILABLE)
|
|
||||||
.header("Retry-After", "5")
|
|
||||||
.build();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return ResponseEntity.accepted().build();
|
return ResponseEntity.accepted().build();
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import com.cameleer3.server.core.agent.AgentInfo;
|
|||||||
import com.cameleer3.server.core.agent.AgentRegistryService;
|
import com.cameleer3.server.core.agent.AgentRegistryService;
|
||||||
import com.cameleer3.server.core.diagram.DiagramLayout;
|
import com.cameleer3.server.core.diagram.DiagramLayout;
|
||||||
import com.cameleer3.server.core.diagram.DiagramRenderer;
|
import com.cameleer3.server.core.diagram.DiagramRenderer;
|
||||||
import com.cameleer3.server.core.storage.DiagramRepository;
|
import com.cameleer3.server.core.storage.DiagramStore;
|
||||||
import io.swagger.v3.oas.annotations.Operation;
|
import io.swagger.v3.oas.annotations.Operation;
|
||||||
import io.swagger.v3.oas.annotations.media.Content;
|
import io.swagger.v3.oas.annotations.media.Content;
|
||||||
import io.swagger.v3.oas.annotations.media.Schema;
|
import io.swagger.v3.oas.annotations.media.Schema;
|
||||||
@@ -39,14 +39,14 @@ public class DiagramRenderController {
|
|||||||
|
|
||||||
private static final MediaType SVG_MEDIA_TYPE = MediaType.valueOf("image/svg+xml");
|
private static final MediaType SVG_MEDIA_TYPE = MediaType.valueOf("image/svg+xml");
|
||||||
|
|
||||||
private final DiagramRepository diagramRepository;
|
private final DiagramStore diagramStore;
|
||||||
private final DiagramRenderer diagramRenderer;
|
private final DiagramRenderer diagramRenderer;
|
||||||
private final AgentRegistryService registryService;
|
private final AgentRegistryService registryService;
|
||||||
|
|
||||||
public DiagramRenderController(DiagramRepository diagramRepository,
|
public DiagramRenderController(DiagramStore diagramStore,
|
||||||
DiagramRenderer diagramRenderer,
|
DiagramRenderer diagramRenderer,
|
||||||
AgentRegistryService registryService) {
|
AgentRegistryService registryService) {
|
||||||
this.diagramRepository = diagramRepository;
|
this.diagramStore = diagramStore;
|
||||||
this.diagramRenderer = diagramRenderer;
|
this.diagramRenderer = diagramRenderer;
|
||||||
this.registryService = registryService;
|
this.registryService = registryService;
|
||||||
}
|
}
|
||||||
@@ -64,7 +64,7 @@ public class DiagramRenderController {
|
|||||||
@PathVariable String contentHash,
|
@PathVariable String contentHash,
|
||||||
HttpServletRequest request) {
|
HttpServletRequest request) {
|
||||||
|
|
||||||
Optional<RouteGraph> graphOpt = diagramRepository.findByContentHash(contentHash);
|
Optional<RouteGraph> graphOpt = diagramStore.findByContentHash(contentHash);
|
||||||
if (graphOpt.isEmpty()) {
|
if (graphOpt.isEmpty()) {
|
||||||
return ResponseEntity.notFound().build();
|
return ResponseEntity.notFound().build();
|
||||||
}
|
}
|
||||||
@@ -105,12 +105,12 @@ public class DiagramRenderController {
|
|||||||
return ResponseEntity.notFound().build();
|
return ResponseEntity.notFound().build();
|
||||||
}
|
}
|
||||||
|
|
||||||
Optional<String> contentHash = diagramRepository.findContentHashForRouteByAgents(routeId, agentIds);
|
Optional<String> contentHash = diagramStore.findContentHashForRouteByAgents(routeId, agentIds);
|
||||||
if (contentHash.isEmpty()) {
|
if (contentHash.isEmpty()) {
|
||||||
return ResponseEntity.notFound().build();
|
return ResponseEntity.notFound().build();
|
||||||
}
|
}
|
||||||
|
|
||||||
Optional<RouteGraph> graphOpt = diagramRepository.findByContentHash(contentHash.get());
|
Optional<RouteGraph> graphOpt = diagramStore.findByContentHash(contentHash.get());
|
||||||
if (graphOpt.isEmpty()) {
|
if (graphOpt.isEmpty()) {
|
||||||
return ResponseEntity.notFound().build();
|
return ResponseEntity.notFound().build();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,8 +1,9 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.common.model.RouteExecution;
|
import com.cameleer3.common.model.RouteExecution;
|
||||||
|
import com.cameleer3.server.core.agent.AgentInfo;
|
||||||
|
import com.cameleer3.server.core.agent.AgentRegistryService;
|
||||||
import com.cameleer3.server.core.ingestion.IngestionService;
|
import com.cameleer3.server.core.ingestion.IngestionService;
|
||||||
import com.cameleer3.server.core.ingestion.TaggedExecution;
|
|
||||||
import com.fasterxml.jackson.core.JsonProcessingException;
|
import com.fasterxml.jackson.core.JsonProcessingException;
|
||||||
import com.fasterxml.jackson.core.type.TypeReference;
|
import com.fasterxml.jackson.core.type.TypeReference;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
@@ -11,7 +12,6 @@ import io.swagger.v3.oas.annotations.responses.ApiResponse;
|
|||||||
import io.swagger.v3.oas.annotations.tags.Tag;
|
import io.swagger.v3.oas.annotations.tags.Tag;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
import org.springframework.http.HttpStatus;
|
|
||||||
import org.springframework.http.ResponseEntity;
|
import org.springframework.http.ResponseEntity;
|
||||||
import org.springframework.security.core.Authentication;
|
import org.springframework.security.core.Authentication;
|
||||||
import org.springframework.security.core.context.SecurityContextHolder;
|
import org.springframework.security.core.context.SecurityContextHolder;
|
||||||
@@ -25,9 +25,8 @@ import java.util.List;
|
|||||||
/**
|
/**
|
||||||
* Ingestion endpoint for route execution data.
|
* Ingestion endpoint for route execution data.
|
||||||
* <p>
|
* <p>
|
||||||
* Accepts both single {@link RouteExecution} and arrays. Data is buffered
|
* Accepts both single {@link RouteExecution} and arrays. Data is written
|
||||||
* in a {@link com.cameleer3.server.core.ingestion.WriteBuffer} and flushed
|
* synchronously to PostgreSQL via {@link IngestionService}.
|
||||||
* to ClickHouse by the flush scheduler.
|
|
||||||
*/
|
*/
|
||||||
@RestController
|
@RestController
|
||||||
@RequestMapping("/api/v1/data")
|
@RequestMapping("/api/v1/data")
|
||||||
@@ -37,10 +36,14 @@ public class ExecutionController {
|
|||||||
private static final Logger log = LoggerFactory.getLogger(ExecutionController.class);
|
private static final Logger log = LoggerFactory.getLogger(ExecutionController.class);
|
||||||
|
|
||||||
private final IngestionService ingestionService;
|
private final IngestionService ingestionService;
|
||||||
|
private final AgentRegistryService registryService;
|
||||||
private final ObjectMapper objectMapper;
|
private final ObjectMapper objectMapper;
|
||||||
|
|
||||||
public ExecutionController(IngestionService ingestionService, ObjectMapper objectMapper) {
|
public ExecutionController(IngestionService ingestionService,
|
||||||
|
AgentRegistryService registryService,
|
||||||
|
ObjectMapper objectMapper) {
|
||||||
this.ingestionService = ingestionService;
|
this.ingestionService = ingestionService;
|
||||||
|
this.registryService = registryService;
|
||||||
this.objectMapper = objectMapper;
|
this.objectMapper = objectMapper;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -48,26 +51,13 @@ public class ExecutionController {
|
|||||||
@Operation(summary = "Ingest route execution data",
|
@Operation(summary = "Ingest route execution data",
|
||||||
description = "Accepts a single RouteExecution or an array of RouteExecutions")
|
description = "Accepts a single RouteExecution or an array of RouteExecutions")
|
||||||
@ApiResponse(responseCode = "202", description = "Data accepted for processing")
|
@ApiResponse(responseCode = "202", description = "Data accepted for processing")
|
||||||
@ApiResponse(responseCode = "503", description = "Buffer full, retry later")
|
|
||||||
public ResponseEntity<Void> ingestExecutions(@RequestBody String body) throws JsonProcessingException {
|
public ResponseEntity<Void> ingestExecutions(@RequestBody String body) throws JsonProcessingException {
|
||||||
String agentId = extractAgentId();
|
String agentId = extractAgentId();
|
||||||
|
String groupName = resolveGroupName(agentId);
|
||||||
List<RouteExecution> executions = parsePayload(body);
|
List<RouteExecution> executions = parsePayload(body);
|
||||||
List<TaggedExecution> tagged = executions.stream()
|
|
||||||
.map(exec -> new TaggedExecution(agentId, exec))
|
|
||||||
.toList();
|
|
||||||
|
|
||||||
boolean accepted;
|
for (RouteExecution execution : executions) {
|
||||||
if (tagged.size() == 1) {
|
ingestionService.ingestExecution(agentId, groupName, execution);
|
||||||
accepted = ingestionService.acceptExecution(tagged.get(0));
|
|
||||||
} else {
|
|
||||||
accepted = ingestionService.acceptExecutions(tagged);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!accepted) {
|
|
||||||
log.warn("Execution buffer full, returning 503");
|
|
||||||
return ResponseEntity.status(HttpStatus.SERVICE_UNAVAILABLE)
|
|
||||||
.header("Retry-After", "5")
|
|
||||||
.build();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return ResponseEntity.accepted().build();
|
return ResponseEntity.accepted().build();
|
||||||
@@ -78,6 +68,11 @@ public class ExecutionController {
|
|||||||
return auth != null ? auth.getName() : "";
|
return auth != null ? auth.getName() : "";
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private String resolveGroupName(String agentId) {
|
||||||
|
AgentInfo agent = registryService.findById(agentId);
|
||||||
|
return agent != null ? agent.group() : "";
|
||||||
|
}
|
||||||
|
|
||||||
private List<RouteExecution> parsePayload(String body) throws JsonProcessingException {
|
private List<RouteExecution> parsePayload(String body) throws JsonProcessingException {
|
||||||
String trimmed = body.strip();
|
String trimmed = body.strip();
|
||||||
if (trimmed.startsWith("[")) {
|
if (trimmed.startsWith("[")) {
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ import java.util.List;
|
|||||||
* Ingestion endpoint for agent metrics.
|
* Ingestion endpoint for agent metrics.
|
||||||
* <p>
|
* <p>
|
||||||
* Accepts an array of {@link MetricsSnapshot}. Data is buffered
|
* Accepts an array of {@link MetricsSnapshot}. Data is buffered
|
||||||
* and flushed to ClickHouse by the flush scheduler.
|
* and flushed to PostgreSQL by the flush scheduler.
|
||||||
*/
|
*/
|
||||||
@RestController
|
@RestController
|
||||||
@RequestMapping("/api/v1/data")
|
@RequestMapping("/api/v1/data")
|
||||||
|
|||||||
@@ -1,159 +0,0 @@
|
|||||||
package com.cameleer3.server.app.ingestion;
|
|
||||||
|
|
||||||
import com.cameleer3.server.app.config.IngestionConfig;
|
|
||||||
import com.cameleer3.server.core.ingestion.TaggedDiagram;
|
|
||||||
import com.cameleer3.server.core.ingestion.TaggedExecution;
|
|
||||||
import com.cameleer3.server.core.ingestion.WriteBuffer;
|
|
||||||
import com.cameleer3.server.core.storage.DiagramRepository;
|
|
||||||
import com.cameleer3.server.core.storage.ExecutionRepository;
|
|
||||||
import com.cameleer3.server.core.storage.MetricsRepository;
|
|
||||||
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
|
||||||
import org.slf4j.Logger;
|
|
||||||
import org.slf4j.LoggerFactory;
|
|
||||||
import org.springframework.context.SmartLifecycle;
|
|
||||||
import org.springframework.scheduling.annotation.Scheduled;
|
|
||||||
import org.springframework.stereotype.Component;
|
|
||||||
|
|
||||||
import java.util.List;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Scheduled task that drains the write buffers and batch-inserts into ClickHouse.
|
|
||||||
* <p>
|
|
||||||
* Implements {@link SmartLifecycle} to ensure all remaining buffered data is
|
|
||||||
* flushed on application shutdown.
|
|
||||||
*/
|
|
||||||
@Component
|
|
||||||
public class ClickHouseFlushScheduler implements SmartLifecycle {
|
|
||||||
|
|
||||||
private static final Logger log = LoggerFactory.getLogger(ClickHouseFlushScheduler.class);
|
|
||||||
|
|
||||||
private final WriteBuffer<TaggedExecution> executionBuffer;
|
|
||||||
private final WriteBuffer<TaggedDiagram> diagramBuffer;
|
|
||||||
private final WriteBuffer<MetricsSnapshot> metricsBuffer;
|
|
||||||
private final ExecutionRepository executionRepository;
|
|
||||||
private final DiagramRepository diagramRepository;
|
|
||||||
private final MetricsRepository metricsRepository;
|
|
||||||
private final int batchSize;
|
|
||||||
|
|
||||||
private volatile boolean running = false;
|
|
||||||
|
|
||||||
public ClickHouseFlushScheduler(WriteBuffer<TaggedExecution> executionBuffer,
|
|
||||||
WriteBuffer<TaggedDiagram> diagramBuffer,
|
|
||||||
WriteBuffer<MetricsSnapshot> metricsBuffer,
|
|
||||||
ExecutionRepository executionRepository,
|
|
||||||
DiagramRepository diagramRepository,
|
|
||||||
MetricsRepository metricsRepository,
|
|
||||||
IngestionConfig config) {
|
|
||||||
this.executionBuffer = executionBuffer;
|
|
||||||
this.diagramBuffer = diagramBuffer;
|
|
||||||
this.metricsBuffer = metricsBuffer;
|
|
||||||
this.executionRepository = executionRepository;
|
|
||||||
this.diagramRepository = diagramRepository;
|
|
||||||
this.metricsRepository = metricsRepository;
|
|
||||||
this.batchSize = config.getBatchSize();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Scheduled(fixedDelayString = "${ingestion.flush-interval-ms:1000}")
|
|
||||||
public void flushAll() {
|
|
||||||
flushExecutions();
|
|
||||||
flushDiagrams();
|
|
||||||
flushMetrics();
|
|
||||||
}
|
|
||||||
|
|
||||||
private void flushExecutions() {
|
|
||||||
try {
|
|
||||||
List<TaggedExecution> batch = executionBuffer.drain(batchSize);
|
|
||||||
if (!batch.isEmpty()) {
|
|
||||||
executionRepository.insertBatch(batch);
|
|
||||||
log.debug("Flushed {} executions to ClickHouse", batch.size());
|
|
||||||
}
|
|
||||||
} catch (Exception e) {
|
|
||||||
log.error("Failed to flush executions to ClickHouse", e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private void flushDiagrams() {
|
|
||||||
try {
|
|
||||||
List<TaggedDiagram> batch = diagramBuffer.drain(batchSize);
|
|
||||||
for (TaggedDiagram diagram : batch) {
|
|
||||||
diagramRepository.store(diagram);
|
|
||||||
}
|
|
||||||
if (!batch.isEmpty()) {
|
|
||||||
log.debug("Flushed {} diagrams to ClickHouse", batch.size());
|
|
||||||
}
|
|
||||||
} catch (Exception e) {
|
|
||||||
log.error("Failed to flush diagrams to ClickHouse", e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private void flushMetrics() {
|
|
||||||
try {
|
|
||||||
List<MetricsSnapshot> batch = metricsBuffer.drain(batchSize);
|
|
||||||
if (!batch.isEmpty()) {
|
|
||||||
metricsRepository.insertBatch(batch);
|
|
||||||
log.debug("Flushed {} metrics to ClickHouse", batch.size());
|
|
||||||
}
|
|
||||||
} catch (Exception e) {
|
|
||||||
log.error("Failed to flush metrics to ClickHouse", e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// SmartLifecycle -- flush remaining data on shutdown
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void start() {
|
|
||||||
running = true;
|
|
||||||
log.info("ClickHouseFlushScheduler started");
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void stop() {
|
|
||||||
log.info("ClickHouseFlushScheduler stopping -- flushing remaining data");
|
|
||||||
drainAll();
|
|
||||||
running = false;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public boolean isRunning() {
|
|
||||||
return running;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int getPhase() {
|
|
||||||
// Run after most beans but before DataSource shutdown
|
|
||||||
return Integer.MAX_VALUE - 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Drain all buffers completely (loop until empty).
|
|
||||||
*/
|
|
||||||
private void drainAll() {
|
|
||||||
drainBufferCompletely("executions", executionBuffer, batch -> executionRepository.insertBatch(batch));
|
|
||||||
drainBufferCompletely("diagrams", diagramBuffer, batch -> {
|
|
||||||
for (TaggedDiagram d : batch) {
|
|
||||||
diagramRepository.store(d);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
drainBufferCompletely("metrics", metricsBuffer, batch -> metricsRepository.insertBatch(batch));
|
|
||||||
}
|
|
||||||
|
|
||||||
private <T> void drainBufferCompletely(String name, WriteBuffer<T> buffer, java.util.function.Consumer<List<T>> inserter) {
|
|
||||||
int total = 0;
|
|
||||||
while (buffer.size() > 0) {
|
|
||||||
List<T> batch = buffer.drain(batchSize);
|
|
||||||
if (batch.isEmpty()) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
try {
|
|
||||||
inserter.accept(batch);
|
|
||||||
total += batch.size();
|
|
||||||
} catch (Exception e) {
|
|
||||||
log.error("Failed to flush remaining {} during shutdown", name, e);
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (total > 0) {
|
|
||||||
log.info("Flushed {} remaining {} during shutdown", total, name);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,59 @@
|
|||||||
|
package com.cameleer3.server.app.ingestion;
|
||||||
|
|
||||||
|
import com.cameleer3.server.app.config.IngestionConfig;
|
||||||
|
import com.cameleer3.server.core.ingestion.WriteBuffer;
|
||||||
|
import com.cameleer3.server.core.storage.MetricsStore;
|
||||||
|
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
import org.springframework.context.SmartLifecycle;
|
||||||
|
import org.springframework.scheduling.annotation.Scheduled;
|
||||||
|
import org.springframework.stereotype.Component;
|
||||||
|
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
@Component
|
||||||
|
public class MetricsFlushScheduler implements SmartLifecycle {
|
||||||
|
|
||||||
|
private static final Logger log = LoggerFactory.getLogger(MetricsFlushScheduler.class);
|
||||||
|
|
||||||
|
private final WriteBuffer<MetricsSnapshot> metricsBuffer;
|
||||||
|
private final MetricsStore metricsStore;
|
||||||
|
private final int batchSize;
|
||||||
|
private volatile boolean running = false;
|
||||||
|
|
||||||
|
public MetricsFlushScheduler(WriteBuffer<MetricsSnapshot> metricsBuffer,
|
||||||
|
MetricsStore metricsStore,
|
||||||
|
IngestionConfig config) {
|
||||||
|
this.metricsBuffer = metricsBuffer;
|
||||||
|
this.metricsStore = metricsStore;
|
||||||
|
this.batchSize = config.getBatchSize();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Scheduled(fixedDelayString = "${ingestion.flush-interval-ms:1000}")
|
||||||
|
public void flush() {
|
||||||
|
try {
|
||||||
|
List<MetricsSnapshot> batch = metricsBuffer.drain(batchSize);
|
||||||
|
if (!batch.isEmpty()) {
|
||||||
|
metricsStore.insertBatch(batch);
|
||||||
|
log.debug("Flushed {} metrics to PostgreSQL", batch.size());
|
||||||
|
}
|
||||||
|
} catch (Exception e) {
|
||||||
|
log.error("Failed to flush metrics", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override public void start() { running = true; }
|
||||||
|
@Override public void stop() {
|
||||||
|
// Drain remaining on shutdown
|
||||||
|
while (metricsBuffer.size() > 0) {
|
||||||
|
List<MetricsSnapshot> batch = metricsBuffer.drain(batchSize);
|
||||||
|
if (batch.isEmpty()) break;
|
||||||
|
try { metricsStore.insertBatch(batch); }
|
||||||
|
catch (Exception e) { log.error("Failed to flush metrics during shutdown", e); break; }
|
||||||
|
}
|
||||||
|
running = false;
|
||||||
|
}
|
||||||
|
@Override public boolean isRunning() { return running; }
|
||||||
|
@Override public int getPhase() { return Integer.MAX_VALUE - 1; }
|
||||||
|
}
|
||||||
@@ -0,0 +1,48 @@
|
|||||||
|
package com.cameleer3.server.app.retention;
|
||||||
|
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
import org.springframework.beans.factory.annotation.Value;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
import org.springframework.scheduling.annotation.Scheduled;
|
||||||
|
import org.springframework.stereotype.Component;
|
||||||
|
|
||||||
|
@Component
|
||||||
|
public class RetentionScheduler {
|
||||||
|
|
||||||
|
private static final Logger log = LoggerFactory.getLogger(RetentionScheduler.class);
|
||||||
|
|
||||||
|
private final JdbcTemplate jdbc;
|
||||||
|
private final int retentionDays;
|
||||||
|
|
||||||
|
public RetentionScheduler(JdbcTemplate jdbc,
|
||||||
|
@Value("${cameleer.retention-days:30}") int retentionDays) {
|
||||||
|
this.jdbc = jdbc;
|
||||||
|
this.retentionDays = retentionDays;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Scheduled(cron = "0 0 2 * * *") // Daily at 2 AM UTC
|
||||||
|
public void dropExpiredChunks() {
|
||||||
|
String interval = retentionDays + " days";
|
||||||
|
try {
|
||||||
|
// Raw data
|
||||||
|
jdbc.execute("SELECT drop_chunks('executions', INTERVAL '" + interval + "')");
|
||||||
|
jdbc.execute("SELECT drop_chunks('processor_executions', INTERVAL '" + interval + "')");
|
||||||
|
jdbc.execute("SELECT drop_chunks('agent_metrics', INTERVAL '" + interval + "')");
|
||||||
|
|
||||||
|
// Continuous aggregates (keep 3x longer)
|
||||||
|
String caggInterval = (retentionDays * 3) + " days";
|
||||||
|
jdbc.execute("SELECT drop_chunks('stats_1m_all', INTERVAL '" + caggInterval + "')");
|
||||||
|
jdbc.execute("SELECT drop_chunks('stats_1m_app', INTERVAL '" + caggInterval + "')");
|
||||||
|
jdbc.execute("SELECT drop_chunks('stats_1m_route', INTERVAL '" + caggInterval + "')");
|
||||||
|
jdbc.execute("SELECT drop_chunks('stats_1m_processor', INTERVAL '" + caggInterval + "')");
|
||||||
|
|
||||||
|
log.info("Retention: dropped chunks older than {} days (aggregates: {} days)",
|
||||||
|
retentionDays, retentionDays * 3);
|
||||||
|
} catch (Exception e) {
|
||||||
|
log.error("Retention job failed", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Note: OpenSearch daily index deletion should be handled via ILM policy
|
||||||
|
// configured at deployment time, not in application code.
|
||||||
|
}
|
||||||
@@ -1,357 +0,0 @@
|
|||||||
package com.cameleer3.server.app.search;
|
|
||||||
|
|
||||||
import com.cameleer3.server.core.search.ExecutionStats;
|
|
||||||
import com.cameleer3.server.core.search.ExecutionSummary;
|
|
||||||
import com.cameleer3.server.core.search.SearchEngine;
|
|
||||||
import com.cameleer3.server.core.search.SearchRequest;
|
|
||||||
import com.cameleer3.server.core.search.SearchResult;
|
|
||||||
import com.cameleer3.server.core.search.StatsTimeseries;
|
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
|
|
||||||
import java.sql.Timestamp;
|
|
||||||
import java.time.Duration;
|
|
||||||
import java.time.Instant;
|
|
||||||
import java.util.ArrayList;
|
|
||||||
import java.util.Collections;
|
|
||||||
import java.util.List;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* ClickHouse implementation of {@link SearchEngine}.
|
|
||||||
* <p>
|
|
||||||
* Builds dynamic WHERE clauses from non-null {@link SearchRequest} fields
|
|
||||||
* and queries the {@code route_executions} table. LIKE patterns are properly
|
|
||||||
* escaped to prevent injection.
|
|
||||||
*/
|
|
||||||
public class ClickHouseSearchEngine implements SearchEngine {
|
|
||||||
|
|
||||||
/** Per-query memory cap (1 GiB) — prevents a single query from OOMing ClickHouse. */
|
|
||||||
private static final String SETTINGS = " SETTINGS max_memory_usage = 1000000000";
|
|
||||||
|
|
||||||
private final JdbcTemplate jdbcTemplate;
|
|
||||||
|
|
||||||
public ClickHouseSearchEngine(JdbcTemplate jdbcTemplate) {
|
|
||||||
this.jdbcTemplate = jdbcTemplate;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public SearchResult<ExecutionSummary> search(SearchRequest request) {
|
|
||||||
var conditions = new ArrayList<String>();
|
|
||||||
var params = new ArrayList<Object>();
|
|
||||||
|
|
||||||
buildWhereClause(request, conditions, params);
|
|
||||||
|
|
||||||
String where = conditions.isEmpty() ? "" : " WHERE " + String.join(" AND ", conditions);
|
|
||||||
|
|
||||||
// Count query
|
|
||||||
var countParams = params.toArray();
|
|
||||||
Long total = jdbcTemplate.queryForObject(
|
|
||||||
"SELECT count() FROM route_executions" + where + SETTINGS, Long.class, countParams);
|
|
||||||
if (total == null) total = 0L;
|
|
||||||
|
|
||||||
if (total == 0) {
|
|
||||||
return SearchResult.empty(request.offset(), request.limit());
|
|
||||||
}
|
|
||||||
|
|
||||||
// Data query
|
|
||||||
params.add(request.limit());
|
|
||||||
params.add(request.offset());
|
|
||||||
String orderDir = "asc".equalsIgnoreCase(request.sortDir()) ? "ASC" : "DESC";
|
|
||||||
String dataSql = "SELECT execution_id, route_id, agent_id, status, start_time, end_time, " +
|
|
||||||
"duration_ms, correlation_id, error_message, diagram_content_hash " +
|
|
||||||
"FROM route_executions" + where +
|
|
||||||
" ORDER BY " + request.sortColumn() + " " + orderDir + " LIMIT ? OFFSET ?" + SETTINGS;
|
|
||||||
|
|
||||||
List<ExecutionSummary> data = jdbcTemplate.query(dataSql, (rs, rowNum) -> {
|
|
||||||
Timestamp endTs = rs.getTimestamp("end_time");
|
|
||||||
return new ExecutionSummary(
|
|
||||||
rs.getString("execution_id"),
|
|
||||||
rs.getString("route_id"),
|
|
||||||
rs.getString("agent_id"),
|
|
||||||
rs.getString("status"),
|
|
||||||
rs.getTimestamp("start_time").toInstant(),
|
|
||||||
endTs != null ? endTs.toInstant() : null,
|
|
||||||
rs.getLong("duration_ms"),
|
|
||||||
rs.getString("correlation_id"),
|
|
||||||
rs.getString("error_message"),
|
|
||||||
rs.getString("diagram_content_hash")
|
|
||||||
);
|
|
||||||
}, params.toArray());
|
|
||||||
|
|
||||||
return new SearchResult<>(data, total, request.offset(), request.limit());
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public long count(SearchRequest request) {
|
|
||||||
var conditions = new ArrayList<String>();
|
|
||||||
var params = new ArrayList<Object>();
|
|
||||||
buildWhereClause(request, conditions, params);
|
|
||||||
|
|
||||||
String where = conditions.isEmpty() ? "" : " WHERE " + String.join(" AND ", conditions);
|
|
||||||
Long result = jdbcTemplate.queryForObject(
|
|
||||||
"SELECT count() FROM route_executions" + where + SETTINGS, Long.class, params.toArray());
|
|
||||||
return result != null ? result : 0L;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public ExecutionStats stats(Instant from, Instant to) {
|
|
||||||
return stats(from, to, null, null);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public ExecutionStats stats(Instant from, Instant to, String routeId, List<String> agentIds) {
|
|
||||||
// Current period — read from rollup
|
|
||||||
var conditions = new ArrayList<String>();
|
|
||||||
var params = new ArrayList<Object>();
|
|
||||||
conditions.add("bucket >= ?");
|
|
||||||
params.add(bucketTimestamp(floorToFiveMinutes(from)));
|
|
||||||
conditions.add("bucket <= ?");
|
|
||||||
params.add(bucketTimestamp(to));
|
|
||||||
addScopeFilters(routeId, agentIds, conditions, params);
|
|
||||||
|
|
||||||
String where = " WHERE " + String.join(" AND ", conditions);
|
|
||||||
|
|
||||||
String rollupSql = "SELECT " +
|
|
||||||
"countMerge(total_count) AS cnt, " +
|
|
||||||
"countIfMerge(failed_count) AS failed, " +
|
|
||||||
"toInt64(ifNotFinite(sumMerge(duration_sum) / countMerge(total_count), 0)) AS avg_ms, " +
|
|
||||||
"toInt64(ifNotFinite(quantileTDigestMerge(0.99)(p99_duration), 0)) AS p99_ms " +
|
|
||||||
"FROM route_execution_stats_5m" + where + SETTINGS;
|
|
||||||
|
|
||||||
record PeriodStats(long totalCount, long failedCount, long avgDurationMs, long p99LatencyMs) {}
|
|
||||||
PeriodStats current = jdbcTemplate.queryForObject(rollupSql,
|
|
||||||
(rs, rowNum) -> new PeriodStats(
|
|
||||||
rs.getLong("cnt"),
|
|
||||||
rs.getLong("failed"),
|
|
||||||
rs.getLong("avg_ms"),
|
|
||||||
rs.getLong("p99_ms")),
|
|
||||||
params.toArray());
|
|
||||||
|
|
||||||
// Active count — PREWHERE reads only the status column before touching wide rows
|
|
||||||
var scopeConditions = new ArrayList<String>();
|
|
||||||
var activeParams = new ArrayList<Object>();
|
|
||||||
addScopeFilters(routeId, agentIds, scopeConditions, activeParams);
|
|
||||||
String scopeWhere = scopeConditions.isEmpty() ? "" : " WHERE " + String.join(" AND ", scopeConditions);
|
|
||||||
Long activeCount = jdbcTemplate.queryForObject(
|
|
||||||
"SELECT count() FROM route_executions PREWHERE status = 'RUNNING'" + scopeWhere + SETTINGS,
|
|
||||||
Long.class, activeParams.toArray());
|
|
||||||
|
|
||||||
// Previous period (same window shifted back 24h) — read from rollup
|
|
||||||
Duration window = Duration.between(from, to);
|
|
||||||
Instant prevFrom = from.minus(Duration.ofHours(24));
|
|
||||||
Instant prevTo = prevFrom.plus(window);
|
|
||||||
var prevConditions = new ArrayList<String>();
|
|
||||||
var prevParams = new ArrayList<Object>();
|
|
||||||
prevConditions.add("bucket >= ?");
|
|
||||||
prevParams.add(bucketTimestamp(floorToFiveMinutes(prevFrom)));
|
|
||||||
prevConditions.add("bucket <= ?");
|
|
||||||
prevParams.add(bucketTimestamp(prevTo));
|
|
||||||
addScopeFilters(routeId, agentIds, prevConditions, prevParams);
|
|
||||||
String prevWhere = " WHERE " + String.join(" AND ", prevConditions);
|
|
||||||
|
|
||||||
String prevRollupSql = "SELECT " +
|
|
||||||
"countMerge(total_count) AS cnt, " +
|
|
||||||
"countIfMerge(failed_count) AS failed, " +
|
|
||||||
"toInt64(ifNotFinite(sumMerge(duration_sum) / countMerge(total_count), 0)) AS avg_ms, " +
|
|
||||||
"toInt64(ifNotFinite(quantileTDigestMerge(0.99)(p99_duration), 0)) AS p99_ms " +
|
|
||||||
"FROM route_execution_stats_5m" + prevWhere + SETTINGS;
|
|
||||||
|
|
||||||
PeriodStats prev = jdbcTemplate.queryForObject(prevRollupSql,
|
|
||||||
(rs, rowNum) -> new PeriodStats(
|
|
||||||
rs.getLong("cnt"),
|
|
||||||
rs.getLong("failed"),
|
|
||||||
rs.getLong("avg_ms"),
|
|
||||||
rs.getLong("p99_ms")),
|
|
||||||
prevParams.toArray());
|
|
||||||
|
|
||||||
// Today total (midnight UTC to now) — read from rollup with same scope
|
|
||||||
Instant todayStart = Instant.now().truncatedTo(java.time.temporal.ChronoUnit.DAYS);
|
|
||||||
var todayConditions = new ArrayList<String>();
|
|
||||||
var todayParams = new ArrayList<Object>();
|
|
||||||
todayConditions.add("bucket >= ?");
|
|
||||||
todayParams.add(bucketTimestamp(floorToFiveMinutes(todayStart)));
|
|
||||||
addScopeFilters(routeId, agentIds, todayConditions, todayParams);
|
|
||||||
String todayWhere = " WHERE " + String.join(" AND ", todayConditions);
|
|
||||||
|
|
||||||
Long totalToday = jdbcTemplate.queryForObject(
|
|
||||||
"SELECT countMerge(total_count) FROM route_execution_stats_5m" + todayWhere + SETTINGS,
|
|
||||||
Long.class, todayParams.toArray());
|
|
||||||
|
|
||||||
return new ExecutionStats(
|
|
||||||
current.totalCount, current.failedCount, current.avgDurationMs,
|
|
||||||
current.p99LatencyMs, activeCount != null ? activeCount : 0L,
|
|
||||||
totalToday != null ? totalToday : 0L,
|
|
||||||
prev.totalCount, prev.failedCount, prev.avgDurationMs, prev.p99LatencyMs);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public StatsTimeseries timeseries(Instant from, Instant to, int bucketCount) {
|
|
||||||
return timeseries(from, to, bucketCount, null, null);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public StatsTimeseries timeseries(Instant from, Instant to, int bucketCount,
|
|
||||||
String routeId, List<String> agentIds) {
|
|
||||||
long intervalSeconds = Duration.between(from, to).getSeconds() / bucketCount;
|
|
||||||
if (intervalSeconds < 1) intervalSeconds = 1;
|
|
||||||
|
|
||||||
var conditions = new ArrayList<String>();
|
|
||||||
var params = new ArrayList<Object>();
|
|
||||||
conditions.add("bucket >= ?");
|
|
||||||
params.add(bucketTimestamp(floorToFiveMinutes(from)));
|
|
||||||
conditions.add("bucket <= ?");
|
|
||||||
params.add(bucketTimestamp(to));
|
|
||||||
addScopeFilters(routeId, agentIds, conditions, params);
|
|
||||||
|
|
||||||
String where = " WHERE " + String.join(" AND ", conditions);
|
|
||||||
|
|
||||||
// Re-aggregate 5-minute rollup buckets into the requested interval
|
|
||||||
String sql = "SELECT " +
|
|
||||||
"toDateTime(intDiv(toUInt32(bucket), " + intervalSeconds + ") * " + intervalSeconds + ") AS ts_bucket, " +
|
|
||||||
"countMerge(total_count) AS cnt, " +
|
|
||||||
"countIfMerge(failed_count) AS failed, " +
|
|
||||||
"toInt64(ifNotFinite(sumMerge(duration_sum) / countMerge(total_count), 0)) AS avg_ms, " +
|
|
||||||
"toInt64(ifNotFinite(quantileTDigestMerge(0.99)(p99_duration), 0)) AS p99_ms " +
|
|
||||||
"FROM route_execution_stats_5m" + where +
|
|
||||||
" GROUP BY ts_bucket ORDER BY ts_bucket" + SETTINGS;
|
|
||||||
|
|
||||||
List<StatsTimeseries.TimeseriesBucket> buckets = jdbcTemplate.query(sql, (rs, rowNum) ->
|
|
||||||
new StatsTimeseries.TimeseriesBucket(
|
|
||||||
rs.getTimestamp("ts_bucket").toInstant(),
|
|
||||||
rs.getLong("cnt"),
|
|
||||||
rs.getLong("failed"),
|
|
||||||
rs.getLong("avg_ms"),
|
|
||||||
rs.getLong("p99_ms"),
|
|
||||||
0L
|
|
||||||
),
|
|
||||||
params.toArray());
|
|
||||||
|
|
||||||
return new StatsTimeseries(buckets);
|
|
||||||
}
|
|
||||||
|
|
||||||
private void buildWhereClause(SearchRequest req, List<String> conditions, List<Object> params) {
|
|
||||||
if (req.status() != null && !req.status().isBlank()) {
|
|
||||||
String[] statuses = req.status().split(",");
|
|
||||||
if (statuses.length == 1) {
|
|
||||||
conditions.add("status = ?");
|
|
||||||
params.add(statuses[0].trim());
|
|
||||||
} else {
|
|
||||||
String placeholders = String.join(", ", Collections.nCopies(statuses.length, "?"));
|
|
||||||
conditions.add("status IN (" + placeholders + ")");
|
|
||||||
for (String s : statuses) {
|
|
||||||
params.add(s.trim());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (req.timeFrom() != null) {
|
|
||||||
conditions.add("start_time >= ?");
|
|
||||||
params.add(Timestamp.from(req.timeFrom()));
|
|
||||||
}
|
|
||||||
if (req.timeTo() != null) {
|
|
||||||
conditions.add("start_time <= ?");
|
|
||||||
params.add(Timestamp.from(req.timeTo()));
|
|
||||||
}
|
|
||||||
if (req.durationMin() != null) {
|
|
||||||
conditions.add("duration_ms >= ?");
|
|
||||||
params.add(req.durationMin());
|
|
||||||
}
|
|
||||||
if (req.durationMax() != null) {
|
|
||||||
conditions.add("duration_ms <= ?");
|
|
||||||
params.add(req.durationMax());
|
|
||||||
}
|
|
||||||
if (req.correlationId() != null && !req.correlationId().isBlank()) {
|
|
||||||
conditions.add("correlation_id = ?");
|
|
||||||
params.add(req.correlationId());
|
|
||||||
}
|
|
||||||
if (req.routeId() != null && !req.routeId().isBlank()) {
|
|
||||||
conditions.add("route_id = ?");
|
|
||||||
params.add(req.routeId());
|
|
||||||
}
|
|
||||||
if (req.agentId() != null && !req.agentId().isBlank()) {
|
|
||||||
conditions.add("agent_id = ?");
|
|
||||||
params.add(req.agentId());
|
|
||||||
}
|
|
||||||
// agentIds from group resolution (takes precedence when agentId is not set)
|
|
||||||
if ((req.agentId() == null || req.agentId().isBlank())
|
|
||||||
&& req.agentIds() != null && !req.agentIds().isEmpty()) {
|
|
||||||
String placeholders = String.join(", ", Collections.nCopies(req.agentIds().size(), "?"));
|
|
||||||
conditions.add("agent_id IN (" + placeholders + ")");
|
|
||||||
params.addAll(req.agentIds());
|
|
||||||
}
|
|
||||||
if (req.processorType() != null && !req.processorType().isBlank()) {
|
|
||||||
conditions.add("has(processor_types, ?)");
|
|
||||||
params.add(req.processorType());
|
|
||||||
}
|
|
||||||
if (req.text() != null && !req.text().isBlank()) {
|
|
||||||
String pattern = "%" + escapeLike(req.text()) + "%";
|
|
||||||
String[] textColumns = {
|
|
||||||
"execution_id", "route_id", "agent_id",
|
|
||||||
"error_message", "error_stacktrace",
|
|
||||||
"exchange_bodies", "exchange_headers"
|
|
||||||
};
|
|
||||||
var likeClauses = java.util.Arrays.stream(textColumns)
|
|
||||||
.map(col -> col + " LIKE ?")
|
|
||||||
.toList();
|
|
||||||
conditions.add("(" + String.join(" OR ", likeClauses) + ")");
|
|
||||||
for (int i = 0; i < textColumns.length; i++) {
|
|
||||||
params.add(pattern);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (req.textInBody() != null && !req.textInBody().isBlank()) {
|
|
||||||
conditions.add("exchange_bodies LIKE ?");
|
|
||||||
params.add("%" + escapeLike(req.textInBody()) + "%");
|
|
||||||
}
|
|
||||||
if (req.textInHeaders() != null && !req.textInHeaders().isBlank()) {
|
|
||||||
conditions.add("exchange_headers LIKE ?");
|
|
||||||
params.add("%" + escapeLike(req.textInHeaders()) + "%");
|
|
||||||
}
|
|
||||||
if (req.textInErrors() != null && !req.textInErrors().isBlank()) {
|
|
||||||
String pattern = "%" + escapeLike(req.textInErrors()) + "%";
|
|
||||||
conditions.add("(error_message LIKE ? OR error_stacktrace LIKE ?)");
|
|
||||||
params.add(pattern);
|
|
||||||
params.add(pattern);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Add route ID and agent IDs scope filters to conditions/params.
|
|
||||||
*/
|
|
||||||
private void addScopeFilters(String routeId, List<String> agentIds,
|
|
||||||
List<String> conditions, List<Object> params) {
|
|
||||||
if (routeId != null && !routeId.isBlank()) {
|
|
||||||
conditions.add("route_id = ?");
|
|
||||||
params.add(routeId);
|
|
||||||
}
|
|
||||||
if (agentIds != null && !agentIds.isEmpty()) {
|
|
||||||
String placeholders = String.join(", ", Collections.nCopies(agentIds.size(), "?"));
|
|
||||||
conditions.add("agent_id IN (" + placeholders + ")");
|
|
||||||
params.addAll(agentIds);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Floor an Instant to the start of its 5-minute bucket.
|
|
||||||
*/
|
|
||||||
private static Instant floorToFiveMinutes(Instant instant) {
|
|
||||||
long epochSecond = instant.getEpochSecond();
|
|
||||||
return Instant.ofEpochSecond(epochSecond - (epochSecond % 300));
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Create a second-precision Timestamp for rollup bucket comparisons.
|
|
||||||
* The bucket column is DateTime('UTC') (second precision); the JDBC driver
|
|
||||||
* sends java.sql.Timestamp with nanoseconds which ClickHouse rejects.
|
|
||||||
*/
|
|
||||||
private static Timestamp bucketTimestamp(Instant instant) {
|
|
||||||
return Timestamp.from(instant.truncatedTo(java.time.temporal.ChronoUnit.SECONDS));
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Escape special LIKE characters to prevent LIKE injection.
|
|
||||||
*/
|
|
||||||
static String escapeLike(String input) {
|
|
||||||
return input
|
|
||||||
.replace("\\", "\\\\")
|
|
||||||
.replace("%", "\\%")
|
|
||||||
.replace("_", "\\_");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,329 @@
|
|||||||
|
package com.cameleer3.server.app.search;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.search.ExecutionSummary;
|
||||||
|
import com.cameleer3.server.core.search.SearchRequest;
|
||||||
|
import com.cameleer3.server.core.search.SearchResult;
|
||||||
|
import com.cameleer3.server.core.storage.SearchIndex;
|
||||||
|
import com.cameleer3.server.core.storage.model.ExecutionDocument;
|
||||||
|
import com.cameleer3.server.core.storage.model.ExecutionDocument.ProcessorDoc;
|
||||||
|
import jakarta.annotation.PostConstruct;
|
||||||
|
import org.opensearch.client.json.JsonData;
|
||||||
|
import org.opensearch.client.opensearch.OpenSearchClient;
|
||||||
|
import org.opensearch.client.opensearch._types.FieldValue;
|
||||||
|
import org.opensearch.client.opensearch._types.SortOrder;
|
||||||
|
import org.opensearch.client.opensearch._types.query_dsl.*;
|
||||||
|
import org.opensearch.client.opensearch.core.*;
|
||||||
|
import org.opensearch.client.opensearch.core.search.Hit;
|
||||||
|
import org.opensearch.client.opensearch.indices.*;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
import org.springframework.stereotype.Repository;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.time.ZoneOffset;
|
||||||
|
import java.time.format.DateTimeFormatter;
|
||||||
|
import java.util.*;
|
||||||
|
import java.util.stream.Collectors;
|
||||||
|
|
||||||
|
@Repository
|
||||||
|
public class OpenSearchIndex implements SearchIndex {
|
||||||
|
|
||||||
|
private static final Logger log = LoggerFactory.getLogger(OpenSearchIndex.class);
|
||||||
|
private static final String INDEX_PREFIX = "executions-";
|
||||||
|
private static final DateTimeFormatter DAY_FMT = DateTimeFormatter.ofPattern("yyyy-MM-dd")
|
||||||
|
.withZone(ZoneOffset.UTC);
|
||||||
|
|
||||||
|
private final OpenSearchClient client;
|
||||||
|
|
||||||
|
public OpenSearchIndex(OpenSearchClient client) {
|
||||||
|
this.client = client;
|
||||||
|
}
|
||||||
|
|
||||||
|
@PostConstruct
|
||||||
|
void ensureIndexTemplate() {
|
||||||
|
try {
|
||||||
|
boolean exists = client.indices().existsIndexTemplate(
|
||||||
|
ExistsIndexTemplateRequest.of(b -> b.name("executions-template"))).value();
|
||||||
|
if (!exists) {
|
||||||
|
client.indices().putIndexTemplate(PutIndexTemplateRequest.of(b -> b
|
||||||
|
.name("executions-template")
|
||||||
|
.indexPatterns(List.of("executions-*"))
|
||||||
|
.template(t -> t
|
||||||
|
.settings(s -> s
|
||||||
|
.numberOfShards("3")
|
||||||
|
.numberOfReplicas("1"))
|
||||||
|
.mappings(m -> m
|
||||||
|
.properties("processors", p -> p
|
||||||
|
.nested(n -> n))))));
|
||||||
|
log.info("OpenSearch index template created");
|
||||||
|
}
|
||||||
|
} catch (IOException e) {
|
||||||
|
log.error("Failed to create index template", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void index(ExecutionDocument doc) {
|
||||||
|
String indexName = INDEX_PREFIX + DAY_FMT.format(doc.startTime());
|
||||||
|
try {
|
||||||
|
client.index(IndexRequest.of(b -> b
|
||||||
|
.index(indexName)
|
||||||
|
.id(doc.executionId())
|
||||||
|
.document(toMap(doc))));
|
||||||
|
} catch (IOException e) {
|
||||||
|
log.error("Failed to index execution {}", doc.executionId(), e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public SearchResult<ExecutionSummary> search(SearchRequest request) {
|
||||||
|
try {
|
||||||
|
var searchReq = buildSearchRequest(request, request.limit());
|
||||||
|
var response = client.search(searchReq, Map.class);
|
||||||
|
|
||||||
|
List<ExecutionSummary> items = response.hits().hits().stream()
|
||||||
|
.map(this::hitToSummary)
|
||||||
|
.collect(Collectors.toList());
|
||||||
|
|
||||||
|
long total = response.hits().total() != null ? response.hits().total().value() : 0;
|
||||||
|
return new SearchResult<>(items, total, request.offset(), request.limit());
|
||||||
|
} catch (IOException e) {
|
||||||
|
log.error("Search failed", e);
|
||||||
|
return SearchResult.empty(request.offset(), request.limit());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long count(SearchRequest request) {
|
||||||
|
try {
|
||||||
|
var countReq = CountRequest.of(b -> b
|
||||||
|
.index(INDEX_PREFIX + "*")
|
||||||
|
.query(buildQuery(request)));
|
||||||
|
return client.count(countReq).count();
|
||||||
|
} catch (IOException e) {
|
||||||
|
log.error("Count failed", e);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void delete(String executionId) {
|
||||||
|
try {
|
||||||
|
client.deleteByQuery(DeleteByQueryRequest.of(b -> b
|
||||||
|
.index(List.of(INDEX_PREFIX + "*"))
|
||||||
|
.query(Query.of(q -> q.term(t -> t
|
||||||
|
.field("execution_id")
|
||||||
|
.value(FieldValue.of(executionId)))))));
|
||||||
|
} catch (IOException e) {
|
||||||
|
log.error("Failed to delete execution {}", executionId, e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private org.opensearch.client.opensearch.core.SearchRequest buildSearchRequest(
|
||||||
|
SearchRequest request, int size) {
|
||||||
|
return org.opensearch.client.opensearch.core.SearchRequest.of(b -> {
|
||||||
|
b.index(INDEX_PREFIX + "*")
|
||||||
|
.query(buildQuery(request))
|
||||||
|
.size(size)
|
||||||
|
.from(request.offset())
|
||||||
|
.sort(s -> s.field(f -> f
|
||||||
|
.field(request.sortColumn())
|
||||||
|
.order("asc".equalsIgnoreCase(request.sortDir())
|
||||||
|
? SortOrder.Asc : SortOrder.Desc)));
|
||||||
|
return b;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
private Query buildQuery(SearchRequest request) {
|
||||||
|
List<Query> must = new ArrayList<>();
|
||||||
|
List<Query> filter = new ArrayList<>();
|
||||||
|
|
||||||
|
// Time range
|
||||||
|
if (request.timeFrom() != null || request.timeTo() != null) {
|
||||||
|
filter.add(Query.of(q -> q.range(r -> {
|
||||||
|
r.field("start_time");
|
||||||
|
if (request.timeFrom() != null)
|
||||||
|
r.gte(JsonData.of(request.timeFrom().toString()));
|
||||||
|
if (request.timeTo() != null)
|
||||||
|
r.lte(JsonData.of(request.timeTo().toString()));
|
||||||
|
return r;
|
||||||
|
})));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Keyword filters (use .keyword sub-field for exact matching on dynamically mapped text fields)
|
||||||
|
if (request.status() != null)
|
||||||
|
filter.add(termQuery("status.keyword", request.status()));
|
||||||
|
if (request.routeId() != null)
|
||||||
|
filter.add(termQuery("route_id.keyword", request.routeId()));
|
||||||
|
if (request.agentId() != null)
|
||||||
|
filter.add(termQuery("agent_id.keyword", request.agentId()));
|
||||||
|
if (request.correlationId() != null)
|
||||||
|
filter.add(termQuery("correlation_id.keyword", request.correlationId()));
|
||||||
|
|
||||||
|
// Full-text search across all fields + nested processor fields
|
||||||
|
if (request.text() != null && !request.text().isBlank()) {
|
||||||
|
String text = request.text();
|
||||||
|
String wildcard = "*" + text.toLowerCase() + "*";
|
||||||
|
List<Query> textQueries = new ArrayList<>();
|
||||||
|
|
||||||
|
// Search top-level text fields (analyzed match + wildcard for substring)
|
||||||
|
textQueries.add(Query.of(q -> q.multiMatch(m -> m
|
||||||
|
.query(text)
|
||||||
|
.fields("error_message", "error_stacktrace"))));
|
||||||
|
textQueries.add(Query.of(q -> q.wildcard(w -> w
|
||||||
|
.field("error_message").value(wildcard).caseInsensitive(true))));
|
||||||
|
textQueries.add(Query.of(q -> q.wildcard(w -> w
|
||||||
|
.field("error_stacktrace").value(wildcard).caseInsensitive(true))));
|
||||||
|
|
||||||
|
// Search nested processor fields (analyzed match + wildcard)
|
||||||
|
textQueries.add(Query.of(q -> q.nested(n -> n
|
||||||
|
.path("processors")
|
||||||
|
.query(nq -> nq.multiMatch(m -> m
|
||||||
|
.query(text)
|
||||||
|
.fields("processors.input_body", "processors.output_body",
|
||||||
|
"processors.input_headers", "processors.output_headers",
|
||||||
|
"processors.error_message", "processors.error_stacktrace"))))));
|
||||||
|
textQueries.add(Query.of(q -> q.nested(n -> n
|
||||||
|
.path("processors")
|
||||||
|
.query(nq -> nq.bool(nb -> nb.should(
|
||||||
|
wildcardQuery("processors.input_body", wildcard),
|
||||||
|
wildcardQuery("processors.output_body", wildcard),
|
||||||
|
wildcardQuery("processors.input_headers", wildcard),
|
||||||
|
wildcardQuery("processors.output_headers", wildcard)
|
||||||
|
).minimumShouldMatch("1"))))));
|
||||||
|
|
||||||
|
// Also try keyword fields for exact matches
|
||||||
|
textQueries.add(Query.of(q -> q.multiMatch(m -> m
|
||||||
|
.query(text)
|
||||||
|
.fields("execution_id", "route_id", "agent_id", "correlation_id", "exchange_id"))));
|
||||||
|
|
||||||
|
must.add(Query.of(q -> q.bool(b -> b.should(textQueries).minimumShouldMatch("1"))));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Scoped text searches (multiMatch + wildcard fallback for substring matching)
|
||||||
|
if (request.textInBody() != null && !request.textInBody().isBlank()) {
|
||||||
|
String bodyText = request.textInBody();
|
||||||
|
String bodyWildcard = "*" + bodyText.toLowerCase() + "*";
|
||||||
|
must.add(Query.of(q -> q.nested(n -> n
|
||||||
|
.path("processors")
|
||||||
|
.query(nq -> nq.bool(nb -> nb.should(
|
||||||
|
Query.of(mq -> mq.multiMatch(m -> m
|
||||||
|
.query(bodyText)
|
||||||
|
.fields("processors.input_body", "processors.output_body"))),
|
||||||
|
wildcardQuery("processors.input_body", bodyWildcard),
|
||||||
|
wildcardQuery("processors.output_body", bodyWildcard)
|
||||||
|
).minimumShouldMatch("1"))))));
|
||||||
|
}
|
||||||
|
if (request.textInHeaders() != null && !request.textInHeaders().isBlank()) {
|
||||||
|
String headerText = request.textInHeaders();
|
||||||
|
String headerWildcard = "*" + headerText.toLowerCase() + "*";
|
||||||
|
must.add(Query.of(q -> q.nested(n -> n
|
||||||
|
.path("processors")
|
||||||
|
.query(nq -> nq.bool(nb -> nb.should(
|
||||||
|
Query.of(mq -> mq.multiMatch(m -> m
|
||||||
|
.query(headerText)
|
||||||
|
.fields("processors.input_headers", "processors.output_headers"))),
|
||||||
|
wildcardQuery("processors.input_headers", headerWildcard),
|
||||||
|
wildcardQuery("processors.output_headers", headerWildcard)
|
||||||
|
).minimumShouldMatch("1"))))));
|
||||||
|
}
|
||||||
|
if (request.textInErrors() != null && !request.textInErrors().isBlank()) {
|
||||||
|
String errText = request.textInErrors();
|
||||||
|
String errWildcard = "*" + errText.toLowerCase() + "*";
|
||||||
|
must.add(Query.of(q -> q.bool(b -> b.should(
|
||||||
|
Query.of(sq -> sq.multiMatch(m -> m
|
||||||
|
.query(errText)
|
||||||
|
.fields("error_message", "error_stacktrace"))),
|
||||||
|
wildcardQuery("error_message", errWildcard),
|
||||||
|
wildcardQuery("error_stacktrace", errWildcard),
|
||||||
|
Query.of(sq -> sq.nested(n -> n
|
||||||
|
.path("processors")
|
||||||
|
.query(nq -> nq.bool(nb -> nb.should(
|
||||||
|
Query.of(nmq -> nmq.multiMatch(m -> m
|
||||||
|
.query(errText)
|
||||||
|
.fields("processors.error_message", "processors.error_stacktrace"))),
|
||||||
|
wildcardQuery("processors.error_message", errWildcard),
|
||||||
|
wildcardQuery("processors.error_stacktrace", errWildcard)
|
||||||
|
).minimumShouldMatch("1")))))
|
||||||
|
).minimumShouldMatch("1"))));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Duration range
|
||||||
|
if (request.durationMin() != null || request.durationMax() != null) {
|
||||||
|
filter.add(Query.of(q -> q.range(r -> {
|
||||||
|
r.field("duration_ms");
|
||||||
|
if (request.durationMin() != null)
|
||||||
|
r.gte(JsonData.of(request.durationMin()));
|
||||||
|
if (request.durationMax() != null)
|
||||||
|
r.lte(JsonData.of(request.durationMax()));
|
||||||
|
return r;
|
||||||
|
})));
|
||||||
|
}
|
||||||
|
|
||||||
|
return Query.of(q -> q.bool(b -> {
|
||||||
|
if (!must.isEmpty()) b.must(must);
|
||||||
|
if (!filter.isEmpty()) b.filter(filter);
|
||||||
|
if (must.isEmpty() && filter.isEmpty()) b.must(Query.of(mq -> mq.matchAll(m -> m)));
|
||||||
|
return b;
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
|
private Query termQuery(String field, String value) {
|
||||||
|
return Query.of(q -> q.term(t -> t.field(field).value(FieldValue.of(value))));
|
||||||
|
}
|
||||||
|
|
||||||
|
private Query wildcardQuery(String field, String pattern) {
|
||||||
|
return Query.of(q -> q.wildcard(w -> w.field(field).value(pattern).caseInsensitive(true)));
|
||||||
|
}
|
||||||
|
|
||||||
|
private Map<String, Object> toMap(ExecutionDocument doc) {
|
||||||
|
Map<String, Object> map = new LinkedHashMap<>();
|
||||||
|
map.put("execution_id", doc.executionId());
|
||||||
|
map.put("route_id", doc.routeId());
|
||||||
|
map.put("agent_id", doc.agentId());
|
||||||
|
map.put("group_name", doc.groupName());
|
||||||
|
map.put("status", doc.status());
|
||||||
|
map.put("correlation_id", doc.correlationId());
|
||||||
|
map.put("exchange_id", doc.exchangeId());
|
||||||
|
map.put("start_time", doc.startTime() != null ? doc.startTime().toString() : null);
|
||||||
|
map.put("end_time", doc.endTime() != null ? doc.endTime().toString() : null);
|
||||||
|
map.put("duration_ms", doc.durationMs());
|
||||||
|
map.put("error_message", doc.errorMessage());
|
||||||
|
map.put("error_stacktrace", doc.errorStacktrace());
|
||||||
|
if (doc.processors() != null) {
|
||||||
|
map.put("processors", doc.processors().stream().map(p -> {
|
||||||
|
Map<String, Object> pm = new LinkedHashMap<>();
|
||||||
|
pm.put("processor_id", p.processorId());
|
||||||
|
pm.put("processor_type", p.processorType());
|
||||||
|
pm.put("status", p.status());
|
||||||
|
pm.put("error_message", p.errorMessage());
|
||||||
|
pm.put("error_stacktrace", p.errorStacktrace());
|
||||||
|
pm.put("input_body", p.inputBody());
|
||||||
|
pm.put("output_body", p.outputBody());
|
||||||
|
pm.put("input_headers", p.inputHeaders());
|
||||||
|
pm.put("output_headers", p.outputHeaders());
|
||||||
|
return pm;
|
||||||
|
}).toList());
|
||||||
|
}
|
||||||
|
return map;
|
||||||
|
}
|
||||||
|
|
||||||
|
@SuppressWarnings("unchecked")
|
||||||
|
private ExecutionSummary hitToSummary(Hit<Map> hit) {
|
||||||
|
Map<String, Object> src = hit.source();
|
||||||
|
if (src == null) return null;
|
||||||
|
return new ExecutionSummary(
|
||||||
|
(String) src.get("execution_id"),
|
||||||
|
(String) src.get("route_id"),
|
||||||
|
(String) src.get("agent_id"),
|
||||||
|
(String) src.get("status"),
|
||||||
|
src.get("start_time") != null ? Instant.parse((String) src.get("start_time")) : null,
|
||||||
|
src.get("end_time") != null ? Instant.parse((String) src.get("end_time")) : null,
|
||||||
|
src.get("duration_ms") != null ? ((Number) src.get("duration_ms")).longValue() : 0L,
|
||||||
|
(String) src.get("correlation_id"),
|
||||||
|
(String) src.get("error_message"),
|
||||||
|
null // diagramContentHash not stored in index
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -16,7 +16,7 @@ import java.util.List;
|
|||||||
* that required security properties are set.
|
* that required security properties are set.
|
||||||
* <p>
|
* <p>
|
||||||
* Fails fast on startup if {@code CAMELEER_AUTH_TOKEN} is not set.
|
* Fails fast on startup if {@code CAMELEER_AUTH_TOKEN} is not set.
|
||||||
* Seeds OIDC config from env vars into ClickHouse if DB is empty.
|
* Seeds OIDC config from env vars into the database if DB is empty.
|
||||||
*/
|
*/
|
||||||
@Configuration
|
@Configuration
|
||||||
@EnableConfigurationProperties(SecurityProperties.class)
|
@EnableConfigurationProperties(SecurityProperties.class)
|
||||||
|
|||||||
@@ -1,418 +0,0 @@
|
|||||||
package com.cameleer3.server.app.storage;
|
|
||||||
|
|
||||||
import com.cameleer3.common.model.ExchangeSnapshot;
|
|
||||||
import com.cameleer3.common.model.ProcessorExecution;
|
|
||||||
import com.cameleer3.common.model.RouteExecution;
|
|
||||||
import com.cameleer3.server.core.detail.RawExecutionRow;
|
|
||||||
import com.cameleer3.server.core.ingestion.TaggedExecution;
|
|
||||||
import com.cameleer3.server.core.storage.DiagramRepository;
|
|
||||||
import com.cameleer3.server.core.storage.ExecutionRepository;
|
|
||||||
import com.fasterxml.jackson.core.JsonProcessingException;
|
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
|
||||||
import org.slf4j.Logger;
|
|
||||||
import org.slf4j.LoggerFactory;
|
|
||||||
import org.springframework.jdbc.core.BatchPreparedStatementSetter;
|
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
import org.springframework.stereotype.Repository;
|
|
||||||
|
|
||||||
import java.sql.PreparedStatement;
|
|
||||||
import java.sql.SQLException;
|
|
||||||
import java.sql.Timestamp;
|
|
||||||
import java.time.Instant;
|
|
||||||
import java.util.ArrayList;
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Map;
|
|
||||||
import java.util.Optional;
|
|
||||||
import java.util.UUID;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* ClickHouse implementation of {@link ExecutionRepository}.
|
|
||||||
* <p>
|
|
||||||
* Performs batch inserts into the {@code route_executions} table.
|
|
||||||
* Processor executions are flattened into parallel arrays with tree metadata
|
|
||||||
* (depth, parent index) for reconstruction.
|
|
||||||
*/
|
|
||||||
@Repository
|
|
||||||
public class ClickHouseExecutionRepository implements ExecutionRepository {
|
|
||||||
|
|
||||||
private static final Logger log = LoggerFactory.getLogger(ClickHouseExecutionRepository.class);
|
|
||||||
|
|
||||||
private static final ObjectMapper OBJECT_MAPPER = new ObjectMapper();
|
|
||||||
|
|
||||||
private static final String INSERT_SQL = """
|
|
||||||
INSERT INTO route_executions (
|
|
||||||
execution_id, route_id, agent_id, status, start_time, end_time,
|
|
||||||
duration_ms, correlation_id, exchange_id, error_message, error_stacktrace,
|
|
||||||
processor_ids, processor_types, processor_starts, processor_ends,
|
|
||||||
processor_durations, processor_statuses,
|
|
||||||
exchange_bodies, exchange_headers,
|
|
||||||
processor_depths, processor_parent_indexes,
|
|
||||||
processor_error_messages, processor_error_stacktraces,
|
|
||||||
processor_input_bodies, processor_output_bodies,
|
|
||||||
processor_input_headers, processor_output_headers,
|
|
||||||
processor_diagram_node_ids, diagram_content_hash
|
|
||||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
|
||||||
""";
|
|
||||||
|
|
||||||
private final JdbcTemplate jdbcTemplate;
|
|
||||||
private final DiagramRepository diagramRepository;
|
|
||||||
|
|
||||||
public ClickHouseExecutionRepository(JdbcTemplate jdbcTemplate, DiagramRepository diagramRepository) {
|
|
||||||
this.jdbcTemplate = jdbcTemplate;
|
|
||||||
this.diagramRepository = diagramRepository;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void insertBatch(List<TaggedExecution> executions) {
|
|
||||||
if (executions.isEmpty()) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
jdbcTemplate.batchUpdate(INSERT_SQL, new BatchPreparedStatementSetter() {
|
|
||||||
@Override
|
|
||||||
public void setValues(PreparedStatement ps, int i) throws SQLException {
|
|
||||||
TaggedExecution tagged = executions.get(i);
|
|
||||||
RouteExecution exec = tagged.execution();
|
|
||||||
String agentId = tagged.agentId() != null ? tagged.agentId() : "";
|
|
||||||
List<FlatProcessor> flatProcessors = flattenWithMetadata(exec.getProcessors());
|
|
||||||
|
|
||||||
int col = 1;
|
|
||||||
ps.setString(col++, UUID.randomUUID().toString());
|
|
||||||
ps.setString(col++, nullSafe(exec.getRouteId()));
|
|
||||||
ps.setString(col++, agentId);
|
|
||||||
ps.setString(col++, exec.getStatus() != null ? exec.getStatus().name() : "RUNNING");
|
|
||||||
ps.setObject(col++, toTimestamp(exec.getStartTime()));
|
|
||||||
ps.setObject(col++, toTimestamp(exec.getEndTime()));
|
|
||||||
ps.setLong(col++, exec.getDurationMs());
|
|
||||||
ps.setString(col++, nullSafe(exec.getCorrelationId()));
|
|
||||||
ps.setString(col++, nullSafe(exec.getExchangeId()));
|
|
||||||
ps.setString(col++, nullSafe(exec.getErrorMessage()));
|
|
||||||
ps.setString(col++, nullSafe(exec.getErrorStackTrace()));
|
|
||||||
|
|
||||||
// Original parallel arrays
|
|
||||||
ps.setObject(col++, flatProcessors.stream().map(fp -> nullSafe(fp.proc.getProcessorId())).toArray(String[]::new));
|
|
||||||
ps.setObject(col++, flatProcessors.stream().map(fp -> nullSafe(fp.proc.getProcessorType())).toArray(String[]::new));
|
|
||||||
ps.setObject(col++, flatProcessors.stream().map(fp -> toTimestamp(fp.proc.getStartTime())).toArray(Timestamp[]::new));
|
|
||||||
ps.setObject(col++, flatProcessors.stream().map(fp -> toTimestamp(fp.proc.getEndTime())).toArray(Timestamp[]::new));
|
|
||||||
ps.setObject(col++, flatProcessors.stream().mapToLong(fp -> fp.proc.getDurationMs()).boxed().toArray(Long[]::new));
|
|
||||||
ps.setObject(col++, flatProcessors.stream().map(fp -> fp.proc.getStatus() != null ? fp.proc.getStatus().name() : "RUNNING").toArray(String[]::new));
|
|
||||||
|
|
||||||
// Phase 2: exchange bodies and headers (concatenated for search)
|
|
||||||
StringBuilder allBodies = new StringBuilder();
|
|
||||||
StringBuilder allHeaders = new StringBuilder();
|
|
||||||
|
|
||||||
String[] inputBodies = new String[flatProcessors.size()];
|
|
||||||
String[] outputBodies = new String[flatProcessors.size()];
|
|
||||||
String[] inputHeaders = new String[flatProcessors.size()];
|
|
||||||
String[] outputHeaders = new String[flatProcessors.size()];
|
|
||||||
String[] errorMessages = new String[flatProcessors.size()];
|
|
||||||
String[] errorStacktraces = new String[flatProcessors.size()];
|
|
||||||
String[] diagramNodeIds = new String[flatProcessors.size()];
|
|
||||||
Short[] depths = new Short[flatProcessors.size()];
|
|
||||||
Integer[] parentIndexes = new Integer[flatProcessors.size()];
|
|
||||||
|
|
||||||
for (int j = 0; j < flatProcessors.size(); j++) {
|
|
||||||
FlatProcessor fp = flatProcessors.get(j);
|
|
||||||
ProcessorExecution p = fp.proc;
|
|
||||||
|
|
||||||
inputBodies[j] = nullSafe(p.getInputBody());
|
|
||||||
outputBodies[j] = nullSafe(p.getOutputBody());
|
|
||||||
inputHeaders[j] = mapToJson(p.getInputHeaders());
|
|
||||||
outputHeaders[j] = mapToJson(p.getOutputHeaders());
|
|
||||||
errorMessages[j] = nullSafe(p.getErrorMessage());
|
|
||||||
errorStacktraces[j] = nullSafe(p.getErrorStackTrace());
|
|
||||||
diagramNodeIds[j] = nullSafe(p.getDiagramNodeId());
|
|
||||||
depths[j] = (short) fp.depth;
|
|
||||||
parentIndexes[j] = fp.parentIndex;
|
|
||||||
|
|
||||||
allBodies.append(inputBodies[j]).append(' ').append(outputBodies[j]).append(' ');
|
|
||||||
allHeaders.append(inputHeaders[j]).append(' ').append(outputHeaders[j]).append(' ');
|
|
||||||
}
|
|
||||||
|
|
||||||
// Include route-level input/output snapshot in searchable text
|
|
||||||
appendSnapshotText(exec.getInputSnapshot(), allBodies, allHeaders);
|
|
||||||
appendSnapshotText(exec.getOutputSnapshot(), allBodies, allHeaders);
|
|
||||||
|
|
||||||
ps.setString(col++, allBodies.toString().trim()); // exchange_bodies
|
|
||||||
ps.setString(col++, allHeaders.toString().trim()); // exchange_headers
|
|
||||||
ps.setObject(col++, depths); // processor_depths
|
|
||||||
ps.setObject(col++, parentIndexes); // processor_parent_indexes
|
|
||||||
ps.setObject(col++, errorMessages); // processor_error_messages
|
|
||||||
ps.setObject(col++, errorStacktraces); // processor_error_stacktraces
|
|
||||||
ps.setObject(col++, inputBodies); // processor_input_bodies
|
|
||||||
ps.setObject(col++, outputBodies); // processor_output_bodies
|
|
||||||
ps.setObject(col++, inputHeaders); // processor_input_headers
|
|
||||||
ps.setObject(col++, outputHeaders); // processor_output_headers
|
|
||||||
ps.setObject(col++, diagramNodeIds); // processor_diagram_node_ids
|
|
||||||
String diagramHash = diagramRepository
|
|
||||||
.findContentHashForRoute(exec.getRouteId(), agentId)
|
|
||||||
.orElse("");
|
|
||||||
ps.setString(col++, diagramHash); // diagram_content_hash
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int getBatchSize() {
|
|
||||||
return executions.size();
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
log.debug("Inserted batch of {} route executions into ClickHouse", executions.size());
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Optional<RawExecutionRow> findRawById(String executionId) {
|
|
||||||
String sql = """
|
|
||||||
SELECT execution_id, route_id, agent_id, status, start_time, end_time,
|
|
||||||
duration_ms, correlation_id, exchange_id, error_message, error_stacktrace,
|
|
||||||
diagram_content_hash,
|
|
||||||
processor_ids, processor_types, processor_statuses,
|
|
||||||
processor_starts, processor_ends, processor_durations,
|
|
||||||
processor_diagram_node_ids,
|
|
||||||
processor_error_messages, processor_error_stacktraces,
|
|
||||||
processor_depths, processor_parent_indexes
|
|
||||||
FROM route_executions
|
|
||||||
WHERE execution_id = ?
|
|
||||||
LIMIT 1
|
|
||||||
""";
|
|
||||||
|
|
||||||
List<RawExecutionRow> results = jdbcTemplate.query(sql, (rs, rowNum) -> {
|
|
||||||
// Extract parallel arrays from ClickHouse
|
|
||||||
String[] processorIds = toStringArray(rs.getArray("processor_ids"));
|
|
||||||
String[] processorTypes = toStringArray(rs.getArray("processor_types"));
|
|
||||||
String[] processorStatuses = toStringArray(rs.getArray("processor_statuses"));
|
|
||||||
Instant[] processorStarts = toInstantArray(rs.getArray("processor_starts"));
|
|
||||||
Instant[] processorEnds = toInstantArray(rs.getArray("processor_ends"));
|
|
||||||
long[] processorDurations = toLongArray(rs.getArray("processor_durations"));
|
|
||||||
String[] processorDiagramNodeIds = toStringArray(rs.getArray("processor_diagram_node_ids"));
|
|
||||||
String[] processorErrorMessages = toStringArray(rs.getArray("processor_error_messages"));
|
|
||||||
String[] processorErrorStacktraces = toStringArray(rs.getArray("processor_error_stacktraces"));
|
|
||||||
int[] processorDepths = toIntArrayFromShort(rs.getArray("processor_depths"));
|
|
||||||
int[] processorParentIndexes = toIntArray(rs.getArray("processor_parent_indexes"));
|
|
||||||
|
|
||||||
Timestamp endTs = rs.getTimestamp("end_time");
|
|
||||||
return new RawExecutionRow(
|
|
||||||
rs.getString("execution_id"),
|
|
||||||
rs.getString("route_id"),
|
|
||||||
rs.getString("agent_id"),
|
|
||||||
rs.getString("status"),
|
|
||||||
rs.getTimestamp("start_time").toInstant(),
|
|
||||||
endTs != null ? endTs.toInstant() : null,
|
|
||||||
rs.getLong("duration_ms"),
|
|
||||||
rs.getString("correlation_id"),
|
|
||||||
rs.getString("exchange_id"),
|
|
||||||
rs.getString("error_message"),
|
|
||||||
rs.getString("error_stacktrace"),
|
|
||||||
rs.getString("diagram_content_hash"),
|
|
||||||
processorIds, processorTypes, processorStatuses,
|
|
||||||
processorStarts, processorEnds, processorDurations,
|
|
||||||
processorDiagramNodeIds,
|
|
||||||
processorErrorMessages, processorErrorStacktraces,
|
|
||||||
processorDepths, processorParentIndexes
|
|
||||||
);
|
|
||||||
}, executionId);
|
|
||||||
|
|
||||||
return results.isEmpty() ? Optional.empty() : Optional.of(results.get(0));
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Find exchange snapshot data for a specific processor by index.
|
|
||||||
*
|
|
||||||
* @param executionId the execution ID
|
|
||||||
* @param processorIndex 0-based processor index
|
|
||||||
* @return map with inputBody, outputBody, inputHeaders, outputHeaders or empty if not found
|
|
||||||
*/
|
|
||||||
public Optional<java.util.Map<String, String>> findProcessorSnapshot(String executionId, int processorIndex) {
|
|
||||||
// ClickHouse arrays are 1-indexed in SQL
|
|
||||||
int chIndex = processorIndex + 1;
|
|
||||||
String sql = """
|
|
||||||
SELECT
|
|
||||||
processor_input_bodies[?] AS input_body,
|
|
||||||
processor_output_bodies[?] AS output_body,
|
|
||||||
processor_input_headers[?] AS input_headers,
|
|
||||||
processor_output_headers[?] AS output_headers,
|
|
||||||
length(processor_ids) AS proc_count
|
|
||||||
FROM route_executions
|
|
||||||
WHERE execution_id = ?
|
|
||||||
LIMIT 1
|
|
||||||
""";
|
|
||||||
|
|
||||||
List<java.util.Map<String, String>> results = jdbcTemplate.query(sql, (rs, rowNum) -> {
|
|
||||||
int procCount = rs.getInt("proc_count");
|
|
||||||
if (processorIndex < 0 || processorIndex >= procCount) {
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
var snapshot = new java.util.LinkedHashMap<String, String>();
|
|
||||||
snapshot.put("inputBody", rs.getString("input_body"));
|
|
||||||
snapshot.put("outputBody", rs.getString("output_body"));
|
|
||||||
snapshot.put("inputHeaders", rs.getString("input_headers"));
|
|
||||||
snapshot.put("outputHeaders", rs.getString("output_headers"));
|
|
||||||
return snapshot;
|
|
||||||
}, chIndex, chIndex, chIndex, chIndex, executionId);
|
|
||||||
|
|
||||||
if (results.isEmpty() || results.get(0) == null) {
|
|
||||||
return Optional.empty();
|
|
||||||
}
|
|
||||||
return Optional.of(results.get(0));
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- Array extraction helpers ---
|
|
||||||
|
|
||||||
private static String[] toStringArray(java.sql.Array sqlArray) throws SQLException {
|
|
||||||
if (sqlArray == null) return new String[0];
|
|
||||||
Object arr = sqlArray.getArray();
|
|
||||||
if (arr instanceof String[] sa) return sa;
|
|
||||||
if (arr instanceof Object[] oa) {
|
|
||||||
String[] result = new String[oa.length];
|
|
||||||
for (int i = 0; i < oa.length; i++) {
|
|
||||||
result[i] = oa[i] != null ? oa[i].toString() : "";
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
return new String[0];
|
|
||||||
}
|
|
||||||
|
|
||||||
private static Instant[] toInstantArray(java.sql.Array sqlArray) throws SQLException {
|
|
||||||
if (sqlArray == null) return new Instant[0];
|
|
||||||
Object arr = sqlArray.getArray();
|
|
||||||
if (arr instanceof Timestamp[] ts) {
|
|
||||||
Instant[] result = new Instant[ts.length];
|
|
||||||
for (int i = 0; i < ts.length; i++) {
|
|
||||||
result[i] = ts[i] != null ? ts[i].toInstant() : Instant.EPOCH;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
if (arr instanceof Object[] oa) {
|
|
||||||
Instant[] result = new Instant[oa.length];
|
|
||||||
for (int i = 0; i < oa.length; i++) {
|
|
||||||
if (oa[i] instanceof Timestamp ts) {
|
|
||||||
result[i] = ts.toInstant();
|
|
||||||
} else {
|
|
||||||
result[i] = Instant.EPOCH;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
return new Instant[0];
|
|
||||||
}
|
|
||||||
|
|
||||||
private static long[] toLongArray(java.sql.Array sqlArray) throws SQLException {
|
|
||||||
if (sqlArray == null) return new long[0];
|
|
||||||
Object arr = sqlArray.getArray();
|
|
||||||
if (arr instanceof long[] la) return la;
|
|
||||||
if (arr instanceof Long[] la) {
|
|
||||||
long[] result = new long[la.length];
|
|
||||||
for (int i = 0; i < la.length; i++) {
|
|
||||||
result[i] = la[i] != null ? la[i] : 0;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
if (arr instanceof Object[] oa) {
|
|
||||||
long[] result = new long[oa.length];
|
|
||||||
for (int i = 0; i < oa.length; i++) {
|
|
||||||
result[i] = oa[i] instanceof Number n ? n.longValue() : 0;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
return new long[0];
|
|
||||||
}
|
|
||||||
|
|
||||||
private static int[] toIntArray(java.sql.Array sqlArray) throws SQLException {
|
|
||||||
if (sqlArray == null) return new int[0];
|
|
||||||
Object arr = sqlArray.getArray();
|
|
||||||
if (arr instanceof int[] ia) return ia;
|
|
||||||
if (arr instanceof Integer[] ia) {
|
|
||||||
int[] result = new int[ia.length];
|
|
||||||
for (int i = 0; i < ia.length; i++) {
|
|
||||||
result[i] = ia[i] != null ? ia[i] : 0;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
if (arr instanceof Object[] oa) {
|
|
||||||
int[] result = new int[oa.length];
|
|
||||||
for (int i = 0; i < oa.length; i++) {
|
|
||||||
result[i] = oa[i] instanceof Number n ? n.intValue() : 0;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
return new int[0];
|
|
||||||
}
|
|
||||||
|
|
||||||
private static int[] toIntArrayFromShort(java.sql.Array sqlArray) throws SQLException {
|
|
||||||
if (sqlArray == null) return new int[0];
|
|
||||||
Object arr = sqlArray.getArray();
|
|
||||||
if (arr instanceof short[] sa) {
|
|
||||||
int[] result = new int[sa.length];
|
|
||||||
for (int i = 0; i < sa.length; i++) {
|
|
||||||
result[i] = sa[i];
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
if (arr instanceof int[] ia) return ia;
|
|
||||||
if (arr instanceof Object[] oa) {
|
|
||||||
int[] result = new int[oa.length];
|
|
||||||
for (int i = 0; i < oa.length; i++) {
|
|
||||||
result[i] = oa[i] instanceof Number n ? n.intValue() : 0;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
return new int[0];
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Internal record for a flattened processor with tree metadata.
|
|
||||||
*/
|
|
||||||
private record FlatProcessor(ProcessorExecution proc, int depth, int parentIndex) {}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Flatten the processor tree with depth and parent index metadata (DFS order).
|
|
||||||
*/
|
|
||||||
private List<FlatProcessor> flattenWithMetadata(List<ProcessorExecution> processors) {
|
|
||||||
if (processors == null || processors.isEmpty()) {
|
|
||||||
return List.of();
|
|
||||||
}
|
|
||||||
var result = new ArrayList<FlatProcessor>();
|
|
||||||
for (ProcessorExecution p : processors) {
|
|
||||||
flattenRecursive(p, 0, -1, result);
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
private void flattenRecursive(ProcessorExecution processor, int depth, int parentIdx,
|
|
||||||
List<FlatProcessor> result) {
|
|
||||||
int myIndex = result.size();
|
|
||||||
result.add(new FlatProcessor(processor, depth, parentIdx));
|
|
||||||
if (processor.getChildren() != null) {
|
|
||||||
for (ProcessorExecution child : processor.getChildren()) {
|
|
||||||
flattenRecursive(child, depth + 1, myIndex, result);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private void appendSnapshotText(ExchangeSnapshot snapshot,
|
|
||||||
StringBuilder allBodies, StringBuilder allHeaders) {
|
|
||||||
if (snapshot != null) {
|
|
||||||
allBodies.append(nullSafe(snapshot.getBody())).append(' ');
|
|
||||||
allHeaders.append(mapToJson(snapshot.getHeaders())).append(' ');
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private static String mapToJson(Map<String, String> map) {
|
|
||||||
if (map == null || map.isEmpty()) {
|
|
||||||
return "{}";
|
|
||||||
}
|
|
||||||
try {
|
|
||||||
return OBJECT_MAPPER.writeValueAsString(map);
|
|
||||||
} catch (JsonProcessingException e) {
|
|
||||||
log.warn("Failed to serialize headers map to JSON", e);
|
|
||||||
return "{}";
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private static String nullSafe(String value) {
|
|
||||||
return value != null ? value : "";
|
|
||||||
}
|
|
||||||
|
|
||||||
private static Timestamp toTimestamp(Instant instant) {
|
|
||||||
return instant != null ? Timestamp.from(instant) : Timestamp.from(Instant.EPOCH);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,67 +0,0 @@
|
|||||||
package com.cameleer3.server.app.storage;
|
|
||||||
|
|
||||||
import com.cameleer3.server.core.storage.MetricsRepository;
|
|
||||||
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
|
||||||
import org.slf4j.Logger;
|
|
||||||
import org.slf4j.LoggerFactory;
|
|
||||||
import org.springframework.jdbc.core.BatchPreparedStatementSetter;
|
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
import org.springframework.stereotype.Repository;
|
|
||||||
|
|
||||||
import java.sql.PreparedStatement;
|
|
||||||
import java.sql.SQLException;
|
|
||||||
import java.sql.Timestamp;
|
|
||||||
import java.time.Instant;
|
|
||||||
import java.util.HashMap;
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Map;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* ClickHouse implementation of {@link MetricsRepository}.
|
|
||||||
* <p>
|
|
||||||
* Performs batch inserts into the {@code agent_metrics} table.
|
|
||||||
*/
|
|
||||||
@Repository
|
|
||||||
public class ClickHouseMetricsRepository implements MetricsRepository {
|
|
||||||
|
|
||||||
private static final Logger log = LoggerFactory.getLogger(ClickHouseMetricsRepository.class);
|
|
||||||
|
|
||||||
private static final String INSERT_SQL = """
|
|
||||||
INSERT INTO agent_metrics (agent_id, collected_at, metric_name, metric_value, tags)
|
|
||||||
VALUES (?, ?, ?, ?, ?)
|
|
||||||
""";
|
|
||||||
|
|
||||||
private final JdbcTemplate jdbcTemplate;
|
|
||||||
|
|
||||||
public ClickHouseMetricsRepository(JdbcTemplate jdbcTemplate) {
|
|
||||||
this.jdbcTemplate = jdbcTemplate;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void insertBatch(List<MetricsSnapshot> metrics) {
|
|
||||||
if (metrics.isEmpty()) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
jdbcTemplate.batchUpdate(INSERT_SQL, new BatchPreparedStatementSetter() {
|
|
||||||
@Override
|
|
||||||
public void setValues(PreparedStatement ps, int i) throws SQLException {
|
|
||||||
MetricsSnapshot m = metrics.get(i);
|
|
||||||
ps.setString(1, m.agentId() != null ? m.agentId() : "");
|
|
||||||
ps.setObject(2, m.collectedAt() != null ? Timestamp.from(m.collectedAt()) : Timestamp.from(Instant.EPOCH));
|
|
||||||
ps.setString(3, m.metricName() != null ? m.metricName() : "");
|
|
||||||
ps.setDouble(4, m.metricValue());
|
|
||||||
// ClickHouse Map(String, String) -- pass as a java.util.Map
|
|
||||||
Map<String, String> tags = m.tags() != null ? m.tags() : new HashMap<>();
|
|
||||||
ps.setObject(5, tags);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int getBatchSize() {
|
|
||||||
return metrics.size();
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
log.debug("Inserted batch of {} metrics into ClickHouse", metrics.size());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,71 +0,0 @@
|
|||||||
package com.cameleer3.server.app.storage;
|
|
||||||
|
|
||||||
import com.cameleer3.server.core.security.OidcConfig;
|
|
||||||
import com.cameleer3.server.core.security.OidcConfigRepository;
|
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
import org.springframework.stereotype.Repository;
|
|
||||||
|
|
||||||
import java.sql.ResultSet;
|
|
||||||
import java.sql.SQLException;
|
|
||||||
import java.util.Arrays;
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Optional;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* ClickHouse implementation of {@link OidcConfigRepository}.
|
|
||||||
* Singleton row with {@code config_id = 'default'}, using ReplacingMergeTree.
|
|
||||||
*/
|
|
||||||
@Repository
|
|
||||||
public class ClickHouseOidcConfigRepository implements OidcConfigRepository {
|
|
||||||
|
|
||||||
private final JdbcTemplate jdbc;
|
|
||||||
|
|
||||||
public ClickHouseOidcConfigRepository(JdbcTemplate jdbc) {
|
|
||||||
this.jdbc = jdbc;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Optional<OidcConfig> find() {
|
|
||||||
List<OidcConfig> results = jdbc.query(
|
|
||||||
"SELECT enabled, issuer_uri, client_id, client_secret, roles_claim, default_roles, auto_signup, display_name_claim "
|
|
||||||
+ "FROM oidc_config FINAL WHERE config_id = 'default'",
|
|
||||||
this::mapRow
|
|
||||||
);
|
|
||||||
return results.isEmpty() ? Optional.empty() : Optional.of(results.get(0));
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void save(OidcConfig config) {
|
|
||||||
jdbc.update(
|
|
||||||
"INSERT INTO oidc_config (config_id, enabled, issuer_uri, client_id, client_secret, roles_claim, default_roles, auto_signup, display_name_claim, updated_at) "
|
|
||||||
+ "VALUES ('default', ?, ?, ?, ?, ?, ?, ?, ?, now64(3, 'UTC'))",
|
|
||||||
config.enabled(),
|
|
||||||
config.issuerUri(),
|
|
||||||
config.clientId(),
|
|
||||||
config.clientSecret(),
|
|
||||||
config.rolesClaim(),
|
|
||||||
config.defaultRoles().toArray(new String[0]),
|
|
||||||
config.autoSignup(),
|
|
||||||
config.displayNameClaim()
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void delete() {
|
|
||||||
jdbc.update("DELETE FROM oidc_config WHERE config_id = 'default'");
|
|
||||||
}
|
|
||||||
|
|
||||||
private OidcConfig mapRow(ResultSet rs, int rowNum) throws SQLException {
|
|
||||||
String[] rolesArray = (String[]) rs.getArray("default_roles").getArray();
|
|
||||||
return new OidcConfig(
|
|
||||||
rs.getBoolean("enabled"),
|
|
||||||
rs.getString("issuer_uri"),
|
|
||||||
rs.getString("client_id"),
|
|
||||||
rs.getString("client_secret"),
|
|
||||||
rs.getString("roles_claim"),
|
|
||||||
Arrays.asList(rolesArray),
|
|
||||||
rs.getBoolean("auto_signup"),
|
|
||||||
rs.getString("display_name_claim")
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,112 +0,0 @@
|
|||||||
package com.cameleer3.server.app.storage;
|
|
||||||
|
|
||||||
import com.cameleer3.server.core.security.UserInfo;
|
|
||||||
import com.cameleer3.server.core.security.UserRepository;
|
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
import org.springframework.stereotype.Repository;
|
|
||||||
|
|
||||||
import java.sql.ResultSet;
|
|
||||||
import java.sql.SQLException;
|
|
||||||
import java.time.Instant;
|
|
||||||
import java.util.Arrays;
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Optional;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* ClickHouse implementation of {@link UserRepository}.
|
|
||||||
* <p>
|
|
||||||
* Uses ReplacingMergeTree — reads use {@code FINAL} to get the latest version.
|
|
||||||
*/
|
|
||||||
@Repository
|
|
||||||
public class ClickHouseUserRepository implements UserRepository {
|
|
||||||
|
|
||||||
private final JdbcTemplate jdbc;
|
|
||||||
|
|
||||||
public ClickHouseUserRepository(JdbcTemplate jdbc) {
|
|
||||||
this.jdbc = jdbc;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Optional<UserInfo> findById(String userId) {
|
|
||||||
List<UserInfo> results = jdbc.query(
|
|
||||||
"SELECT user_id, provider, email, display_name, roles, created_at "
|
|
||||||
+ "FROM users FINAL WHERE user_id = ?",
|
|
||||||
this::mapRow,
|
|
||||||
userId
|
|
||||||
);
|
|
||||||
return results.isEmpty() ? Optional.empty() : Optional.of(results.get(0));
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public List<UserInfo> findAll() {
|
|
||||||
return jdbc.query(
|
|
||||||
"SELECT user_id, provider, email, display_name, roles, created_at FROM users FINAL ORDER BY user_id",
|
|
||||||
this::mapRow
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void upsert(UserInfo user) {
|
|
||||||
Optional<UserInfo> existing = findById(user.userId());
|
|
||||||
if (existing.isPresent()) {
|
|
||||||
UserInfo ex = existing.get();
|
|
||||||
// Skip write if nothing changed — avoids accumulating un-merged rows
|
|
||||||
if (ex.provider().equals(user.provider())
|
|
||||||
&& ex.email().equals(user.email())
|
|
||||||
&& ex.displayName().equals(user.displayName())
|
|
||||||
&& ex.roles().equals(user.roles())) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
jdbc.update(
|
|
||||||
"INSERT INTO users (user_id, provider, email, display_name, roles, created_at, updated_at) "
|
|
||||||
+ "SELECT user_id, ?, ?, ?, ?, created_at, now64(3, 'UTC') "
|
|
||||||
+ "FROM users FINAL WHERE user_id = ?",
|
|
||||||
user.provider(),
|
|
||||||
user.email(),
|
|
||||||
user.displayName(),
|
|
||||||
user.roles().toArray(new String[0]),
|
|
||||||
user.userId()
|
|
||||||
);
|
|
||||||
} else {
|
|
||||||
jdbc.update(
|
|
||||||
"INSERT INTO users (user_id, provider, email, display_name, roles, updated_at) "
|
|
||||||
+ "VALUES (?, ?, ?, ?, ?, now64(3, 'UTC'))",
|
|
||||||
user.userId(),
|
|
||||||
user.provider(),
|
|
||||||
user.email(),
|
|
||||||
user.displayName(),
|
|
||||||
user.roles().toArray(new String[0])
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void updateRoles(String userId, List<String> roles) {
|
|
||||||
// ReplacingMergeTree: insert a new row with updated_at to supersede the old one.
|
|
||||||
// Copy existing fields, update roles.
|
|
||||||
jdbc.update(
|
|
||||||
"INSERT INTO users (user_id, provider, email, display_name, roles, created_at, updated_at) "
|
|
||||||
+ "SELECT user_id, provider, email, display_name, ?, created_at, now64(3, 'UTC') "
|
|
||||||
+ "FROM users FINAL WHERE user_id = ?",
|
|
||||||
roles.toArray(new String[0]),
|
|
||||||
userId
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void delete(String userId) {
|
|
||||||
jdbc.update("DELETE FROM users WHERE user_id = ?", userId);
|
|
||||||
}
|
|
||||||
|
|
||||||
private UserInfo mapRow(ResultSet rs, int rowNum) throws SQLException {
|
|
||||||
String[] rolesArray = (String[]) rs.getArray("roles").getArray();
|
|
||||||
return new UserInfo(
|
|
||||||
rs.getString("user_id"),
|
|
||||||
rs.getString("provider"),
|
|
||||||
rs.getString("email"),
|
|
||||||
rs.getString("display_name"),
|
|
||||||
Arrays.asList(rolesArray),
|
|
||||||
rs.getTimestamp("created_at").toInstant()
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -2,7 +2,7 @@ package com.cameleer3.server.app.storage;
|
|||||||
|
|
||||||
import com.cameleer3.common.graph.RouteGraph;
|
import com.cameleer3.common.graph.RouteGraph;
|
||||||
import com.cameleer3.server.core.ingestion.TaggedDiagram;
|
import com.cameleer3.server.core.ingestion.TaggedDiagram;
|
||||||
import com.cameleer3.server.core.storage.DiagramRepository;
|
import com.cameleer3.server.core.storage.DiagramStore;
|
||||||
import com.fasterxml.jackson.core.JsonProcessingException;
|
import com.fasterxml.jackson.core.JsonProcessingException;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
import com.fasterxml.jackson.datatype.jsr310.JavaTimeModule;
|
import com.fasterxml.jackson.datatype.jsr310.JavaTimeModule;
|
||||||
@@ -22,19 +22,20 @@ import java.util.Map;
|
|||||||
import java.util.Optional;
|
import java.util.Optional;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* ClickHouse implementation of {@link DiagramRepository}.
|
* PostgreSQL implementation of {@link DiagramStore}.
|
||||||
* <p>
|
* <p>
|
||||||
* Stores route graphs as JSON with SHA-256 content-hash deduplication.
|
* Stores route graphs as JSON with SHA-256 content-hash deduplication.
|
||||||
* The underlying table uses ReplacingMergeTree keyed on content_hash.
|
* Uses {@code ON CONFLICT (content_hash) DO NOTHING} for idempotent inserts.
|
||||||
*/
|
*/
|
||||||
@Repository
|
@Repository
|
||||||
public class ClickHouseDiagramRepository implements DiagramRepository {
|
public class PostgresDiagramStore implements DiagramStore {
|
||||||
|
|
||||||
private static final Logger log = LoggerFactory.getLogger(ClickHouseDiagramRepository.class);
|
private static final Logger log = LoggerFactory.getLogger(PostgresDiagramStore.class);
|
||||||
|
|
||||||
private static final String INSERT_SQL = """
|
private static final String INSERT_SQL = """
|
||||||
INSERT INTO route_diagrams (content_hash, route_id, agent_id, definition)
|
INSERT INTO route_diagrams (content_hash, route_id, agent_id, definition)
|
||||||
VALUES (?, ?, ?, ?)
|
VALUES (?, ?, ?, ?::jsonb)
|
||||||
|
ON CONFLICT (content_hash) DO NOTHING
|
||||||
""";
|
""";
|
||||||
|
|
||||||
private static final String SELECT_BY_HASH = """
|
private static final String SELECT_BY_HASH = """
|
||||||
@@ -50,7 +51,7 @@ public class ClickHouseDiagramRepository implements DiagramRepository {
|
|||||||
private final JdbcTemplate jdbcTemplate;
|
private final JdbcTemplate jdbcTemplate;
|
||||||
private final ObjectMapper objectMapper;
|
private final ObjectMapper objectMapper;
|
||||||
|
|
||||||
public ClickHouseDiagramRepository(JdbcTemplate jdbcTemplate) {
|
public PostgresDiagramStore(JdbcTemplate jdbcTemplate) {
|
||||||
this.jdbcTemplate = jdbcTemplate;
|
this.jdbcTemplate = jdbcTemplate;
|
||||||
this.objectMapper = new ObjectMapper();
|
this.objectMapper = new ObjectMapper();
|
||||||
this.objectMapper.registerModule(new JavaTimeModule());
|
this.objectMapper.registerModule(new JavaTimeModule());
|
||||||
@@ -82,7 +83,7 @@ public class ClickHouseDiagramRepository implements DiagramRepository {
|
|||||||
try {
|
try {
|
||||||
return Optional.of(objectMapper.readValue(json, RouteGraph.class));
|
return Optional.of(objectMapper.readValue(json, RouteGraph.class));
|
||||||
} catch (JsonProcessingException e) {
|
} catch (JsonProcessingException e) {
|
||||||
log.error("Failed to deserialize RouteGraph from ClickHouse", e);
|
log.error("Failed to deserialize RouteGraph from PostgreSQL", e);
|
||||||
return Optional.empty();
|
return Optional.empty();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -0,0 +1,131 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
import org.springframework.jdbc.core.RowMapper;
|
||||||
|
import org.springframework.stereotype.Repository;
|
||||||
|
|
||||||
|
import java.sql.ResultSet;
|
||||||
|
import java.sql.SQLException;
|
||||||
|
import java.sql.Timestamp;
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Optional;
|
||||||
|
|
||||||
|
@Repository
|
||||||
|
public class PostgresExecutionStore implements ExecutionStore {
|
||||||
|
|
||||||
|
private final JdbcTemplate jdbc;
|
||||||
|
|
||||||
|
public PostgresExecutionStore(JdbcTemplate jdbc) {
|
||||||
|
this.jdbc = jdbc;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void upsert(ExecutionRecord execution) {
|
||||||
|
jdbc.update("""
|
||||||
|
INSERT INTO executions (execution_id, route_id, agent_id, group_name,
|
||||||
|
status, correlation_id, exchange_id, start_time, end_time,
|
||||||
|
duration_ms, error_message, error_stacktrace, diagram_content_hash,
|
||||||
|
created_at, updated_at)
|
||||||
|
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, now(), now())
|
||||||
|
ON CONFLICT (execution_id, start_time) DO UPDATE SET
|
||||||
|
status = CASE
|
||||||
|
WHEN EXCLUDED.status IN ('COMPLETED', 'FAILED')
|
||||||
|
AND executions.status = 'RUNNING'
|
||||||
|
THEN EXCLUDED.status
|
||||||
|
WHEN EXCLUDED.status = executions.status THEN executions.status
|
||||||
|
ELSE EXCLUDED.status
|
||||||
|
END,
|
||||||
|
end_time = COALESCE(EXCLUDED.end_time, executions.end_time),
|
||||||
|
duration_ms = COALESCE(EXCLUDED.duration_ms, executions.duration_ms),
|
||||||
|
error_message = COALESCE(EXCLUDED.error_message, executions.error_message),
|
||||||
|
error_stacktrace = COALESCE(EXCLUDED.error_stacktrace, executions.error_stacktrace),
|
||||||
|
diagram_content_hash = COALESCE(EXCLUDED.diagram_content_hash, executions.diagram_content_hash),
|
||||||
|
updated_at = now()
|
||||||
|
""",
|
||||||
|
execution.executionId(), execution.routeId(), execution.agentId(),
|
||||||
|
execution.groupName(), execution.status(), execution.correlationId(),
|
||||||
|
execution.exchangeId(),
|
||||||
|
Timestamp.from(execution.startTime()),
|
||||||
|
execution.endTime() != null ? Timestamp.from(execution.endTime()) : null,
|
||||||
|
execution.durationMs(), execution.errorMessage(),
|
||||||
|
execution.errorStacktrace(), execution.diagramContentHash());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void upsertProcessors(String executionId, Instant startTime,
|
||||||
|
String groupName, String routeId,
|
||||||
|
List<ProcessorRecord> processors) {
|
||||||
|
jdbc.batchUpdate("""
|
||||||
|
INSERT INTO processor_executions (execution_id, processor_id, processor_type,
|
||||||
|
diagram_node_id, group_name, route_id, depth, parent_processor_id,
|
||||||
|
status, start_time, end_time, duration_ms, error_message, error_stacktrace,
|
||||||
|
input_body, output_body, input_headers, output_headers)
|
||||||
|
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?::jsonb, ?::jsonb)
|
||||||
|
ON CONFLICT (execution_id, processor_id, start_time) DO UPDATE SET
|
||||||
|
status = EXCLUDED.status,
|
||||||
|
end_time = COALESCE(EXCLUDED.end_time, processor_executions.end_time),
|
||||||
|
duration_ms = COALESCE(EXCLUDED.duration_ms, processor_executions.duration_ms),
|
||||||
|
error_message = COALESCE(EXCLUDED.error_message, processor_executions.error_message),
|
||||||
|
error_stacktrace = COALESCE(EXCLUDED.error_stacktrace, processor_executions.error_stacktrace),
|
||||||
|
input_body = COALESCE(EXCLUDED.input_body, processor_executions.input_body),
|
||||||
|
output_body = COALESCE(EXCLUDED.output_body, processor_executions.output_body),
|
||||||
|
input_headers = COALESCE(EXCLUDED.input_headers, processor_executions.input_headers),
|
||||||
|
output_headers = COALESCE(EXCLUDED.output_headers, processor_executions.output_headers)
|
||||||
|
""",
|
||||||
|
processors.stream().map(p -> new Object[]{
|
||||||
|
p.executionId(), p.processorId(), p.processorType(),
|
||||||
|
p.diagramNodeId(), p.groupName(), p.routeId(),
|
||||||
|
p.depth(), p.parentProcessorId(), p.status(),
|
||||||
|
Timestamp.from(p.startTime()),
|
||||||
|
p.endTime() != null ? Timestamp.from(p.endTime()) : null,
|
||||||
|
p.durationMs(), p.errorMessage(), p.errorStacktrace(),
|
||||||
|
p.inputBody(), p.outputBody(), p.inputHeaders(), p.outputHeaders()
|
||||||
|
}).toList());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Optional<ExecutionRecord> findById(String executionId) {
|
||||||
|
List<ExecutionRecord> results = jdbc.query(
|
||||||
|
"SELECT * FROM executions WHERE execution_id = ? ORDER BY start_time DESC LIMIT 1",
|
||||||
|
EXECUTION_MAPPER, executionId);
|
||||||
|
return results.isEmpty() ? Optional.empty() : Optional.of(results.get(0));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<ProcessorRecord> findProcessors(String executionId) {
|
||||||
|
return jdbc.query(
|
||||||
|
"SELECT * FROM processor_executions WHERE execution_id = ? ORDER BY depth, start_time",
|
||||||
|
PROCESSOR_MAPPER, executionId);
|
||||||
|
}
|
||||||
|
|
||||||
|
private static final RowMapper<ExecutionRecord> EXECUTION_MAPPER = (rs, rowNum) ->
|
||||||
|
new ExecutionRecord(
|
||||||
|
rs.getString("execution_id"), rs.getString("route_id"),
|
||||||
|
rs.getString("agent_id"), rs.getString("group_name"),
|
||||||
|
rs.getString("status"), rs.getString("correlation_id"),
|
||||||
|
rs.getString("exchange_id"),
|
||||||
|
toInstant(rs, "start_time"), toInstant(rs, "end_time"),
|
||||||
|
rs.getObject("duration_ms") != null ? rs.getLong("duration_ms") : null,
|
||||||
|
rs.getString("error_message"), rs.getString("error_stacktrace"),
|
||||||
|
rs.getString("diagram_content_hash"));
|
||||||
|
|
||||||
|
private static final RowMapper<ProcessorRecord> PROCESSOR_MAPPER = (rs, rowNum) ->
|
||||||
|
new ProcessorRecord(
|
||||||
|
rs.getString("execution_id"), rs.getString("processor_id"),
|
||||||
|
rs.getString("processor_type"), rs.getString("diagram_node_id"),
|
||||||
|
rs.getString("group_name"), rs.getString("route_id"),
|
||||||
|
rs.getInt("depth"), rs.getString("parent_processor_id"),
|
||||||
|
rs.getString("status"),
|
||||||
|
toInstant(rs, "start_time"), toInstant(rs, "end_time"),
|
||||||
|
rs.getObject("duration_ms") != null ? rs.getLong("duration_ms") : null,
|
||||||
|
rs.getString("error_message"), rs.getString("error_stacktrace"),
|
||||||
|
rs.getString("input_body"), rs.getString("output_body"),
|
||||||
|
rs.getString("input_headers"), rs.getString("output_headers"));
|
||||||
|
|
||||||
|
private static Instant toInstant(ResultSet rs, String column) throws SQLException {
|
||||||
|
Timestamp ts = rs.getTimestamp(column);
|
||||||
|
return ts != null ? ts.toInstant() : null;
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,42 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.storage.MetricsStore;
|
||||||
|
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
||||||
|
import com.fasterxml.jackson.core.JsonProcessingException;
|
||||||
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
import org.springframework.stereotype.Repository;
|
||||||
|
|
||||||
|
import java.sql.Timestamp;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
@Repository
|
||||||
|
public class PostgresMetricsStore implements MetricsStore {
|
||||||
|
|
||||||
|
private static final ObjectMapper MAPPER = new ObjectMapper();
|
||||||
|
private final JdbcTemplate jdbc;
|
||||||
|
|
||||||
|
public PostgresMetricsStore(JdbcTemplate jdbc) {
|
||||||
|
this.jdbc = jdbc;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void insertBatch(List<MetricsSnapshot> snapshots) {
|
||||||
|
jdbc.batchUpdate("""
|
||||||
|
INSERT INTO agent_metrics (agent_id, metric_name, metric_value, tags,
|
||||||
|
collected_at, server_received_at)
|
||||||
|
VALUES (?, ?, ?, ?::jsonb, ?, now())
|
||||||
|
""",
|
||||||
|
snapshots.stream().map(s -> new Object[]{
|
||||||
|
s.agentId(), s.metricName(), s.metricValue(),
|
||||||
|
tagsToJson(s.tags()),
|
||||||
|
Timestamp.from(s.collectedAt())
|
||||||
|
}).toList());
|
||||||
|
}
|
||||||
|
|
||||||
|
private String tagsToJson(java.util.Map<String, String> tags) {
|
||||||
|
if (tags == null || tags.isEmpty()) return null;
|
||||||
|
try { return MAPPER.writeValueAsString(tags); }
|
||||||
|
catch (JsonProcessingException e) { return null; }
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,59 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.security.OidcConfig;
|
||||||
|
import com.cameleer3.server.core.security.OidcConfigRepository;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
import org.springframework.stereotype.Repository;
|
||||||
|
|
||||||
|
import java.sql.Array;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Optional;
|
||||||
|
|
||||||
|
@Repository
|
||||||
|
public class PostgresOidcConfigRepository implements OidcConfigRepository {
|
||||||
|
|
||||||
|
private final JdbcTemplate jdbc;
|
||||||
|
|
||||||
|
public PostgresOidcConfigRepository(JdbcTemplate jdbc) {
|
||||||
|
this.jdbc = jdbc;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Optional<OidcConfig> find() {
|
||||||
|
var results = jdbc.query(
|
||||||
|
"SELECT * FROM oidc_config WHERE config_id = 'default'",
|
||||||
|
(rs, rowNum) -> {
|
||||||
|
Array arr = rs.getArray("default_roles");
|
||||||
|
String[] roles = arr != null ? (String[]) arr.getArray() : new String[0];
|
||||||
|
return new OidcConfig(
|
||||||
|
rs.getBoolean("enabled"), rs.getString("issuer_uri"),
|
||||||
|
rs.getString("client_id"), rs.getString("client_secret"),
|
||||||
|
rs.getString("roles_claim"), List.of(roles),
|
||||||
|
rs.getBoolean("auto_signup"), rs.getString("display_name_claim"));
|
||||||
|
});
|
||||||
|
return results.isEmpty() ? Optional.empty() : Optional.of(results.get(0));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void save(OidcConfig config) {
|
||||||
|
jdbc.update("""
|
||||||
|
INSERT INTO oidc_config (config_id, enabled, issuer_uri, client_id, client_secret,
|
||||||
|
roles_claim, default_roles, auto_signup, display_name_claim, updated_at)
|
||||||
|
VALUES ('default', ?, ?, ?, ?, ?, ?, ?, ?, now())
|
||||||
|
ON CONFLICT (config_id) DO UPDATE SET
|
||||||
|
enabled = EXCLUDED.enabled, issuer_uri = EXCLUDED.issuer_uri,
|
||||||
|
client_id = EXCLUDED.client_id, client_secret = EXCLUDED.client_secret,
|
||||||
|
roles_claim = EXCLUDED.roles_claim, default_roles = EXCLUDED.default_roles,
|
||||||
|
auto_signup = EXCLUDED.auto_signup, display_name_claim = EXCLUDED.display_name_claim,
|
||||||
|
updated_at = now()
|
||||||
|
""",
|
||||||
|
config.enabled(), config.issuerUri(), config.clientId(), config.clientSecret(),
|
||||||
|
config.rolesClaim(), config.defaultRoles().toArray(new String[0]),
|
||||||
|
config.autoSignup(), config.displayNameClaim());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void delete() {
|
||||||
|
jdbc.update("DELETE FROM oidc_config WHERE config_id = 'default'");
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,187 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.search.ExecutionStats;
|
||||||
|
import com.cameleer3.server.core.search.StatsTimeseries;
|
||||||
|
import com.cameleer3.server.core.search.StatsTimeseries.TimeseriesBucket;
|
||||||
|
import com.cameleer3.server.core.storage.StatsStore;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
import org.springframework.stereotype.Repository;
|
||||||
|
|
||||||
|
import java.sql.Timestamp;
|
||||||
|
import java.time.Duration;
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.time.temporal.ChronoUnit;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
@Repository
|
||||||
|
public class PostgresStatsStore implements StatsStore {
|
||||||
|
|
||||||
|
private final JdbcTemplate jdbc;
|
||||||
|
|
||||||
|
public PostgresStatsStore(JdbcTemplate jdbc) {
|
||||||
|
this.jdbc = jdbc;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ExecutionStats stats(Instant from, Instant to) {
|
||||||
|
return queryStats("stats_1m_all", from, to, List.of());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ExecutionStats statsForApp(Instant from, Instant to, String groupName) {
|
||||||
|
return queryStats("stats_1m_app", from, to, List.of(
|
||||||
|
new Filter("group_name", groupName)));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ExecutionStats statsForRoute(Instant from, Instant to, String routeId, List<String> agentIds) {
|
||||||
|
// Note: agentIds is accepted for interface compatibility but not filterable
|
||||||
|
// on the continuous aggregate (it groups by route_id, not agent_id).
|
||||||
|
// All agents for the same route contribute to the same aggregate.
|
||||||
|
return queryStats("stats_1m_route", from, to, List.of(
|
||||||
|
new Filter("route_id", routeId)));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ExecutionStats statsForProcessor(Instant from, Instant to, String routeId, String processorType) {
|
||||||
|
return queryStats("stats_1m_processor", from, to, List.of(
|
||||||
|
new Filter("route_id", routeId),
|
||||||
|
new Filter("processor_type", processorType)));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public StatsTimeseries timeseries(Instant from, Instant to, int bucketCount) {
|
||||||
|
return queryTimeseries("stats_1m_all", from, to, bucketCount, List.of(), true);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public StatsTimeseries timeseriesForApp(Instant from, Instant to, int bucketCount, String groupName) {
|
||||||
|
return queryTimeseries("stats_1m_app", from, to, bucketCount, List.of(
|
||||||
|
new Filter("group_name", groupName)), true);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public StatsTimeseries timeseriesForRoute(Instant from, Instant to, int bucketCount,
|
||||||
|
String routeId, List<String> agentIds) {
|
||||||
|
return queryTimeseries("stats_1m_route", from, to, bucketCount, List.of(
|
||||||
|
new Filter("route_id", routeId)), true);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public StatsTimeseries timeseriesForProcessor(Instant from, Instant to, int bucketCount,
|
||||||
|
String routeId, String processorType) {
|
||||||
|
// stats_1m_processor does NOT have running_count column
|
||||||
|
return queryTimeseries("stats_1m_processor", from, to, bucketCount, List.of(
|
||||||
|
new Filter("route_id", routeId),
|
||||||
|
new Filter("processor_type", processorType)), false);
|
||||||
|
}
|
||||||
|
|
||||||
|
private record Filter(String column, String value) {}
|
||||||
|
|
||||||
|
private ExecutionStats queryStats(String view, Instant from, Instant to, List<Filter> filters) {
|
||||||
|
// running_count only exists on execution-level aggregates, not processor
|
||||||
|
boolean hasRunning = !view.equals("stats_1m_processor");
|
||||||
|
String runningCol = hasRunning ? "COALESCE(SUM(running_count), 0)" : "0";
|
||||||
|
|
||||||
|
String sql = "SELECT COALESCE(SUM(total_count), 0) AS total_count, " +
|
||||||
|
"COALESCE(SUM(failed_count), 0) AS failed_count, " +
|
||||||
|
"CASE WHEN SUM(total_count) > 0 THEN SUM(duration_sum) / SUM(total_count) ELSE 0 END AS avg_duration, " +
|
||||||
|
"COALESCE(MAX(p99_duration), 0) AS p99_duration, " +
|
||||||
|
runningCol + " AS active_count " +
|
||||||
|
"FROM " + view + " WHERE bucket >= ? AND bucket < ?";
|
||||||
|
|
||||||
|
List<Object> params = new ArrayList<>();
|
||||||
|
params.add(Timestamp.from(from));
|
||||||
|
params.add(Timestamp.from(to));
|
||||||
|
for (Filter f : filters) {
|
||||||
|
sql += " AND " + f.column() + " = ?";
|
||||||
|
params.add(f.value());
|
||||||
|
}
|
||||||
|
|
||||||
|
long totalCount = 0, failedCount = 0, avgDuration = 0, p99Duration = 0, activeCount = 0;
|
||||||
|
var currentResult = jdbc.query(sql, (rs, rowNum) -> new long[]{
|
||||||
|
rs.getLong("total_count"), rs.getLong("failed_count"),
|
||||||
|
rs.getLong("avg_duration"), rs.getLong("p99_duration"),
|
||||||
|
rs.getLong("active_count")
|
||||||
|
}, params.toArray());
|
||||||
|
if (!currentResult.isEmpty()) {
|
||||||
|
long[] r = currentResult.get(0);
|
||||||
|
totalCount = r[0]; failedCount = r[1]; avgDuration = r[2];
|
||||||
|
p99Duration = r[3]; activeCount = r[4];
|
||||||
|
}
|
||||||
|
|
||||||
|
// Previous period (shifted back 24h)
|
||||||
|
Instant prevFrom = from.minus(Duration.ofHours(24));
|
||||||
|
Instant prevTo = to.minus(Duration.ofHours(24));
|
||||||
|
List<Object> prevParams = new ArrayList<>();
|
||||||
|
prevParams.add(Timestamp.from(prevFrom));
|
||||||
|
prevParams.add(Timestamp.from(prevTo));
|
||||||
|
for (Filter f : filters) prevParams.add(f.value());
|
||||||
|
String prevSql = sql; // same shape, different time params
|
||||||
|
|
||||||
|
long prevTotal = 0, prevFailed = 0, prevAvg = 0, prevP99 = 0;
|
||||||
|
var prevResult = jdbc.query(prevSql, (rs, rowNum) -> new long[]{
|
||||||
|
rs.getLong("total_count"), rs.getLong("failed_count"),
|
||||||
|
rs.getLong("avg_duration"), rs.getLong("p99_duration")
|
||||||
|
}, prevParams.toArray());
|
||||||
|
if (!prevResult.isEmpty()) {
|
||||||
|
long[] r = prevResult.get(0);
|
||||||
|
prevTotal = r[0]; prevFailed = r[1]; prevAvg = r[2]; prevP99 = r[3];
|
||||||
|
}
|
||||||
|
|
||||||
|
// Today total (from midnight UTC)
|
||||||
|
Instant todayStart = Instant.now().truncatedTo(ChronoUnit.DAYS);
|
||||||
|
List<Object> todayParams = new ArrayList<>();
|
||||||
|
todayParams.add(Timestamp.from(todayStart));
|
||||||
|
todayParams.add(Timestamp.from(Instant.now()));
|
||||||
|
for (Filter f : filters) todayParams.add(f.value());
|
||||||
|
String todaySql = sql;
|
||||||
|
|
||||||
|
long totalToday = 0;
|
||||||
|
var todayResult = jdbc.query(todaySql, (rs, rowNum) -> rs.getLong("total_count"),
|
||||||
|
todayParams.toArray());
|
||||||
|
if (!todayResult.isEmpty()) totalToday = todayResult.get(0);
|
||||||
|
|
||||||
|
return new ExecutionStats(
|
||||||
|
totalCount, failedCount, avgDuration, p99Duration, activeCount,
|
||||||
|
totalToday, prevTotal, prevFailed, prevAvg, prevP99);
|
||||||
|
}
|
||||||
|
|
||||||
|
private StatsTimeseries queryTimeseries(String view, Instant from, Instant to,
|
||||||
|
int bucketCount, List<Filter> filters,
|
||||||
|
boolean hasRunningCount) {
|
||||||
|
long intervalSeconds = Duration.between(from, to).toSeconds() / Math.max(bucketCount, 1);
|
||||||
|
if (intervalSeconds < 60) intervalSeconds = 60;
|
||||||
|
|
||||||
|
String runningCol = hasRunningCount ? "COALESCE(SUM(running_count), 0)" : "0";
|
||||||
|
|
||||||
|
String sql = "SELECT time_bucket(? * INTERVAL '1 second', bucket) AS period, " +
|
||||||
|
"COALESCE(SUM(total_count), 0) AS total_count, " +
|
||||||
|
"COALESCE(SUM(failed_count), 0) AS failed_count, " +
|
||||||
|
"CASE WHEN SUM(total_count) > 0 THEN SUM(duration_sum) / SUM(total_count) ELSE 0 END AS avg_duration, " +
|
||||||
|
"COALESCE(MAX(p99_duration), 0) AS p99_duration, " +
|
||||||
|
runningCol + " AS active_count " +
|
||||||
|
"FROM " + view + " WHERE bucket >= ? AND bucket < ?";
|
||||||
|
|
||||||
|
List<Object> params = new ArrayList<>();
|
||||||
|
params.add(intervalSeconds);
|
||||||
|
params.add(Timestamp.from(from));
|
||||||
|
params.add(Timestamp.from(to));
|
||||||
|
for (Filter f : filters) {
|
||||||
|
sql += " AND " + f.column() + " = ?";
|
||||||
|
params.add(f.value());
|
||||||
|
}
|
||||||
|
sql += " GROUP BY period ORDER BY period";
|
||||||
|
|
||||||
|
List<TimeseriesBucket> buckets = jdbc.query(sql, (rs, rowNum) ->
|
||||||
|
new TimeseriesBucket(
|
||||||
|
rs.getTimestamp("period").toInstant(),
|
||||||
|
rs.getLong("total_count"), rs.getLong("failed_count"),
|
||||||
|
rs.getLong("avg_duration"), rs.getLong("p99_duration"),
|
||||||
|
rs.getLong("active_count")
|
||||||
|
), params.toArray());
|
||||||
|
|
||||||
|
return new StatsTimeseries(buckets);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,71 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.security.UserInfo;
|
||||||
|
import com.cameleer3.server.core.security.UserRepository;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
import org.springframework.stereotype.Repository;
|
||||||
|
|
||||||
|
import java.sql.Array;
|
||||||
|
import java.sql.Timestamp;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Optional;
|
||||||
|
|
||||||
|
@Repository
|
||||||
|
public class PostgresUserRepository implements UserRepository {
|
||||||
|
|
||||||
|
private final JdbcTemplate jdbc;
|
||||||
|
|
||||||
|
public PostgresUserRepository(JdbcTemplate jdbc) {
|
||||||
|
this.jdbc = jdbc;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Optional<UserInfo> findById(String userId) {
|
||||||
|
var results = jdbc.query(
|
||||||
|
"SELECT * FROM users WHERE user_id = ?",
|
||||||
|
(rs, rowNum) -> mapUser(rs), userId);
|
||||||
|
return results.isEmpty() ? Optional.empty() : Optional.of(results.get(0));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<UserInfo> findAll() {
|
||||||
|
return jdbc.query("SELECT * FROM users ORDER BY user_id",
|
||||||
|
(rs, rowNum) -> mapUser(rs));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void upsert(UserInfo user) {
|
||||||
|
jdbc.update("""
|
||||||
|
INSERT INTO users (user_id, provider, email, display_name, roles, created_at, updated_at)
|
||||||
|
VALUES (?, ?, ?, ?, ?, now(), now())
|
||||||
|
ON CONFLICT (user_id) DO UPDATE SET
|
||||||
|
provider = EXCLUDED.provider, email = EXCLUDED.email,
|
||||||
|
display_name = EXCLUDED.display_name, roles = EXCLUDED.roles,
|
||||||
|
updated_at = now()
|
||||||
|
""",
|
||||||
|
user.userId(), user.provider(), user.email(), user.displayName(),
|
||||||
|
user.roles().toArray(new String[0]));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void updateRoles(String userId, List<String> roles) {
|
||||||
|
jdbc.update("UPDATE users SET roles = ?, updated_at = now() WHERE user_id = ?",
|
||||||
|
roles.toArray(new String[0]), userId);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void delete(String userId) {
|
||||||
|
jdbc.update("DELETE FROM users WHERE user_id = ?", userId);
|
||||||
|
}
|
||||||
|
|
||||||
|
private UserInfo mapUser(java.sql.ResultSet rs) throws java.sql.SQLException {
|
||||||
|
Array rolesArray = rs.getArray("roles");
|
||||||
|
String[] roles = rolesArray != null ? (String[]) rolesArray.getArray() : new String[0];
|
||||||
|
java.sql.Timestamp ts = rs.getTimestamp("created_at");
|
||||||
|
java.time.Instant createdAt = ts != null ? ts.toInstant() : null;
|
||||||
|
return new UserInfo(
|
||||||
|
rs.getString("user_id"), rs.getString("provider"),
|
||||||
|
rs.getString("email"), rs.getString("display_name"),
|
||||||
|
List.of(roles), createdAt);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -3,10 +3,13 @@ server:
|
|||||||
|
|
||||||
spring:
|
spring:
|
||||||
datasource:
|
datasource:
|
||||||
url: jdbc:ch://localhost:8123/cameleer3
|
url: jdbc:postgresql://localhost:5432/cameleer3
|
||||||
username: cameleer
|
username: cameleer
|
||||||
password: cameleer_dev
|
password: ${CAMELEER_DB_PASSWORD:cameleer_dev}
|
||||||
driver-class-name: com.clickhouse.jdbc.ClickHouseDriver
|
driver-class-name: org.postgresql.Driver
|
||||||
|
flyway:
|
||||||
|
enabled: true
|
||||||
|
locations: classpath:db/migration
|
||||||
mvc:
|
mvc:
|
||||||
async:
|
async:
|
||||||
request-timeout: -1
|
request-timeout: -1
|
||||||
@@ -29,8 +32,14 @@ ingestion:
|
|||||||
batch-size: 5000
|
batch-size: 5000
|
||||||
flush-interval-ms: 1000
|
flush-interval-ms: 1000
|
||||||
|
|
||||||
clickhouse:
|
opensearch:
|
||||||
ttl-days: 30
|
url: ${OPENSEARCH_URL:http://localhost:9200}
|
||||||
|
queue-size: ${CAMELEER_OPENSEARCH_QUEUE_SIZE:10000}
|
||||||
|
debounce-ms: ${CAMELEER_OPENSEARCH_DEBOUNCE_MS:2000}
|
||||||
|
|
||||||
|
cameleer:
|
||||||
|
body-size-limit: ${CAMELEER_BODY_SIZE_LIMIT:16384}
|
||||||
|
retention-days: ${CAMELEER_RETENTION_DAYS:30}
|
||||||
|
|
||||||
security:
|
security:
|
||||||
access-token-expiry-ms: 3600000
|
access-token-expiry-ms: 3600000
|
||||||
|
|||||||
@@ -1,57 +0,0 @@
|
|||||||
-- Cameleer3 ClickHouse Schema
|
|
||||||
-- Tables for route executions, route diagrams, and agent metrics.
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS route_executions (
|
|
||||||
execution_id String,
|
|
||||||
route_id LowCardinality(String),
|
|
||||||
agent_id LowCardinality(String),
|
|
||||||
status LowCardinality(String),
|
|
||||||
start_time DateTime64(3, 'UTC'),
|
|
||||||
end_time Nullable(DateTime64(3, 'UTC')),
|
|
||||||
duration_ms UInt64,
|
|
||||||
correlation_id String,
|
|
||||||
exchange_id String,
|
|
||||||
error_message String DEFAULT '',
|
|
||||||
error_stacktrace String DEFAULT '',
|
|
||||||
-- Nested processor executions stored as parallel arrays
|
|
||||||
processor_ids Array(String),
|
|
||||||
processor_types Array(LowCardinality(String)),
|
|
||||||
processor_starts Array(DateTime64(3, 'UTC')),
|
|
||||||
processor_ends Array(DateTime64(3, 'UTC')),
|
|
||||||
processor_durations Array(UInt64),
|
|
||||||
processor_statuses Array(LowCardinality(String)),
|
|
||||||
-- Metadata
|
|
||||||
server_received_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC'),
|
|
||||||
-- Skip indexes
|
|
||||||
INDEX idx_correlation correlation_id TYPE bloom_filter GRANULARITY 4,
|
|
||||||
INDEX idx_error error_message TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4
|
|
||||||
)
|
|
||||||
ENGINE = MergeTree()
|
|
||||||
PARTITION BY toYYYYMMDD(start_time)
|
|
||||||
ORDER BY (agent_id, status, start_time, execution_id)
|
|
||||||
TTL toDateTime(start_time) + toIntervalDay(30)
|
|
||||||
SETTINGS ttl_only_drop_parts = 1;
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS route_diagrams (
|
|
||||||
content_hash String,
|
|
||||||
route_id LowCardinality(String),
|
|
||||||
agent_id LowCardinality(String),
|
|
||||||
definition String,
|
|
||||||
created_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC')
|
|
||||||
)
|
|
||||||
ENGINE = ReplacingMergeTree(created_at)
|
|
||||||
ORDER BY (content_hash);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS agent_metrics (
|
|
||||||
agent_id LowCardinality(String),
|
|
||||||
collected_at DateTime64(3, 'UTC'),
|
|
||||||
metric_name LowCardinality(String),
|
|
||||||
metric_value Float64,
|
|
||||||
tags Map(String, String),
|
|
||||||
server_received_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC')
|
|
||||||
)
|
|
||||||
ENGINE = MergeTree()
|
|
||||||
PARTITION BY toYYYYMMDD(collected_at)
|
|
||||||
ORDER BY (agent_id, metric_name, collected_at)
|
|
||||||
TTL toDateTime(collected_at) + toIntervalDay(30)
|
|
||||||
SETTINGS ttl_only_drop_parts = 1;
|
|
||||||
@@ -1,25 +0,0 @@
|
|||||||
-- Phase 2: Schema extension for search, detail, and diagram linking columns.
|
|
||||||
-- Adds exchange snapshot data, processor tree metadata, and diagram content hash.
|
|
||||||
|
|
||||||
ALTER TABLE route_executions
|
|
||||||
ADD COLUMN IF NOT EXISTS exchange_bodies String DEFAULT '',
|
|
||||||
ADD COLUMN IF NOT EXISTS exchange_headers String DEFAULT '',
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_depths Array(UInt16) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_parent_indexes Array(Int32) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_error_messages Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_error_stacktraces Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_input_bodies Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_output_bodies Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_input_headers Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_output_headers Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_diagram_node_ids Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS diagram_content_hash String DEFAULT '';
|
|
||||||
|
|
||||||
-- Skip indexes for full-text search on new text columns
|
|
||||||
ALTER TABLE route_executions
|
|
||||||
ADD INDEX IF NOT EXISTS idx_exchange_bodies exchange_bodies TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4,
|
|
||||||
ADD INDEX IF NOT EXISTS idx_exchange_headers exchange_headers TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4;
|
|
||||||
|
|
||||||
-- Skip index on error_stacktrace (not indexed in 01-schema.sql, needed for SRCH-05)
|
|
||||||
ALTER TABLE route_executions
|
|
||||||
ADD INDEX IF NOT EXISTS idx_error_stacktrace error_stacktrace TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4;
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
CREATE TABLE IF NOT EXISTS users (
|
|
||||||
user_id String,
|
|
||||||
provider LowCardinality(String),
|
|
||||||
email String DEFAULT '',
|
|
||||||
display_name String DEFAULT '',
|
|
||||||
roles Array(LowCardinality(String)),
|
|
||||||
created_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC'),
|
|
||||||
updated_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC')
|
|
||||||
) ENGINE = ReplacingMergeTree(updated_at)
|
|
||||||
ORDER BY (user_id);
|
|
||||||
@@ -1,13 +0,0 @@
|
|||||||
CREATE TABLE IF NOT EXISTS oidc_config (
|
|
||||||
config_id String DEFAULT 'default',
|
|
||||||
enabled Bool DEFAULT false,
|
|
||||||
issuer_uri String DEFAULT '',
|
|
||||||
client_id String DEFAULT '',
|
|
||||||
client_secret String DEFAULT '',
|
|
||||||
roles_claim String DEFAULT 'realm_access.roles',
|
|
||||||
default_roles Array(LowCardinality(String)),
|
|
||||||
auto_signup Bool DEFAULT true,
|
|
||||||
display_name_claim String DEFAULT 'name',
|
|
||||||
updated_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC')
|
|
||||||
) ENGINE = ReplacingMergeTree(updated_at)
|
|
||||||
ORDER BY (config_id);
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
ALTER TABLE oidc_config ADD COLUMN IF NOT EXISTS auto_signup Bool DEFAULT true;
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
ALTER TABLE oidc_config ADD COLUMN IF NOT EXISTS display_name_claim String DEFAULT 'name';
|
|
||||||
@@ -1,35 +0,0 @@
|
|||||||
-- Pre-aggregated 5-minute stats rollup for route executions.
|
|
||||||
-- Uses AggregatingMergeTree with -State/-Merge combinators so intermediate
|
|
||||||
-- aggregates can be merged across arbitrary time windows and dimensions.
|
|
||||||
|
|
||||||
-- Drop existing objects to allow schema changes (MV must be dropped before table)
|
|
||||||
DROP VIEW IF EXISTS route_execution_stats_5m_mv;
|
|
||||||
DROP TABLE IF EXISTS route_execution_stats_5m;
|
|
||||||
|
|
||||||
CREATE TABLE route_execution_stats_5m (
|
|
||||||
bucket DateTime('UTC'),
|
|
||||||
route_id LowCardinality(String),
|
|
||||||
agent_id LowCardinality(String),
|
|
||||||
total_count AggregateFunction(count),
|
|
||||||
failed_count AggregateFunction(countIf, UInt8),
|
|
||||||
duration_sum AggregateFunction(sum, UInt64),
|
|
||||||
p99_duration AggregateFunction(quantileTDigest(0.99), UInt64)
|
|
||||||
)
|
|
||||||
ENGINE = AggregatingMergeTree()
|
|
||||||
PARTITION BY toYYYYMMDD(bucket)
|
|
||||||
ORDER BY (agent_id, route_id, bucket)
|
|
||||||
TTL bucket + toIntervalDay(30)
|
|
||||||
SETTINGS ttl_only_drop_parts = 1;
|
|
||||||
|
|
||||||
CREATE MATERIALIZED VIEW route_execution_stats_5m_mv
|
|
||||||
TO route_execution_stats_5m
|
|
||||||
AS SELECT
|
|
||||||
toStartOfFiveMinutes(start_time) AS bucket,
|
|
||||||
route_id,
|
|
||||||
agent_id,
|
|
||||||
countState() AS total_count,
|
|
||||||
countIfState(status = 'FAILED') AS failed_count,
|
|
||||||
sumState(duration_ms) AS duration_sum,
|
|
||||||
quantileTDigestState(0.99)(duration_ms) AS p99_duration
|
|
||||||
FROM route_executions
|
|
||||||
GROUP BY bucket, route_id, agent_id;
|
|
||||||
@@ -1,16 +0,0 @@
|
|||||||
-- One-time idempotent backfill of existing route_executions into the
|
|
||||||
-- 5-minute stats rollup table. Safe for repeated execution — the WHERE
|
|
||||||
-- clause skips the INSERT if the target table already contains data.
|
|
||||||
|
|
||||||
INSERT INTO route_execution_stats_5m
|
|
||||||
SELECT
|
|
||||||
toStartOfFiveMinutes(start_time) AS bucket,
|
|
||||||
route_id,
|
|
||||||
agent_id,
|
|
||||||
countState() AS total_count,
|
|
||||||
countIfState(status = 'FAILED') AS failed_count,
|
|
||||||
sumState(duration_ms) AS duration_sum,
|
|
||||||
quantileTDigestState(0.99)(duration_ms) AS p99_duration
|
|
||||||
FROM route_executions
|
|
||||||
WHERE (SELECT count() FROM route_execution_stats_5m) = 0
|
|
||||||
GROUP BY bucket, route_id, agent_id;
|
|
||||||
@@ -0,0 +1,2 @@
|
|||||||
|
CREATE EXTENSION IF NOT EXISTS timescaledb;
|
||||||
|
CREATE EXTENSION IF NOT EXISTS timescaledb_toolkit;
|
||||||
@@ -0,0 +1,25 @@
|
|||||||
|
CREATE TABLE executions (
|
||||||
|
execution_id TEXT NOT NULL,
|
||||||
|
route_id TEXT NOT NULL,
|
||||||
|
agent_id TEXT NOT NULL,
|
||||||
|
group_name TEXT NOT NULL,
|
||||||
|
status TEXT NOT NULL,
|
||||||
|
correlation_id TEXT,
|
||||||
|
exchange_id TEXT,
|
||||||
|
start_time TIMESTAMPTZ NOT NULL,
|
||||||
|
end_time TIMESTAMPTZ,
|
||||||
|
duration_ms BIGINT,
|
||||||
|
error_message TEXT,
|
||||||
|
error_stacktrace TEXT,
|
||||||
|
diagram_content_hash TEXT,
|
||||||
|
created_at TIMESTAMPTZ NOT NULL DEFAULT now(),
|
||||||
|
updated_at TIMESTAMPTZ NOT NULL DEFAULT now(),
|
||||||
|
PRIMARY KEY (execution_id, start_time)
|
||||||
|
);
|
||||||
|
|
||||||
|
SELECT create_hypertable('executions', 'start_time', chunk_time_interval => INTERVAL '1 day');
|
||||||
|
|
||||||
|
CREATE INDEX idx_executions_agent_time ON executions (agent_id, start_time DESC);
|
||||||
|
CREATE INDEX idx_executions_route_time ON executions (route_id, start_time DESC);
|
||||||
|
CREATE INDEX idx_executions_group_time ON executions (group_name, start_time DESC);
|
||||||
|
CREATE INDEX idx_executions_correlation ON executions (correlation_id);
|
||||||
@@ -0,0 +1,28 @@
|
|||||||
|
CREATE TABLE processor_executions (
|
||||||
|
id BIGSERIAL,
|
||||||
|
execution_id TEXT NOT NULL,
|
||||||
|
processor_id TEXT NOT NULL,
|
||||||
|
processor_type TEXT NOT NULL,
|
||||||
|
diagram_node_id TEXT,
|
||||||
|
group_name TEXT NOT NULL,
|
||||||
|
route_id TEXT NOT NULL,
|
||||||
|
depth INT NOT NULL,
|
||||||
|
parent_processor_id TEXT,
|
||||||
|
status TEXT NOT NULL,
|
||||||
|
start_time TIMESTAMPTZ NOT NULL,
|
||||||
|
end_time TIMESTAMPTZ,
|
||||||
|
duration_ms BIGINT,
|
||||||
|
error_message TEXT,
|
||||||
|
error_stacktrace TEXT,
|
||||||
|
input_body TEXT,
|
||||||
|
output_body TEXT,
|
||||||
|
input_headers JSONB,
|
||||||
|
output_headers JSONB,
|
||||||
|
created_at TIMESTAMPTZ NOT NULL DEFAULT now(),
|
||||||
|
UNIQUE (execution_id, processor_id, start_time)
|
||||||
|
);
|
||||||
|
|
||||||
|
SELECT create_hypertable('processor_executions', 'start_time', chunk_time_interval => INTERVAL '1 day');
|
||||||
|
|
||||||
|
CREATE INDEX idx_proc_exec_execution ON processor_executions (execution_id);
|
||||||
|
CREATE INDEX idx_proc_exec_type_time ON processor_executions (processor_type, start_time DESC);
|
||||||
@@ -0,0 +1,12 @@
|
|||||||
|
CREATE TABLE agent_metrics (
|
||||||
|
agent_id TEXT NOT NULL,
|
||||||
|
metric_name TEXT NOT NULL,
|
||||||
|
metric_value DOUBLE PRECISION NOT NULL,
|
||||||
|
tags JSONB,
|
||||||
|
collected_at TIMESTAMPTZ NOT NULL,
|
||||||
|
server_received_at TIMESTAMPTZ NOT NULL DEFAULT now()
|
||||||
|
);
|
||||||
|
|
||||||
|
SELECT create_hypertable('agent_metrics', 'collected_at', chunk_time_interval => INTERVAL '1 day');
|
||||||
|
|
||||||
|
CREATE INDEX idx_metrics_agent_name ON agent_metrics (agent_id, metric_name, collected_at DESC);
|
||||||
@@ -0,0 +1,9 @@
|
|||||||
|
CREATE TABLE route_diagrams (
|
||||||
|
content_hash TEXT PRIMARY KEY,
|
||||||
|
route_id TEXT NOT NULL,
|
||||||
|
agent_id TEXT NOT NULL,
|
||||||
|
definition TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMPTZ NOT NULL DEFAULT now()
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE INDEX idx_diagrams_route_agent ON route_diagrams (route_id, agent_id);
|
||||||
@@ -0,0 +1,9 @@
|
|||||||
|
CREATE TABLE users (
|
||||||
|
user_id TEXT PRIMARY KEY,
|
||||||
|
provider TEXT NOT NULL,
|
||||||
|
email TEXT,
|
||||||
|
display_name TEXT,
|
||||||
|
roles TEXT[] NOT NULL DEFAULT '{}',
|
||||||
|
created_at TIMESTAMPTZ NOT NULL DEFAULT now(),
|
||||||
|
updated_at TIMESTAMPTZ NOT NULL DEFAULT now()
|
||||||
|
);
|
||||||
@@ -0,0 +1,12 @@
|
|||||||
|
CREATE TABLE oidc_config (
|
||||||
|
config_id TEXT PRIMARY KEY DEFAULT 'default',
|
||||||
|
enabled BOOLEAN NOT NULL DEFAULT false,
|
||||||
|
issuer_uri TEXT,
|
||||||
|
client_id TEXT,
|
||||||
|
client_secret TEXT,
|
||||||
|
roles_claim TEXT,
|
||||||
|
default_roles TEXT[] NOT NULL DEFAULT '{}',
|
||||||
|
auto_signup BOOLEAN DEFAULT false,
|
||||||
|
display_name_claim TEXT,
|
||||||
|
updated_at TIMESTAMPTZ NOT NULL DEFAULT now()
|
||||||
|
);
|
||||||
@@ -0,0 +1,87 @@
|
|||||||
|
-- Global stats
|
||||||
|
CREATE MATERIALIZED VIEW stats_1m_all
|
||||||
|
WITH (timescaledb.continuous, timescaledb.materialized_only = false) AS
|
||||||
|
SELECT
|
||||||
|
time_bucket('1 minute', start_time) AS bucket,
|
||||||
|
COUNT(*) AS total_count,
|
||||||
|
COUNT(*) FILTER (WHERE status = 'FAILED') AS failed_count,
|
||||||
|
COUNT(*) FILTER (WHERE status = 'RUNNING') AS running_count,
|
||||||
|
SUM(duration_ms) AS duration_sum,
|
||||||
|
MAX(duration_ms) AS duration_max,
|
||||||
|
approx_percentile(0.99, percentile_agg(duration_ms::DOUBLE PRECISION)) AS p99_duration
|
||||||
|
FROM executions
|
||||||
|
WHERE status IS NOT NULL
|
||||||
|
GROUP BY bucket
|
||||||
|
WITH NO DATA;
|
||||||
|
|
||||||
|
SELECT add_continuous_aggregate_policy('stats_1m_all',
|
||||||
|
start_offset => INTERVAL '1 hour',
|
||||||
|
end_offset => INTERVAL '1 minute',
|
||||||
|
schedule_interval => INTERVAL '1 minute');
|
||||||
|
|
||||||
|
-- Per-application stats
|
||||||
|
CREATE MATERIALIZED VIEW stats_1m_app
|
||||||
|
WITH (timescaledb.continuous, timescaledb.materialized_only = false) AS
|
||||||
|
SELECT
|
||||||
|
time_bucket('1 minute', start_time) AS bucket,
|
||||||
|
group_name,
|
||||||
|
COUNT(*) AS total_count,
|
||||||
|
COUNT(*) FILTER (WHERE status = 'FAILED') AS failed_count,
|
||||||
|
COUNT(*) FILTER (WHERE status = 'RUNNING') AS running_count,
|
||||||
|
SUM(duration_ms) AS duration_sum,
|
||||||
|
MAX(duration_ms) AS duration_max,
|
||||||
|
approx_percentile(0.99, percentile_agg(duration_ms::DOUBLE PRECISION)) AS p99_duration
|
||||||
|
FROM executions
|
||||||
|
WHERE status IS NOT NULL
|
||||||
|
GROUP BY bucket, group_name
|
||||||
|
WITH NO DATA;
|
||||||
|
|
||||||
|
SELECT add_continuous_aggregate_policy('stats_1m_app',
|
||||||
|
start_offset => INTERVAL '1 hour',
|
||||||
|
end_offset => INTERVAL '1 minute',
|
||||||
|
schedule_interval => INTERVAL '1 minute');
|
||||||
|
|
||||||
|
-- Per-route stats
|
||||||
|
CREATE MATERIALIZED VIEW stats_1m_route
|
||||||
|
WITH (timescaledb.continuous, timescaledb.materialized_only = false) AS
|
||||||
|
SELECT
|
||||||
|
time_bucket('1 minute', start_time) AS bucket,
|
||||||
|
group_name,
|
||||||
|
route_id,
|
||||||
|
COUNT(*) AS total_count,
|
||||||
|
COUNT(*) FILTER (WHERE status = 'FAILED') AS failed_count,
|
||||||
|
COUNT(*) FILTER (WHERE status = 'RUNNING') AS running_count,
|
||||||
|
SUM(duration_ms) AS duration_sum,
|
||||||
|
MAX(duration_ms) AS duration_max,
|
||||||
|
approx_percentile(0.99, percentile_agg(duration_ms::DOUBLE PRECISION)) AS p99_duration
|
||||||
|
FROM executions
|
||||||
|
WHERE status IS NOT NULL
|
||||||
|
GROUP BY bucket, group_name, route_id
|
||||||
|
WITH NO DATA;
|
||||||
|
|
||||||
|
SELECT add_continuous_aggregate_policy('stats_1m_route',
|
||||||
|
start_offset => INTERVAL '1 hour',
|
||||||
|
end_offset => INTERVAL '1 minute',
|
||||||
|
schedule_interval => INTERVAL '1 minute');
|
||||||
|
|
||||||
|
-- Per-processor stats (uses denormalized group_name/route_id on processor_executions)
|
||||||
|
CREATE MATERIALIZED VIEW stats_1m_processor
|
||||||
|
WITH (timescaledb.continuous, timescaledb.materialized_only = false) AS
|
||||||
|
SELECT
|
||||||
|
time_bucket('1 minute', start_time) AS bucket,
|
||||||
|
group_name,
|
||||||
|
route_id,
|
||||||
|
processor_type,
|
||||||
|
COUNT(*) AS total_count,
|
||||||
|
COUNT(*) FILTER (WHERE status = 'FAILED') AS failed_count,
|
||||||
|
SUM(duration_ms) AS duration_sum,
|
||||||
|
MAX(duration_ms) AS duration_max,
|
||||||
|
approx_percentile(0.99, percentile_agg(duration_ms::DOUBLE PRECISION)) AS p99_duration
|
||||||
|
FROM processor_executions
|
||||||
|
GROUP BY bucket, group_name, route_id, processor_type
|
||||||
|
WITH NO DATA;
|
||||||
|
|
||||||
|
SELECT add_continuous_aggregate_policy('stats_1m_processor',
|
||||||
|
start_offset => INTERVAL '1 hour',
|
||||||
|
end_offset => INTERVAL '1 minute',
|
||||||
|
schedule_interval => INTERVAL '1 minute');
|
||||||
@@ -1,82 +0,0 @@
|
|||||||
package com.cameleer3.server.app;
|
|
||||||
|
|
||||||
import org.springframework.beans.factory.annotation.Autowired;
|
|
||||||
import org.springframework.boot.test.context.SpringBootTest;
|
|
||||||
import org.springframework.jdbc.core.JdbcTemplate;
|
|
||||||
import org.springframework.test.context.ActiveProfiles;
|
|
||||||
import org.springframework.test.context.DynamicPropertyRegistry;
|
|
||||||
import org.springframework.test.context.DynamicPropertySource;
|
|
||||||
import org.testcontainers.clickhouse.ClickHouseContainer;
|
|
||||||
|
|
||||||
import org.junit.jupiter.api.BeforeAll;
|
|
||||||
|
|
||||||
import java.nio.charset.StandardCharsets;
|
|
||||||
import java.nio.file.Files;
|
|
||||||
import java.nio.file.Path;
|
|
||||||
import java.sql.Connection;
|
|
||||||
import java.sql.DriverManager;
|
|
||||||
import java.sql.Statement;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Base class for integration tests requiring a ClickHouse instance.
|
|
||||||
* <p>
|
|
||||||
* Uses Testcontainers to spin up a ClickHouse server and initializes the schema
|
|
||||||
* from {@code clickhouse/init/01-schema.sql} before the first test runs.
|
|
||||||
* Subclasses get a {@link JdbcTemplate} for direct database assertions.
|
|
||||||
* <p>
|
|
||||||
* Container lifecycle is managed manually (started once, shared across all test classes).
|
|
||||||
*/
|
|
||||||
@SpringBootTest(webEnvironment = SpringBootTest.WebEnvironment.RANDOM_PORT)
|
|
||||||
@ActiveProfiles("test")
|
|
||||||
public abstract class AbstractClickHouseIT {
|
|
||||||
|
|
||||||
protected static final ClickHouseContainer CLICKHOUSE;
|
|
||||||
|
|
||||||
static {
|
|
||||||
CLICKHOUSE = new ClickHouseContainer("clickhouse/clickhouse-server:25.3");
|
|
||||||
CLICKHOUSE.start();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Autowired
|
|
||||||
protected JdbcTemplate jdbcTemplate;
|
|
||||||
|
|
||||||
@DynamicPropertySource
|
|
||||||
static void overrideProperties(DynamicPropertyRegistry registry) {
|
|
||||||
registry.add("spring.datasource.url", CLICKHOUSE::getJdbcUrl);
|
|
||||||
registry.add("spring.datasource.username", CLICKHOUSE::getUsername);
|
|
||||||
registry.add("spring.datasource.password", CLICKHOUSE::getPassword);
|
|
||||||
}
|
|
||||||
|
|
||||||
@BeforeAll
|
|
||||||
static void initSchema() throws Exception {
|
|
||||||
// Surefire runs from the module directory; schema is in the project root
|
|
||||||
Path baseDir = Path.of("clickhouse/init");
|
|
||||||
if (!Files.exists(baseDir)) {
|
|
||||||
baseDir = Path.of("../clickhouse/init");
|
|
||||||
}
|
|
||||||
|
|
||||||
// Load all schema files in order
|
|
||||||
String[] schemaFiles = {"01-schema.sql", "02-search-columns.sql", "03-users.sql", "04-oidc-config.sql", "05-oidc-auto-signup.sql"};
|
|
||||||
|
|
||||||
try (Connection conn = DriverManager.getConnection(
|
|
||||||
CLICKHOUSE.getJdbcUrl(),
|
|
||||||
CLICKHOUSE.getUsername(),
|
|
||||||
CLICKHOUSE.getPassword());
|
|
||||||
Statement stmt = conn.createStatement()) {
|
|
||||||
|
|
||||||
for (String schemaFile : schemaFiles) {
|
|
||||||
Path schemaPath = baseDir.resolve(schemaFile);
|
|
||||||
if (Files.exists(schemaPath)) {
|
|
||||||
String sql = Files.readString(schemaPath, StandardCharsets.UTF_8);
|
|
||||||
// Execute each statement separately (separated by semicolons)
|
|
||||||
for (String statement : sql.split(";")) {
|
|
||||||
String trimmed = statement.trim();
|
|
||||||
if (!trimmed.isEmpty()) {
|
|
||||||
stmt.execute(trimmed);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,47 @@
|
|||||||
|
package com.cameleer3.server.app;
|
||||||
|
|
||||||
|
import org.opensearch.testcontainers.OpensearchContainer;
|
||||||
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
|
import org.springframework.boot.test.context.SpringBootTest;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
import org.springframework.test.context.ActiveProfiles;
|
||||||
|
import org.springframework.test.context.DynamicPropertyRegistry;
|
||||||
|
import org.springframework.test.context.DynamicPropertySource;
|
||||||
|
import org.testcontainers.containers.PostgreSQLContainer;
|
||||||
|
import org.testcontainers.utility.DockerImageName;
|
||||||
|
|
||||||
|
@SpringBootTest(webEnvironment = SpringBootTest.WebEnvironment.RANDOM_PORT)
|
||||||
|
@ActiveProfiles("test")
|
||||||
|
public abstract class AbstractPostgresIT {
|
||||||
|
|
||||||
|
private static final DockerImageName TIMESCALEDB_IMAGE =
|
||||||
|
DockerImageName.parse("timescale/timescaledb-ha:pg16")
|
||||||
|
.asCompatibleSubstituteFor("postgres");
|
||||||
|
|
||||||
|
static final PostgreSQLContainer<?> postgres;
|
||||||
|
static final OpensearchContainer<?> opensearch;
|
||||||
|
|
||||||
|
static {
|
||||||
|
postgres = new PostgreSQLContainer<>(TIMESCALEDB_IMAGE)
|
||||||
|
.withDatabaseName("cameleer3")
|
||||||
|
.withUsername("cameleer")
|
||||||
|
.withPassword("test");
|
||||||
|
postgres.start();
|
||||||
|
|
||||||
|
opensearch = new OpensearchContainer<>("opensearchproject/opensearch:2.19.0");
|
||||||
|
opensearch.start();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Autowired
|
||||||
|
protected JdbcTemplate jdbcTemplate;
|
||||||
|
|
||||||
|
@DynamicPropertySource
|
||||||
|
static void configureProperties(DynamicPropertyRegistry registry) {
|
||||||
|
registry.add("spring.datasource.url", postgres::getJdbcUrl);
|
||||||
|
registry.add("spring.datasource.username", postgres::getUsername);
|
||||||
|
registry.add("spring.datasource.password", postgres::getPassword);
|
||||||
|
registry.add("spring.datasource.driver-class-name", () -> "org.postgresql.Driver");
|
||||||
|
registry.add("spring.flyway.enabled", () -> "true");
|
||||||
|
registry.add("opensearch.url", opensearch::getHttpHostAddress);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -27,11 +27,40 @@ public class TestSecurityHelper {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Registers a test agent and returns a valid JWT access token for it.
|
* Registers a test agent and returns a valid JWT access token with AGENT role.
|
||||||
*/
|
*/
|
||||||
public String registerTestAgent(String agentId) {
|
public String registerTestAgent(String agentId) {
|
||||||
agentRegistryService.register(agentId, "test", "test-group", "1.0", List.of(), Map.of());
|
agentRegistryService.register(agentId, "test", "test-group", "1.0", List.of(), Map.of());
|
||||||
return jwtService.createAccessToken(agentId, "test-group");
|
return jwtService.createAccessToken(agentId, "test-group", List.of("AGENT"));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns a valid JWT access token with the given roles (no agent registration).
|
||||||
|
*/
|
||||||
|
public String createToken(String subject, String group, List<String> roles) {
|
||||||
|
return jwtService.createAccessToken(subject, group, roles);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns a valid JWT access token with OPERATOR role.
|
||||||
|
*/
|
||||||
|
public String operatorToken() {
|
||||||
|
// Subject must start with "user:" for JwtAuthenticationFilter to treat it as a UI user token
|
||||||
|
return jwtService.createAccessToken("user:test-operator", "user", List.of("OPERATOR"));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns a valid JWT access token with ADMIN role.
|
||||||
|
*/
|
||||||
|
public String adminToken() {
|
||||||
|
return jwtService.createAccessToken("user:test-admin", "user", List.of("ADMIN"));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns a valid JWT access token with VIEWER role.
|
||||||
|
*/
|
||||||
|
public String viewerToken() {
|
||||||
|
return jwtService.createAccessToken("user:test-viewer", "user", List.of("VIEWER"));
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
@@ -18,7 +18,7 @@ import java.util.UUID;
|
|||||||
|
|
||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
|
|
||||||
class AgentCommandControllerIT extends AbstractClickHouseIT {
|
class AgentCommandControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -29,11 +29,13 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
@Autowired
|
@Autowired
|
||||||
private TestSecurityHelper securityHelper;
|
private TestSecurityHelper securityHelper;
|
||||||
|
|
||||||
private String jwt;
|
private String agentJwt;
|
||||||
|
private String operatorJwt;
|
||||||
|
|
||||||
@BeforeEach
|
@BeforeEach
|
||||||
void setUp() {
|
void setUp() {
|
||||||
jwt = securityHelper.registerTestAgent("test-agent-command-it");
|
agentJwt = securityHelper.registerTestAgent("test-agent-command-it");
|
||||||
|
operatorJwt = securityHelper.operatorToken();
|
||||||
}
|
}
|
||||||
|
|
||||||
private ResponseEntity<String> registerAgent(String agentId, String name, String group) {
|
private ResponseEntity<String> registerAgent(String agentId, String name, String group) {
|
||||||
@@ -65,7 +67,7 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.postForEntity(
|
ResponseEntity<String> response = restTemplate.postForEntity(
|
||||||
"/api/v1/agents/" + agentId + "/commands",
|
"/api/v1/agents/" + agentId + "/commands",
|
||||||
new HttpEntity<>(commandJson, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(commandJson, securityHelper.authHeaders(operatorJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
@@ -88,7 +90,7 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.postForEntity(
|
ResponseEntity<String> response = restTemplate.postForEntity(
|
||||||
"/api/v1/agents/groups/" + group + "/commands",
|
"/api/v1/agents/groups/" + group + "/commands",
|
||||||
new HttpEntity<>(commandJson, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(commandJson, securityHelper.authHeaders(operatorJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
@@ -110,7 +112,7 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.postForEntity(
|
ResponseEntity<String> response = restTemplate.postForEntity(
|
||||||
"/api/v1/agents/commands",
|
"/api/v1/agents/commands",
|
||||||
new HttpEntity<>(commandJson, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(commandJson, securityHelper.authHeaders(operatorJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
@@ -131,7 +133,7 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
ResponseEntity<String> cmdResponse = restTemplate.postForEntity(
|
ResponseEntity<String> cmdResponse = restTemplate.postForEntity(
|
||||||
"/api/v1/agents/" + agentId + "/commands",
|
"/api/v1/agents/" + agentId + "/commands",
|
||||||
new HttpEntity<>(commandJson, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(commandJson, securityHelper.authHeaders(operatorJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
JsonNode cmdBody = objectMapper.readTree(cmdResponse.getBody());
|
JsonNode cmdBody = objectMapper.readTree(cmdResponse.getBody());
|
||||||
@@ -140,7 +142,7 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
ResponseEntity<Void> ackResponse = restTemplate.exchange(
|
ResponseEntity<Void> ackResponse = restTemplate.exchange(
|
||||||
"/api/v1/agents/" + agentId + "/commands/" + commandId + "/ack",
|
"/api/v1/agents/" + agentId + "/commands/" + commandId + "/ack",
|
||||||
HttpMethod.POST,
|
HttpMethod.POST,
|
||||||
new HttpEntity<>(securityHelper.authHeadersNoBody(jwt)),
|
new HttpEntity<>(securityHelper.authHeadersNoBody(agentJwt)),
|
||||||
Void.class);
|
Void.class);
|
||||||
|
|
||||||
assertThat(ackResponse.getStatusCode()).isEqualTo(HttpStatus.OK);
|
assertThat(ackResponse.getStatusCode()).isEqualTo(HttpStatus.OK);
|
||||||
@@ -154,7 +156,7 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
ResponseEntity<Void> response = restTemplate.exchange(
|
ResponseEntity<Void> response = restTemplate.exchange(
|
||||||
"/api/v1/agents/" + agentId + "/commands/nonexistent-cmd-id/ack",
|
"/api/v1/agents/" + agentId + "/commands/nonexistent-cmd-id/ack",
|
||||||
HttpMethod.POST,
|
HttpMethod.POST,
|
||||||
new HttpEntity<>(securityHelper.authHeadersNoBody(jwt)),
|
new HttpEntity<>(securityHelper.authHeadersNoBody(agentJwt)),
|
||||||
Void.class);
|
Void.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.NOT_FOUND);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.NOT_FOUND);
|
||||||
@@ -168,7 +170,7 @@ class AgentCommandControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.postForEntity(
|
ResponseEntity<String> response = restTemplate.postForEntity(
|
||||||
"/api/v1/agents/nonexistent-agent-xyz/commands",
|
"/api/v1/agents/nonexistent-agent-xyz/commands",
|
||||||
new HttpEntity<>(commandJson, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(commandJson, securityHelper.authHeaders(operatorJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.NOT_FOUND);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.NOT_FOUND);
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
@@ -16,7 +16,7 @@ import org.springframework.http.ResponseEntity;
|
|||||||
|
|
||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
|
|
||||||
class AgentRegistrationControllerIT extends AbstractClickHouseIT {
|
class AgentRegistrationControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -28,10 +28,12 @@ class AgentRegistrationControllerIT extends AbstractClickHouseIT {
|
|||||||
private TestSecurityHelper securityHelper;
|
private TestSecurityHelper securityHelper;
|
||||||
|
|
||||||
private String jwt;
|
private String jwt;
|
||||||
|
private String viewerJwt;
|
||||||
|
|
||||||
@BeforeEach
|
@BeforeEach
|
||||||
void setUp() {
|
void setUp() {
|
||||||
jwt = securityHelper.registerTestAgent("test-agent-registration-it");
|
jwt = securityHelper.registerTestAgent("test-agent-registration-it");
|
||||||
|
viewerJwt = securityHelper.viewerToken();
|
||||||
}
|
}
|
||||||
|
|
||||||
private ResponseEntity<String> registerAgent(String agentId, String name) {
|
private ResponseEntity<String> registerAgent(String agentId, String name) {
|
||||||
@@ -114,7 +116,7 @@ class AgentRegistrationControllerIT extends AbstractClickHouseIT {
|
|||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
"/api/v1/agents",
|
"/api/v1/agents",
|
||||||
HttpMethod.GET,
|
HttpMethod.GET,
|
||||||
new HttpEntity<>(securityHelper.authHeadersNoBody(jwt)),
|
new HttpEntity<>(securityHelper.authHeadersNoBody(viewerJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.OK);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.OK);
|
||||||
@@ -131,7 +133,7 @@ class AgentRegistrationControllerIT extends AbstractClickHouseIT {
|
|||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
"/api/v1/agents?status=LIVE",
|
"/api/v1/agents?status=LIVE",
|
||||||
HttpMethod.GET,
|
HttpMethod.GET,
|
||||||
new HttpEntity<>(securityHelper.authHeadersNoBody(jwt)),
|
new HttpEntity<>(securityHelper.authHeadersNoBody(viewerJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.OK);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.OK);
|
||||||
@@ -148,7 +150,7 @@ class AgentRegistrationControllerIT extends AbstractClickHouseIT {
|
|||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
"/api/v1/agents?status=INVALID",
|
"/api/v1/agents?status=INVALID",
|
||||||
HttpMethod.GET,
|
HttpMethod.GET,
|
||||||
new HttpEntity<>(securityHelper.authHeadersNoBody(jwt)),
|
new HttpEntity<>(securityHelper.authHeadersNoBody(viewerJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.BAD_REQUEST);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.BAD_REQUEST);
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
@@ -30,7 +30,7 @@ import java.util.concurrent.atomic.AtomicInteger;
|
|||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
import static org.awaitility.Awaitility.await;
|
import static org.awaitility.Awaitility.await;
|
||||||
|
|
||||||
class AgentSseControllerIT extends AbstractClickHouseIT {
|
class AgentSseControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -45,10 +45,12 @@ class AgentSseControllerIT extends AbstractClickHouseIT {
|
|||||||
private int port;
|
private int port;
|
||||||
|
|
||||||
private String jwt;
|
private String jwt;
|
||||||
|
private String operatorJwt;
|
||||||
|
|
||||||
@BeforeEach
|
@BeforeEach
|
||||||
void setUp() {
|
void setUp() {
|
||||||
jwt = securityHelper.registerTestAgent("test-agent-sse-it");
|
jwt = securityHelper.registerTestAgent("test-agent-sse-it");
|
||||||
|
operatorJwt = securityHelper.operatorToken();
|
||||||
}
|
}
|
||||||
|
|
||||||
private ResponseEntity<String> registerAgent(String agentId, String name, String group) {
|
private ResponseEntity<String> registerAgent(String agentId, String name, String group) {
|
||||||
@@ -76,7 +78,7 @@ class AgentSseControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
return restTemplate.postForEntity(
|
return restTemplate.postForEntity(
|
||||||
"/api/v1/agents/" + agentId + "/commands",
|
"/api/v1/agents/" + agentId + "/commands",
|
||||||
new HttpEntity<>(json, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(json, securityHelper.authHeaders(operatorJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.cameleer3.server.core.ingestion.IngestionService;
|
import com.cameleer3.server.core.ingestion.IngestionService;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
@@ -13,21 +13,20 @@ import org.springframework.http.HttpStatus;
|
|||||||
import org.springframework.http.ResponseEntity;
|
import org.springframework.http.ResponseEntity;
|
||||||
import org.springframework.test.context.TestPropertySource;
|
import org.springframework.test.context.TestPropertySource;
|
||||||
|
|
||||||
import static java.util.concurrent.TimeUnit.SECONDS;
|
|
||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
import static org.awaitility.Awaitility.await;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests backpressure behavior when write buffers are full.
|
* Tests backpressure behavior when the metrics write buffer is full.
|
||||||
* Uses a tiny buffer (capacity=5) and a very long flush interval
|
* <p>
|
||||||
* to prevent the scheduler from draining the buffer during the test.
|
* Execution and diagram ingestion are now synchronous (no buffers).
|
||||||
|
* Only the metrics pipeline still uses a write buffer with backpressure.
|
||||||
*/
|
*/
|
||||||
@TestPropertySource(properties = {
|
@TestPropertySource(properties = {
|
||||||
"ingestion.buffer-capacity=5",
|
"ingestion.buffer-capacity=5",
|
||||||
"ingestion.batch-size=5",
|
"ingestion.batch-size=5",
|
||||||
"ingestion.flush-interval-ms=60000" // 60s -- effectively no flush during test
|
"ingestion.flush-interval-ms=60000" // 60s -- effectively no flush during test
|
||||||
})
|
})
|
||||||
class BackpressureIT extends AbstractClickHouseIT {
|
class BackpressureIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -47,34 +46,31 @@ class BackpressureIT extends AbstractClickHouseIT {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void whenBufferFull_returns503WithRetryAfter() {
|
void whenMetricsBufferFull_returns503WithRetryAfter() {
|
||||||
// Wait for any initial scheduled flush to complete, then fill buffer via batch POST
|
// Fill the metrics buffer completely with a batch of 5
|
||||||
await().atMost(5, SECONDS).until(() -> ingestionService.getExecutionBufferDepth() == 0);
|
|
||||||
|
|
||||||
// Fill the buffer completely with a batch of 5
|
|
||||||
String batchJson = """
|
String batchJson = """
|
||||||
[
|
[
|
||||||
{"routeId":"bp-0","exchangeId":"bp-e0","status":"COMPLETED","startTime":"2026-03-11T10:00:00Z","durationMs":100,"processors":[]},
|
{"agentId":"bp-agent","timestamp":"2026-03-11T10:00:00Z","metrics":{}},
|
||||||
{"routeId":"bp-1","exchangeId":"bp-e1","status":"COMPLETED","startTime":"2026-03-11T10:00:00Z","durationMs":100,"processors":[]},
|
{"agentId":"bp-agent","timestamp":"2026-03-11T10:00:01Z","metrics":{}},
|
||||||
{"routeId":"bp-2","exchangeId":"bp-e2","status":"COMPLETED","startTime":"2026-03-11T10:00:00Z","durationMs":100,"processors":[]},
|
{"agentId":"bp-agent","timestamp":"2026-03-11T10:00:02Z","metrics":{}},
|
||||||
{"routeId":"bp-3","exchangeId":"bp-e3","status":"COMPLETED","startTime":"2026-03-11T10:00:00Z","durationMs":100,"processors":[]},
|
{"agentId":"bp-agent","timestamp":"2026-03-11T10:00:03Z","metrics":{}},
|
||||||
{"routeId":"bp-4","exchangeId":"bp-e4","status":"COMPLETED","startTime":"2026-03-11T10:00:00Z","durationMs":100,"processors":[]}
|
{"agentId":"bp-agent","timestamp":"2026-03-11T10:00:04Z","metrics":{}}
|
||||||
]
|
]
|
||||||
""";
|
""";
|
||||||
|
|
||||||
ResponseEntity<String> batchResponse = restTemplate.postForEntity(
|
ResponseEntity<String> batchResponse = restTemplate.postForEntity(
|
||||||
"/api/v1/data/executions",
|
"/api/v1/data/metrics",
|
||||||
new HttpEntity<>(batchJson, authHeaders),
|
new HttpEntity<>(batchJson, authHeaders),
|
||||||
String.class);
|
String.class);
|
||||||
assertThat(batchResponse.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(batchResponse.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
|
|
||||||
// Now buffer should be full -- next POST should get 503
|
// Now buffer should be full -- next POST should get 503
|
||||||
String overflowJson = """
|
String overflowJson = """
|
||||||
{"routeId":"bp-overflow","exchangeId":"bp-overflow-e","status":"COMPLETED","startTime":"2026-03-11T10:00:00Z","durationMs":100,"processors":[]}
|
[{"agentId":"bp-agent","timestamp":"2026-03-11T10:00:05Z","metrics":{}}]
|
||||||
""";
|
""";
|
||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.postForEntity(
|
ResponseEntity<String> response = restTemplate.postForEntity(
|
||||||
"/api/v1/data/executions",
|
"/api/v1/data/metrics",
|
||||||
new HttpEntity<>(overflowJson, authHeaders),
|
new HttpEntity<>(overflowJson, authHeaders),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
@@ -83,25 +79,17 @@ class BackpressureIT extends AbstractClickHouseIT {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void bufferedDataNotLost_afterBackpressure() {
|
void executionIngestion_isSynchronous_returnsAccepted() {
|
||||||
// Post data to the diagram buffer (separate from executions used above)
|
String json = """
|
||||||
for (int i = 0; i < 3; i++) {
|
{"routeId":"bp-sync","exchangeId":"bp-sync-e","status":"COMPLETED","startTime":"2026-03-11T10:00:00Z","durationMs":100,"processors":[]}
|
||||||
String json = String.format("""
|
""";
|
||||||
{
|
|
||||||
"routeId": "bp-persist-diagram-%d",
|
|
||||||
"version": 1,
|
|
||||||
"nodes": [],
|
|
||||||
"edges": []
|
|
||||||
}
|
|
||||||
""", i);
|
|
||||||
|
|
||||||
restTemplate.postForEntity(
|
ResponseEntity<String> response = restTemplate.postForEntity(
|
||||||
"/api/v1/data/diagrams",
|
"/api/v1/data/executions",
|
||||||
new HttpEntity<>(json, authHeaders),
|
new HttpEntity<>(json, authHeaders),
|
||||||
String.class);
|
String.class);
|
||||||
}
|
|
||||||
|
|
||||||
// Data is in the buffer. Verify the buffer has data.
|
// Synchronous ingestion always returns 202 (no buffer to overflow)
|
||||||
assertThat(ingestionService.getDiagramBufferDepth()).isGreaterThanOrEqualTo(3);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
@@ -23,7 +23,7 @@ import static org.awaitility.Awaitility.await;
|
|||||||
* Integration tests for the detail and processor snapshot endpoints.
|
* Integration tests for the detail and processor snapshot endpoints.
|
||||||
*/
|
*/
|
||||||
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
||||||
class DetailControllerIT extends AbstractClickHouseIT {
|
class DetailControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -34,6 +34,7 @@ class DetailControllerIT extends AbstractClickHouseIT {
|
|||||||
private final ObjectMapper objectMapper = new ObjectMapper();
|
private final ObjectMapper objectMapper = new ObjectMapper();
|
||||||
|
|
||||||
private String jwt;
|
private String jwt;
|
||||||
|
private String viewerJwt;
|
||||||
private String seededExecutionId;
|
private String seededExecutionId;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -43,6 +44,7 @@ class DetailControllerIT extends AbstractClickHouseIT {
|
|||||||
@BeforeAll
|
@BeforeAll
|
||||||
void seedTestData() {
|
void seedTestData() {
|
||||||
jwt = securityHelper.registerTestAgent("test-agent-detail-it");
|
jwt = securityHelper.registerTestAgent("test-agent-detail-it");
|
||||||
|
viewerJwt = securityHelper.viewerToken();
|
||||||
|
|
||||||
String json = """
|
String json = """
|
||||||
{
|
{
|
||||||
@@ -121,13 +123,13 @@ class DetailControllerIT extends AbstractClickHouseIT {
|
|||||||
// Wait for flush and get the execution_id
|
// Wait for flush and get the execution_id
|
||||||
await().atMost(10, SECONDS).untilAsserted(() -> {
|
await().atMost(10, SECONDS).untilAsserted(() -> {
|
||||||
Integer count = jdbcTemplate.queryForObject(
|
Integer count = jdbcTemplate.queryForObject(
|
||||||
"SELECT count() FROM route_executions WHERE route_id = 'detail-test-route'",
|
"SELECT count(*) FROM executions WHERE route_id = 'detail-test-route'",
|
||||||
Integer.class);
|
Integer.class);
|
||||||
assertThat(count).isGreaterThanOrEqualTo(1);
|
assertThat(count).isGreaterThanOrEqualTo(1);
|
||||||
});
|
});
|
||||||
|
|
||||||
seededExecutionId = jdbcTemplate.queryForObject(
|
seededExecutionId = jdbcTemplate.queryForObject(
|
||||||
"SELECT execution_id FROM route_executions WHERE route_id = 'detail-test-route' LIMIT 1",
|
"SELECT execution_id FROM executions WHERE route_id = 'detail-test-route' LIMIT 1",
|
||||||
String.class);
|
String.class);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -217,7 +219,7 @@ class DetailControllerIT extends AbstractClickHouseIT {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private ResponseEntity<String> detailGet(String path) {
|
private ResponseEntity<String> detailGet(String path) {
|
||||||
HttpHeaders headers = securityHelper.authHeadersNoBody(jwt);
|
HttpHeaders headers = securityHelper.authHeadersNoBody(viewerJwt);
|
||||||
return restTemplate.exchange(
|
return restTemplate.exchange(
|
||||||
"/api/v1/executions" + path,
|
"/api/v1/executions" + path,
|
||||||
HttpMethod.GET,
|
HttpMethod.GET,
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -15,7 +15,7 @@ import static java.util.concurrent.TimeUnit.SECONDS;
|
|||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
import static org.awaitility.Awaitility.await;
|
import static org.awaitility.Awaitility.await;
|
||||||
|
|
||||||
class DiagramControllerIT extends AbstractClickHouseIT {
|
class DiagramControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -53,7 +53,7 @@ class DiagramControllerIT extends AbstractClickHouseIT {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void postDiagram_dataAppearsInClickHouseAfterFlush() {
|
void postDiagram_dataAppearsAfterFlush() {
|
||||||
String json = """
|
String json = """
|
||||||
{
|
{
|
||||||
"routeId": "diagram-flush-route",
|
"routeId": "diagram-flush-route",
|
||||||
@@ -72,7 +72,7 @@ class DiagramControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
await().atMost(10, SECONDS).untilAsserted(() -> {
|
await().atMost(10, SECONDS).untilAsserted(() -> {
|
||||||
Integer count = jdbcTemplate.queryForObject(
|
Integer count = jdbcTemplate.queryForObject(
|
||||||
"SELECT count() FROM route_diagrams WHERE route_id = 'diagram-flush-route'",
|
"SELECT count(*) FROM route_diagrams WHERE route_id = 'diagram-flush-route'",
|
||||||
Integer.class);
|
Integer.class);
|
||||||
assertThat(count).isGreaterThanOrEqualTo(1);
|
assertThat(count).isGreaterThanOrEqualTo(1);
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -20,7 +20,7 @@ import static org.awaitility.Awaitility.await;
|
|||||||
* Integration tests for {@link DiagramRenderController}.
|
* Integration tests for {@link DiagramRenderController}.
|
||||||
* Seeds a diagram via the ingestion endpoint, then tests rendering.
|
* Seeds a diagram via the ingestion endpoint, then tests rendering.
|
||||||
*/
|
*/
|
||||||
class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
class DiagramRenderControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -29,6 +29,7 @@ class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
|||||||
private TestSecurityHelper securityHelper;
|
private TestSecurityHelper securityHelper;
|
||||||
|
|
||||||
private String jwt;
|
private String jwt;
|
||||||
|
private String viewerJwt;
|
||||||
private String contentHash;
|
private String contentHash;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -37,6 +38,7 @@ class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
|||||||
@BeforeEach
|
@BeforeEach
|
||||||
void seedDiagram() {
|
void seedDiagram() {
|
||||||
jwt = securityHelper.registerTestAgent("test-agent-diagram-render-it");
|
jwt = securityHelper.registerTestAgent("test-agent-diagram-render-it");
|
||||||
|
viewerJwt = securityHelper.viewerToken();
|
||||||
|
|
||||||
String json = """
|
String json = """
|
||||||
{
|
{
|
||||||
@@ -61,7 +63,7 @@ class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
|||||||
new HttpEntity<>(json, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(json, securityHelper.authHeaders(jwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
// Wait for flush to ClickHouse and retrieve the content hash
|
// Wait for flush to storage and retrieve the content hash
|
||||||
await().atMost(10, SECONDS).untilAsserted(() -> {
|
await().atMost(10, SECONDS).untilAsserted(() -> {
|
||||||
String hash = jdbcTemplate.queryForObject(
|
String hash = jdbcTemplate.queryForObject(
|
||||||
"SELECT content_hash FROM route_diagrams WHERE route_id = 'render-test-route' LIMIT 1",
|
"SELECT content_hash FROM route_diagrams WHERE route_id = 'render-test-route' LIMIT 1",
|
||||||
@@ -73,7 +75,7 @@ class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
void getSvg_withAcceptHeader_returnsSvg() {
|
void getSvg_withAcceptHeader_returnsSvg() {
|
||||||
HttpHeaders headers = securityHelper.authHeadersNoBody(jwt);
|
HttpHeaders headers = securityHelper.authHeadersNoBody(viewerJwt);
|
||||||
headers.set("Accept", "image/svg+xml");
|
headers.set("Accept", "image/svg+xml");
|
||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
@@ -90,7 +92,7 @@ class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
void getJson_withAcceptHeader_returnsJson() {
|
void getJson_withAcceptHeader_returnsJson() {
|
||||||
HttpHeaders headers = securityHelper.authHeadersNoBody(jwt);
|
HttpHeaders headers = securityHelper.authHeadersNoBody(viewerJwt);
|
||||||
headers.set("Accept", "application/json");
|
headers.set("Accept", "application/json");
|
||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
@@ -107,7 +109,7 @@ class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
void getNonExistentHash_returns404() {
|
void getNonExistentHash_returns404() {
|
||||||
HttpHeaders headers = securityHelper.authHeadersNoBody(jwt);
|
HttpHeaders headers = securityHelper.authHeadersNoBody(viewerJwt);
|
||||||
headers.set("Accept", "image/svg+xml");
|
headers.set("Accept", "image/svg+xml");
|
||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
@@ -122,7 +124,7 @@ class DiagramRenderControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
void getWithNoAcceptHeader_defaultsToSvg() {
|
void getWithNoAcceptHeader_defaultsToSvg() {
|
||||||
HttpHeaders headers = securityHelper.authHeadersNoBody(jwt);
|
HttpHeaders headers = securityHelper.authHeadersNoBody(viewerJwt);
|
||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
"/api/v1/diagrams/{hash}/render",
|
"/api/v1/diagrams/{hash}/render",
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -16,7 +16,7 @@ import static java.util.concurrent.TimeUnit.SECONDS;
|
|||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
import static org.awaitility.Awaitility.await;
|
import static org.awaitility.Awaitility.await;
|
||||||
|
|
||||||
class ExecutionControllerIT extends AbstractClickHouseIT {
|
class ExecutionControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -90,7 +90,7 @@ class ExecutionControllerIT extends AbstractClickHouseIT {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void postExecution_dataAppearsInClickHouseAfterFlush() {
|
void postExecution_dataAppearsAfterFlush() {
|
||||||
String json = """
|
String json = """
|
||||||
{
|
{
|
||||||
"routeId": "flush-test-route",
|
"routeId": "flush-test-route",
|
||||||
@@ -111,7 +111,7 @@ class ExecutionControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
await().atMost(10, SECONDS).untilAsserted(() -> {
|
await().atMost(10, SECONDS).untilAsserted(() -> {
|
||||||
Integer count = jdbcTemplate.queryForObject(
|
Integer count = jdbcTemplate.queryForObject(
|
||||||
"SELECT count() FROM route_executions WHERE route_id = 'flush-test-route'",
|
"SELECT count(*) FROM executions WHERE route_id = 'flush-test-route'",
|
||||||
Integer.class);
|
Integer.class);
|
||||||
assertThat(count).isGreaterThanOrEqualTo(1);
|
assertThat(count).isGreaterThanOrEqualTo(1);
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -16,7 +16,7 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
* Integration test for forward compatibility (API-05).
|
* Integration test for forward compatibility (API-05).
|
||||||
* Verifies that unknown JSON fields in request bodies do not cause deserialization errors.
|
* Verifies that unknown JSON fields in request bodies do not cause deserialization errors.
|
||||||
*/
|
*/
|
||||||
class ForwardCompatIT extends AbstractClickHouseIT {
|
class ForwardCompatIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
import org.springframework.beans.factory.annotation.Autowired;
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
import org.springframework.boot.test.web.client.TestRestTemplate;
|
import org.springframework.boot.test.web.client.TestRestTemplate;
|
||||||
@@ -8,9 +8,9 @@ import org.springframework.boot.test.web.client.TestRestTemplate;
|
|||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Integration tests for the health endpoint and ClickHouse TTL verification.
|
* Integration tests for the health endpoint.
|
||||||
*/
|
*/
|
||||||
class HealthControllerIT extends AbstractClickHouseIT {
|
class HealthControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -28,20 +28,4 @@ class HealthControllerIT extends AbstractClickHouseIT {
|
|||||||
var response = restTemplate.getForEntity("/api/v1/health", String.class);
|
var response = restTemplate.getForEntity("/api/v1/health", String.class);
|
||||||
assertThat(response.getStatusCode().value()).isEqualTo(200);
|
assertThat(response.getStatusCode().value()).isEqualTo(200);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
|
||||||
void ttlConfiguredOnRouteExecutions() {
|
|
||||||
String createTable = jdbcTemplate.queryForObject(
|
|
||||||
"SHOW CREATE TABLE route_executions", String.class);
|
|
||||||
assertThat(createTable).containsIgnoringCase("TTL");
|
|
||||||
assertThat(createTable).contains("toIntervalDay(30)");
|
|
||||||
}
|
|
||||||
|
|
||||||
@Test
|
|
||||||
void ttlConfiguredOnAgentMetrics() {
|
|
||||||
String createTable = jdbcTemplate.queryForObject(
|
|
||||||
"SHOW CREATE TABLE agent_metrics", String.class);
|
|
||||||
assertThat(createTable).containsIgnoringCase("TTL");
|
|
||||||
assertThat(createTable).contains("toIntervalDay(30)");
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -15,7 +15,7 @@ import static java.util.concurrent.TimeUnit.SECONDS;
|
|||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
import static org.awaitility.Awaitility.await;
|
import static org.awaitility.Awaitility.await;
|
||||||
|
|
||||||
class MetricsControllerIT extends AbstractClickHouseIT {
|
class MetricsControllerIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -52,7 +52,7 @@ class MetricsControllerIT extends AbstractClickHouseIT {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void postMetrics_dataAppearsInClickHouseAfterFlush() {
|
void postMetrics_dataAppearsAfterFlush() {
|
||||||
String json = """
|
String json = """
|
||||||
[{
|
[{
|
||||||
"agentId": "agent-flush-test",
|
"agentId": "agent-flush-test",
|
||||||
@@ -70,7 +70,7 @@ class MetricsControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
await().atMost(10, SECONDS).untilAsserted(() -> {
|
await().atMost(10, SECONDS).untilAsserted(() -> {
|
||||||
Integer count = jdbcTemplate.queryForObject(
|
Integer count = jdbcTemplate.queryForObject(
|
||||||
"SELECT count() FROM agent_metrics WHERE agent_id = 'agent-flush-test'",
|
"SELECT count(*) FROM agent_metrics WHERE agent_id = 'agent-flush-test'",
|
||||||
Integer.class);
|
Integer.class);
|
||||||
assertThat(count).isGreaterThanOrEqualTo(1);
|
assertThat(count).isGreaterThanOrEqualTo(1);
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
import org.springframework.beans.factory.annotation.Autowired;
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
import org.springframework.boot.test.web.client.TestRestTemplate;
|
import org.springframework.boot.test.web.client.TestRestTemplate;
|
||||||
@@ -10,7 +10,7 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
/**
|
/**
|
||||||
* Integration tests for OpenAPI documentation endpoints.
|
* Integration tests for OpenAPI documentation endpoints.
|
||||||
*/
|
*/
|
||||||
class OpenApiIT extends AbstractClickHouseIT {
|
class OpenApiIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
|
|||||||
@@ -1,12 +1,11 @@
|
|||||||
package com.cameleer3.server.app.controller;
|
package com.cameleer3.server.app.controller;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
import org.junit.jupiter.api.BeforeAll;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
import org.junit.jupiter.api.TestInstance;
|
|
||||||
import org.springframework.beans.factory.annotation.Autowired;
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
import org.springframework.boot.test.web.client.TestRestTemplate;
|
import org.springframework.boot.test.web.client.TestRestTemplate;
|
||||||
import org.springframework.http.HttpEntity;
|
import org.springframework.http.HttpEntity;
|
||||||
@@ -23,8 +22,7 @@ import static org.awaitility.Awaitility.await;
|
|||||||
* Integration tests for the search controller endpoints.
|
* Integration tests for the search controller endpoints.
|
||||||
* Tests all filter types independently and in combination.
|
* Tests all filter types independently and in combination.
|
||||||
*/
|
*/
|
||||||
@TestInstance(TestInstance.Lifecycle.PER_CLASS)
|
class SearchControllerIT extends AbstractPostgresIT {
|
||||||
class SearchControllerIT extends AbstractClickHouseIT {
|
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -34,15 +32,20 @@ class SearchControllerIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
private final ObjectMapper objectMapper = new ObjectMapper();
|
private final ObjectMapper objectMapper = new ObjectMapper();
|
||||||
|
|
||||||
private String jwt;
|
private static String jwt;
|
||||||
|
private static String viewerJwt;
|
||||||
|
private static boolean seeded;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Seed test data: Insert executions with varying statuses, times, durations,
|
* Seed test data: Insert executions with varying statuses, times, durations,
|
||||||
* correlationIds, error messages, and exchange snapshot data.
|
* correlationIds, error messages, and exchange snapshot data.
|
||||||
*/
|
*/
|
||||||
@BeforeAll
|
@BeforeEach
|
||||||
void seedTestData() {
|
void seedTestData() {
|
||||||
|
if (seeded) return;
|
||||||
|
seeded = true;
|
||||||
jwt = securityHelper.registerTestAgent("test-agent-search-it");
|
jwt = securityHelper.registerTestAgent("test-agent-search-it");
|
||||||
|
viewerJwt = securityHelper.viewerToken();
|
||||||
|
|
||||||
// Execution 1: COMPLETED, short duration, no errors
|
// Execution 1: COMPLETED, short duration, no errors
|
||||||
ingest("""
|
ingest("""
|
||||||
@@ -152,12 +155,18 @@ class SearchControllerIT extends AbstractClickHouseIT {
|
|||||||
""", i, i, i, i, i));
|
""", i, i, i, i, i));
|
||||||
}
|
}
|
||||||
|
|
||||||
// Wait for all data to flush
|
// Verify all data is in PostgreSQL (synchronous writes)
|
||||||
await().atMost(10, SECONDS).untilAsserted(() -> {
|
Integer count = jdbcTemplate.queryForObject(
|
||||||
Integer count = jdbcTemplate.queryForObject(
|
"SELECT count(*) FROM executions WHERE route_id LIKE 'search-route-%'",
|
||||||
"SELECT count() FROM route_executions WHERE route_id LIKE 'search-route-%'",
|
Integer.class);
|
||||||
Integer.class);
|
assertThat(count).isEqualTo(10);
|
||||||
assertThat(count).isEqualTo(10);
|
|
||||||
|
// Wait for async OpenSearch indexing (debounce + index time)
|
||||||
|
// Check for last seeded execution specifically to avoid false positives from other test classes
|
||||||
|
await().atMost(30, SECONDS).untilAsserted(() -> {
|
||||||
|
ResponseEntity<String> r = searchGet("?correlationId=corr-page-10");
|
||||||
|
JsonNode body = objectMapper.readTree(r.getBody());
|
||||||
|
assertThat(body.get("total").asLong()).isGreaterThanOrEqualTo(1);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -376,7 +385,7 @@ class SearchControllerIT extends AbstractClickHouseIT {
|
|||||||
return restTemplate.exchange(
|
return restTemplate.exchange(
|
||||||
"/api/v1/search/executions",
|
"/api/v1/search/executions",
|
||||||
HttpMethod.POST,
|
HttpMethod.POST,
|
||||||
new HttpEntity<>(jsonBody, securityHelper.authHeaders(jwt)),
|
new HttpEntity<>(jsonBody, securityHelper.authHeaders(viewerJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.interceptor;
|
package com.cameleer3.server.app.interceptor;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -18,7 +18,7 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
* With security enabled, requests to protected endpoints need JWT auth
|
* With security enabled, requests to protected endpoints need JWT auth
|
||||||
* to reach the interceptor layer.
|
* to reach the interceptor layer.
|
||||||
*/
|
*/
|
||||||
class ProtocolVersionIT extends AbstractClickHouseIT {
|
class ProtocolVersionIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
|
|||||||
@@ -0,0 +1,82 @@
|
|||||||
|
package com.cameleer3.server.app.search;
|
||||||
|
|
||||||
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
|
import com.cameleer3.server.core.search.ExecutionSummary;
|
||||||
|
import com.cameleer3.server.core.search.SearchRequest;
|
||||||
|
import com.cameleer3.server.core.search.SearchResult;
|
||||||
|
import com.cameleer3.server.core.storage.SearchIndex;
|
||||||
|
import com.cameleer3.server.core.storage.model.ExecutionDocument;
|
||||||
|
import com.cameleer3.server.core.storage.model.ExecutionDocument.ProcessorDoc;
|
||||||
|
import org.junit.jupiter.api.Test;
|
||||||
|
import org.opensearch.client.opensearch.OpenSearchClient;
|
||||||
|
import org.opensearch.client.opensearch.indices.RefreshRequest;
|
||||||
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
import static org.junit.jupiter.api.Assertions.*;
|
||||||
|
|
||||||
|
// Extends AbstractPostgresIT which provides both PostgreSQL and OpenSearch testcontainers
|
||||||
|
class OpenSearchIndexIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
|
@Autowired
|
||||||
|
SearchIndex searchIndex;
|
||||||
|
|
||||||
|
@Autowired
|
||||||
|
OpenSearchClient openSearchClient;
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void indexAndSearchByText() throws Exception {
|
||||||
|
Instant now = Instant.now();
|
||||||
|
ExecutionDocument doc = new ExecutionDocument(
|
||||||
|
"search-1", "route-a", "agent-1", "app-1",
|
||||||
|
"FAILED", "corr-1", "exch-1",
|
||||||
|
now, now.plusMillis(100), 100L,
|
||||||
|
"OrderNotFoundException: order-12345 not found", null,
|
||||||
|
List.of(new ProcessorDoc("proc-1", "log", "COMPLETED",
|
||||||
|
null, null, "request body with customer-99", null, null, null)));
|
||||||
|
|
||||||
|
searchIndex.index(doc);
|
||||||
|
refreshOpenSearchIndices();
|
||||||
|
|
||||||
|
SearchRequest request = new SearchRequest(
|
||||||
|
null, now.minusSeconds(60), now.plusSeconds(60),
|
||||||
|
null, null, null,
|
||||||
|
"OrderNotFoundException", null, null, null,
|
||||||
|
null, null, null, null, null,
|
||||||
|
0, 50, "startTime", "desc");
|
||||||
|
|
||||||
|
SearchResult<ExecutionSummary> result = searchIndex.search(request);
|
||||||
|
assertTrue(result.total() > 0);
|
||||||
|
assertEquals("search-1", result.data().get(0).executionId());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void wildcardSearchFindsSubstring() throws Exception {
|
||||||
|
Instant now = Instant.now();
|
||||||
|
ExecutionDocument doc = new ExecutionDocument(
|
||||||
|
"wild-1", "route-b", "agent-1", "app-1",
|
||||||
|
"COMPLETED", null, null,
|
||||||
|
now, now.plusMillis(50), 50L, null, null,
|
||||||
|
List.of(new ProcessorDoc("proc-1", "bean", "COMPLETED",
|
||||||
|
null, null, "UniquePayloadIdentifier12345", null, null, null)));
|
||||||
|
|
||||||
|
searchIndex.index(doc);
|
||||||
|
refreshOpenSearchIndices();
|
||||||
|
|
||||||
|
SearchRequest request = new SearchRequest(
|
||||||
|
null, now.minusSeconds(60), now.plusSeconds(60),
|
||||||
|
null, null, null,
|
||||||
|
"PayloadIdentifier", null, null, null,
|
||||||
|
null, null, null, null, null,
|
||||||
|
0, 50, "startTime", "desc");
|
||||||
|
|
||||||
|
SearchResult<ExecutionSummary> result = searchIndex.search(request);
|
||||||
|
assertTrue(result.total() > 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void refreshOpenSearchIndices() throws Exception {
|
||||||
|
openSearchClient.indices().refresh(RefreshRequest.of(r -> r.index("executions-*")));
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.security;
|
package com.cameleer3.server.app.security;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -17,7 +17,7 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
/**
|
/**
|
||||||
* Integration tests verifying bootstrap token validation on the registration endpoint.
|
* Integration tests verifying bootstrap token validation on the registration endpoint.
|
||||||
*/
|
*/
|
||||||
class BootstrapTokenIT extends AbstractClickHouseIT {
|
class BootstrapTokenIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.security;
|
package com.cameleer3.server.app.security;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.cameleer3.server.core.security.JwtService;
|
import com.cameleer3.server.core.security.JwtService;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
@@ -20,7 +20,7 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
/**
|
/**
|
||||||
* Integration tests for the JWT refresh flow.
|
* Integration tests for the JWT refresh flow.
|
||||||
*/
|
*/
|
||||||
class JwtRefreshIT extends AbstractClickHouseIT {
|
class JwtRefreshIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -153,13 +153,13 @@ class JwtRefreshIT extends AbstractClickHouseIT {
|
|||||||
JsonNode refreshBody2 = objectMapper.readTree(refreshResponse.getBody());
|
JsonNode refreshBody2 = objectMapper.readTree(refreshResponse.getBody());
|
||||||
String newAccessToken = refreshBody2.get("accessToken").asText();
|
String newAccessToken = refreshBody2.get("accessToken").asText();
|
||||||
|
|
||||||
// Use the new access token to hit a protected endpoint
|
// Use the new access token to hit a protected endpoint accessible by AGENT role
|
||||||
HttpHeaders authHeaders = new HttpHeaders();
|
HttpHeaders authHeaders = new HttpHeaders();
|
||||||
authHeaders.set("Authorization", "Bearer " + newAccessToken);
|
authHeaders.set("Authorization", "Bearer " + newAccessToken);
|
||||||
authHeaders.set("X-Cameleer-Protocol-Version", "1");
|
authHeaders.set("X-Cameleer-Protocol-Version", "1");
|
||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
"/api/v1/agents",
|
"/api/v1/search/executions",
|
||||||
HttpMethod.GET,
|
HttpMethod.GET,
|
||||||
new HttpEntity<>(authHeaders),
|
new HttpEntity<>(authHeaders),
|
||||||
String.class);
|
String.class);
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.security;
|
package com.cameleer3.server.app.security;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -19,7 +19,7 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
* Integration tests verifying that registration returns security credentials
|
* Integration tests verifying that registration returns security credentials
|
||||||
* and that those credentials can be used to access protected endpoints.
|
* and that those credentials can be used to access protected endpoints.
|
||||||
*/
|
*/
|
||||||
class RegistrationSecurityIT extends AbstractClickHouseIT {
|
class RegistrationSecurityIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -81,13 +81,13 @@ class RegistrationSecurityIT extends AbstractClickHouseIT {
|
|||||||
JsonNode regBody = objectMapper.readTree(regResponse.getBody());
|
JsonNode regBody = objectMapper.readTree(regResponse.getBody());
|
||||||
String accessToken = regBody.get("accessToken").asText();
|
String accessToken = regBody.get("accessToken").asText();
|
||||||
|
|
||||||
// Use the access token to hit a protected endpoint
|
// Use the access token to hit a protected endpoint accessible by AGENT role
|
||||||
HttpHeaders headers = new HttpHeaders();
|
HttpHeaders headers = new HttpHeaders();
|
||||||
headers.set("Authorization", "Bearer " + accessToken);
|
headers.set("Authorization", "Bearer " + accessToken);
|
||||||
headers.set("X-Cameleer-Protocol-Version", "1");
|
headers.set("X-Cameleer-Protocol-Version", "1");
|
||||||
|
|
||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
"/api/v1/agents",
|
"/api/v1/search/executions",
|
||||||
HttpMethod.GET,
|
HttpMethod.GET,
|
||||||
new HttpEntity<>(headers),
|
new HttpEntity<>(headers),
|
||||||
String.class);
|
String.class);
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.security;
|
package com.cameleer3.server.app.security;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -19,7 +19,7 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
* Integration tests verifying that the SecurityFilterChain correctly
|
* Integration tests verifying that the SecurityFilterChain correctly
|
||||||
* protects endpoints and allows public access where configured.
|
* protects endpoints and allows public access where configured.
|
||||||
*/
|
*/
|
||||||
class SecurityFilterIT extends AbstractClickHouseIT {
|
class SecurityFilterIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -28,10 +28,12 @@ class SecurityFilterIT extends AbstractClickHouseIT {
|
|||||||
private TestSecurityHelper securityHelper;
|
private TestSecurityHelper securityHelper;
|
||||||
|
|
||||||
private String jwt;
|
private String jwt;
|
||||||
|
private String viewerJwt;
|
||||||
|
|
||||||
@BeforeEach
|
@BeforeEach
|
||||||
void setUp() {
|
void setUp() {
|
||||||
jwt = securityHelper.registerTestAgent("test-agent-security-filter-it");
|
jwt = securityHelper.registerTestAgent("test-agent-security-filter-it");
|
||||||
|
viewerJwt = securityHelper.viewerToken();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@@ -53,7 +55,7 @@ class SecurityFilterIT extends AbstractClickHouseIT {
|
|||||||
ResponseEntity<String> response = restTemplate.exchange(
|
ResponseEntity<String> response = restTemplate.exchange(
|
||||||
"/api/v1/agents",
|
"/api/v1/agents",
|
||||||
HttpMethod.GET,
|
HttpMethod.GET,
|
||||||
new HttpEntity<>(securityHelper.authHeadersNoBody(jwt)),
|
new HttpEntity<>(securityHelper.authHeadersNoBody(viewerJwt)),
|
||||||
String.class);
|
String.class);
|
||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.OK);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.OK);
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
package com.cameleer3.server.app.security;
|
package com.cameleer3.server.app.security;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import com.cameleer3.server.core.security.Ed25519SigningService;
|
import com.cameleer3.server.core.security.Ed25519SigningService;
|
||||||
import com.fasterxml.jackson.databind.JsonNode;
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
import com.fasterxml.jackson.databind.ObjectMapper;
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
@@ -44,7 +45,7 @@ import static org.awaitility.Awaitility.await;
|
|||||||
* open SSE stream (with JWT query param) -> push config-update command (with JWT) ->
|
* open SSE stream (with JWT query param) -> push config-update command (with JWT) ->
|
||||||
* receive SSE event -> verify signature field against server's Ed25519 public key.
|
* receive SSE event -> verify signature field against server's Ed25519 public key.
|
||||||
*/
|
*/
|
||||||
class SseSigningIT extends AbstractClickHouseIT {
|
class SseSigningIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -52,6 +53,9 @@ class SseSigningIT extends AbstractClickHouseIT {
|
|||||||
@Autowired
|
@Autowired
|
||||||
private ObjectMapper objectMapper;
|
private ObjectMapper objectMapper;
|
||||||
|
|
||||||
|
@Autowired
|
||||||
|
private TestSecurityHelper securityHelper;
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private Ed25519SigningService ed25519SigningService;
|
private Ed25519SigningService ed25519SigningService;
|
||||||
|
|
||||||
@@ -165,6 +169,7 @@ class SseSigningIT extends AbstractClickHouseIT {
|
|||||||
String agentId = "sse-sign-it-" + UUID.randomUUID().toString().substring(0, 8);
|
String agentId = "sse-sign-it-" + UUID.randomUUID().toString().substring(0, 8);
|
||||||
JsonNode registration = registerAgentWithAuth(agentId);
|
JsonNode registration = registerAgentWithAuth(agentId);
|
||||||
String accessToken = registration.get("accessToken").asText();
|
String accessToken = registration.get("accessToken").asText();
|
||||||
|
String operatorToken = securityHelper.operatorToken();
|
||||||
String serverPublicKey = registration.get("serverPublicKey").asText();
|
String serverPublicKey = registration.get("serverPublicKey").asText();
|
||||||
|
|
||||||
SseStream stream = openSseStream(agentId, accessToken);
|
SseStream stream = openSseStream(agentId, accessToken);
|
||||||
@@ -177,7 +182,7 @@ class SseSigningIT extends AbstractClickHouseIT {
|
|||||||
await().atMost(10, TimeUnit.SECONDS).pollInterval(200, TimeUnit.MILLISECONDS)
|
await().atMost(10, TimeUnit.SECONDS).pollInterval(200, TimeUnit.MILLISECONDS)
|
||||||
.ignoreExceptions()
|
.ignoreExceptions()
|
||||||
.until(() -> {
|
.until(() -> {
|
||||||
sendCommand(agentId, "config-update", originalPayload, accessToken);
|
sendCommand(agentId, "config-update", originalPayload, operatorToken);
|
||||||
List<String> lines = stream.snapshot();
|
List<String> lines = stream.snapshot();
|
||||||
return lines.stream().anyMatch(l -> l.contains("event:config-update"));
|
return lines.stream().anyMatch(l -> l.contains("event:config-update"));
|
||||||
});
|
});
|
||||||
@@ -221,6 +226,7 @@ class SseSigningIT extends AbstractClickHouseIT {
|
|||||||
String agentId = "sse-sign-trace-" + UUID.randomUUID().toString().substring(0, 8);
|
String agentId = "sse-sign-trace-" + UUID.randomUUID().toString().substring(0, 8);
|
||||||
JsonNode registration = registerAgentWithAuth(agentId);
|
JsonNode registration = registerAgentWithAuth(agentId);
|
||||||
String accessToken = registration.get("accessToken").asText();
|
String accessToken = registration.get("accessToken").asText();
|
||||||
|
String operatorToken = securityHelper.operatorToken();
|
||||||
String serverPublicKey = registration.get("serverPublicKey").asText();
|
String serverPublicKey = registration.get("serverPublicKey").asText();
|
||||||
|
|
||||||
SseStream stream = openSseStream(agentId, accessToken);
|
SseStream stream = openSseStream(agentId, accessToken);
|
||||||
@@ -232,7 +238,7 @@ class SseSigningIT extends AbstractClickHouseIT {
|
|||||||
await().atMost(10, TimeUnit.SECONDS).pollInterval(200, TimeUnit.MILLISECONDS)
|
await().atMost(10, TimeUnit.SECONDS).pollInterval(200, TimeUnit.MILLISECONDS)
|
||||||
.ignoreExceptions()
|
.ignoreExceptions()
|
||||||
.until(() -> {
|
.until(() -> {
|
||||||
sendCommand(agentId, "deep-trace", originalPayload, accessToken);
|
sendCommand(agentId, "deep-trace", originalPayload, operatorToken);
|
||||||
List<String> lines = stream.snapshot();
|
List<String> lines = stream.snapshot();
|
||||||
return lines.stream().anyMatch(l -> l.contains("event:deep-trace"));
|
return lines.stream().anyMatch(l -> l.contains("event:deep-trace"));
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.storage;
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -11,15 +11,13 @@ import org.springframework.http.HttpHeaders;
|
|||||||
import org.springframework.http.HttpStatus;
|
import org.springframework.http.HttpStatus;
|
||||||
import org.springframework.http.ResponseEntity;
|
import org.springframework.http.ResponseEntity;
|
||||||
|
|
||||||
import static java.util.concurrent.TimeUnit.SECONDS;
|
|
||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
import static org.awaitility.Awaitility.await;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Integration test proving that diagram_content_hash is populated during
|
* Integration test proving that diagram_content_hash is populated during
|
||||||
* execution ingestion when a RouteGraph exists for the same route+agent.
|
* execution ingestion when a RouteGraph exists for the same route+agent.
|
||||||
*/
|
*/
|
||||||
class DiagramLinkingIT extends AbstractClickHouseIT {
|
class DiagramLinkingIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -59,12 +57,10 @@ class DiagramLinkingIT extends AbstractClickHouseIT {
|
|||||||
String.class);
|
String.class);
|
||||||
assertThat(diagramResponse.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(diagramResponse.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
|
|
||||||
await().atMost(10, SECONDS).untilAsserted(() -> {
|
String diagramHash = jdbcTemplate.queryForObject(
|
||||||
String hash = jdbcTemplate.queryForObject(
|
"SELECT content_hash FROM route_diagrams WHERE route_id = 'diagram-link-route' LIMIT 1",
|
||||||
"SELECT content_hash FROM route_diagrams WHERE route_id = 'diagram-link-route' LIMIT 1",
|
String.class);
|
||||||
String.class);
|
assertThat(diagramHash).isNotNull().isNotEmpty();
|
||||||
assertThat(hash).isNotNull().isNotEmpty();
|
|
||||||
});
|
|
||||||
|
|
||||||
String executionJson = """
|
String executionJson = """
|
||||||
{
|
{
|
||||||
@@ -95,16 +91,14 @@ class DiagramLinkingIT extends AbstractClickHouseIT {
|
|||||||
String.class);
|
String.class);
|
||||||
assertThat(execResponse.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(execResponse.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
|
|
||||||
await().atMost(10, SECONDS).ignoreExceptions().untilAsserted(() -> {
|
String hash = jdbcTemplate.queryForObject(
|
||||||
String hash = jdbcTemplate.queryForObject(
|
"SELECT diagram_content_hash FROM executions WHERE route_id = 'diagram-link-route'",
|
||||||
"SELECT diagram_content_hash FROM route_executions WHERE route_id = 'diagram-link-route'",
|
String.class);
|
||||||
String.class);
|
assertThat(hash)
|
||||||
assertThat(hash)
|
.isNotNull()
|
||||||
.isNotNull()
|
.isNotEmpty()
|
||||||
.isNotEmpty()
|
.hasSize(64)
|
||||||
.hasSize(64)
|
.matches("[a-f0-9]{64}");
|
||||||
.matches("[a-f0-9]{64}");
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@@ -138,13 +132,11 @@ class DiagramLinkingIT extends AbstractClickHouseIT {
|
|||||||
String.class);
|
String.class);
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
|
|
||||||
await().atMost(10, SECONDS).ignoreExceptions().untilAsserted(() -> {
|
String hash = jdbcTemplate.queryForObject(
|
||||||
String hash = jdbcTemplate.queryForObject(
|
"SELECT diagram_content_hash FROM executions WHERE route_id = 'no-diagram-route'",
|
||||||
"SELECT diagram_content_hash FROM route_executions WHERE route_id = 'no-diagram-route'",
|
String.class);
|
||||||
String.class);
|
assertThat(hash)
|
||||||
assertThat(hash)
|
.isNotNull()
|
||||||
.isNotNull()
|
.isEmpty();
|
||||||
.isEmpty();
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,36 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
|
import org.junit.jupiter.api.Test;
|
||||||
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
|
||||||
|
import static org.junit.jupiter.api.Assertions.*;
|
||||||
|
|
||||||
|
class FlywayMigrationIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
|
@Autowired
|
||||||
|
JdbcTemplate jdbcTemplate;
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void allMigrationsApplySuccessfully() {
|
||||||
|
// Verify core tables exist
|
||||||
|
Integer execCount = jdbcTemplate.queryForObject(
|
||||||
|
"SELECT COUNT(*) FROM executions", Integer.class);
|
||||||
|
assertEquals(0, execCount);
|
||||||
|
|
||||||
|
Integer procCount = jdbcTemplate.queryForObject(
|
||||||
|
"SELECT COUNT(*) FROM processor_executions", Integer.class);
|
||||||
|
assertEquals(0, procCount);
|
||||||
|
|
||||||
|
Integer userCount = jdbcTemplate.queryForObject(
|
||||||
|
"SELECT COUNT(*) FROM users", Integer.class);
|
||||||
|
assertEquals(0, userCount);
|
||||||
|
|
||||||
|
// Verify continuous aggregates exist
|
||||||
|
Integer caggCount = jdbcTemplate.queryForObject(
|
||||||
|
"SELECT COUNT(*) FROM timescaledb_information.continuous_aggregates",
|
||||||
|
Integer.class);
|
||||||
|
assertEquals(4, caggCount);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
package com.cameleer3.server.app.storage;
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
import com.cameleer3.server.app.AbstractClickHouseIT;
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
import com.cameleer3.server.app.TestSecurityHelper;
|
import com.cameleer3.server.app.TestSecurityHelper;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
@@ -11,18 +11,16 @@ import org.springframework.http.HttpHeaders;
|
|||||||
import org.springframework.http.HttpStatus;
|
import org.springframework.http.HttpStatus;
|
||||||
import org.springframework.http.ResponseEntity;
|
import org.springframework.http.ResponseEntity;
|
||||||
|
|
||||||
import java.util.Arrays;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
import static java.util.concurrent.TimeUnit.SECONDS;
|
|
||||||
import static org.assertj.core.api.Assertions.assertThat;
|
import static org.assertj.core.api.Assertions.assertThat;
|
||||||
import static org.awaitility.Awaitility.await;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Integration test verifying that Phase 2 schema columns are correctly populated
|
* Integration test verifying that processor execution data is correctly populated
|
||||||
* during ingestion of route executions with nested processors and exchange data.
|
* during ingestion of route executions with nested processors and exchange data.
|
||||||
*/
|
*/
|
||||||
class IngestionSchemaIT extends AbstractClickHouseIT {
|
class IngestionSchemaIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
@Autowired
|
@Autowired
|
||||||
private TestRestTemplate restTemplate;
|
private TestRestTemplate restTemplate;
|
||||||
@@ -39,7 +37,7 @@ class IngestionSchemaIT extends AbstractClickHouseIT {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void processorTreeMetadata_depthsAndParentIndexesCorrect() {
|
void processorTreeMetadata_depthsAndParentIdsCorrect() {
|
||||||
String json = """
|
String json = """
|
||||||
{
|
{
|
||||||
"routeId": "schema-test-tree",
|
"routeId": "schema-test-tree",
|
||||||
@@ -94,44 +92,46 @@ class IngestionSchemaIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
postExecution(json);
|
postExecution(json);
|
||||||
|
|
||||||
await().atMost(30, SECONDS).ignoreExceptions().untilAsserted(() -> {
|
// Verify execution row exists
|
||||||
var depths = queryArray(
|
Integer execCount = jdbcTemplate.queryForObject(
|
||||||
"SELECT processor_depths FROM route_executions WHERE route_id = 'schema-test-tree'");
|
"SELECT count(*) FROM executions WHERE execution_id = 'ex-tree-1'",
|
||||||
assertThat(depths).containsExactly("0", "1", "2");
|
Integer.class);
|
||||||
|
assertThat(execCount).isEqualTo(1);
|
||||||
|
|
||||||
var parentIndexes = queryArray(
|
// Verify processors were flattened into processor_executions
|
||||||
"SELECT processor_parent_indexes FROM route_executions WHERE route_id = 'schema-test-tree'");
|
List<Map<String, Object>> processors = jdbcTemplate.queryForList(
|
||||||
assertThat(parentIndexes).containsExactly("-1", "0", "1");
|
"SELECT processor_id, processor_type, depth, parent_processor_id, " +
|
||||||
|
"diagram_node_id, input_body, output_body, input_headers " +
|
||||||
|
"FROM processor_executions WHERE execution_id = 'ex-tree-1' " +
|
||||||
|
"ORDER BY depth, processor_id");
|
||||||
|
assertThat(processors).hasSize(3);
|
||||||
|
|
||||||
var diagramNodeIds = queryArray(
|
// Root processor: depth=0, no parent
|
||||||
"SELECT processor_diagram_node_ids FROM route_executions WHERE route_id = 'schema-test-tree'");
|
assertThat(processors.get(0).get("processor_id")).isEqualTo("root-proc");
|
||||||
assertThat(diagramNodeIds).containsExactly("node-root", "node-child", "node-grandchild");
|
assertThat(((Number) processors.get(0).get("depth")).intValue()).isEqualTo(0);
|
||||||
|
assertThat(processors.get(0).get("parent_processor_id")).isNull();
|
||||||
|
assertThat(processors.get(0).get("diagram_node_id")).isEqualTo("node-root");
|
||||||
|
assertThat(processors.get(0).get("input_body")).isEqualTo("root-input");
|
||||||
|
assertThat(processors.get(0).get("output_body")).isEqualTo("root-output");
|
||||||
|
assertThat(processors.get(0).get("input_headers").toString()).contains("Content-Type");
|
||||||
|
|
||||||
String bodies = jdbcTemplate.queryForObject(
|
// Child processor: depth=1, parent=root-proc
|
||||||
"SELECT exchange_bodies FROM route_executions WHERE route_id = 'schema-test-tree'",
|
assertThat(processors.get(1).get("processor_id")).isEqualTo("child-proc");
|
||||||
String.class);
|
assertThat(((Number) processors.get(1).get("depth")).intValue()).isEqualTo(1);
|
||||||
assertThat(bodies).contains("root-input");
|
assertThat(processors.get(1).get("parent_processor_id")).isEqualTo("root-proc");
|
||||||
assertThat(bodies).contains("root-output");
|
assertThat(processors.get(1).get("diagram_node_id")).isEqualTo("node-child");
|
||||||
assertThat(bodies).contains("child-input");
|
assertThat(processors.get(1).get("input_body")).isEqualTo("child-input");
|
||||||
assertThat(bodies).contains("child-output");
|
assertThat(processors.get(1).get("output_body")).isEqualTo("child-output");
|
||||||
|
|
||||||
var inputBodies = queryArray(
|
// Grandchild processor: depth=2, parent=child-proc
|
||||||
"SELECT processor_input_bodies FROM route_executions WHERE route_id = 'schema-test-tree'");
|
assertThat(processors.get(2).get("processor_id")).isEqualTo("grandchild-proc");
|
||||||
assertThat(inputBodies).containsExactly("root-input", "child-input", "");
|
assertThat(((Number) processors.get(2).get("depth")).intValue()).isEqualTo(2);
|
||||||
|
assertThat(processors.get(2).get("parent_processor_id")).isEqualTo("child-proc");
|
||||||
var outputBodies = queryArray(
|
assertThat(processors.get(2).get("diagram_node_id")).isEqualTo("node-grandchild");
|
||||||
"SELECT processor_output_bodies FROM route_executions WHERE route_id = 'schema-test-tree'");
|
|
||||||
assertThat(outputBodies).containsExactly("root-output", "child-output", "");
|
|
||||||
|
|
||||||
var inputHeaders = queryArray(
|
|
||||||
"SELECT processor_input_headers FROM route_executions WHERE route_id = 'schema-test-tree'");
|
|
||||||
assertThat(inputHeaders.get(0)).contains("Content-Type");
|
|
||||||
assertThat(inputHeaders.get(0)).contains("application/json");
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void exchangeBodiesContainsConcatenatedText() {
|
void exchangeBodiesStored() {
|
||||||
String json = """
|
String json = """
|
||||||
{
|
{
|
||||||
"routeId": "schema-test-bodies",
|
"routeId": "schema-test-bodies",
|
||||||
@@ -140,14 +140,6 @@ class IngestionSchemaIT extends AbstractClickHouseIT {
|
|||||||
"startTime": "2026-03-11T10:00:00Z",
|
"startTime": "2026-03-11T10:00:00Z",
|
||||||
"endTime": "2026-03-11T10:00:01Z",
|
"endTime": "2026-03-11T10:00:01Z",
|
||||||
"durationMs": 1000,
|
"durationMs": 1000,
|
||||||
"inputSnapshot": {
|
|
||||||
"body": "route-level-input-body",
|
|
||||||
"headers": {"X-Route": "header-value"}
|
|
||||||
},
|
|
||||||
"outputSnapshot": {
|
|
||||||
"body": "route-level-output-body",
|
|
||||||
"headers": {}
|
|
||||||
},
|
|
||||||
"processors": [
|
"processors": [
|
||||||
{
|
{
|
||||||
"processorId": "proc-1",
|
"processorId": "proc-1",
|
||||||
@@ -166,21 +158,13 @@ class IngestionSchemaIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
postExecution(json);
|
postExecution(json);
|
||||||
|
|
||||||
await().atMost(30, SECONDS).ignoreExceptions().untilAsserted(() -> {
|
// Verify processor body data
|
||||||
String bodies = jdbcTemplate.queryForObject(
|
List<Map<String, Object>> processors = jdbcTemplate.queryForList(
|
||||||
"SELECT exchange_bodies FROM route_executions WHERE route_id = 'schema-test-bodies'",
|
"SELECT input_body, output_body FROM processor_executions " +
|
||||||
String.class);
|
"WHERE execution_id = 'ex-bodies-1'");
|
||||||
assertThat(bodies).contains("processor-body-text");
|
assertThat(processors).hasSize(1);
|
||||||
assertThat(bodies).contains("processor-output-text");
|
assertThat(processors.get(0).get("input_body")).isEqualTo("processor-body-text");
|
||||||
assertThat(bodies).contains("route-level-input-body");
|
assertThat(processors.get(0).get("output_body")).isEqualTo("processor-output-text");
|
||||||
assertThat(bodies).contains("route-level-output-body");
|
|
||||||
|
|
||||||
String headers = jdbcTemplate.queryForObject(
|
|
||||||
"SELECT exchange_headers FROM route_executions WHERE route_id = 'schema-test-bodies'",
|
|
||||||
String.class);
|
|
||||||
assertThat(headers).contains("X-Route");
|
|
||||||
assertThat(headers).contains("header-value");
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
@@ -209,20 +193,19 @@ class IngestionSchemaIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
postExecution(json);
|
postExecution(json);
|
||||||
|
|
||||||
await().atMost(30, SECONDS).ignoreExceptions().untilAsserted(() -> {
|
// Verify execution exists
|
||||||
String bodies = jdbcTemplate.queryForObject(
|
Integer count = jdbcTemplate.queryForObject(
|
||||||
"SELECT exchange_bodies FROM route_executions WHERE route_id = 'schema-test-null-snap'",
|
"SELECT count(*) FROM executions WHERE execution_id = 'ex-null-1'",
|
||||||
String.class);
|
Integer.class);
|
||||||
assertThat(bodies).isNotNull();
|
assertThat(count).isEqualTo(1);
|
||||||
|
|
||||||
var depths = queryArray(
|
// Verify processor with null bodies inserted successfully
|
||||||
"SELECT processor_depths FROM route_executions WHERE route_id = 'schema-test-null-snap'");
|
List<Map<String, Object>> processors = jdbcTemplate.queryForList(
|
||||||
assertThat(depths).containsExactly("0");
|
"SELECT depth, parent_processor_id, input_body, output_body " +
|
||||||
|
"FROM processor_executions WHERE execution_id = 'ex-null-1'");
|
||||||
var parentIndexes = queryArray(
|
assertThat(processors).hasSize(1);
|
||||||
"SELECT processor_parent_indexes FROM route_executions WHERE route_id = 'schema-test-null-snap'");
|
assertThat(((Number) processors.get(0).get("depth")).intValue()).isEqualTo(0);
|
||||||
assertThat(parentIndexes).containsExactly("-1");
|
assertThat(processors.get(0).get("parent_processor_id")).isNull();
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private void postExecution(String json) {
|
private void postExecution(String json) {
|
||||||
@@ -233,22 +216,4 @@ class IngestionSchemaIT extends AbstractClickHouseIT {
|
|||||||
|
|
||||||
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
assertThat(response.getStatusCode()).isEqualTo(HttpStatus.ACCEPTED);
|
||||||
}
|
}
|
||||||
|
|
||||||
private List<String> queryArray(String sql) {
|
|
||||||
return jdbcTemplate.query(sql, (rs, rowNum) -> {
|
|
||||||
Object arr = rs.getArray(1).getArray();
|
|
||||||
if (arr instanceof Object[] objects) {
|
|
||||||
return Arrays.stream(objects).map(Object::toString).toList();
|
|
||||||
} else if (arr instanceof short[] shorts) {
|
|
||||||
var result = new java.util.ArrayList<String>();
|
|
||||||
for (short s : shorts) result.add(String.valueOf(s));
|
|
||||||
return result;
|
|
||||||
} else if (arr instanceof int[] ints) {
|
|
||||||
var result = new java.util.ArrayList<String>();
|
|
||||||
for (int v : ints) result.add(String.valueOf(v));
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
return List.<String>of();
|
|
||||||
}).get(0);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,83 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ExecutionRecord;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ProcessorRecord;
|
||||||
|
import org.junit.jupiter.api.Test;
|
||||||
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Optional;
|
||||||
|
|
||||||
|
import static org.junit.jupiter.api.Assertions.*;
|
||||||
|
|
||||||
|
class PostgresExecutionStoreIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
|
@Autowired
|
||||||
|
ExecutionStore executionStore;
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void upsertAndFindById() {
|
||||||
|
Instant now = Instant.now();
|
||||||
|
ExecutionRecord record = new ExecutionRecord(
|
||||||
|
"exec-1", "route-a", "agent-1", "app-1",
|
||||||
|
"COMPLETED", "corr-1", "exchange-1",
|
||||||
|
now, now.plusMillis(100), 100L,
|
||||||
|
null, null, null);
|
||||||
|
|
||||||
|
executionStore.upsert(record);
|
||||||
|
Optional<ExecutionRecord> found = executionStore.findById("exec-1");
|
||||||
|
|
||||||
|
assertTrue(found.isPresent());
|
||||||
|
assertEquals("exec-1", found.get().executionId());
|
||||||
|
assertEquals("COMPLETED", found.get().status());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void upsertDeduplicatesByExecutionId() {
|
||||||
|
Instant now = Instant.now();
|
||||||
|
ExecutionRecord first = new ExecutionRecord(
|
||||||
|
"exec-dup", "route-a", "agent-1", "app-1",
|
||||||
|
"RUNNING", null, null, now, null, null, null, null, null);
|
||||||
|
ExecutionRecord second = new ExecutionRecord(
|
||||||
|
"exec-dup", "route-a", "agent-1", "app-1",
|
||||||
|
"COMPLETED", null, null, now, now.plusMillis(200), 200L, null, null, null);
|
||||||
|
|
||||||
|
executionStore.upsert(first);
|
||||||
|
executionStore.upsert(second);
|
||||||
|
|
||||||
|
Optional<ExecutionRecord> found = executionStore.findById("exec-dup");
|
||||||
|
assertTrue(found.isPresent());
|
||||||
|
assertEquals("COMPLETED", found.get().status());
|
||||||
|
assertEquals(200L, found.get().durationMs());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void upsertProcessorsAndFind() {
|
||||||
|
Instant now = Instant.now();
|
||||||
|
ExecutionRecord exec = new ExecutionRecord(
|
||||||
|
"exec-proc", "route-a", "agent-1", "app-1",
|
||||||
|
"COMPLETED", null, null, now, now.plusMillis(50), 50L, null, null, null);
|
||||||
|
executionStore.upsert(exec);
|
||||||
|
|
||||||
|
List<ProcessorRecord> processors = List.of(
|
||||||
|
new ProcessorRecord("exec-proc", "proc-1", "log", null,
|
||||||
|
"app-1", "route-a", 0, null, "COMPLETED",
|
||||||
|
now, now.plusMillis(10), 10L, null, null,
|
||||||
|
"input body", "output body", null, null),
|
||||||
|
new ProcessorRecord("exec-proc", "proc-2", "to", null,
|
||||||
|
"app-1", "route-a", 1, "proc-1", "COMPLETED",
|
||||||
|
now.plusMillis(10), now.plusMillis(30), 20L, null, null,
|
||||||
|
null, null, null, null)
|
||||||
|
);
|
||||||
|
executionStore.upsertProcessors("exec-proc", now, "app-1", "route-a", processors);
|
||||||
|
|
||||||
|
List<ProcessorRecord> found = executionStore.findProcessors("exec-proc");
|
||||||
|
assertEquals(2, found.size());
|
||||||
|
assertEquals("proc-1", found.get(0).processorId());
|
||||||
|
assertEquals("proc-2", found.get(1).processorId());
|
||||||
|
assertEquals("proc-1", found.get(1).parentProcessorId());
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,64 @@
|
|||||||
|
package com.cameleer3.server.app.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.app.AbstractPostgresIT;
|
||||||
|
import com.cameleer3.server.core.search.ExecutionStats;
|
||||||
|
import com.cameleer3.server.core.search.StatsTimeseries;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ExecutionRecord;
|
||||||
|
import com.cameleer3.server.core.storage.StatsStore;
|
||||||
|
import org.junit.jupiter.api.Test;
|
||||||
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
|
import org.springframework.jdbc.core.JdbcTemplate;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.time.temporal.ChronoUnit;
|
||||||
|
|
||||||
|
import static org.junit.jupiter.api.Assertions.*;
|
||||||
|
|
||||||
|
class PostgresStatsStoreIT extends AbstractPostgresIT {
|
||||||
|
|
||||||
|
@Autowired StatsStore statsStore;
|
||||||
|
@Autowired ExecutionStore executionStore;
|
||||||
|
@Autowired JdbcTemplate jdbc;
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void statsReturnsCountsForTimeWindow() {
|
||||||
|
// Use a unique route + statsForRoute to avoid data contamination from other tests
|
||||||
|
String uniqueRoute = "stats-route-" + System.nanoTime();
|
||||||
|
Instant base = Instant.now().minus(5, ChronoUnit.MINUTES).truncatedTo(ChronoUnit.SECONDS);
|
||||||
|
insertExecution("stats-1-" + uniqueRoute, uniqueRoute, "app-stats", "COMPLETED", base, 100L);
|
||||||
|
insertExecution("stats-2-" + uniqueRoute, uniqueRoute, "app-stats", "FAILED", base.plusSeconds(10), 200L);
|
||||||
|
insertExecution("stats-3-" + uniqueRoute, uniqueRoute, "app-stats", "COMPLETED", base.plusSeconds(20), 50L);
|
||||||
|
|
||||||
|
// Force continuous aggregate refresh
|
||||||
|
jdbc.execute("CALL refresh_continuous_aggregate('stats_1m_route', NOW() - INTERVAL '1 hour', NOW() + INTERVAL '1 hour')");
|
||||||
|
|
||||||
|
ExecutionStats stats = statsStore.statsForRoute(base.minusSeconds(60), base.plusSeconds(60), uniqueRoute, null);
|
||||||
|
assertEquals(3, stats.totalCount());
|
||||||
|
assertEquals(1, stats.failedCount());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
void timeseriesReturnsBuckets() {
|
||||||
|
String uniqueRoute = "ts-route-" + System.nanoTime();
|
||||||
|
Instant base = Instant.now().minus(10, ChronoUnit.MINUTES).truncatedTo(ChronoUnit.MINUTES);
|
||||||
|
for (int i = 0; i < 10; i++) {
|
||||||
|
insertExecution("ts-" + i + "-" + uniqueRoute, uniqueRoute, "app-ts", "COMPLETED",
|
||||||
|
base.plusSeconds(i * 30), 100L + i);
|
||||||
|
}
|
||||||
|
|
||||||
|
jdbc.execute("CALL refresh_continuous_aggregate('stats_1m_route', NOW() - INTERVAL '1 hour', NOW() + INTERVAL '1 hour')");
|
||||||
|
|
||||||
|
StatsTimeseries ts = statsStore.timeseriesForRoute(base.minus(1, ChronoUnit.MINUTES), base.plus(10, ChronoUnit.MINUTES), 5, uniqueRoute, null);
|
||||||
|
assertNotNull(ts);
|
||||||
|
assertFalse(ts.buckets().isEmpty());
|
||||||
|
}
|
||||||
|
|
||||||
|
private void insertExecution(String id, String routeId, String groupName,
|
||||||
|
String status, Instant startTime, long durationMs) {
|
||||||
|
executionStore.upsert(new ExecutionRecord(
|
||||||
|
id, routeId, "agent-1", groupName, status, null, null,
|
||||||
|
startTime, startTime.plusMillis(durationMs), durationMs,
|
||||||
|
status.equals("FAILED") ? "error" : null, null, null));
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,9 +1,10 @@
|
|||||||
spring:
|
spring:
|
||||||
datasource:
|
flyway:
|
||||||
url: jdbc:ch://placeholder:8123/cameleer3
|
enabled: true
|
||||||
username: default
|
|
||||||
password: ""
|
opensearch:
|
||||||
driver-class-name: com.clickhouse.jdbc.ClickHouseDriver
|
url: http://localhost:9200
|
||||||
|
debounce-ms: 100
|
||||||
|
|
||||||
ingestion:
|
ingestion:
|
||||||
buffer-capacity: 100
|
buffer-capacity: 100
|
||||||
|
|||||||
@@ -1,104 +1,61 @@
|
|||||||
package com.cameleer3.server.core.detail;
|
package com.cameleer3.server.core.detail;
|
||||||
|
|
||||||
import com.cameleer3.server.core.storage.ExecutionRepository;
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ProcessorRecord;
|
||||||
|
|
||||||
import java.util.ArrayList;
|
import java.util.*;
|
||||||
import java.util.List;
|
|
||||||
import java.util.Optional;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Provides execution detail with reconstructed processor tree.
|
|
||||||
* <p>
|
|
||||||
* This is a plain class (no Spring annotations) -- it lives in the core module
|
|
||||||
* and is wired as a bean by the app module configuration.
|
|
||||||
*/
|
|
||||||
public class DetailService {
|
public class DetailService {
|
||||||
|
|
||||||
private final ExecutionRepository repository;
|
private final ExecutionStore executionStore;
|
||||||
|
|
||||||
public DetailService(ExecutionRepository repository) {
|
public DetailService(ExecutionStore executionStore) {
|
||||||
this.repository = repository;
|
this.executionStore = executionStore;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Get the full detail of a route execution, including the nested processor tree.
|
|
||||||
*
|
|
||||||
* @param executionId the execution ID to look up
|
|
||||||
* @return the execution detail, or empty if not found
|
|
||||||
*/
|
|
||||||
public Optional<ExecutionDetail> getDetail(String executionId) {
|
public Optional<ExecutionDetail> getDetail(String executionId) {
|
||||||
return repository.findRawById(executionId)
|
return executionStore.findById(executionId)
|
||||||
.map(this::toDetail);
|
.map(exec -> {
|
||||||
|
List<ProcessorRecord> processors = executionStore.findProcessors(executionId);
|
||||||
|
List<ProcessorNode> roots = buildTree(processors);
|
||||||
|
return new ExecutionDetail(
|
||||||
|
exec.executionId(), exec.routeId(), exec.agentId(),
|
||||||
|
exec.status(), exec.startTime(), exec.endTime(),
|
||||||
|
exec.durationMs() != null ? exec.durationMs() : 0L,
|
||||||
|
exec.correlationId(), exec.exchangeId(),
|
||||||
|
exec.errorMessage(), exec.errorStacktrace(),
|
||||||
|
exec.diagramContentHash(), roots
|
||||||
|
);
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
private ExecutionDetail toDetail(RawExecutionRow row) {
|
List<ProcessorNode> buildTree(List<ProcessorRecord> processors) {
|
||||||
List<ProcessorNode> roots = reconstructTree(
|
if (processors.isEmpty()) return List.of();
|
||||||
row.processorIds(),
|
|
||||||
row.processorTypes(),
|
|
||||||
row.processorStatuses(),
|
|
||||||
row.processorStarts(),
|
|
||||||
row.processorEnds(),
|
|
||||||
row.processorDurations(),
|
|
||||||
row.processorDiagramNodeIds(),
|
|
||||||
row.processorErrorMessages(),
|
|
||||||
row.processorErrorStacktraces(),
|
|
||||||
row.processorDepths(),
|
|
||||||
row.processorParentIndexes()
|
|
||||||
);
|
|
||||||
|
|
||||||
return new ExecutionDetail(
|
Map<String, ProcessorNode> nodeMap = new LinkedHashMap<>();
|
||||||
row.executionId(),
|
for (ProcessorRecord p : processors) {
|
||||||
row.routeId(),
|
nodeMap.put(p.processorId(), new ProcessorNode(
|
||||||
row.agentId(),
|
p.processorId(), p.processorType(), p.status(),
|
||||||
row.status(),
|
p.startTime(), p.endTime(),
|
||||||
row.startTime(),
|
p.durationMs() != null ? p.durationMs() : 0L,
|
||||||
row.endTime(),
|
p.diagramNodeId(), p.errorMessage(), p.errorStacktrace()
|
||||||
row.durationMs(),
|
));
|
||||||
row.correlationId(),
|
|
||||||
row.exchangeId(),
|
|
||||||
row.errorMessage(),
|
|
||||||
row.errorStackTrace(),
|
|
||||||
row.diagramContentHash(),
|
|
||||||
roots
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Reconstruct the nested processor tree from flat parallel arrays.
|
|
||||||
* <p>
|
|
||||||
* Uses parentIndexes to wire children: parentIndex == -1 means the node is a root.
|
|
||||||
* Otherwise, parentIndex is the array index of the parent node.
|
|
||||||
*/
|
|
||||||
List<ProcessorNode> reconstructTree(
|
|
||||||
String[] ids, String[] types, String[] statuses,
|
|
||||||
java.time.Instant[] starts, java.time.Instant[] ends, long[] durations,
|
|
||||||
String[] diagramNodeIds, String[] errorMessages, String[] errorStacktraces,
|
|
||||||
int[] depths, int[] parentIndexes) {
|
|
||||||
|
|
||||||
if (ids == null || ids.length == 0) {
|
|
||||||
return List.of();
|
|
||||||
}
|
|
||||||
|
|
||||||
int len = ids.length;
|
|
||||||
ProcessorNode[] nodes = new ProcessorNode[len];
|
|
||||||
|
|
||||||
for (int i = 0; i < len; i++) {
|
|
||||||
nodes[i] = new ProcessorNode(
|
|
||||||
ids[i], types[i], statuses[i],
|
|
||||||
starts[i], ends[i], durations[i],
|
|
||||||
diagramNodeIds[i], errorMessages[i], errorStacktraces[i]
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
List<ProcessorNode> roots = new ArrayList<>();
|
List<ProcessorNode> roots = new ArrayList<>();
|
||||||
for (int i = 0; i < len; i++) {
|
for (ProcessorRecord p : processors) {
|
||||||
if (parentIndexes[i] == -1) {
|
ProcessorNode node = nodeMap.get(p.processorId());
|
||||||
roots.add(nodes[i]);
|
if (p.parentProcessorId() == null) {
|
||||||
|
roots.add(node);
|
||||||
} else {
|
} else {
|
||||||
nodes[parentIndexes[i]].addChild(nodes[i]);
|
ProcessorNode parent = nodeMap.get(p.parentProcessorId());
|
||||||
|
if (parent != null) {
|
||||||
|
parent.addChild(node);
|
||||||
|
} else {
|
||||||
|
roots.add(node); // orphan safety
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return roots;
|
return roots;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import java.util.List;
|
|||||||
* Full detail of a route execution, including the nested processor tree.
|
* Full detail of a route execution, including the nested processor tree.
|
||||||
* <p>
|
* <p>
|
||||||
* This is the rich detail model returned by the detail endpoint. The processor
|
* This is the rich detail model returned by the detail endpoint. The processor
|
||||||
* tree is reconstructed from flat parallel arrays stored in ClickHouse.
|
* tree is reconstructed from individual processor records stored in PostgreSQL.
|
||||||
*
|
*
|
||||||
* @param executionId unique execution identifier
|
* @param executionId unique execution identifier
|
||||||
* @param routeId Camel route ID
|
* @param routeId Camel route ID
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import java.util.List;
|
|||||||
/**
|
/**
|
||||||
* Nested tree node representing a single processor execution within a route.
|
* Nested tree node representing a single processor execution within a route.
|
||||||
* <p>
|
* <p>
|
||||||
* The tree structure is reconstructed from flat parallel arrays stored in ClickHouse.
|
* The tree structure is reconstructed from individual processor records stored in PostgreSQL.
|
||||||
* Each node may have children (e.g., processors inside a split or try-catch block).
|
* Each node may have children (e.g., processors inside a split or try-catch block).
|
||||||
*/
|
*/
|
||||||
public final class ProcessorNode {
|
public final class ProcessorNode {
|
||||||
|
|||||||
@@ -1,59 +0,0 @@
|
|||||||
package com.cameleer3.server.core.detail;
|
|
||||||
|
|
||||||
import java.time.Instant;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Raw execution data from ClickHouse, including all parallel arrays needed
|
|
||||||
* for tree reconstruction. This is the intermediate representation between
|
|
||||||
* the database and the {@link ExecutionDetail} domain object.
|
|
||||||
*
|
|
||||||
* @param executionId unique execution identifier
|
|
||||||
* @param routeId Camel route ID
|
|
||||||
* @param agentId agent instance
|
|
||||||
* @param status execution status
|
|
||||||
* @param startTime execution start time
|
|
||||||
* @param endTime execution end time
|
|
||||||
* @param durationMs execution duration in milliseconds
|
|
||||||
* @param correlationId correlation ID
|
|
||||||
* @param exchangeId Camel exchange ID
|
|
||||||
* @param errorMessage execution-level error message
|
|
||||||
* @param errorStackTrace execution-level error stack trace
|
|
||||||
* @param diagramContentHash content hash for diagram linking
|
|
||||||
* @param processorIds processor IDs (parallel array)
|
|
||||||
* @param processorTypes processor types (parallel array)
|
|
||||||
* @param processorStatuses processor statuses (parallel array)
|
|
||||||
* @param processorStarts processor start times (parallel array)
|
|
||||||
* @param processorEnds processor end times (parallel array)
|
|
||||||
* @param processorDurations processor durations in ms (parallel array)
|
|
||||||
* @param processorDiagramNodeIds processor diagram node IDs (parallel array)
|
|
||||||
* @param processorErrorMessages processor error messages (parallel array)
|
|
||||||
* @param processorErrorStacktraces processor error stack traces (parallel array)
|
|
||||||
* @param processorDepths processor tree depths (parallel array)
|
|
||||||
* @param processorParentIndexes processor parent indexes, -1 for roots (parallel array)
|
|
||||||
*/
|
|
||||||
public record RawExecutionRow(
|
|
||||||
String executionId,
|
|
||||||
String routeId,
|
|
||||||
String agentId,
|
|
||||||
String status,
|
|
||||||
Instant startTime,
|
|
||||||
Instant endTime,
|
|
||||||
long durationMs,
|
|
||||||
String correlationId,
|
|
||||||
String exchangeId,
|
|
||||||
String errorMessage,
|
|
||||||
String errorStackTrace,
|
|
||||||
String diagramContentHash,
|
|
||||||
String[] processorIds,
|
|
||||||
String[] processorTypes,
|
|
||||||
String[] processorStatuses,
|
|
||||||
Instant[] processorStarts,
|
|
||||||
Instant[] processorEnds,
|
|
||||||
long[] processorDurations,
|
|
||||||
String[] processorDiagramNodeIds,
|
|
||||||
String[] processorErrorMessages,
|
|
||||||
String[] processorErrorStacktraces,
|
|
||||||
int[] processorDepths,
|
|
||||||
int[] processorParentIndexes
|
|
||||||
) {
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,5 @@
|
|||||||
|
package com.cameleer3.server.core.indexing;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
|
|
||||||
|
public record ExecutionUpdatedEvent(String executionId, Instant startTime) {}
|
||||||
@@ -0,0 +1,79 @@
|
|||||||
|
package com.cameleer3.server.core.indexing;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ExecutionRecord;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ProcessorRecord;
|
||||||
|
import com.cameleer3.server.core.storage.SearchIndex;
|
||||||
|
import com.cameleer3.server.core.storage.model.ExecutionDocument;
|
||||||
|
import com.cameleer3.server.core.storage.model.ExecutionDocument.ProcessorDoc;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.concurrent.*;
|
||||||
|
|
||||||
|
public class SearchIndexer {
|
||||||
|
|
||||||
|
private static final Logger log = LoggerFactory.getLogger(SearchIndexer.class);
|
||||||
|
|
||||||
|
private final ExecutionStore executionStore;
|
||||||
|
private final SearchIndex searchIndex;
|
||||||
|
private final long debounceMs;
|
||||||
|
private final int queueCapacity;
|
||||||
|
|
||||||
|
private final Map<String, ScheduledFuture<?>> pending = new ConcurrentHashMap<>();
|
||||||
|
private final ScheduledExecutorService scheduler = Executors.newSingleThreadScheduledExecutor(
|
||||||
|
r -> { Thread t = new Thread(r, "search-indexer"); t.setDaemon(true); return t; });
|
||||||
|
|
||||||
|
public SearchIndexer(ExecutionStore executionStore, SearchIndex searchIndex,
|
||||||
|
long debounceMs, int queueCapacity) {
|
||||||
|
this.executionStore = executionStore;
|
||||||
|
this.searchIndex = searchIndex;
|
||||||
|
this.debounceMs = debounceMs;
|
||||||
|
this.queueCapacity = queueCapacity;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void onExecutionUpdated(ExecutionUpdatedEvent event) {
|
||||||
|
if (pending.size() >= queueCapacity) {
|
||||||
|
log.warn("Search indexer queue full, dropping event for {}", event.executionId());
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
ScheduledFuture<?> existing = pending.put(event.executionId(),
|
||||||
|
scheduler.schedule(() -> indexExecution(event.executionId()),
|
||||||
|
debounceMs, TimeUnit.MILLISECONDS));
|
||||||
|
if (existing != null) {
|
||||||
|
existing.cancel(false);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void indexExecution(String executionId) {
|
||||||
|
pending.remove(executionId);
|
||||||
|
try {
|
||||||
|
ExecutionRecord exec = executionStore.findById(executionId).orElse(null);
|
||||||
|
if (exec == null) return;
|
||||||
|
|
||||||
|
List<ProcessorRecord> processors = executionStore.findProcessors(executionId);
|
||||||
|
List<ProcessorDoc> processorDocs = processors.stream()
|
||||||
|
.map(p -> new ProcessorDoc(
|
||||||
|
p.processorId(), p.processorType(), p.status(),
|
||||||
|
p.errorMessage(), p.errorStacktrace(),
|
||||||
|
p.inputBody(), p.outputBody(),
|
||||||
|
p.inputHeaders(), p.outputHeaders()))
|
||||||
|
.toList();
|
||||||
|
|
||||||
|
searchIndex.index(new ExecutionDocument(
|
||||||
|
exec.executionId(), exec.routeId(), exec.agentId(), exec.groupName(),
|
||||||
|
exec.status(), exec.correlationId(), exec.exchangeId(),
|
||||||
|
exec.startTime(), exec.endTime(), exec.durationMs(),
|
||||||
|
exec.errorMessage(), exec.errorStacktrace(), processorDocs));
|
||||||
|
} catch (Exception e) {
|
||||||
|
log.error("Failed to index execution {}", executionId, e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public void shutdown() {
|
||||||
|
scheduler.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,113 +1,132 @@
|
|||||||
package com.cameleer3.server.core.ingestion;
|
package com.cameleer3.server.core.ingestion;
|
||||||
|
|
||||||
|
import com.cameleer3.common.model.ProcessorExecution;
|
||||||
|
import com.cameleer3.common.model.RouteExecution;
|
||||||
|
import com.cameleer3.server.core.indexing.ExecutionUpdatedEvent;
|
||||||
|
import com.cameleer3.server.core.storage.DiagramStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ExecutionRecord;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ProcessorRecord;
|
||||||
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
||||||
|
import com.fasterxml.jackson.core.JsonProcessingException;
|
||||||
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
|
|
||||||
|
import java.util.ArrayList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.function.Consumer;
|
||||||
|
|
||||||
/**
|
|
||||||
* Routes incoming data to the appropriate {@link WriteBuffer} instances.
|
|
||||||
* <p>
|
|
||||||
* This is a plain class (no Spring annotations) -- it lives in the core module
|
|
||||||
* and is wired as a bean by the app module configuration.
|
|
||||||
*/
|
|
||||||
public class IngestionService {
|
public class IngestionService {
|
||||||
|
|
||||||
private final WriteBuffer<TaggedExecution> executionBuffer;
|
private static final ObjectMapper JSON = new ObjectMapper();
|
||||||
private final WriteBuffer<TaggedDiagram> diagramBuffer;
|
|
||||||
|
private final ExecutionStore executionStore;
|
||||||
|
private final DiagramStore diagramStore;
|
||||||
private final WriteBuffer<MetricsSnapshot> metricsBuffer;
|
private final WriteBuffer<MetricsSnapshot> metricsBuffer;
|
||||||
|
private final Consumer<ExecutionUpdatedEvent> eventPublisher;
|
||||||
|
private final int bodySizeLimit;
|
||||||
|
|
||||||
public IngestionService(WriteBuffer<TaggedExecution> executionBuffer,
|
public IngestionService(ExecutionStore executionStore,
|
||||||
WriteBuffer<TaggedDiagram> diagramBuffer,
|
DiagramStore diagramStore,
|
||||||
WriteBuffer<MetricsSnapshot> metricsBuffer) {
|
WriteBuffer<MetricsSnapshot> metricsBuffer,
|
||||||
this.executionBuffer = executionBuffer;
|
Consumer<ExecutionUpdatedEvent> eventPublisher,
|
||||||
this.diagramBuffer = diagramBuffer;
|
int bodySizeLimit) {
|
||||||
|
this.executionStore = executionStore;
|
||||||
|
this.diagramStore = diagramStore;
|
||||||
this.metricsBuffer = metricsBuffer;
|
this.metricsBuffer = metricsBuffer;
|
||||||
|
this.eventPublisher = eventPublisher;
|
||||||
|
this.bodySizeLimit = bodySizeLimit;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
public void ingestExecution(String agentId, String groupName, RouteExecution execution) {
|
||||||
* Accept a batch of tagged route executions into the buffer.
|
ExecutionRecord record = toExecutionRecord(agentId, groupName, execution);
|
||||||
*
|
executionStore.upsert(record);
|
||||||
* @return true if all items were buffered, false if buffer is full (backpressure)
|
|
||||||
*/
|
if (execution.getProcessors() != null && !execution.getProcessors().isEmpty()) {
|
||||||
public boolean acceptExecutions(List<TaggedExecution> executions) {
|
List<ProcessorRecord> processors = flattenProcessors(
|
||||||
return executionBuffer.offerBatch(executions);
|
execution.getProcessors(), record.executionId(),
|
||||||
|
record.startTime(), groupName, execution.getRouteId(),
|
||||||
|
null, 0);
|
||||||
|
executionStore.upsertProcessors(
|
||||||
|
record.executionId(), record.startTime(),
|
||||||
|
groupName, execution.getRouteId(), processors);
|
||||||
|
}
|
||||||
|
|
||||||
|
eventPublisher.accept(new ExecutionUpdatedEvent(
|
||||||
|
record.executionId(), record.startTime()));
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
public void ingestDiagram(TaggedDiagram diagram) {
|
||||||
* Accept a single tagged route execution into the buffer.
|
diagramStore.store(diagram);
|
||||||
*
|
|
||||||
* @return true if the item was buffered, false if buffer is full (backpressure)
|
|
||||||
*/
|
|
||||||
public boolean acceptExecution(TaggedExecution execution) {
|
|
||||||
return executionBuffer.offer(execution);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Accept a single tagged route diagram into the buffer.
|
|
||||||
*
|
|
||||||
* @return true if the item was buffered, false if buffer is full (backpressure)
|
|
||||||
*/
|
|
||||||
public boolean acceptDiagram(TaggedDiagram diagram) {
|
|
||||||
return diagramBuffer.offer(diagram);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Accept a batch of tagged route diagrams into the buffer.
|
|
||||||
*
|
|
||||||
* @return true if all items were buffered, false if buffer is full (backpressure)
|
|
||||||
*/
|
|
||||||
public boolean acceptDiagrams(List<TaggedDiagram> diagrams) {
|
|
||||||
return diagramBuffer.offerBatch(diagrams);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Accept a batch of metrics snapshots into the buffer.
|
|
||||||
*
|
|
||||||
* @return true if all items were buffered, false if buffer is full (backpressure)
|
|
||||||
*/
|
|
||||||
public boolean acceptMetrics(List<MetricsSnapshot> metrics) {
|
public boolean acceptMetrics(List<MetricsSnapshot> metrics) {
|
||||||
return metricsBuffer.offerBatch(metrics);
|
return metricsBuffer.offerBatch(metrics);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* @return current number of items in the execution buffer
|
|
||||||
*/
|
|
||||||
public int getExecutionBufferDepth() {
|
|
||||||
return executionBuffer.size();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @return current number of items in the diagram buffer
|
|
||||||
*/
|
|
||||||
public int getDiagramBufferDepth() {
|
|
||||||
return diagramBuffer.size();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @return current number of items in the metrics buffer
|
|
||||||
*/
|
|
||||||
public int getMetricsBufferDepth() {
|
public int getMetricsBufferDepth() {
|
||||||
return metricsBuffer.size();
|
return metricsBuffer.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* @return the execution write buffer (for use by flush scheduler)
|
|
||||||
*/
|
|
||||||
public WriteBuffer<TaggedExecution> getExecutionBuffer() {
|
|
||||||
return executionBuffer;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @return the diagram write buffer (for use by flush scheduler)
|
|
||||||
*/
|
|
||||||
public WriteBuffer<TaggedDiagram> getDiagramBuffer() {
|
|
||||||
return diagramBuffer;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @return the metrics write buffer (for use by flush scheduler)
|
|
||||||
*/
|
|
||||||
public WriteBuffer<MetricsSnapshot> getMetricsBuffer() {
|
public WriteBuffer<MetricsSnapshot> getMetricsBuffer() {
|
||||||
return metricsBuffer;
|
return metricsBuffer;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private ExecutionRecord toExecutionRecord(String agentId, String groupName,
|
||||||
|
RouteExecution exec) {
|
||||||
|
String diagramHash = diagramStore
|
||||||
|
.findContentHashForRoute(exec.getRouteId(), agentId)
|
||||||
|
.orElse("");
|
||||||
|
return new ExecutionRecord(
|
||||||
|
exec.getExchangeId(), exec.getRouteId(), agentId, groupName,
|
||||||
|
exec.getStatus() != null ? exec.getStatus().name() : "RUNNING",
|
||||||
|
exec.getCorrelationId(), exec.getExchangeId(),
|
||||||
|
exec.getStartTime(), exec.getEndTime(),
|
||||||
|
exec.getDurationMs(),
|
||||||
|
exec.getErrorMessage(), exec.getErrorStackTrace(),
|
||||||
|
diagramHash
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
private List<ProcessorRecord> flattenProcessors(
|
||||||
|
List<ProcessorExecution> processors, String executionId,
|
||||||
|
java.time.Instant execStartTime, String groupName, String routeId,
|
||||||
|
String parentProcessorId, int depth) {
|
||||||
|
List<ProcessorRecord> flat = new ArrayList<>();
|
||||||
|
for (ProcessorExecution p : processors) {
|
||||||
|
flat.add(new ProcessorRecord(
|
||||||
|
executionId, p.getProcessorId(), p.getProcessorType(),
|
||||||
|
p.getDiagramNodeId(), groupName, routeId,
|
||||||
|
depth, parentProcessorId,
|
||||||
|
p.getStatus() != null ? p.getStatus().name() : "RUNNING",
|
||||||
|
p.getStartTime() != null ? p.getStartTime() : execStartTime,
|
||||||
|
p.getEndTime(),
|
||||||
|
p.getDurationMs(),
|
||||||
|
p.getErrorMessage(), p.getErrorStackTrace(),
|
||||||
|
truncateBody(p.getInputBody()), truncateBody(p.getOutputBody()),
|
||||||
|
toJson(p.getInputHeaders()), toJson(p.getOutputHeaders())
|
||||||
|
));
|
||||||
|
if (p.getChildren() != null) {
|
||||||
|
flat.addAll(flattenProcessors(
|
||||||
|
p.getChildren(), executionId, execStartTime,
|
||||||
|
groupName, routeId, p.getProcessorId(), depth + 1));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return flat;
|
||||||
|
}
|
||||||
|
|
||||||
|
private String truncateBody(String body) {
|
||||||
|
if (body == null) return null;
|
||||||
|
if (body.length() > bodySizeLimit) return body.substring(0, bodySizeLimit);
|
||||||
|
return body;
|
||||||
|
}
|
||||||
|
|
||||||
|
private static String toJson(Map<String, String> headers) {
|
||||||
|
if (headers == null) return null;
|
||||||
|
try {
|
||||||
|
return JSON.writeValueAsString(headers);
|
||||||
|
} catch (JsonProcessingException e) {
|
||||||
|
return "{}";
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import java.util.concurrent.ArrayBlockingQueue;
|
|||||||
import java.util.concurrent.BlockingQueue;
|
import java.util.concurrent.BlockingQueue;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Bounded write buffer that decouples HTTP ingestion from ClickHouse batch inserts.
|
* Bounded write buffer that decouples HTTP ingestion from database batch inserts.
|
||||||
* <p>
|
* <p>
|
||||||
* Items are offered to the buffer by controllers and drained in batches by a
|
* Items are offered to the buffer by controllers and drained in batches by a
|
||||||
* scheduled flush task. When the buffer is full, {@link #offer} returns false,
|
* scheduled flush task. When the buffer is full, {@link #offer} returns false,
|
||||||
|
|||||||
@@ -1,72 +0,0 @@
|
|||||||
package com.cameleer3.server.core.search;
|
|
||||||
|
|
||||||
import java.util.List;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Swappable search backend abstraction.
|
|
||||||
* <p>
|
|
||||||
* The current implementation uses ClickHouse for search. This interface allows
|
|
||||||
* replacing the search backend (e.g., with OpenSearch) without changing the
|
|
||||||
* service layer or controllers.
|
|
||||||
*/
|
|
||||||
public interface SearchEngine {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Search for route executions matching the given criteria.
|
|
||||||
*
|
|
||||||
* @param request search filters and pagination
|
|
||||||
* @return paginated search results with total count
|
|
||||||
*/
|
|
||||||
SearchResult<ExecutionSummary> search(SearchRequest request);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Count route executions matching the given criteria (without fetching data).
|
|
||||||
*
|
|
||||||
* @param request search filters
|
|
||||||
* @return total number of matching executions
|
|
||||||
*/
|
|
||||||
long count(SearchRequest request);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Compute aggregate stats: P99 latency and count of currently running executions.
|
|
||||||
*
|
|
||||||
* @param from start of the time window
|
|
||||||
* @param to end of the time window
|
|
||||||
* @return execution stats
|
|
||||||
*/
|
|
||||||
ExecutionStats stats(java.time.Instant from, java.time.Instant to);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Compute aggregate stats scoped to specific routes and agents.
|
|
||||||
*
|
|
||||||
* @param from start of the time window
|
|
||||||
* @param to end of the time window
|
|
||||||
* @param routeId optional route ID filter
|
|
||||||
* @param agentIds optional agent ID filter (from group resolution)
|
|
||||||
* @return execution stats
|
|
||||||
*/
|
|
||||||
ExecutionStats stats(java.time.Instant from, java.time.Instant to, String routeId, List<String> agentIds);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Compute bucketed time-series stats over a time window.
|
|
||||||
*
|
|
||||||
* @param from start of the time window
|
|
||||||
* @param to end of the time window
|
|
||||||
* @param bucketCount number of buckets to divide the window into
|
|
||||||
* @return bucketed stats
|
|
||||||
*/
|
|
||||||
StatsTimeseries timeseries(java.time.Instant from, java.time.Instant to, int bucketCount);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Compute bucketed time-series stats scoped to specific routes and agents.
|
|
||||||
*
|
|
||||||
* @param from start of the time window
|
|
||||||
* @param to end of the time window
|
|
||||||
* @param bucketCount number of buckets to divide the window into
|
|
||||||
* @param routeId optional route ID filter
|
|
||||||
* @param agentIds optional agent ID filter (from group resolution)
|
|
||||||
* @return bucketed stats
|
|
||||||
*/
|
|
||||||
StatsTimeseries timeseries(java.time.Instant from, java.time.Instant to, int bucketCount,
|
|
||||||
String routeId, List<String> agentIds);
|
|
||||||
}
|
|
||||||
@@ -75,7 +75,7 @@ public record SearchRequest(
|
|||||||
if (!"asc".equalsIgnoreCase(sortDir)) sortDir = "desc";
|
if (!"asc".equalsIgnoreCase(sortDir)) sortDir = "desc";
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Returns the validated ClickHouse column name for ORDER BY. */
|
/** Returns the validated database column name for ORDER BY. */
|
||||||
public String sortColumn() {
|
public String sortColumn() {
|
||||||
return SORT_FIELD_TO_COLUMN.getOrDefault(sortField, "start_time");
|
return SORT_FIELD_TO_COLUMN.getOrDefault(sortField, "start_time");
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,63 +1,43 @@
|
|||||||
package com.cameleer3.server.core.search;
|
package com.cameleer3.server.core.search;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.storage.SearchIndex;
|
||||||
|
import com.cameleer3.server.core.storage.StatsStore;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
/**
|
|
||||||
* Orchestrates search operations, delegating to a {@link SearchEngine} backend.
|
|
||||||
* <p>
|
|
||||||
* This is a plain class (no Spring annotations) -- it lives in the core module
|
|
||||||
* and is wired as a bean by the app module configuration. The thin orchestration
|
|
||||||
* layer allows adding cross-cutting concerns (logging, caching, metrics) later.
|
|
||||||
*/
|
|
||||||
public class SearchService {
|
public class SearchService {
|
||||||
|
|
||||||
private final SearchEngine engine;
|
private final SearchIndex searchIndex;
|
||||||
|
private final StatsStore statsStore;
|
||||||
|
|
||||||
public SearchService(SearchEngine engine) {
|
public SearchService(SearchIndex searchIndex, StatsStore statsStore) {
|
||||||
this.engine = engine;
|
this.searchIndex = searchIndex;
|
||||||
|
this.statsStore = statsStore;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Search for route executions matching the given criteria.
|
|
||||||
*/
|
|
||||||
public SearchResult<ExecutionSummary> search(SearchRequest request) {
|
public SearchResult<ExecutionSummary> search(SearchRequest request) {
|
||||||
return engine.search(request);
|
return searchIndex.search(request);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Count route executions matching the given criteria.
|
|
||||||
*/
|
|
||||||
public long count(SearchRequest request) {
|
public long count(SearchRequest request) {
|
||||||
return engine.count(request);
|
return searchIndex.count(request);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
public ExecutionStats stats(Instant from, Instant to) {
|
||||||
* Compute aggregate execution stats (P99 latency, active count).
|
return statsStore.stats(from, to);
|
||||||
*/
|
|
||||||
public ExecutionStats stats(java.time.Instant from, java.time.Instant to) {
|
|
||||||
return engine.stats(from, to);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
public ExecutionStats stats(Instant from, Instant to, String routeId, List<String> agentIds) {
|
||||||
* Compute aggregate execution stats scoped to specific routes and agents.
|
return statsStore.statsForRoute(from, to, routeId, agentIds);
|
||||||
*/
|
|
||||||
public ExecutionStats stats(java.time.Instant from, java.time.Instant to,
|
|
||||||
String routeId, List<String> agentIds) {
|
|
||||||
return engine.stats(from, to, routeId, agentIds);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
public StatsTimeseries timeseries(Instant from, Instant to, int bucketCount) {
|
||||||
* Compute bucketed time-series stats over a time window.
|
return statsStore.timeseries(from, to, bucketCount);
|
||||||
*/
|
|
||||||
public StatsTimeseries timeseries(java.time.Instant from, java.time.Instant to, int bucketCount) {
|
|
||||||
return engine.timeseries(from, to, bucketCount);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
public StatsTimeseries timeseries(Instant from, Instant to, int bucketCount,
|
||||||
* Compute bucketed time-series stats scoped to specific routes and agents.
|
|
||||||
*/
|
|
||||||
public StatsTimeseries timeseries(java.time.Instant from, java.time.Instant to, int bucketCount,
|
|
||||||
String routeId, List<String> agentIds) {
|
String routeId, List<String> agentIds) {
|
||||||
return engine.timeseries(from, to, bucketCount, routeId, agentIds);
|
return statsStore.timeseriesForRoute(from, to, bucketCount, routeId, agentIds);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,35 +0,0 @@
|
|||||||
package com.cameleer3.server.core.storage;
|
|
||||||
|
|
||||||
import com.cameleer3.common.graph.RouteGraph;
|
|
||||||
import com.cameleer3.server.core.ingestion.TaggedDiagram;
|
|
||||||
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Optional;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Repository for route diagram storage with content-hash deduplication.
|
|
||||||
*/
|
|
||||||
public interface DiagramRepository {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Store a tagged route graph. Uses content-hash deduplication via ReplacingMergeTree.
|
|
||||||
*/
|
|
||||||
void store(TaggedDiagram diagram);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Find a route graph by its content hash.
|
|
||||||
*/
|
|
||||||
Optional<RouteGraph> findByContentHash(String contentHash);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Find the content hash for the latest diagram of a given route and agent.
|
|
||||||
*/
|
|
||||||
Optional<String> findContentHashForRoute(String routeId, String agentId);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Find the content hash for the latest diagram of a route across any agent in the given list.
|
|
||||||
* All instances of the same application produce the same route graph, so any agent's
|
|
||||||
* diagram for the same route will have the same content hash.
|
|
||||||
*/
|
|
||||||
Optional<String> findContentHashForRouteByAgents(String routeId, List<String> agentIds);
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,18 @@
|
|||||||
|
package com.cameleer3.server.core.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.common.graph.RouteGraph;
|
||||||
|
import com.cameleer3.server.core.ingestion.TaggedDiagram;
|
||||||
|
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Optional;
|
||||||
|
|
||||||
|
public interface DiagramStore {
|
||||||
|
|
||||||
|
void store(TaggedDiagram diagram);
|
||||||
|
|
||||||
|
Optional<RouteGraph> findByContentHash(String contentHash);
|
||||||
|
|
||||||
|
Optional<String> findContentHashForRoute(String routeId, String agentId);
|
||||||
|
|
||||||
|
Optional<String> findContentHashForRouteByAgents(String routeId, List<String> agentIds);
|
||||||
|
}
|
||||||
@@ -1,28 +0,0 @@
|
|||||||
package com.cameleer3.server.core.storage;
|
|
||||||
|
|
||||||
import com.cameleer3.server.core.detail.RawExecutionRow;
|
|
||||||
import com.cameleer3.server.core.ingestion.TaggedExecution;
|
|
||||||
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Optional;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Repository for route execution storage and retrieval.
|
|
||||||
*/
|
|
||||||
public interface ExecutionRepository {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Insert a batch of tagged route executions.
|
|
||||||
* Implementations must perform a single batch insert for efficiency.
|
|
||||||
*/
|
|
||||||
void insertBatch(List<TaggedExecution> executions);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Find a raw execution row by execution ID, including all parallel arrays
|
|
||||||
* needed for processor tree reconstruction.
|
|
||||||
*
|
|
||||||
* @param executionId the execution ID to look up
|
|
||||||
* @return the raw execution row, or empty if not found
|
|
||||||
*/
|
|
||||||
Optional<RawExecutionRow> findRawById(String executionId);
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,34 @@
|
|||||||
|
package com.cameleer3.server.core.storage;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Optional;
|
||||||
|
|
||||||
|
public interface ExecutionStore {
|
||||||
|
|
||||||
|
void upsert(ExecutionRecord execution);
|
||||||
|
|
||||||
|
void upsertProcessors(String executionId, Instant startTime,
|
||||||
|
String groupName, String routeId,
|
||||||
|
List<ProcessorRecord> processors);
|
||||||
|
|
||||||
|
Optional<ExecutionRecord> findById(String executionId);
|
||||||
|
|
||||||
|
List<ProcessorRecord> findProcessors(String executionId);
|
||||||
|
|
||||||
|
record ExecutionRecord(
|
||||||
|
String executionId, String routeId, String agentId, String groupName,
|
||||||
|
String status, String correlationId, String exchangeId,
|
||||||
|
Instant startTime, Instant endTime, Long durationMs,
|
||||||
|
String errorMessage, String errorStacktrace, String diagramContentHash
|
||||||
|
) {}
|
||||||
|
|
||||||
|
record ProcessorRecord(
|
||||||
|
String executionId, String processorId, String processorType,
|
||||||
|
String diagramNodeId, String groupName, String routeId,
|
||||||
|
int depth, String parentProcessorId, String status,
|
||||||
|
Instant startTime, Instant endTime, Long durationMs,
|
||||||
|
String errorMessage, String errorStacktrace,
|
||||||
|
String inputBody, String outputBody, String inputHeaders, String outputHeaders
|
||||||
|
) {}
|
||||||
|
}
|
||||||
@@ -1,17 +0,0 @@
|
|||||||
package com.cameleer3.server.core.storage;
|
|
||||||
|
|
||||||
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
|
||||||
|
|
||||||
import java.util.List;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Repository for agent metrics batch inserts into ClickHouse.
|
|
||||||
*/
|
|
||||||
public interface MetricsRepository {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Insert a batch of metrics snapshots.
|
|
||||||
* Implementations must perform a single batch insert for efficiency.
|
|
||||||
*/
|
|
||||||
void insertBatch(List<MetricsSnapshot> metrics);
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,10 @@
|
|||||||
|
package com.cameleer3.server.core.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.storage.model.MetricsSnapshot;
|
||||||
|
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
public interface MetricsStore {
|
||||||
|
|
||||||
|
void insertBatch(List<MetricsSnapshot> snapshots);
|
||||||
|
}
|
||||||
@@ -0,0 +1,17 @@
|
|||||||
|
package com.cameleer3.server.core.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.search.ExecutionSummary;
|
||||||
|
import com.cameleer3.server.core.search.SearchRequest;
|
||||||
|
import com.cameleer3.server.core.search.SearchResult;
|
||||||
|
import com.cameleer3.server.core.storage.model.ExecutionDocument;
|
||||||
|
|
||||||
|
public interface SearchIndex {
|
||||||
|
|
||||||
|
SearchResult<ExecutionSummary> search(SearchRequest request);
|
||||||
|
|
||||||
|
long count(SearchRequest request);
|
||||||
|
|
||||||
|
void index(ExecutionDocument document);
|
||||||
|
|
||||||
|
void delete(String executionId);
|
||||||
|
}
|
||||||
@@ -0,0 +1,36 @@
|
|||||||
|
package com.cameleer3.server.core.storage;
|
||||||
|
|
||||||
|
import com.cameleer3.server.core.search.ExecutionStats;
|
||||||
|
import com.cameleer3.server.core.search.StatsTimeseries;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
public interface StatsStore {
|
||||||
|
|
||||||
|
// Global stats (stats_1m_all)
|
||||||
|
ExecutionStats stats(Instant from, Instant to);
|
||||||
|
|
||||||
|
// Per-app stats (stats_1m_app)
|
||||||
|
ExecutionStats statsForApp(Instant from, Instant to, String groupName);
|
||||||
|
|
||||||
|
// Per-route stats (stats_1m_route), optionally scoped to specific agents
|
||||||
|
ExecutionStats statsForRoute(Instant from, Instant to, String routeId, List<String> agentIds);
|
||||||
|
|
||||||
|
// Per-processor stats (stats_1m_processor)
|
||||||
|
ExecutionStats statsForProcessor(Instant from, Instant to, String routeId, String processorType);
|
||||||
|
|
||||||
|
// Global timeseries
|
||||||
|
StatsTimeseries timeseries(Instant from, Instant to, int bucketCount);
|
||||||
|
|
||||||
|
// Per-app timeseries
|
||||||
|
StatsTimeseries timeseriesForApp(Instant from, Instant to, int bucketCount, String groupName);
|
||||||
|
|
||||||
|
// Per-route timeseries, optionally scoped to specific agents
|
||||||
|
StatsTimeseries timeseriesForRoute(Instant from, Instant to, int bucketCount,
|
||||||
|
String routeId, List<String> agentIds);
|
||||||
|
|
||||||
|
// Per-processor timeseries
|
||||||
|
StatsTimeseries timeseriesForProcessor(Instant from, Instant to, int bucketCount,
|
||||||
|
String routeId, String processorType);
|
||||||
|
}
|
||||||
@@ -0,0 +1,19 @@
|
|||||||
|
package com.cameleer3.server.core.storage.model;
|
||||||
|
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
public record ExecutionDocument(
|
||||||
|
String executionId, String routeId, String agentId, String groupName,
|
||||||
|
String status, String correlationId, String exchangeId,
|
||||||
|
Instant startTime, Instant endTime, Long durationMs,
|
||||||
|
String errorMessage, String errorStacktrace,
|
||||||
|
List<ProcessorDoc> processors
|
||||||
|
) {
|
||||||
|
public record ProcessorDoc(
|
||||||
|
String processorId, String processorType, String status,
|
||||||
|
String errorMessage, String errorStacktrace,
|
||||||
|
String inputBody, String outputBody,
|
||||||
|
String inputHeaders, String outputHeaders
|
||||||
|
) {}
|
||||||
|
}
|
||||||
@@ -1,6 +1,7 @@
|
|||||||
package com.cameleer3.server.core.detail;
|
package com.cameleer3.server.core.detail;
|
||||||
|
|
||||||
import com.cameleer3.server.core.storage.ExecutionRepository;
|
import com.cameleer3.server.core.storage.ExecutionStore;
|
||||||
|
import com.cameleer3.server.core.storage.ExecutionStore.ProcessorRecord;
|
||||||
import org.junit.jupiter.api.Test;
|
import org.junit.jupiter.api.Test;
|
||||||
|
|
||||||
import java.time.Instant;
|
import java.time.Instant;
|
||||||
@@ -10,33 +11,36 @@ import static org.assertj.core.api.Assertions.assertThat;
|
|||||||
import static org.mockito.Mockito.mock;
|
import static org.mockito.Mockito.mock;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Unit tests for {@link DetailService#reconstructTree} logic.
|
* Unit tests for {@link DetailService#buildTree} logic.
|
||||||
* <p>
|
* <p>
|
||||||
* Verifies correct parent-child wiring from flat parallel arrays.
|
* Verifies correct parent-child wiring from flat ProcessorRecord lists.
|
||||||
*/
|
*/
|
||||||
class TreeReconstructionTest {
|
class TreeReconstructionTest {
|
||||||
|
|
||||||
private final DetailService detailService = new DetailService(mock(ExecutionRepository.class));
|
private final DetailService detailService = new DetailService(mock(ExecutionStore.class));
|
||||||
|
|
||||||
private static final Instant NOW = Instant.parse("2026-03-10T10:00:00Z");
|
private static final Instant NOW = Instant.parse("2026-03-10T10:00:00Z");
|
||||||
|
|
||||||
|
private ProcessorRecord proc(String id, String type, String status,
|
||||||
|
int depth, String parentId) {
|
||||||
|
return new ProcessorRecord(
|
||||||
|
"exec-1", id, type, "node-" + id,
|
||||||
|
"default", "route1", depth, parentId,
|
||||||
|
status, NOW, NOW, 10L,
|
||||||
|
null, null, null, null, null, null
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void linearChain_rootChildGrandchild() {
|
void linearChain_rootChildGrandchild() {
|
||||||
// [root, child, grandchild], depths=[0,1,2], parents=[-1,0,1]
|
List<ProcessorRecord> processors = List.of(
|
||||||
List<ProcessorNode> roots = detailService.reconstructTree(
|
proc("root", "log", "COMPLETED", 0, null),
|
||||||
new String[]{"root", "child", "grandchild"},
|
proc("child", "bean", "COMPLETED", 1, "root"),
|
||||||
new String[]{"log", "bean", "to"},
|
proc("grandchild", "to", "COMPLETED", 2, "child")
|
||||||
new String[]{"COMPLETED", "COMPLETED", "COMPLETED"},
|
|
||||||
new Instant[]{NOW, NOW, NOW},
|
|
||||||
new Instant[]{NOW, NOW, NOW},
|
|
||||||
new long[]{10, 20, 30},
|
|
||||||
new String[]{"n1", "n2", "n3"},
|
|
||||||
new String[]{"", "", ""},
|
|
||||||
new String[]{"", "", ""},
|
|
||||||
new int[]{0, 1, 2},
|
|
||||||
new int[]{-1, 0, 1}
|
|
||||||
);
|
);
|
||||||
|
|
||||||
|
List<ProcessorNode> roots = detailService.buildTree(processors);
|
||||||
|
|
||||||
assertThat(roots).hasSize(1);
|
assertThat(roots).hasSize(1);
|
||||||
ProcessorNode root = roots.get(0);
|
ProcessorNode root = roots.get(0);
|
||||||
assertThat(root.getProcessorId()).isEqualTo("root");
|
assertThat(root.getProcessorId()).isEqualTo("root");
|
||||||
@@ -53,21 +57,14 @@ class TreeReconstructionTest {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
void multipleRoots_noNesting() {
|
void multipleRoots_noNesting() {
|
||||||
// [A, B, C], depths=[0,0,0], parents=[-1,-1,-1]
|
List<ProcessorRecord> processors = List.of(
|
||||||
List<ProcessorNode> roots = detailService.reconstructTree(
|
proc("A", "log", "COMPLETED", 0, null),
|
||||||
new String[]{"A", "B", "C"},
|
proc("B", "log", "COMPLETED", 0, null),
|
||||||
new String[]{"log", "log", "log"},
|
proc("C", "log", "COMPLETED", 0, null)
|
||||||
new String[]{"COMPLETED", "COMPLETED", "COMPLETED"},
|
|
||||||
new Instant[]{NOW, NOW, NOW},
|
|
||||||
new Instant[]{NOW, NOW, NOW},
|
|
||||||
new long[]{10, 20, 30},
|
|
||||||
new String[]{"n1", "n2", "n3"},
|
|
||||||
new String[]{"", "", ""},
|
|
||||||
new String[]{"", "", ""},
|
|
||||||
new int[]{0, 0, 0},
|
|
||||||
new int[]{-1, -1, -1}
|
|
||||||
);
|
);
|
||||||
|
|
||||||
|
List<ProcessorNode> roots = detailService.buildTree(processors);
|
||||||
|
|
||||||
assertThat(roots).hasSize(3);
|
assertThat(roots).hasSize(3);
|
||||||
assertThat(roots.get(0).getProcessorId()).isEqualTo("A");
|
assertThat(roots.get(0).getProcessorId()).isEqualTo("A");
|
||||||
assertThat(roots.get(1).getProcessorId()).isEqualTo("B");
|
assertThat(roots.get(1).getProcessorId()).isEqualTo("B");
|
||||||
@@ -77,21 +74,15 @@ class TreeReconstructionTest {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
void branchingTree_parentWithTwoChildren_secondChildHasGrandchild() {
|
void branchingTree_parentWithTwoChildren_secondChildHasGrandchild() {
|
||||||
// [parent, child1, child2, grandchild], depths=[0,1,1,2], parents=[-1,0,0,2]
|
List<ProcessorRecord> processors = List.of(
|
||||||
List<ProcessorNode> roots = detailService.reconstructTree(
|
proc("parent", "split", "COMPLETED", 0, null),
|
||||||
new String[]{"parent", "child1", "child2", "grandchild"},
|
proc("child1", "log", "COMPLETED", 1, "parent"),
|
||||||
new String[]{"split", "log", "bean", "to"},
|
proc("child2", "bean", "COMPLETED", 1, "parent"),
|
||||||
new String[]{"COMPLETED", "COMPLETED", "COMPLETED", "COMPLETED"},
|
proc("grandchild", "to", "COMPLETED", 2, "child2")
|
||||||
new Instant[]{NOW, NOW, NOW, NOW},
|
|
||||||
new Instant[]{NOW, NOW, NOW, NOW},
|
|
||||||
new long[]{100, 20, 30, 5},
|
|
||||||
new String[]{"n1", "n2", "n3", "n4"},
|
|
||||||
new String[]{"", "", "", ""},
|
|
||||||
new String[]{"", "", "", ""},
|
|
||||||
new int[]{0, 1, 1, 2},
|
|
||||||
new int[]{-1, 0, 0, 2}
|
|
||||||
);
|
);
|
||||||
|
|
||||||
|
List<ProcessorNode> roots = detailService.buildTree(processors);
|
||||||
|
|
||||||
assertThat(roots).hasSize(1);
|
assertThat(roots).hasSize(1);
|
||||||
ProcessorNode parent = roots.get(0);
|
ProcessorNode parent = roots.get(0);
|
||||||
assertThat(parent.getProcessorId()).isEqualTo("parent");
|
assertThat(parent.getProcessorId()).isEqualTo("parent");
|
||||||
@@ -111,30 +102,8 @@ class TreeReconstructionTest {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
void emptyArrays_producesEmptyList() {
|
void emptyList_producesEmptyRoots() {
|
||||||
List<ProcessorNode> roots = detailService.reconstructTree(
|
List<ProcessorNode> roots = detailService.buildTree(List.of());
|
||||||
new String[]{},
|
|
||||||
new String[]{},
|
|
||||||
new String[]{},
|
|
||||||
new Instant[]{},
|
|
||||||
new Instant[]{},
|
|
||||||
new long[]{},
|
|
||||||
new String[]{},
|
|
||||||
new String[]{},
|
|
||||||
new String[]{},
|
|
||||||
new int[]{},
|
|
||||||
new int[]{}
|
|
||||||
);
|
|
||||||
|
|
||||||
assertThat(roots).isEmpty();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Test
|
|
||||||
void nullArrays_producesEmptyList() {
|
|
||||||
List<ProcessorNode> roots = detailService.reconstructTree(
|
|
||||||
null, null, null, null, null, null, null, null, null, null, null
|
|
||||||
);
|
|
||||||
|
|
||||||
assertThat(roots).isEmpty();
|
assertThat(roots).isEmpty();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,57 +0,0 @@
|
|||||||
-- Cameleer3 ClickHouse Schema
|
|
||||||
-- Tables for route executions, route diagrams, and agent metrics.
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS route_executions (
|
|
||||||
execution_id String,
|
|
||||||
route_id LowCardinality(String),
|
|
||||||
agent_id LowCardinality(String),
|
|
||||||
status LowCardinality(String),
|
|
||||||
start_time DateTime64(3, 'UTC'),
|
|
||||||
end_time Nullable(DateTime64(3, 'UTC')),
|
|
||||||
duration_ms UInt64,
|
|
||||||
correlation_id String,
|
|
||||||
exchange_id String,
|
|
||||||
error_message String DEFAULT '',
|
|
||||||
error_stacktrace String DEFAULT '',
|
|
||||||
-- Nested processor executions stored as parallel arrays
|
|
||||||
processor_ids Array(String),
|
|
||||||
processor_types Array(LowCardinality(String)),
|
|
||||||
processor_starts Array(DateTime64(3, 'UTC')),
|
|
||||||
processor_ends Array(DateTime64(3, 'UTC')),
|
|
||||||
processor_durations Array(UInt64),
|
|
||||||
processor_statuses Array(LowCardinality(String)),
|
|
||||||
-- Metadata
|
|
||||||
server_received_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC'),
|
|
||||||
-- Skip indexes
|
|
||||||
INDEX idx_correlation correlation_id TYPE bloom_filter GRANULARITY 4,
|
|
||||||
INDEX idx_error error_message TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4
|
|
||||||
)
|
|
||||||
ENGINE = MergeTree()
|
|
||||||
PARTITION BY toYYYYMMDD(start_time)
|
|
||||||
ORDER BY (agent_id, status, start_time, execution_id)
|
|
||||||
TTL toDateTime(start_time) + toIntervalDay(30)
|
|
||||||
SETTINGS ttl_only_drop_parts = 1;
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS route_diagrams (
|
|
||||||
content_hash String,
|
|
||||||
route_id LowCardinality(String),
|
|
||||||
agent_id LowCardinality(String),
|
|
||||||
definition String,
|
|
||||||
created_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC')
|
|
||||||
)
|
|
||||||
ENGINE = ReplacingMergeTree(created_at)
|
|
||||||
ORDER BY (content_hash);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS agent_metrics (
|
|
||||||
agent_id LowCardinality(String),
|
|
||||||
collected_at DateTime64(3, 'UTC'),
|
|
||||||
metric_name LowCardinality(String),
|
|
||||||
metric_value Float64,
|
|
||||||
tags Map(String, String),
|
|
||||||
server_received_at DateTime64(3, 'UTC') DEFAULT now64(3, 'UTC')
|
|
||||||
)
|
|
||||||
ENGINE = MergeTree()
|
|
||||||
PARTITION BY toYYYYMMDD(collected_at)
|
|
||||||
ORDER BY (agent_id, metric_name, collected_at)
|
|
||||||
TTL toDateTime(collected_at) + toIntervalDay(30)
|
|
||||||
SETTINGS ttl_only_drop_parts = 1;
|
|
||||||
@@ -1,25 +0,0 @@
|
|||||||
-- Phase 2: Schema extension for search, detail, and diagram linking columns.
|
|
||||||
-- Adds exchange snapshot data, processor tree metadata, and diagram content hash.
|
|
||||||
|
|
||||||
ALTER TABLE route_executions
|
|
||||||
ADD COLUMN IF NOT EXISTS exchange_bodies String DEFAULT '',
|
|
||||||
ADD COLUMN IF NOT EXISTS exchange_headers String DEFAULT '',
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_depths Array(UInt16) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_parent_indexes Array(Int32) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_error_messages Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_error_stacktraces Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_input_bodies Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_output_bodies Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_input_headers Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_output_headers Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS processor_diagram_node_ids Array(String) DEFAULT [],
|
|
||||||
ADD COLUMN IF NOT EXISTS diagram_content_hash String DEFAULT '';
|
|
||||||
|
|
||||||
-- Skip indexes for full-text search on new text columns
|
|
||||||
ALTER TABLE route_executions
|
|
||||||
ADD INDEX IF NOT EXISTS idx_exchange_bodies exchange_bodies TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4,
|
|
||||||
ADD INDEX IF NOT EXISTS idx_exchange_headers exchange_headers TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4;
|
|
||||||
|
|
||||||
-- Skip index on error_stacktrace (not indexed in 01-schema.sql, needed for SRCH-05)
|
|
||||||
ALTER TABLE route_executions
|
|
||||||
ADD INDEX IF NOT EXISTS idx_error_stacktrace error_stacktrace TYPE tokenbf_v1(32768, 3, 0) GRANULARITY 4;
|
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user