Rename Java packages from com.cameleer3 to com.cameleer, module directories from cameleer3-* to cameleer-*, and all references throughout workflows, Dockerfiles, docs, migrations, and pom.xml. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
973 lines
33 KiB
Markdown
973 lines
33 KiB
Markdown
# Container Startup Log Capture — Implementation Plan
|
|
|
|
> **For agentic workers:** REQUIRED SUB-SKILL: Use superpowers:subagent-driven-development (recommended) or superpowers:executing-plans to implement this plan task-by-task. Steps use checkbox (`- [ ]`) syntax for tracking.
|
|
|
|
**Goal:** Capture Docker container stdout/stderr from startup until the Cameleer agent registers via SSE, storing logs in ClickHouse for display in deployment views and general log search.
|
|
|
|
**Architecture:** Extend `RuntimeOrchestrator` with log capture methods. A new `ContainerLogForwarder` component (Docker-specific) streams container output using `docker logs --follow`, batches lines, and flushes them to ClickHouse via the existing `insertBufferedBatch()` path with `source = 'container'`. Capture stops on SSE connect or container death. UI adds a startup log panel below `DeploymentProgress` and source badges in log views.
|
|
|
|
**Tech Stack:** Java 17, Spring Boot 3.4, Docker Java client (zerodep), ClickHouse, React, TanStack Query
|
|
|
|
**Spec:** `docs/superpowers/specs/2026-04-14-container-startup-log-capture-design.md`
|
|
|
|
---
|
|
|
|
## File Map
|
|
|
|
### New Files
|
|
|
|
| File | Responsibility |
|
|
|------|----------------|
|
|
| `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/ContainerLogForwarder.java` | Docker log streaming, batching, ClickHouse flush, session lifecycle |
|
|
| `ui/src/components/StartupLogPanel.tsx` | Collapsible log panel for deployment startup logs |
|
|
| `ui/src/components/StartupLogPanel.module.css` | Styles for startup log panel |
|
|
|
|
### Modified Files
|
|
|
|
| File | Change |
|
|
|------|--------|
|
|
| `cameleer-server-core/.../RuntimeOrchestrator.java` | Add 3 default no-op methods |
|
|
| `cameleer-server-app/.../DockerRuntimeOrchestrator.java` | Implement log capture, delegate to forwarder |
|
|
| `cameleer-server-app/.../DisabledRuntimeOrchestrator.java` | Override no-ops (explicit) |
|
|
| `cameleer-server-app/.../RuntimeOrchestratorAutoConfig.java` | Create and wire `ContainerLogForwarder` bean |
|
|
| `cameleer-server-app/.../DeploymentExecutor.java` | Call `startLogCapture()` after container start |
|
|
| `cameleer-server-app/.../SseConnectionManager.java` | Call `stopLogCaptureByApp()` on SSE connect |
|
|
| `cameleer-server-app/.../DockerEventMonitor.java` | Call `stopLogCapture()` on die/oom |
|
|
| `ui/src/pages/AppsTab/AppsTab.tsx` | Render `StartupLogPanel` below `DeploymentProgress` |
|
|
| `ui/src/pages/AgentInstance/AgentInstance.tsx` | Add `container` to `LOG_SOURCE_ITEMS` |
|
|
| `ui/src/api/queries/logs.ts` | Add `useStartupLogs` hook |
|
|
|
|
---
|
|
|
|
### Task 1: Extend RuntimeOrchestrator Interface
|
|
|
|
**Files:**
|
|
- Modify: `cameleer-server-core/src/main/java/com/cameleer/server/core/runtime/RuntimeOrchestrator.java`
|
|
- Modify: `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DisabledRuntimeOrchestrator.java`
|
|
|
|
- [ ] **Step 1: Add default methods to RuntimeOrchestrator**
|
|
|
|
In `cameleer-server-core/src/main/java/com/cameleer/server/core/runtime/RuntimeOrchestrator.java`, add three default no-op methods after the existing `getLogs` method (line 11):
|
|
|
|
```java
|
|
package com.cameleer.server.core.runtime;
|
|
|
|
import java.util.stream.Stream;
|
|
|
|
public interface RuntimeOrchestrator {
|
|
boolean isEnabled();
|
|
String startContainer(ContainerRequest request);
|
|
void stopContainer(String containerId);
|
|
void removeContainer(String containerId);
|
|
ContainerStatus getContainerStatus(String containerId);
|
|
Stream<String> getLogs(String containerId, int tailLines);
|
|
|
|
/** Start streaming container logs to ClickHouse. */
|
|
default void startLogCapture(String containerId, String appSlug, String envSlug, String tenantId) {}
|
|
|
|
/** Stop log capture for a specific container (e.g., on die/oom). */
|
|
default void stopLogCapture(String containerId) {}
|
|
|
|
/** Stop log capture for all containers matching this app+env (e.g., on agent SSE connect). */
|
|
default void stopLogCaptureByApp(String appSlug, String envSlug) {}
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 2: Add explicit overrides to DisabledRuntimeOrchestrator**
|
|
|
|
In `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DisabledRuntimeOrchestrator.java`, add explicit no-op overrides after line 15:
|
|
|
|
```java
|
|
@Override public void startLogCapture(String containerId, String appSlug, String envSlug, String tenantId) {}
|
|
@Override public void stopLogCapture(String containerId) {}
|
|
@Override public void stopLogCaptureByApp(String appSlug, String envSlug) {}
|
|
```
|
|
|
|
- [ ] **Step 3: Verify compilation**
|
|
|
|
Run: `mvn clean compile -pl cameleer-server-core,cameleer-server-app -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 4: Commit**
|
|
|
|
```bash
|
|
git add cameleer-server-core/src/main/java/com/cameleer/server/core/runtime/RuntimeOrchestrator.java \
|
|
cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DisabledRuntimeOrchestrator.java
|
|
git commit -m "feat: add log capture methods to RuntimeOrchestrator interface"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 2: Create ContainerLogForwarder
|
|
|
|
**Files:**
|
|
- Create: `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/ContainerLogForwarder.java`
|
|
|
|
- [ ] **Step 1: Create ContainerLogForwarder class**
|
|
|
|
```java
|
|
package com.cameleer.server.app.runtime;
|
|
|
|
import com.cameleer.common.model.LogEntry;
|
|
import com.cameleer.server.app.search.ClickHouseLogStore;
|
|
import com.cameleer.server.core.ingestion.BufferedLogEntry;
|
|
import com.github.dockerjava.api.DockerClient;
|
|
import com.github.dockerjava.api.async.ResultCallback;
|
|
import com.github.dockerjava.api.model.Frame;
|
|
import jakarta.annotation.PreDestroy;
|
|
import org.slf4j.Logger;
|
|
import org.slf4j.LoggerFactory;
|
|
|
|
import java.time.Instant;
|
|
import java.time.format.DateTimeParseException;
|
|
import java.util.ArrayList;
|
|
import java.util.Collections;
|
|
import java.util.List;
|
|
import java.util.Map;
|
|
import java.util.concurrent.*;
|
|
import java.util.regex.Matcher;
|
|
import java.util.regex.Pattern;
|
|
|
|
/**
|
|
* Streams Docker container stdout/stderr to ClickHouse until the agent registers via SSE.
|
|
* One capture session per container, managed by container ID.
|
|
*/
|
|
public class ContainerLogForwarder {
|
|
|
|
private static final Logger log = LoggerFactory.getLogger(ContainerLogForwarder.class);
|
|
|
|
private static final int FLUSH_BATCH_SIZE = 50;
|
|
private static final long FLUSH_INTERVAL_MS = 2_000;
|
|
private static final long MAX_CAPTURE_DURATION_MS = 5 * 60 * 1_000;
|
|
private static final long CLEANUP_INTERVAL_MS = 30_000;
|
|
|
|
// Pattern to match Docker timestamp prefix: "2026-04-14T14:23:01.234567890Z "
|
|
private static final Pattern DOCKER_TS_PATTERN = Pattern.compile(
|
|
"^(\\d{4}-\\d{2}-\\d{2}T\\d{2}:\\d{2}:\\d{2}\\.\\d+Z)\\s(.*)$", Pattern.DOTALL);
|
|
|
|
// Pattern to infer log level from Java log output
|
|
private static final Pattern LEVEL_PATTERN = Pattern.compile(
|
|
"\\b(ERROR|WARN|INFO|DEBUG|TRACE)\\b");
|
|
|
|
private final DockerClient dockerClient;
|
|
private final ClickHouseLogStore logStore;
|
|
private final ConcurrentHashMap<String, CaptureSession> sessions = new ConcurrentHashMap<>();
|
|
private final ExecutorService executor = Executors.newFixedThreadPool(10,
|
|
r -> { Thread t = new Thread(r, "log-capture"); t.setDaemon(true); return t; });
|
|
private final ScheduledExecutorService scheduler = Executors.newSingleThreadScheduledExecutor(
|
|
r -> { Thread t = new Thread(r, "log-capture-cleanup"); t.setDaemon(true); return t; });
|
|
|
|
public ContainerLogForwarder(DockerClient dockerClient, ClickHouseLogStore logStore) {
|
|
this.dockerClient = dockerClient;
|
|
this.logStore = logStore;
|
|
scheduler.scheduleAtFixedRate(this::cleanupExpiredSessions,
|
|
CLEANUP_INTERVAL_MS, CLEANUP_INTERVAL_MS, TimeUnit.MILLISECONDS);
|
|
}
|
|
|
|
public void startCapture(String containerId, String appSlug, String envSlug, String tenantId) {
|
|
if (sessions.containsKey(containerId)) {
|
|
log.debug("Already capturing logs for container {}", containerId.substring(0, 12));
|
|
return;
|
|
}
|
|
|
|
CaptureSession session = new CaptureSession(containerId, appSlug, envSlug, tenantId);
|
|
if (sessions.putIfAbsent(containerId, session) != null) {
|
|
return; // another thread beat us
|
|
}
|
|
|
|
Future<?> future = executor.submit(() -> streamLogs(session));
|
|
session.future = future;
|
|
log.info("Started log capture for container {} (app={}, env={})",
|
|
containerId.substring(0, 12), appSlug, envSlug);
|
|
}
|
|
|
|
public void stopCapture(String containerId) {
|
|
CaptureSession session = sessions.remove(containerId);
|
|
if (session == null) return;
|
|
|
|
session.cancel();
|
|
flushBuffer(session);
|
|
log.info("Stopped log capture for container {} ({} lines captured)",
|
|
containerId.substring(0, 12), session.lineCount);
|
|
}
|
|
|
|
public void stopCaptureByApp(String appSlug, String envSlug) {
|
|
List<String> toRemove = new ArrayList<>();
|
|
for (Map.Entry<String, CaptureSession> entry : sessions.entrySet()) {
|
|
CaptureSession s = entry.getValue();
|
|
if (appSlug.equals(s.appSlug) && envSlug.equals(s.envSlug)) {
|
|
toRemove.add(entry.getKey());
|
|
}
|
|
}
|
|
for (String containerId : toRemove) {
|
|
stopCapture(containerId);
|
|
}
|
|
if (!toRemove.isEmpty()) {
|
|
log.info("Stopped log capture for app={} env={} ({} containers)",
|
|
appSlug, envSlug, toRemove.size());
|
|
}
|
|
}
|
|
|
|
@PreDestroy
|
|
public void shutdown() {
|
|
for (String containerId : new ArrayList<>(sessions.keySet())) {
|
|
stopCapture(containerId);
|
|
}
|
|
scheduler.shutdownNow();
|
|
executor.shutdownNow();
|
|
}
|
|
|
|
private void streamLogs(CaptureSession session) {
|
|
try {
|
|
session.callback = dockerClient.logContainerCmd(session.containerId)
|
|
.withFollowStream(true)
|
|
.withStdOut(true)
|
|
.withStdErr(true)
|
|
.withTimestamps(true)
|
|
.exec(new ResultCallback.Adapter<Frame>() {
|
|
@Override
|
|
public void onNext(Frame frame) {
|
|
if (session.cancelled) return;
|
|
String line = new String(frame.getPayload()).trim();
|
|
if (line.isEmpty()) return;
|
|
|
|
session.buffer.add(line);
|
|
session.lineCount++;
|
|
|
|
if (session.buffer.size() >= FLUSH_BATCH_SIZE
|
|
|| System.currentTimeMillis() - session.lastFlush >= FLUSH_INTERVAL_MS) {
|
|
flushBuffer(session);
|
|
}
|
|
}
|
|
|
|
@Override
|
|
public void onComplete() {
|
|
flushBuffer(session);
|
|
sessions.remove(session.containerId);
|
|
log.debug("Log stream completed for container {}",
|
|
session.containerId.substring(0, 12));
|
|
}
|
|
|
|
@Override
|
|
public void onError(Throwable throwable) {
|
|
flushBuffer(session);
|
|
sessions.remove(session.containerId);
|
|
log.debug("Log stream error for container {}: {}",
|
|
session.containerId.substring(0, 12), throwable.getMessage());
|
|
}
|
|
});
|
|
} catch (Exception e) {
|
|
sessions.remove(session.containerId);
|
|
log.warn("Failed to start log capture for container {}: {}",
|
|
session.containerId.substring(0, 12), e.getMessage());
|
|
}
|
|
}
|
|
|
|
private synchronized void flushBuffer(CaptureSession session) {
|
|
List<String> lines;
|
|
synchronized (session.buffer) {
|
|
if (session.buffer.isEmpty()) return;
|
|
lines = new ArrayList<>(session.buffer);
|
|
session.buffer.clear();
|
|
}
|
|
session.lastFlush = System.currentTimeMillis();
|
|
|
|
List<BufferedLogEntry> entries = new ArrayList<>(lines.size());
|
|
for (String line : lines) {
|
|
Instant timestamp = Instant.now();
|
|
String message = line;
|
|
|
|
Matcher tsMatcher = DOCKER_TS_PATTERN.matcher(line);
|
|
if (tsMatcher.matches()) {
|
|
try {
|
|
timestamp = Instant.parse(tsMatcher.group(1));
|
|
} catch (DateTimeParseException e) {
|
|
// keep Instant.now()
|
|
}
|
|
message = tsMatcher.group(2);
|
|
}
|
|
|
|
String level = inferLevel(message);
|
|
|
|
LogEntry logEntry = new LogEntry();
|
|
logEntry.setTimestamp(timestamp);
|
|
logEntry.setLevel(level);
|
|
logEntry.setMessage(message);
|
|
logEntry.setLoggerName("");
|
|
logEntry.setThreadName("");
|
|
logEntry.setStackTrace("");
|
|
logEntry.setMdc(Collections.emptyMap());
|
|
logEntry.setSource("container");
|
|
|
|
entries.add(new BufferedLogEntry(
|
|
session.tenantId, session.envSlug, session.containerId.substring(0, 12),
|
|
session.appSlug, logEntry));
|
|
}
|
|
|
|
try {
|
|
logStore.insertBufferedBatch(entries);
|
|
} catch (Exception e) {
|
|
log.warn("Failed to flush {} container log entries for {}: {}",
|
|
entries.size(), session.appSlug, e.getMessage());
|
|
}
|
|
}
|
|
|
|
private String inferLevel(String message) {
|
|
if (message.startsWith("\tat ") || message.startsWith("Caused by:")) {
|
|
return "ERROR";
|
|
}
|
|
Matcher m = LEVEL_PATTERN.matcher(message);
|
|
if (m.find()) {
|
|
return m.group(1);
|
|
}
|
|
return "INFO";
|
|
}
|
|
|
|
private void cleanupExpiredSessions() {
|
|
long now = System.currentTimeMillis();
|
|
for (Map.Entry<String, CaptureSession> entry : sessions.entrySet()) {
|
|
CaptureSession session = entry.getValue();
|
|
if (now - session.startedAt > MAX_CAPTURE_DURATION_MS) {
|
|
log.info("Log capture timeout for container {} (app={}), stopping",
|
|
entry.getKey().substring(0, 12), session.appSlug);
|
|
stopCapture(entry.getKey());
|
|
}
|
|
}
|
|
}
|
|
|
|
private static class CaptureSession {
|
|
final String containerId;
|
|
final String appSlug;
|
|
final String envSlug;
|
|
final String tenantId;
|
|
final long startedAt = System.currentTimeMillis();
|
|
final List<String> buffer = Collections.synchronizedList(new ArrayList<>());
|
|
volatile long lastFlush = System.currentTimeMillis();
|
|
volatile long lineCount = 0;
|
|
volatile boolean cancelled = false;
|
|
volatile Future<?> future;
|
|
volatile ResultCallback.Adapter<Frame> callback;
|
|
|
|
CaptureSession(String containerId, String appSlug, String envSlug, String tenantId) {
|
|
this.containerId = containerId;
|
|
this.appSlug = appSlug;
|
|
this.envSlug = envSlug;
|
|
this.tenantId = tenantId;
|
|
}
|
|
|
|
void cancel() {
|
|
cancelled = true;
|
|
if (callback != null) {
|
|
try { callback.close(); } catch (Exception e) { /* ignore */ }
|
|
}
|
|
if (future != null) {
|
|
future.cancel(true);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 2: Verify compilation**
|
|
|
|
Run: `mvn clean compile -pl cameleer-server-app -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 3: Commit**
|
|
|
|
```bash
|
|
git add cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/ContainerLogForwarder.java
|
|
git commit -m "feat: add ContainerLogForwarder for Docker log streaming to ClickHouse"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 3: Wire ContainerLogForwarder and Implement in DockerRuntimeOrchestrator
|
|
|
|
**Files:**
|
|
- Modify: `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DockerRuntimeOrchestrator.java`
|
|
- Modify: `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/RuntimeOrchestratorAutoConfig.java`
|
|
|
|
- [ ] **Step 1: Add ContainerLogForwarder field and log capture methods to DockerRuntimeOrchestrator**
|
|
|
|
In `DockerRuntimeOrchestrator.java`, add a field and setter after the `dockerClient` field (line 32), and implement the three methods after `getLogs()` (line 198):
|
|
|
|
After line 32 (`private DockerClient dockerClient;`), add:
|
|
|
|
```java
|
|
private ContainerLogForwarder logForwarder;
|
|
|
|
public void setLogForwarder(ContainerLogForwarder logForwarder) {
|
|
this.logForwarder = logForwarder;
|
|
}
|
|
```
|
|
|
|
After line 198 (closing brace of `getLogs()`), add:
|
|
|
|
```java
|
|
@Override
|
|
public void startLogCapture(String containerId, String appSlug, String envSlug, String tenantId) {
|
|
if (logForwarder != null) {
|
|
logForwarder.startCapture(containerId, appSlug, envSlug, tenantId);
|
|
}
|
|
}
|
|
|
|
@Override
|
|
public void stopLogCapture(String containerId) {
|
|
if (logForwarder != null) {
|
|
logForwarder.stopCapture(containerId);
|
|
}
|
|
}
|
|
|
|
@Override
|
|
public void stopLogCaptureByApp(String appSlug, String envSlug) {
|
|
if (logForwarder != null) {
|
|
logForwarder.stopCaptureByApp(appSlug, envSlug);
|
|
}
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 2: Wire ContainerLogForwarder in RuntimeOrchestratorAutoConfig**
|
|
|
|
In `RuntimeOrchestratorAutoConfig.java`, add a new bean method and inject it into the orchestrator. Replace the `runtimeOrchestrator()` bean (lines 19-29):
|
|
|
|
```java
|
|
@Bean
|
|
public RuntimeOrchestrator runtimeOrchestrator(
|
|
@org.springframework.beans.factory.annotation.Autowired(required = false)
|
|
ContainerLogForwarder logForwarder) {
|
|
// Auto-detect: Docker socket available?
|
|
if (Files.exists(Path.of("/var/run/docker.sock"))) {
|
|
log.info("Docker socket detected - enabling Docker runtime orchestrator");
|
|
DockerRuntimeOrchestrator orchestrator = new DockerRuntimeOrchestrator();
|
|
if (logForwarder != null) {
|
|
orchestrator.setLogForwarder(logForwarder);
|
|
}
|
|
return orchestrator;
|
|
}
|
|
// TODO: K8s detection (check for service account token)
|
|
log.info("No Docker socket or K8s detected - runtime management disabled (observability-only mode)");
|
|
return new DisabledRuntimeOrchestrator();
|
|
}
|
|
```
|
|
|
|
Add the `ContainerLogForwarder` bean after `dockerEventMonitor()` (after line 46):
|
|
|
|
```java
|
|
@Bean
|
|
public ContainerLogForwarder containerLogForwarder(RuntimeOrchestrator orchestrator,
|
|
ClickHouseLogStore logStore) {
|
|
if (orchestrator instanceof DockerRuntimeOrchestrator docker) {
|
|
return new ContainerLogForwarder(docker.getDockerClient(), logStore);
|
|
}
|
|
return null;
|
|
}
|
|
```
|
|
|
|
Add the import at the top:
|
|
|
|
```java
|
|
import com.cameleer.server.app.search.ClickHouseLogStore;
|
|
```
|
|
|
|
- [ ] **Step 3: Verify compilation**
|
|
|
|
Run: `mvn clean compile -pl cameleer-server-app -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 4: Commit**
|
|
|
|
```bash
|
|
git add cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DockerRuntimeOrchestrator.java \
|
|
cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/RuntimeOrchestratorAutoConfig.java
|
|
git commit -m "feat: wire ContainerLogForwarder into DockerRuntimeOrchestrator"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 4: Integrate Start Capture in DeploymentExecutor
|
|
|
|
**Files:**
|
|
- Modify: `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DeploymentExecutor.java`
|
|
|
|
- [ ] **Step 1: Add startLogCapture call after each replica starts**
|
|
|
|
In `DeploymentExecutor.java`, inside the replica loop (after line 200, after the `connectContainer` loop), add a call to start log capture. Insert after the closing brace of the `for (String net : additionalNets)` block (line 200) and before the `replicaStates.add(...)` call (line 202):
|
|
|
|
```java
|
|
orchestrator.startLogCapture(containerId, app.slug(), env.slug(), tenantId);
|
|
```
|
|
|
|
This goes right between the network connection loop and the `replicaStates.add(Map.of(...))` call, so the full context looks like:
|
|
|
|
```java
|
|
// Connect to additional networks after container is started
|
|
for (String net : additionalNets) {
|
|
if (networkManager != null) {
|
|
networkManager.connectContainer(containerId, net);
|
|
}
|
|
}
|
|
|
|
orchestrator.startLogCapture(containerId, app.slug(), env.slug(), tenantId);
|
|
|
|
replicaStates.add(Map.of(
|
|
"index", i,
|
|
"containerId", containerId,
|
|
"containerName", containerName,
|
|
"status", "STARTING"
|
|
));
|
|
```
|
|
|
|
- [ ] **Step 2: Verify compilation**
|
|
|
|
Run: `mvn clean compile -pl cameleer-server-app -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 3: Commit**
|
|
|
|
```bash
|
|
git add cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DeploymentExecutor.java
|
|
git commit -m "feat: start log capture when deployment replicas are created"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 5: Integrate Stop Capture in SseConnectionManager
|
|
|
|
**Files:**
|
|
- Modify: `cameleer-server-app/src/main/java/com/cameleer/server/app/agent/SseConnectionManager.java`
|
|
|
|
- [ ] **Step 1: Inject RuntimeOrchestrator and call stopLogCaptureByApp on SSE connect**
|
|
|
|
In `SseConnectionManager.java`, add the `RuntimeOrchestrator` dependency. Modify the constructor (lines 39-45) to accept it:
|
|
|
|
```java
|
|
private final RuntimeOrchestrator runtimeOrchestrator;
|
|
|
|
public SseConnectionManager(AgentRegistryService registryService, AgentRegistryConfig config,
|
|
SsePayloadSigner ssePayloadSigner, ObjectMapper objectMapper,
|
|
RuntimeOrchestrator runtimeOrchestrator) {
|
|
this.registryService = registryService;
|
|
this.config = config;
|
|
this.ssePayloadSigner = ssePayloadSigner;
|
|
this.objectMapper = objectMapper;
|
|
this.runtimeOrchestrator = runtimeOrchestrator;
|
|
}
|
|
```
|
|
|
|
Add the import at the top:
|
|
|
|
```java
|
|
import com.cameleer.server.core.runtime.RuntimeOrchestrator;
|
|
import com.cameleer.server.core.agent.AgentInfo;
|
|
```
|
|
|
|
In the `connect()` method (line 60), after the `log.info("SSE connection established...")` (line 83), add the stop capture call:
|
|
|
|
```java
|
|
// Stop container log capture — agent is now online and will send its own logs
|
|
AgentInfo agent = registryService.findById(agentId);
|
|
if (agent != null) {
|
|
runtimeOrchestrator.stopLogCaptureByApp(agent.applicationId(), agent.environmentId());
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 2: Verify compilation**
|
|
|
|
Run: `mvn clean compile -pl cameleer-server-app -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 3: Commit**
|
|
|
|
```bash
|
|
git add cameleer-server-app/src/main/java/com/cameleer/server/app/agent/SseConnectionManager.java
|
|
git commit -m "feat: stop container log capture when agent SSE connects"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 6: Integrate Stop Capture in DockerEventMonitor
|
|
|
|
**Files:**
|
|
- Modify: `cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DockerEventMonitor.java`
|
|
|
|
- [ ] **Step 1: Call stopLogCapture on die/oom/stop events**
|
|
|
|
In `DockerEventMonitor.java`, inside the `handleEvent()` method, after the replica state is updated and before the `changed` flag check (after line 102, inside the `switch` block), add a call to stop log capture. The cleanest place is right after the `replicas.set(i, updated)` line (line 99), within the same `if` block:
|
|
|
|
Insert after line 101 (`break;`) and before line 102 (`}`):
|
|
|
|
```java
|
|
// Stop log capture for this container — it's dead or stopped
|
|
if ("die".equals(action) || "oom".equals(action) || "stop".equals(action)) {
|
|
runtimeOrchestrator.stopLogCapture(containerId);
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 2: Verify compilation**
|
|
|
|
Run: `mvn clean compile -pl cameleer-server-app -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 3: Commit**
|
|
|
|
```bash
|
|
git add cameleer-server-app/src/main/java/com/cameleer/server/app/runtime/DockerEventMonitor.java
|
|
git commit -m "feat: stop container log capture on Docker die/oom events"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 7: Full Backend Compilation and Verification
|
|
|
|
- [ ] **Step 1: Run full compile across all modules**
|
|
|
|
Run: `mvn clean compile test-compile -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 2: Verify no missing imports or type errors**
|
|
|
|
If compilation fails, fix issues and re-run. Common issues to watch for:
|
|
- Circular bean dependency between `RuntimeOrchestrator` and `ContainerLogForwarder` — the `@Autowired(required = false)` pattern in the auto-config handles this
|
|
- Missing `AgentInfo` import or `findById()` method on `AgentRegistryService` — verify the method exists
|
|
|
|
- [ ] **Step 3: Commit any fixes**
|
|
|
|
Only if Step 2 required changes.
|
|
|
|
---
|
|
|
|
### Task 8: Add useStartupLogs Hook
|
|
|
|
**Files:**
|
|
- Modify: `ui/src/api/queries/logs.ts`
|
|
|
|
- [ ] **Step 1: Add environment to LogSearchParams and fetchLogs**
|
|
|
|
In `ui/src/api/queries/logs.ts`, add `environment` to the `LogSearchParams` interface (after the `source` field, line 33):
|
|
|
|
```typescript
|
|
export interface LogSearchParams {
|
|
q?: string;
|
|
level?: string;
|
|
application?: string;
|
|
agentId?: string;
|
|
source?: string;
|
|
environment?: string;
|
|
exchangeId?: string;
|
|
logger?: string;
|
|
from?: string;
|
|
to?: string;
|
|
cursor?: string;
|
|
limit?: number;
|
|
sort?: 'asc' | 'desc';
|
|
}
|
|
```
|
|
|
|
In the `fetchLogs` function, add the environment param after the source line (after line 50):
|
|
|
|
```typescript
|
|
if (params.environment) urlParams.set('environment', params.environment);
|
|
```
|
|
|
|
- [ ] **Step 2: Add useStartupLogs hook**
|
|
|
|
In `ui/src/api/queries/logs.ts`, add the following hook after the `useApplicationLogs` function (after line 128):
|
|
|
|
```typescript
|
|
/**
|
|
* Fetches container startup logs for a deployment.
|
|
* Polls every 3s while the deployment is STARTING, stops when RUNNING/FAILED.
|
|
*/
|
|
export function useStartupLogs(
|
|
application: string | undefined,
|
|
environment: string | undefined,
|
|
deployCreatedAt: string | undefined,
|
|
isStarting: boolean,
|
|
) {
|
|
const params: LogSearchParams = {
|
|
application: application || undefined,
|
|
environment: environment || undefined,
|
|
source: 'container',
|
|
from: deployCreatedAt || undefined,
|
|
sort: 'asc',
|
|
limit: 500,
|
|
};
|
|
|
|
return useLogs(params, {
|
|
enabled: !!application && !!deployCreatedAt,
|
|
refetchInterval: isStarting ? 3_000 : false,
|
|
});
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 2: Commit**
|
|
|
|
```bash
|
|
git add ui/src/api/queries/logs.ts
|
|
git commit -m "feat: add useStartupLogs hook for container startup log polling"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 9: Create StartupLogPanel Component
|
|
|
|
**Files:**
|
|
- Create: `ui/src/components/StartupLogPanel.module.css`
|
|
- Create: `ui/src/components/StartupLogPanel.tsx`
|
|
|
|
- [ ] **Step 1: Create CSS module**
|
|
|
|
Create `ui/src/components/StartupLogPanel.module.css`:
|
|
|
|
```css
|
|
.panel {
|
|
background: var(--bg-secondary);
|
|
border: 1px solid var(--border);
|
|
border-radius: 6px;
|
|
overflow: hidden;
|
|
margin-top: 8px;
|
|
}
|
|
|
|
.header {
|
|
display: flex;
|
|
justify-content: space-between;
|
|
align-items: center;
|
|
padding: 8px 12px;
|
|
border-bottom: 1px solid var(--border);
|
|
}
|
|
|
|
.headerLeft {
|
|
display: flex;
|
|
align-items: center;
|
|
gap: 8px;
|
|
}
|
|
|
|
.title {
|
|
font-size: 13px;
|
|
font-weight: 600;
|
|
color: var(--text-primary);
|
|
}
|
|
|
|
.badge {
|
|
font-size: 11px;
|
|
padding: 1px 8px;
|
|
border-radius: 10px;
|
|
}
|
|
|
|
.badgeLive {
|
|
background: var(--success-muted);
|
|
color: var(--success);
|
|
}
|
|
|
|
.badgeStopped {
|
|
background: var(--error-muted);
|
|
color: var(--error);
|
|
}
|
|
|
|
.pollingHint {
|
|
font-size: 11px;
|
|
color: var(--text-muted);
|
|
}
|
|
|
|
.lineCount {
|
|
font-size: 12px;
|
|
color: var(--text-muted);
|
|
}
|
|
|
|
.empty {
|
|
padding: 16px;
|
|
text-align: center;
|
|
font-size: 13px;
|
|
color: var(--text-muted);
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 2: Create StartupLogPanel component**
|
|
|
|
Create `ui/src/components/StartupLogPanel.tsx`:
|
|
|
|
```tsx
|
|
import { LogViewer } from '@cameleer/design-system';
|
|
import { useStartupLogs } from '../api/queries/logs';
|
|
import type { Deployment } from '../api/queries/admin/apps';
|
|
import styles from './StartupLogPanel.module.css';
|
|
|
|
interface StartupLogPanelProps {
|
|
deployment: Deployment;
|
|
appSlug: string;
|
|
envSlug: string;
|
|
}
|
|
|
|
export function StartupLogPanel({ deployment, appSlug, envSlug }: StartupLogPanelProps) {
|
|
const isStarting = deployment.status === 'STARTING';
|
|
const isFailed = deployment.status === 'FAILED';
|
|
|
|
const { data } = useStartupLogs(appSlug, envSlug, deployment.createdAt, isStarting);
|
|
|
|
const entries = data?.data ?? [];
|
|
|
|
if (entries.length === 0 && !isStarting) return null;
|
|
|
|
return (
|
|
<div className={styles.panel}>
|
|
<div className={styles.header}>
|
|
<div className={styles.headerLeft}>
|
|
<span className={styles.title}>Startup Logs</span>
|
|
{isStarting && (
|
|
<>
|
|
<span className={`${styles.badge} ${styles.badgeLive}`}>● live</span>
|
|
<span className={styles.pollingHint}>polling every 3s</span>
|
|
</>
|
|
)}
|
|
{isFailed && (
|
|
<span className={`${styles.badge} ${styles.badgeStopped}`}>stopped</span>
|
|
)}
|
|
</div>
|
|
<span className={styles.lineCount}>{entries.length} lines</span>
|
|
</div>
|
|
{entries.length > 0 ? (
|
|
<LogViewer entries={entries} maxHeight={300} />
|
|
) : (
|
|
<div className={styles.empty}>Waiting for container output...</div>
|
|
)}
|
|
</div>
|
|
);
|
|
}
|
|
```
|
|
|
|
- [ ] **Step 3: Commit**
|
|
|
|
```bash
|
|
git add ui/src/components/StartupLogPanel.tsx ui/src/components/StartupLogPanel.module.css
|
|
git commit -m "feat: add StartupLogPanel component for deployment startup logs"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 10: Integrate StartupLogPanel into AppsTab
|
|
|
|
**Files:**
|
|
- Modify: `ui/src/pages/AppsTab/AppsTab.tsx`
|
|
|
|
- [ ] **Step 1: Import StartupLogPanel**
|
|
|
|
In `AppsTab.tsx`, add the import after the `DeploymentProgress` import (line 41):
|
|
|
|
```typescript
|
|
import { StartupLogPanel } from '../../components/StartupLogPanel';
|
|
```
|
|
|
|
- [ ] **Step 2: Render StartupLogPanel below DeploymentProgress**
|
|
|
|
In `AppsTab.tsx`, find the block that renders `DeploymentProgress` (lines 774-779):
|
|
|
|
```tsx
|
|
{deployments.filter((d) => d.deployStage).map((d) => (
|
|
<div key={`progress-${d.id}`} style={{ marginBottom: 8 }}>
|
|
<span className={styles.cellMeta}>{d.containerName}</span>
|
|
<DeploymentProgress currentStage={d.deployStage} failed={d.status === 'FAILED'} />
|
|
</div>
|
|
))}
|
|
```
|
|
|
|
Replace with:
|
|
|
|
```tsx
|
|
{deployments.filter((d) => d.deployStage || d.status === 'FAILED').map((d) => (
|
|
<div key={`progress-${d.id}`} style={{ marginBottom: 8 }}>
|
|
<span className={styles.cellMeta}>{d.containerName}</span>
|
|
<DeploymentProgress currentStage={d.deployStage} failed={d.status === 'FAILED'} />
|
|
<StartupLogPanel deployment={d} appSlug={app.slug} envSlug={environments.find(e => e.id === d.environmentId)?.slug ?? ''} />
|
|
</div>
|
|
))}
|
|
```
|
|
|
|
Note: the filter changes from `d.deployStage` to `d.deployStage || d.status === 'FAILED'` so that failed deployments still show their startup logs even after the deploy stage is cleared.
|
|
|
|
- [ ] **Step 3: Verify the UI compiles**
|
|
|
|
Run: `cd ui && npx tsc --noEmit`
|
|
Expected: No errors
|
|
|
|
- [ ] **Step 4: Commit**
|
|
|
|
```bash
|
|
git add ui/src/pages/AppsTab/AppsTab.tsx
|
|
git commit -m "feat: show startup logs panel below deployment progress"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 11: Add Container Source to Log Source Filters
|
|
|
|
**Files:**
|
|
- Modify: `ui/src/pages/AgentInstance/AgentInstance.tsx`
|
|
|
|
- [ ] **Step 1: Add container to LOG_SOURCE_ITEMS**
|
|
|
|
In `AgentInstance.tsx`, find the `LOG_SOURCE_ITEMS` constant (lines 28-31):
|
|
|
|
```typescript
|
|
const LOG_SOURCE_ITEMS: ButtonGroupItem[] = [
|
|
{ value: 'app', label: 'App' },
|
|
{ value: 'agent', label: 'Agent' },
|
|
];
|
|
```
|
|
|
|
Replace with:
|
|
|
|
```typescript
|
|
const LOG_SOURCE_ITEMS: ButtonGroupItem[] = [
|
|
{ value: 'app', label: 'App' },
|
|
{ value: 'agent', label: 'Agent' },
|
|
{ value: 'container', label: 'Container' },
|
|
];
|
|
```
|
|
|
|
- [ ] **Step 2: Check if AgentHealth has the same filter**
|
|
|
|
Check `ui/src/pages/AgentHealth/AgentHealth.tsx` for `LOG_SOURCE_ITEMS` — if it has the same constant, add `container` there too.
|
|
|
|
- [ ] **Step 3: Commit**
|
|
|
|
```bash
|
|
git add ui/src/pages/AgentInstance/AgentInstance.tsx
|
|
# Also add AgentHealth.tsx if modified
|
|
git commit -m "feat: add container source option to log source filters"
|
|
```
|
|
|
|
---
|
|
|
|
### Task 12: End-to-End Verification
|
|
|
|
- [ ] **Step 1: Full backend build**
|
|
|
|
Run: `mvn clean compile test-compile -q`
|
|
Expected: BUILD SUCCESS
|
|
|
|
- [ ] **Step 2: Full UI type check**
|
|
|
|
Run: `cd ui && npx tsc --noEmit`
|
|
Expected: No errors
|
|
|
|
- [ ] **Step 3: Start dev server and verify UI**
|
|
|
|
Run: `cd ui && npm run dev`
|
|
|
|
Open in browser and verify:
|
|
1. Navigate to an app in the Deployments tab
|
|
2. The deployment progress section renders without errors
|
|
3. The source filter in Agent Instance page shows App / Agent / Container options
|
|
4. No console errors
|
|
|
|
- [ ] **Step 4: Commit any fixes from verification**
|
|
|
|
Only if prior steps required changes.
|
|
|
|
---
|
|
|
|
## Out of Scope (Design System)
|
|
|
|
**Source badge on log lines:** The spec calls for a small `container` / `app` badge on each log line in the `LogViewer` component. `LogViewer` lives in `@cameleer/design-system` — a separate repo we don't build here. This requires a design system update to render the `source` field from `LogEntryResponse`. The data is already flowing (`source` field exists on all log entries); the rendering change belongs in the DS repo. Track as a follow-up.
|