Compare commits

4 Commits

Author SHA1 Message Date
Timi
3cc371c53e v1.0.2
All checks were successful
CI / build-deploy (pull_request) Successful in 27s
CI / notify-on-failure (pull_request) Has been skipped
2026-04-13 18:27:29 +08:00
Timi
b46e9079d5 v1.0.1
All checks were successful
CI / build-deploy (pull_request) Successful in 26s
CI / notify-on-failure (pull_request) Has been skipped
2026-04-13 00:09:55 +08:00
Timi
dc20070bf8 update status api 2026-04-13 00:09:48 +08:00
Timi
78163441dd v1.0.0
All checks were successful
CI / build-deploy (pull_request) Successful in 16s
CI / notify-on-failure (pull_request) Has been skipped
2026-04-09 13:14:20 +08:00
12 changed files with 142 additions and 97 deletions

View File

@@ -57,20 +57,17 @@ jobs:
- name: Deploy service - name: Deploy service
if: success() if: success()
env: env:
HOST: host.docker.internal CONTAINER_NAME: ${{ vars.CONTAINER_NAME }}
APP_PATH: ${{ vars.APP_PATH }} CONTAINER_TARGET_PATH: ${{ vars.CONTAINER_TARGET_PATH }}
DOCKER_CONTAINER_NAME: ${{ vars.DOCKER_CONTAINER_NAME }}
SSHPASS: ${{ secrets.TIMI_SERVER_SSH_PWD }}
MAX_RETRIES: 3 MAX_RETRIES: 3
RETRY_DELAY: 10 RETRY_DELAY: 10
run: | run: |
if [ -z "$HOST" ] || [ -z "$APP_PATH" ] || [ -z "DOCKER_CONTAINER_NAME" ] || [ -z "$SSHPASS" ]; then if [ -z "$CONTAINER_NAME" ] || [ -z "$CONTAINER_TARGET_PATH" ]; then
echo "Missing production environment variables" echo "Missing production environment variables"
echo "Required: APP_PATH, DOCKER_CONTAINER_NAME, TIMI_SERVER_SSH_PWD" echo "Required: CONTAINER_NAME, CONTAINER_TARGET_PATH"
exit 1 exit 1
fi fi
# 重试函数
retry_command() { retry_command() {
local cmd="$1" local cmd="$1"
local desc="$2" local desc="$2"
@@ -79,10 +76,10 @@ jobs:
while [ $attempt -le $MAX_RETRIES ]; do while [ $attempt -le $MAX_RETRIES ]; do
echo "[$desc] Attempt $attempt/$MAX_RETRIES..." echo "[$desc] Attempt $attempt/$MAX_RETRIES..."
if eval "$cmd"; then if eval "$cmd"; then
echo " $desc succeeded" echo "OK: $desc succeeded"
return 0 return 0
fi fi
echo " $desc failed (attempt $attempt/$MAX_RETRIES)" echo "FAIL: $desc failed (attempt $attempt/$MAX_RETRIES)"
if [ $attempt -lt $MAX_RETRIES ]; then if [ $attempt -lt $MAX_RETRIES ]; then
echo "Retrying in ${RETRY_DELAY}s..." echo "Retrying in ${RETRY_DELAY}s..."
sleep $RETRY_DELAY sleep $RETRY_DELAY
@@ -90,16 +87,10 @@ jobs:
attempt=$((attempt + 1)) attempt=$((attempt + 1))
done done
echo " $desc failed after $MAX_RETRIES attempts" echo "FAIL: $desc failed after $MAX_RETRIES attempts"
return 1 return 1
} }
# SSH 配置(使用密码认证)
SSH_PORT="22"
SSH_OPTS="-o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=30 -o ServerAliveInterval=10 -o ServerAliveCountMax=3 -p $SSH_PORT"
SCP_OPTS="-o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=30 -o ServerAliveInterval=10 -o ServerAliveCountMax=3 -P $SSH_PORT"
# 获取构建产物信息
version=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.version) version=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.version)
artifact_id=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.artifactId) artifact_id=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.artifactId)
jar_file="target/${artifact_id}-${version}.jar" jar_file="target/${artifact_id}-${version}.jar"
@@ -109,18 +100,26 @@ jobs:
exit 1 exit 1
fi fi
# 目标文件名(去掉版本号) if ! command -v docker >/dev/null 2>&1; then
target_jar="${artifact_id}.jar" echo "docker command not found in runner environment"
echo "Deploying $jar_file to $HOST:$APP_PATH/$target_jar"
# 上传文件(带重试)
if ! retry_command "sshpass -e scp $SCP_OPTS \"$jar_file\" \"root@$HOST:$APP_PATH/$target_jar\"" "SCP upload"; then
exit 1 exit 1
fi fi
# 重启 Docker 服务(带重试) if ! docker inspect "$CONTAINER_NAME" >/dev/null 2>&1; then
echo "Restarting Docker service: $DOCKER_SERVICE_NAME" echo "Docker container not found: $CONTAINER_NAME"
if ! retry_command "sshpass -e ssh $SSH_OPTS \"root@$HOST\" \"docker restart $DOCKER_SERVICE_NAME\"" "Docker restart"; then exit 1
fi
target_jar="${artifact_id}.jar"
container_target="${CONTAINER_TARGET_PATH%/}/$target_jar"
echo "Deploying $jar_file to container $CONTAINER_NAME:$container_target"
if ! retry_command "docker cp \"$jar_file\" \"$CONTAINER_NAME:$container_target\"" "Docker copy"; then
exit 1
fi
echo "Restarting Docker container: $CONTAINER_NAME"
if ! retry_command "docker restart \"$CONTAINER_NAME\"" "Docker restart"; then
exit 1 exit 1
fi fi
echo "Deployment completed successfully" echo "Deployment completed successfully"
@@ -142,7 +141,6 @@ jobs:
exit 1 exit 1
fi fi
# Use internal URL if available, fallback to public URL
if [ -n "$GITEA_INTERNAL_URL" ]; then if [ -n "$GITEA_INTERNAL_URL" ]; then
api_base_url="$GITEA_INTERNAL_URL" api_base_url="$GITEA_INTERNAL_URL"
echo "Using internal Gitea URL: $api_base_url" echo "Using internal Gitea URL: $api_base_url"
@@ -151,7 +149,6 @@ jobs:
echo "Using public Gitea URL: $api_base_url" echo "Using public Gitea URL: $api_base_url"
fi fi
# 获取构建产物信息
version=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.version) version=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.version)
artifact_id=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.artifactId) artifact_id=$(mvn -q -DforceStdout help:evaluate -Dexpression=project.artifactId)
jar_file="target/${artifact_id}-${version}.jar" jar_file="target/${artifact_id}-${version}.jar"
@@ -179,17 +176,14 @@ jobs:
echo "API URL: $api_url" echo "API URL: $api_url"
echo "Target commit: $RELEASE_TARGET" echo "Target commit: $RELEASE_TARGET"
# 使用唯一临时文件避免跨 job 污染
release_response_file=$(mktemp /tmp/release_response_XXXXXX.json) release_response_file=$(mktemp /tmp/release_response_XXXXXX.json)
trap "rm -f $release_response_file" EXIT trap "rm -f $release_response_file" EXIT
# 创建 release带重试处理幂等性
release_id="" release_id=""
attempt=1 attempt=1
while [ $attempt -le $MAX_RETRIES ] && [ -z "$release_id" ]; do while [ $attempt -le $MAX_RETRIES ] && [ -z "$release_id" ]; do
echo "[Create release] Attempt $attempt/$MAX_RETRIES..." echo "[Create release] Attempt $attempt/$MAX_RETRIES..."
# 清空临时文件
> "$release_response_file" > "$release_response_file"
http_code=$(curl -sS -w "%{http_code}" -o "$release_response_file" -X POST "$api_url" \ http_code=$(curl -sS -w "%{http_code}" -o "$release_response_file" -X POST "$api_url" \
@@ -203,30 +197,27 @@ jobs:
echo "HTTP Status: $http_code" echo "HTTP Status: $http_code"
if [ "$http_code" = "201" ]; then if [ "$http_code" = "201" ]; then
# 提取第一个 id 字段的值,确保去除换行符
if command -v jq >/dev/null 2>&1; then if command -v jq >/dev/null 2>&1; then
release_id=$(echo "$response" | jq -r '.id' 2>/dev/null) release_id=$(echo "$response" | jq -r '.id' 2>/dev/null)
else else
release_id=$(echo "$response" | grep -o '"id":[0-9]*' | head -1 | cut -d: -f2 | tr -d '\n\r') release_id=$(echo "$response" | grep -o '"id":[0-9]*' | head -1 | cut -d: -f2 | tr -d '\n\r')
fi fi
echo " Release created: id=$release_id" echo "OK: Release created: id=$release_id"
elif [ "$http_code" = "409" ]; then elif [ "$http_code" = "409" ]; then
# HTTP 409 Conflict: Release 已存在,获取现有的 release_id
echo "Release already exists (HTTP 409), fetching existing release..." echo "Release already exists (HTTP 409), fetching existing release..."
existing=$(curl -sS "$api_url" -H "Authorization: token $GITEA_TOKEN" --connect-timeout 30 2>/dev/null || echo "[]") existing=$(curl -sS "$api_url" -H "Authorization: token $GITEA_TOKEN" --connect-timeout 30 2>/dev/null || echo "[]")
# 使用 jq 解析 JSON如果没有 jq 则用 grep
if command -v jq >/dev/null 2>&1; then if command -v jq >/dev/null 2>&1; then
release_id=$(echo "$existing" | jq -r ".[] | select(.tag_name==\"$RELEASE_TAG\") | .id" 2>/dev/null | head -1) release_id=$(echo "$existing" | jq -r ".[] | select(.tag_name==\"$RELEASE_TAG\") | .id" 2>/dev/null | head -1)
else else
release_id=$(echo "$existing" | grep -o '"id":[0-9]*' | head -1 | cut -d: -f2 | tr -d '\n\r') release_id=$(echo "$existing" | grep -o '"id":[0-9]*' | head -1 | cut -d: -f2 | tr -d '\n\r')
fi fi
if [ -n "$release_id" ]; then if [ -n "$release_id" ]; then
echo " Found existing release: id=$release_id" echo "OK: Found existing release: id=$release_id"
else else
echo " Could not find existing release id" echo "FAIL: Could not find existing release id"
fi fi
else else
echo "✗ Failed (HTTP $http_code)" echo "FAIL: Create release failed (HTTP $http_code)"
if [ $attempt -lt $MAX_RETRIES ]; then if [ $attempt -lt $MAX_RETRIES ]; then
echo "Retrying in ${RETRY_DELAY}s..." echo "Retrying in ${RETRY_DELAY}s..."
sleep $RETRY_DELAY sleep $RETRY_DELAY
@@ -236,17 +227,15 @@ jobs:
done done
if [ -z "$release_id" ]; then if [ -z "$release_id" ]; then
echo " Failed to create/find release after $MAX_RETRIES attempts" echo "FAIL: Failed to create or find release after $MAX_RETRIES attempts"
exit 1 exit 1
fi fi
# 上传 fat jar带重试
asset_name=$(basename "$jar_file") asset_name=$(basename "$jar_file")
echo "Uploading asset: $asset_name (size: $file_size bytes)" echo "Uploading asset: $asset_name (size: $file_size bytes)"
upload_url="$api_url/$release_id/assets?name=$asset_name" upload_url="$api_url/$release_id/assets?name=$asset_name"
echo "Upload URL: $upload_url" echo "Upload URL: $upload_url"
# 使用唯一临时文件避免跨 job 污染
asset_response_file=$(mktemp /tmp/asset_response_XXXXXX.json) asset_response_file=$(mktemp /tmp/asset_response_XXXXXX.json)
trap "rm -f $release_response_file $asset_response_file" EXIT trap "rm -f $release_response_file $asset_response_file" EXIT
@@ -255,10 +244,8 @@ jobs:
while [ $attempt -le $MAX_RETRIES ] && [ "$upload_success" = "false" ]; do while [ $attempt -le $MAX_RETRIES ] && [ "$upload_success" = "false" ]; do
echo "[Upload asset] Attempt $attempt/$MAX_RETRIES..." echo "[Upload asset] Attempt $attempt/$MAX_RETRIES..."
# 清空临时文件
> "$asset_response_file" > "$asset_response_file"
# Gitea API 要求使用 multipart/form-data 格式上传文件
http_code=$(curl -sS -w "%{http_code}" -o "$asset_response_file" -X POST "$upload_url" \ http_code=$(curl -sS -w "%{http_code}" -o "$asset_response_file" -X POST "$upload_url" \
-H "Authorization: token $GITEA_TOKEN" \ -H "Authorization: token $GITEA_TOKEN" \
--connect-timeout 30 \ --connect-timeout 30 \
@@ -267,9 +254,9 @@ jobs:
if [ "$http_code" = "201" ]; then if [ "$http_code" = "201" ]; then
upload_success=true upload_success=true
echo " Successfully uploaded: $asset_name" echo "OK: Successfully uploaded: $asset_name"
else else
echo " Upload failed (HTTP $http_code)" echo "FAIL: Upload failed (HTTP $http_code)"
cat "$asset_response_file" 2>/dev/null || true cat "$asset_response_file" 2>/dev/null || true
fi fi
@@ -281,7 +268,7 @@ jobs:
done done
if [ "$upload_success" = "false" ]; then if [ "$upload_success" = "false" ]; then
echo " Failed to upload asset after $MAX_RETRIES attempts" echo "FAIL: Failed to upload asset after $MAX_RETRIES attempts"
exit 1 exit 1
fi fi
@@ -306,7 +293,6 @@ jobs:
COMMIT_SHA: ${{ github.sha }} COMMIT_SHA: ${{ github.sha }}
REPO: ${{ github.repository }} REPO: ${{ github.repository }}
SERVER_URL: ${{ github.server_url }} SERVER_URL: ${{ github.server_url }}
# 通知配置(按需启用)
WEBHOOK_URL: ${{ vars.NOTIFY_WEBHOOK_URL }} WEBHOOK_URL: ${{ vars.NOTIFY_WEBHOOK_URL }}
run: | run: |
echo "=========================================" echo "========================================="
@@ -324,11 +310,9 @@ jobs:
echo "" echo ""
echo "=========================================" echo "========================================="
# 发送 Webhook 通知(钉钉/企业微信/Slack 等)
if [ -n "$WEBHOOK_URL" ]; then if [ -n "$WEBHOOK_URL" ]; then
message="🚨 CI 部署失败\n\nPR: #$PR_NUMBER - $PR_TITLE\n分支: $SOURCE_BRANCH\n提交者: $AUTHOR\n\n请检查并决定:\n 重试 CI\n 回滚合并" message="CI 部署失败\n\nPR: #$PR_NUMBER - $PR_TITLE\n分支: $SOURCE_BRANCH\n提交者: $AUTHOR\n\n请检查并决定:\n- 重试 CI\n- 回滚合并"
# 通用 JSON 格式(适配大多数 Webhook
payload=$(cat <<EOF payload=$(cat <<EOF
{ {
"msgtype": "text", "msgtype": "text",
@@ -343,7 +327,7 @@ jobs:
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d "$payload" || echo "Warning: Failed to send notification" -d "$payload" || echo "Warning: Failed to send notification"
echo " Notification sent" echo "OK: Notification sent"
else else
echo "Note: Set vars.NOTIFY_WEBHOOK_URL to enable webhook notifications" echo "Note: Set vars.NOTIFY_WEBHOOK_URL to enable webhook notifications"
fi fi

View File

@@ -11,7 +11,7 @@
<groupId>com.imyeyu.timiserverapi</groupId> <groupId>com.imyeyu.timiserverapi</groupId>
<artifactId>TimiServerAPI</artifactId> <artifactId>TimiServerAPI</artifactId>
<version>1.0.0</version> <version>1.0.1</version>
<packaging>jar</packaging> <packaging>jar</packaging>
<name>TimiServerAPI</name> <name>TimiServerAPI</name>
<description>imyeyu.com API</description> <description>imyeyu.com API</description>

View File

@@ -348,10 +348,10 @@ public class ServerStatus implements TimiJava {
private String mountPoint; private String mountPoint;
/** 分区总空间 */ /** 分区总空间 */
private long totalBytes; private long total;
/** 分区已用空间 */ /** 分区已用空间 */
private Long usedBytes; private Long used;
/** 磁盘传输耗时 */ /** 磁盘传输耗时 */
private long transferTimeMs; private long transferTimeMs;

View File

@@ -26,17 +26,17 @@ public class DockerController {
private final DockerService dockerService; private final DockerService dockerService;
@GetMapping("/containers") @GetMapping("/container")
public List<DockerContainerSummaryView> listContainers() { public List<DockerContainerSummaryView> listContainers() {
return dockerService.listContainers(); return dockerService.listContainers();
} }
@GetMapping("/containers/{containerId}/status") @GetMapping("/container/{containerId}/status")
public DockerContainerStatusView getContainerStatus(@PathVariable String containerId) { public DockerContainerStatusView getContainerStatus(@PathVariable String containerId) {
return dockerService.getContainerStatus(containerId); return dockerService.getContainerStatus(containerId);
} }
@GetMapping("/containers/{containerId}/history") @GetMapping("/container/{containerId}/history")
public DockerContainerHistoryView getContainerHistory(@PathVariable String containerId, @RequestParam(required = false) String window) { public DockerContainerHistoryView getContainerHistory(@PathVariable String containerId, @RequestParam(required = false) String window) {
return dockerService.getContainerHistory(containerId, window); return dockerService.getContainerHistory(containerId, window);
} }

View File

@@ -72,7 +72,6 @@ public class StatusServiceImplement implements StatusService {
SystemStatusDataView.OS os = new SystemStatusDataView.OS(); SystemStatusDataView.OS os = new SystemStatusDataView.OS();
os.setName(serverStatus.getOs().getName()); os.setName(serverStatus.getOs().getName());
os.setBootAt(serverStatus.getOs().getBootAt()); os.setBootAt(serverStatus.getOs().getBootAt());
os.setUptimeMs(Math.max(0, serverTime - serverStatus.getOs().getBootAt()));
snapshot.setOs(os); snapshot.setOs(os);
} }
if (selectedMetrics.contains(Metric.CPU)) { if (selectedMetrics.contains(Metric.CPU)) {
@@ -80,8 +79,8 @@ public class StatusServiceImplement implements StatusService {
cpu.setModel(serverStatus.getCpu().getName()); cpu.setModel(serverStatus.getCpu().getName());
cpu.setPhysicalCores(serverStatus.getCpu().getCoreCount()); cpu.setPhysicalCores(serverStatus.getCpu().getCoreCount());
cpu.setLogicalCores(serverStatus.getCpu().getLogicalCount()); cpu.setLogicalCores(serverStatus.getCpu().getLogicalCount());
cpu.setUsagePercent(lastDouble(serverStatus.getCpu().getUsed())); cpu.setUsageTotal(lastDouble(serverStatus.getCpu().getUsed()));
cpu.setSystemPercent(lastDouble(serverStatus.getCpu().getSystem())); cpu.setUsageSystem(lastDouble(serverStatus.getCpu().getSystem()));
cpu.setTemperatureCelsius(serverStatus.getCpu().getTemperature()); cpu.setTemperatureCelsius(serverStatus.getCpu().getTemperature());
snapshot.setCpu(cpu); snapshot.setCpu(cpu);
} }
@@ -91,7 +90,6 @@ public class StatusServiceImplement implements StatusService {
Long swapUsedBytes = lastLong(serverStatus.getMemory().getSwapUsed()); Long swapUsedBytes = lastLong(serverStatus.getMemory().getSwapUsed());
memory.setTotalBytes(serverStatus.getMemory().getSize()); memory.setTotalBytes(serverStatus.getMemory().getSize());
memory.setUsedBytes(usedBytes); memory.setUsedBytes(usedBytes);
memory.setUsagePercent(toPercent(usedBytes, serverStatus.getMemory().getSize()));
memory.setSwapTotalBytes(serverStatus.getMemory().getSwapSize()); memory.setSwapTotalBytes(serverStatus.getMemory().getSwapSize());
memory.setSwapUsedBytes(swapUsedBytes); memory.setSwapUsedBytes(swapUsedBytes);
snapshot.setMemory(memory); snapshot.setMemory(memory);
@@ -160,9 +158,9 @@ public class StatusServiceImplement implements StatusService {
item.setPartitionType(partition.getPartitionType()); item.setPartitionType(partition.getPartitionType());
item.setUuid(partition.getUuid()); item.setUuid(partition.getUuid());
item.setMountPoint(partition.getMountPoint()); item.setMountPoint(partition.getMountPoint());
item.setTotalBytes(partition.getTotalBytes()); item.setTotal(partition.getTotal());
item.setUsedBytes(partition.getUsedBytes()); item.setUsed(partition.getUsed());
item.setUsagePercent(toPercent(partition.getUsedBytes(), partition.getTotalBytes())); item.setUsagePercent(toPercent(partition.getUsed(), partition.getTotal()));
item.setTransferTimeMs(partition.getTransferTimeMs()); item.setTransferTimeMs(partition.getTransferTimeMs());
storagePartitions.add(item); storagePartitions.add(item);
} }

View File

@@ -23,7 +23,7 @@ public class UpsServiceImplement implements UpsService {
private final UpsStatusTask upsStatusTask; private final UpsStatusTask upsStatusTask;
private final UpsStatusStore upsStatusStore; private final UpsStatusStore upsStatusStore;
@Value("${ups.collect-rate-ms:60000}") @Value("${ups.collect-rate-ms:3000}")
private long collectRateMs; private long collectRateMs;
@Override @Override

View File

@@ -4,6 +4,7 @@ import com.fasterxml.jackson.databind.JsonNode;
import com.fasterxml.jackson.databind.node.ArrayNode; import com.fasterxml.jackson.databind.node.ArrayNode;
import com.imyeyu.api.modules.system.bean.DockerStatusStore; import com.imyeyu.api.modules.system.bean.DockerStatusStore;
import com.imyeyu.api.modules.system.util.DockerEngineClient; import com.imyeyu.api.modules.system.util.DockerEngineClient;
import com.imyeyu.java.TimiJava;
import com.imyeyu.utils.Time; import com.imyeyu.utils.Time;
import jakarta.validation.constraints.NotNull; import jakarta.validation.constraints.NotNull;
import lombok.RequiredArgsConstructor; import lombok.RequiredArgsConstructor;
@@ -15,12 +16,13 @@ import org.springframework.scheduling.support.PeriodicTrigger;
import org.springframework.stereotype.Service; import org.springframework.stereotype.Service;
import java.util.HashSet; import java.util.HashSet;
import java.util.LinkedHashMap;
import java.util.Map; import java.util.Map;
import java.util.Set; import java.util.Set;
import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeUnit;
/** /**
* Docker 鐘舵€侀噰闆嗕换鍔? * Docker 容器状态采集任务
* *
* @author Codex * @author Codex
* @since 2026-04-06 * @since 2026-04-06
@@ -56,20 +58,28 @@ public class DockerStatusTask implements SchedulingConfigurer {
try { try {
ArrayNode containers = (ArrayNode) dockerEngineClient.getJson("/containers/json", DockerEngineClient.query("all", "true")); ArrayNode containers = (ArrayNode) dockerEngineClient.getJson("/containers/json", DockerEngineClient.query("all", "true"));
long now = Time.now(); long now = Time.now();
synchronized (dockerStatusStore) { Map<String, DockerStatusStore.Container> collectedContainers = new LinkedHashMap<>();
Set<String> activeIds = new HashSet<>();
for (JsonNode summary : containers) { for (JsonNode summary : containers) {
try { try {
String containerId = getAsString(summary, "Id"); String containerId = getAsString(summary, "Id");
activeIds.add(containerId); DockerStatusStore.Container container = new DockerStatusStore.Container();
DockerStatusStore.Container container = dockerStatusStore.getContainers().computeIfAbsent(containerId, key -> new DockerStatusStore.Container()); if (TimiJava.isEmpty(container)) {
continue;
}
updateContainerSummary(container, summary); updateContainerSummary(container, summary);
updateContainerInspect(containerId, container); updateContainerInspect(containerId, container);
updateContainerStats(containerId, container, now); updateContainerStats(containerId, container, now);
collectedContainers.put(container.getId(), container);
} catch (Exception e) { } catch (Exception e) {
log.error("collect docker container item error", e); log.error("collect docker container item error", e);
} }
} }
synchronized (dockerStatusStore) {
Set<String> activeIds = new HashSet<>(collectedContainers.keySet());
for (Map.Entry<String, DockerStatusStore.Container> item : collectedContainers.entrySet()) {
DockerStatusStore.Container container = dockerStatusStore.getContainers().computeIfAbsent(item.getKey(), key -> new DockerStatusStore.Container());
applyCollectedContainer(container, item.getValue());
}
dockerStatusStore.getContainers().entrySet().removeIf(item -> !activeIds.contains(item.getKey())); dockerStatusStore.getContainers().entrySet().removeIf(item -> !activeIds.contains(item.getKey()));
} }
} catch (Exception e) { } catch (Exception e) {
@@ -77,6 +87,39 @@ public class DockerStatusTask implements SchedulingConfigurer {
} }
} }
private void applyCollectedContainer(DockerStatusStore.Container target, DockerStatusStore.Container source) {
target.setId(source.getId());
target.setName(source.getName());
target.setImage(source.getImage());
target.setImageId(source.getImageId());
target.setCreatedAt(source.getCreatedAt());
target.setState(source.getState());
target.setStatus(source.getStatus());
target.setHealthStatus(source.getHealthStatus());
target.setStartedAt(source.getStartedAt());
target.setFinishedAt(source.getFinishedAt());
target.setExitCode(source.getExitCode());
target.setRestartCount(source.getRestartCount());
target.setOomKilled(source.isOomKilled());
target.setCpuPercent(source.getCpuPercent());
target.setMemoryUsageBytes(source.getMemoryUsageBytes());
target.setMemoryLimitBytes(source.getMemoryLimitBytes());
target.setMemoryPercent(source.getMemoryPercent());
target.setNetworkRxBytes(source.getNetworkRxBytes());
target.setNetworkTxBytes(source.getNetworkTxBytes());
target.setBlockReadBytes(source.getBlockReadBytes());
target.setBlockWriteBytes(source.getBlockWriteBytes());
target.setPids(source.getPids());
target.setUpdatedAt(source.getUpdatedAt());
DockerStatusStore.Point point = source.getHistory().peekLast();
if (point != null) {
target.getHistory().addLast(point);
while (historyLimit < target.getHistory().size()) {
target.getHistory().pollFirst();
}
}
}
private void updateContainerSummary(DockerStatusStore.Container container, JsonNode summary) { private void updateContainerSummary(DockerStatusStore.Container container, JsonNode summary) {
container.setId(getAsString(summary, "Id")); container.setId(getAsString(summary, "Id"));
container.setName(trimContainerName(readFirstArrayText(summary, "Names"))); container.setName(trimContainerName(readFirstArrayText(summary, "Names")));
@@ -108,8 +151,8 @@ public class DockerStatusTask implements SchedulingConfigurer {
if (memoryUsageBytes != null && memoryLimitBytes != null && 0 < memoryLimitBytes) { if (memoryUsageBytes != null && memoryLimitBytes != null && 0 < memoryLimitBytes) {
memoryPercent = memoryUsageBytes * 100D / memoryLimitBytes; memoryPercent = memoryUsageBytes * 100D / memoryLimitBytes;
} }
Long networkRxBytes = 0L; long networkRxBytes = 0L;
Long networkTxBytes = 0L; long networkTxBytes = 0L;
JsonNode networks = getAsObject(stats, "networks"); JsonNode networks = getAsObject(stats, "networks");
if (networks != null) { if (networks != null) {
for (Map.Entry<String, JsonNode> item : (Iterable<Map.Entry<String, JsonNode>>) networks::fields) { for (Map.Entry<String, JsonNode> item : (Iterable<Map.Entry<String, JsonNode>>) networks::fields) {
@@ -118,8 +161,8 @@ public class DockerStatusTask implements SchedulingConfigurer {
networkTxBytes += getAsLong(network, "tx_bytes", 0L); networkTxBytes += getAsLong(network, "tx_bytes", 0L);
} }
} }
Long blockReadBytes = 0L; long blockReadBytes = 0L;
Long blockWriteBytes = 0L; long blockWriteBytes = 0L;
JsonNode blkioStats = getAsObject(stats, "blkio_stats"); JsonNode blkioStats = getAsObject(stats, "blkio_stats");
ArrayNode ioServiceBytes = blkioStats == null ? null : getAsArray(blkioStats, "io_service_bytes_recursive"); ArrayNode ioServiceBytes = blkioStats == null ? null : getAsArray(blkioStats, "io_service_bytes_recursive");
if (ioServiceBytes != null) { if (ioServiceBytes != null) {
@@ -134,7 +177,6 @@ public class DockerStatusTask implements SchedulingConfigurer {
} }
} }
Integer pids = getNestedInteger(stats, "pids_stats", "current"); Integer pids = getNestedInteger(stats, "pids_stats", "current");
container.setCpuPercent(cpuPercent); container.setCpuPercent(cpuPercent);
container.setMemoryUsageBytes(memoryUsageBytes); container.setMemoryUsageBytes(memoryUsageBytes);
container.setMemoryLimitBytes(memoryLimitBytes); container.setMemoryLimitBytes(memoryLimitBytes);

View File

@@ -38,10 +38,10 @@ public class CpuStatusCollector extends AbstractDequeStatusCollector {
long irq = ticks[CentralProcessor.TickType.IRQ.getIndex()] - lastCpuTicks[CentralProcessor.TickType.IRQ.getIndex()]; long irq = ticks[CentralProcessor.TickType.IRQ.getIndex()] - lastCpuTicks[CentralProcessor.TickType.IRQ.getIndex()];
long softIrq = ticks[CentralProcessor.TickType.SOFTIRQ.getIndex()] - lastCpuTicks[CentralProcessor.TickType.SOFTIRQ.getIndex()]; long softIrq = ticks[CentralProcessor.TickType.SOFTIRQ.getIndex()] - lastCpuTicks[CentralProcessor.TickType.SOFTIRQ.getIndex()];
long steal = ticks[CentralProcessor.TickType.STEAL.getIndex()] - lastCpuTicks[CentralProcessor.TickType.STEAL.getIndex()]; long steal = ticks[CentralProcessor.TickType.STEAL.getIndex()] - lastCpuTicks[CentralProcessor.TickType.STEAL.getIndex()];
long total = user + nice + sys + idle + ioWait + irq + softIrq + steal; double total = user + nice + sys + idle + ioWait + irq + softIrq + steal;
if (0 < total) { if (0 < total) {
putDeque(context, context.getStatus().getCpu().getSystem(), 100D * sys / total); putDeque(context, context.getStatus().getCpu().getSystem(), sys / total);
putDeque(context, context.getStatus().getCpu().getUsed(), 100 - 100D * idle / total); putDeque(context, context.getStatus().getCpu().getUsed(), 1 - idle / total);
} }
} }
lastCpuTicks = ticks; lastCpuTicks = ticks;

View File

@@ -38,13 +38,13 @@ public class StorageStatusCollector implements StatusCollector {
item.setPartitionType(partition.getType()); item.setPartitionType(partition.getType());
item.setUuid(partition.getUuid()); item.setUuid(partition.getUuid());
item.setMountPoint(partition.getMountPoint()); item.setMountPoint(partition.getMountPoint());
item.setTotalBytes(partition.getSize()); item.setTotal(partition.getSize());
item.setTransferTimeMs(diskStore.getTransferTime()); item.setTransferTimeMs(diskStore.getTransferTime());
OSFileStore fileStore = matchFileStore(partition, fileStoreMap); OSFileStore fileStore = matchFileStore(partition, fileStoreMap);
if (fileStore != null) { if (fileStore != null) {
fileStore.updateAttributes(); fileStore.updateAttributes();
item.setUsedBytes(fileStore.getTotalSpace() - fileStore.getUsableSpace()); item.setUsed(fileStore.getTotalSpace() - fileStore.getUsableSpace());
} }
context.getStatus().getStoragePartitions().add(item); context.getStatus().getStoragePartitions().add(item);
} }

View File

@@ -18,6 +18,7 @@ import java.net.http.HttpResponse;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import java.nio.channels.SocketChannel; import java.nio.channels.SocketChannel;
import java.nio.charset.StandardCharsets; import java.nio.charset.StandardCharsets;
import java.nio.file.Files;
import java.nio.file.Path; import java.nio.file.Path;
import java.time.Duration; import java.time.Duration;
import java.util.LinkedHashMap; import java.util.LinkedHashMap;
@@ -73,6 +74,28 @@ public class DockerEngineClient {
} }
} }
/**
* 判断 Docker Engine 当前是否可访问。
*
* @return 可访问返回 true不可访问返回 false
*/
public boolean isAvailable() {
if (host.startsWith("unix://")) {
String socketPath = host.substring("unix://".length());
return Files.exists(Path.of(socketPath));
}
return true;
}
/**
* 获取 Docker Engine 主机配置。
*
* @return 主机配置字符串
*/
public String getHost() {
return host;
}
private String buildRequestPath(String path, Map<String, String> queryParams) { private String buildRequestPath(String path, Map<String, String> queryParams) {
StringBuilder builder = new StringBuilder(); StringBuilder builder = new StringBuilder();
builder.append("/"); builder.append("/");

View File

@@ -101,9 +101,6 @@ public class SystemStatusDataView {
/** 启动时间 */ /** 启动时间 */
private long bootAt; private long bootAt;
/** 运行时长 */
private long uptimeMs;
} }
/** /**
@@ -125,10 +122,10 @@ public class SystemStatusDataView {
private int logicalCores; private int logicalCores;
/** 总占用 */ /** 总占用 */
private Double usagePercent; private Double usageTotal;
/** 系统占用 */ /** 系统占用 */
private Double systemPercent; private Double usageSystem;
/** 温度 */ /** 温度 */
private double temperatureCelsius; private double temperatureCelsius;
@@ -149,9 +146,6 @@ public class SystemStatusDataView {
/** 已用内存 */ /** 已用内存 */
private Long usedBytes; private Long usedBytes;
/** 使用率 */
private Double usagePercent;
/** 交换分区总量 */ /** 交换分区总量 */
private long swapTotalBytes; private long swapTotalBytes;
@@ -364,10 +358,10 @@ public class SystemStatusDataView {
private String mountPoint; private String mountPoint;
/** 分区总空间 */ /** 分区总空间 */
private long totalBytes; private long total;
/** 已用空间 */ /** 已用空间 */
private Long usedBytes; private Long used;
/** 使用率 */ /** 使用率 */
private Double usagePercent; private Double usagePercent;

View File

@@ -1,5 +1,9 @@
server: server:
shutdown: graceful shutdown: graceful
# 压缩
compression:
enable: true
min-response-size: 10KB
# 开发环境语言,激活开发配置时,多语言系统始终使用此语言环境 # 开发环境语言,激活开发配置时,多语言系统始终使用此语言环境
dev: dev: