diff options
| author | Dan Engelbrecht <[email protected]> | 2026-03-30 13:58:14 +0200 |
|---|---|---|
| committer | GitHub Enterprise <[email protected]> | 2026-03-30 13:58:14 +0200 |
| commit | 6d75696d11aab547bb34ea22ec10fcdc594e5a44 (patch) | |
| tree | 047db726b2c4cfc05fca433561fe09f635ae88a8 /scripts/test_scripts | |
| parent | hub resource limits (#900) (diff) | |
| download | zen-6d75696d11aab547bb34ea22ec10fcdc594e5a44.tar.xz zen-6d75696d11aab547bb34ea22ec10fcdc594e5a44.zip | |
hub s3 hydrate improvements (#902)
- Feature: Added `--hub-hydration-target-config` option to specify the hydration target via a JSON config file (mutually exclusive with `--hub-hydration-target-spec`); supports `file` and `s3` types with structured settings
```json
{
"type": "file",
"settings": {
"path": "/path/to/hydration/storage"
}
}
```
```json
{
"type": "s3",
"settings": {
"uri": "s3://bucket[/prefix]",
"region": "us-east-1",
"endpoint": "http://localhost:9000",
"path-style": true
}
}
```
- Improvement: Hub hydration dehydration skips the `.sentry-native` directory
- Bugfix: Fixed `MakeSafeAbsolutePathInPlace` when a UNC prefix is present but path uses mixed delimiters
Diffstat (limited to 'scripts/test_scripts')
| -rw-r--r-- | scripts/test_scripts/hub_load_test.py | 969 | ||||
| -rw-r--r-- | scripts/test_scripts/hub_provision_perf_test.py | 501 |
2 files changed, 1470 insertions, 0 deletions
diff --git a/scripts/test_scripts/hub_load_test.py b/scripts/test_scripts/hub_load_test.py new file mode 100644 index 000000000..7bff1eb37 --- /dev/null +++ b/scripts/test_scripts/hub_load_test.py @@ -0,0 +1,969 @@ +#!/usr/bin/env python3 +"""Hub sustained load test. + +Keeps ~N modules concurrently provisioned from a pool of 1000 predefined +module names, writing and reading data to each instance, then either letting +the hub watchdog deprovision idle ones or explicitly deprovisioning them. +Runs indefinitely until Ctrl-C. + +Optional --s3: starts a local MinIO server and configures the hub to use it +as the de/hydration backend. + +Requirements: + pip install boto3 (only needed with --s3) +""" + +from __future__ import annotations + +import argparse +import json +import os +import queue +import random +import subprocess +import sys +import threading +import time +import urllib.error +import urllib.request +import webbrowser +from concurrent.futures import Future, ThreadPoolExecutor, wait as futures_wait +from dataclasses import dataclass, field +from pathlib import Path +from typing import Optional + +_EXE_SUFFIX = ".exe" if sys.platform == "win32" else "" +_MINIO_USER = "minioadmin" +_MINIO_PASS = "minioadmin" +_NAMESPACE = "loadtest" +_BUCKET = "bucket" + +# Key sizes to use for activity writes (bytes) - biased toward smaller. +# Blobs are pre-generated at startup; one per size, shared across all requests. +_KEY_SIZES = [512, 512, 2048, 2048, 8192, 32768] +_BLOBS: list[bytes] = [] # populated by _init_blobs() + + +def _init_blobs() -> None: + seen: set[int] = set() + for size in _KEY_SIZES: + if size not in seen: + seen.add(size) + _BLOBS.append(os.urandom(size)) + else: + # reuse the already-generated blob for this size + _BLOBS.append(next(b for b in _BLOBS if len(b) == size)) + + +# --------------------------------------------------------------------------- +# Executable discovery +# --------------------------------------------------------------------------- + +def _find_zenserver(override: Optional[str]) -> Path: + if override: + p = Path(override) / f"zenserver{_EXE_SUFFIX}" + if not p.exists(): + sys.exit(f"zenserver not found at {p}") + return p + + script_dir = Path(__file__).resolve().parent + repo_root = script_dir.parent.parent + candidates = [ + repo_root / "build" / "windows" / "x64" / "release" / f"zenserver{_EXE_SUFFIX}", + repo_root / "build" / "linux" / "x86_64" / "release" / f"zenserver{_EXE_SUFFIX}", + repo_root / "build" / "macosx" / "x86_64" / "release" / f"zenserver{_EXE_SUFFIX}", + ] + for c in candidates: + if c.exists(): + return c + + matches = list(repo_root.glob(f"build/**/release/zenserver{_EXE_SUFFIX}")) + if matches: + return max(matches, key=lambda p: p.stat().st_mtime) + + sys.exit( + "zenserver executable not found in build/. " + "Run: xmake config -y -m release -a x64 && xmake -y\n" + "Or pass --zenserver-dir <dir>." + ) + + +def _find_minio(zenserver_path: Path) -> Path: + p = zenserver_path.parent / f"minio{_EXE_SUFFIX}" + if not p.exists(): + sys.exit( + f"minio executable not found at {p}. " + "Build with: xmake config -y -m release -a x64 && xmake -y" + ) + return p + + +# --------------------------------------------------------------------------- +# MinIO +# --------------------------------------------------------------------------- + +def _start_minio(minio_exe: Path, data_dir: Path, port: int, console_port: int) -> subprocess.Popen: + minio_data = data_dir / "minio" + minio_data.mkdir(parents=True, exist_ok=True) + env = os.environ.copy() + env["MINIO_ROOT_USER"] = _MINIO_USER + env["MINIO_ROOT_PASSWORD"] = _MINIO_PASS + popen_kwargs: dict = {} + if sys.platform == "win32": + popen_kwargs["creationflags"] = subprocess.CREATE_NEW_PROCESS_GROUP + proc = subprocess.Popen( + [str(minio_exe), "server", str(minio_data), + "--address", f":{port}", + "--console-address", f":{console_port}", + "--quiet"], + env=env, + stdout=subprocess.DEVNULL, + stderr=subprocess.DEVNULL, + **popen_kwargs, + ) + print(f"[minio] started (pid {proc.pid}) on port {port}, console on port {console_port}") + return proc + + +def _wait_for_minio(port: int, timeout_s: float = 30.0) -> None: + deadline = time.monotonic() + timeout_s + url = f"http://localhost:{port}/minio/health/live" + while time.monotonic() < deadline: + try: + with urllib.request.urlopen(url, timeout=1): + print("[minio] ready") + return + except Exception: + time.sleep(0.1) + sys.exit(f"[minio] timed out waiting for readiness after {timeout_s}s") + + +def _create_minio_bucket(port: int, bucket: str) -> None: + try: + import boto3 + import botocore.config + import botocore.exceptions + except ImportError: + sys.exit( + "[minio] boto3 is required for --s3.\n" + "Install it with: pip install boto3" + ) + s3 = boto3.client( + "s3", + endpoint_url=f"http://localhost:{port}", + aws_access_key_id=_MINIO_USER, + aws_secret_access_key=_MINIO_PASS, + region_name="us-east-1", + config=botocore.config.Config(signature_version="s3v4"), + ) + try: + s3.create_bucket(Bucket=bucket) + print(f"[minio] created bucket '{bucket}'") + except botocore.exceptions.ClientError as e: + if e.response["Error"]["Code"] in ("BucketAlreadyOwnedByYou", "BucketAlreadyExists"): + print(f"[minio] bucket '{bucket}' already exists") + else: + raise + + +# --------------------------------------------------------------------------- +# Hub lifecycle +# --------------------------------------------------------------------------- + +def _start_hub( + zenserver_exe: Path, + data_dir: Path, + port: int, + log_file: Path, + idle_timeout: int, + extra_args: list[str], + extra_env: Optional[dict[str, str]], +) -> tuple[subprocess.Popen, object]: + data_dir.mkdir(parents=True, exist_ok=True) + cmd = [ + str(zenserver_exe), + "hub", + f"--data-dir={data_dir}", + f"--port={port}", + "--hub-instance-http-threads=8", + "--hub-instance-corelimit=4", + "--hub-provision-disk-limit-percent=99", + "--hub-provision-memory-limit-percent=80", + "--hub-instance-limit=500", + f"--hub-watchdog-provisioned-inactivity-timeout-seconds={idle_timeout}", + "--hub-watchdog-inactivity-check-margin-seconds=5", + "--hub-watchdog-cycle-interval-ms=2000", + "--hub-watchdog-cycle-processing-budget-ms=3000", + "--hub-watchdog-activity-check-connect-timeout-ms=20", + "--hub-watchdog-activity-check-request-timeout-ms=50", + ] + extra_args + + env = os.environ.copy() + if extra_env: + env.update(extra_env) + + popen_kwargs: dict = {} + if sys.platform == "win32": + popen_kwargs["creationflags"] = subprocess.CREATE_NEW_PROCESS_GROUP + log_handle = log_file.open("wb") + try: + proc = subprocess.Popen( + cmd, env=env, stdout=log_handle, stderr=subprocess.STDOUT, + **popen_kwargs, + ) + except Exception: + log_handle.close() + raise + print(f"[hub] started (pid {proc.pid}), log: {log_file}") + return proc, log_handle + + +def _wait_for_hub(proc: subprocess.Popen, port: int, timeout_s: float = 100.0) -> None: + deadline = time.monotonic() + timeout_s + req = urllib.request.Request(f"http://localhost:{port}/hub/status", + headers={"Accept": "application/json"}) + while time.monotonic() < deadline: + if proc.poll() is not None: + sys.exit(f"[hub] process exited unexpectedly (rc={proc.returncode}) - " + f"is another zenserver already running on port {port}?") + try: + with urllib.request.urlopen(req, timeout=2): + print("[hub] ready") + return + except Exception: + time.sleep(0.2) + sys.exit(f"[hub] timed out waiting for readiness after {timeout_s}s") + + +def _stop_process(proc: subprocess.Popen, name: str, timeout_s: float = 10.0) -> None: + if proc.poll() is not None: + return + proc.terminate() + try: + proc.wait(timeout=timeout_s) + except subprocess.TimeoutExpired: + print(f"[{name}] did not exit after {timeout_s}s, killing") + proc.kill() + proc.wait() + + +# --------------------------------------------------------------------------- +# Module state +# --------------------------------------------------------------------------- + +@dataclass +class ModuleState: + state: str = "idle" # idle|provisioning|active|deprovisioning|watchdog-pending + base_uri: Optional[str] = None + written_keys: list[str] = field(default_factory=list) + activity_rounds_left: int = 0 + explicit_deprovision: bool = False + watchdog_pending_since: float = 0.0 + generation: int = 0 # incremented on each provision; queue entries carry this to discard stale entries + + +# --------------------------------------------------------------------------- +# Counters +# --------------------------------------------------------------------------- + +@dataclass +class Counters: + provisions: int = 0 + deprovisions_explicit: int = 0 + deprovisions_watchdog: int = 0 + provision_rejected: int = 0 + activity_rounds: int = 0 + writes: int = 0 + reads: int = 0 + errors: int = 0 + last_reject_time: float = 0.0 + _lock: threading.Lock = field(default_factory=threading.Lock, repr=False, compare=False) + + def inc(self, name: str, n: int = 1) -> None: + with self._lock: + setattr(self, name, getattr(self, name) + n) + + def record_reject(self) -> None: + with self._lock: + self.provision_rejected += 1 + self.last_reject_time = time.monotonic() + + def snapshot(self) -> dict: + with self._lock: + return {k: v for k, v in self.__dict__.items() if not k.startswith("_")} + + +# --------------------------------------------------------------------------- +# Hub API helpers (urllib, no external deps) +# --------------------------------------------------------------------------- + +def _hub_post(port: int, path: str, timeout_s: float = 30.0) -> tuple[int, dict]: + url = f"http://localhost:{port}{path}" + req = urllib.request.Request(url, data=b"{}", method="POST", + headers={"Content-Type": "application/json", + "Accept": "application/json"}) + try: + with urllib.request.urlopen(req, timeout=timeout_s) as resp: + body = json.loads(resp.read()) + return resp.status, body + except urllib.error.HTTPError as e: + try: + body = json.loads(e.read()) + except Exception: + body = {} + return e.code, body + except Exception: + return 0, {} + + +def _hub_status(port: int, timeout_s: float = 5.0) -> Optional[list[dict]]: + try: + req = urllib.request.Request(f"http://localhost:{port}/hub/status", + headers={"Accept": "application/json"}) + with urllib.request.urlopen(req, timeout=timeout_s) as resp: + data = json.loads(resp.read()) + return data.get("modules", []) + except Exception: + return None + + +def _instance_put(base_uri: str, key: str, data: bytes, timeout_s: float = 10.0) -> int: + url = f"{base_uri}/z$/{_NAMESPACE}/{_BUCKET}/{key}" + req = urllib.request.Request(url, data=data, method="PUT", + headers={"Content-Type": "application/octet-stream", + "Accept": "application/json"}) + try: + with urllib.request.urlopen(req, timeout=timeout_s) as resp: + return resp.status + except urllib.error.HTTPError as e: + return e.code + except Exception: + return 0 + + +def _instance_get(base_uri: str, key: str, timeout_s: float = 10.0) -> int: + url = f"{base_uri}/z$/{_NAMESPACE}/{_BUCKET}/{key}" + req = urllib.request.Request(url, headers={"Accept": "application/octet-stream"}) + try: + with urllib.request.urlopen(req, timeout=timeout_s) as resp: + resp.read() + return resp.status + except urllib.error.HTTPError as e: + return e.code + except Exception: + return 0 + + +# --------------------------------------------------------------------------- +# Worker tasks (run in thread pool - no sleeping) +# --------------------------------------------------------------------------- + +def _task_provision( + module_id: str, + port: int, + state_map: dict[str, ModuleState], + state_lock: threading.Lock, + activity_queue: queue.PriorityQueue, + counters: Counters, + explicit_deprovision_rate: float, + stop_event: threading.Event, +) -> None: + status, body = _hub_post(port, f"/hub/modules/{module_id}/provision") + + generation = 0 + schedule_burst = False + with state_lock: + mod = state_map[module_id] + if mod.state != "provisioning": + # Shutdown changed state while provision was in-flight; leave it alone + return + if status in (200, 202): + instance_port = body.get("port") + base_uri = f"http://localhost:{instance_port}" if instance_port else None + if status == 200: + # Instance is immediately ready + mod.generation += 1 + generation = mod.generation + mod.state = "active" + mod.base_uri = base_uri + mod.written_keys = [] + mod.activity_rounds_left = random.randint(5, 20) + mod.explicit_deprovision = random.random() < explicit_deprovision_rate + schedule_burst = not stop_event.is_set() + else: + # 202: async provision - instance is still starting up. + # Stay in "provisioning"; the hub status poll activates it when ready. + # _shutdown_deprovision_all also covers "provisioning", so shutdown is safe. + mod.state = "provisioning" + mod.base_uri = base_uri + elif status == 409: + mod.state = "idle" + counters.record_reject() + return + else: + mod.state = "idle" + counters.inc("errors") + return + + counters.inc("provisions") + if schedule_burst: + activity_queue.put((time.monotonic() + random.uniform(0.1, 0.3), generation, module_id)) + + +def _task_deprovision( + module_id: str, + port: int, + state_map: dict[str, ModuleState], + state_lock: threading.Lock, + counters: Counters, + retries: int = 5, +) -> None: + succeeded = False + for attempt in range(retries + 1): + status, _ = _hub_post(port, f"/hub/modules/{module_id}/deprovision") + if status in (200, 202, 404): + succeeded = True + break + if status == 409 and attempt < retries: + time.sleep(0.2) + continue + counters.inc("errors") + break + + with state_lock: + state_map[module_id].state = "idle" + state_map[module_id].base_uri = None + state_map[module_id].written_keys = [] + + if succeeded: + counters.inc("deprovisions_explicit") + + +def _task_activity_burst( + module_id: str, + generation: int, + state_map: dict[str, ModuleState], + state_lock: threading.Lock, + activity_queue: queue.PriorityQueue, + counters: Counters, + pool: ThreadPoolExecutor, + port: int, +) -> None: + with state_lock: + mod = state_map[module_id] + if mod.state != "active" or mod.generation != generation: + return + base_uri = mod.base_uri + existing_keys = list(mod.written_keys) + + if not base_uri: + with state_lock: + state_map[module_id].state = "idle" + return + + # Write 3-8 new keys per burst + num_writes = random.randint(3, 8) + new_keys: list[str] = [] + write_errors = 0 + for _ in range(num_writes): + key = f"{random.getrandbits(160):040x}" + data = random.choice(_BLOBS) + status = _instance_put(base_uri, key, data) + if status in (200, 201, 204): + new_keys.append(key) + counters.inc("writes") + else: + write_errors += 1 + counters.inc("errors") + + if write_errors > 0 and not new_keys and not existing_keys: + # Instance unreachable - likely watchdog fired while we were scheduled + with state_lock: + mod = state_map[module_id] + if mod.generation == generation: + mod.state = "idle" + mod.base_uri = None + mod.written_keys = [] + return + + # Read back 1-3 random keys from all known keys + all_keys = existing_keys + new_keys + num_reads = min(random.randint(2, 5), len(all_keys)) + for key in random.sample(all_keys, num_reads): + status = _instance_get(base_uri, key) + if status == 200: + counters.inc("reads") + elif status in (404, 0): + # Key may not exist yet or instance gone; not fatal + pass + else: + counters.inc("errors") + + counters.inc("activity_rounds") + + next_generation = 0 + with state_lock: + mod = state_map[module_id] + if mod.state != "active" or mod.generation != generation: + return + mod.written_keys = (existing_keys + new_keys)[-200:] # cap list size + mod.activity_rounds_left -= 1 + + if mod.activity_rounds_left <= 0: + if mod.explicit_deprovision: + mod.state = "deprovisioning" + mod.base_uri = None + try: + pool.submit( + _task_deprovision, + module_id, port, state_map, state_lock, counters, + ) + except RuntimeError: + pass # pool shutting down; hub watchdog will clean up + else: + mod.state = "watchdog-pending" + mod.watchdog_pending_since = time.monotonic() + return + + next_generation = mod.generation # capture under lock before releasing + + # Schedule next burst soon for semi-continuous activity + next_time = time.monotonic() + random.uniform(0.2, 0.8) + activity_queue.put((next_time, next_generation, module_id)) + + +# --------------------------------------------------------------------------- +# Orchestrator +# --------------------------------------------------------------------------- + +def _orchestrate( + port: int, + state_map: dict[str, ModuleState], + state_lock: threading.Lock, + activity_queue: queue.PriorityQueue, + counters: Counters, + pool: ThreadPoolExecutor, + stop_event: threading.Event, + target_active: int, + explicit_deprovision_rate: float, + idle_timeout: int, +) -> None: + last_status_poll = 0.0 + provision_backoff = False + + while not stop_event.is_set(): + now = time.monotonic() + + # Drain activity queue for due bursts + while True: + try: + next_time, generation, module_id = activity_queue.get_nowait() + except queue.Empty: + break + if next_time <= now: + try: + pool.submit( + _task_activity_burst, + module_id, generation, state_map, state_lock, + activity_queue, counters, pool, port, + ) + except RuntimeError: + break # pool shutting down + else: + activity_queue.put((next_time, generation, module_id)) + break + + # Activate/deactivate backoff based on recent 409 rejections + last_reject = counters.last_reject_time + if last_reject > 0 and now - last_reject < 5.0: + provision_backoff = True + elif provision_backoff and now - last_reject >= 5.0: + provision_backoff = False + + # Count states and submit provision tasks if needed + if not provision_backoff: + with state_lock: + idle_ids = [mid for mid, m in state_map.items() if m.state == "idle"] + working_count = sum( + 1 for m in state_map.values() + if m.state in ("active", "provisioning", "deprovisioning") + ) + watchdog_count = sum( + 1 for m in state_map.values() + if m.state == "watchdog-pending" + ) + + # Provision enough to keep working_count at target, but cap total + # in-flight (working + watchdog-pending) at 2x target to prevent + # runaway accumulation when all modules cycle to watchdog-pending. + inflight_cap = target_active * 2 + deficit = min( + target_active - working_count, + inflight_cap - working_count - watchdog_count, + ) + to_provision = idle_ids[:max(0, deficit)] + for mid in to_provision: + with state_lock: + if state_map[mid].state != "idle": + continue + state_map[mid].state = "provisioning" + try: + pool.submit( + _task_provision, + mid, port, state_map, state_lock, + activity_queue, counters, explicit_deprovision_rate, + stop_event, + ) + except RuntimeError: + with state_lock: + state_map[mid].state = "idle" + + # Poll hub status to detect async provision completions and watchdog-fired modules + if now - last_status_poll >= 5.0: + last_status_poll = now + modules_status = _hub_status(port) + if modules_status is not None: + hub_ids = {m["moduleId"]: m.get("state", "") for m in modules_status} + # Hub watchdog fires at ~idle_timeout from last activity. However, if the + # watchdog's previous visit predates the last burst, it sees a changed + # activity sum and resets LastActivityTime, delaying deprovision by ~173s. + # Explicitly deprovision after idle_timeout + 15s to avoid waiting for that. + timeout_threshold = idle_timeout + 15 + to_deprovision_explicitly: list[str] = [] + with state_lock: + for mid, mod in state_map.items(): + if mod.state == "provisioning" and mod.base_uri is not None: + # Waiting for async (202) provision to complete + hub_state = hub_ids.get(mid, "") + if hub_state == "provisioned": + mod.generation += 1 + mod.state = "active" + mod.written_keys = [] + mod.activity_rounds_left = random.randint(5, 20) + mod.explicit_deprovision = random.random() < explicit_deprovision_rate + activity_queue.put( + (time.monotonic() + random.uniform(0.1, 0.3), mod.generation, mid) + ) + elif mid not in hub_ids: + # Provision failed or was rolled back + mod.state = "idle" + mod.base_uri = None + elif mod.state == "watchdog-pending": + hub_state = hub_ids.get(mid, "") + gone = mid not in hub_ids + timed_out = (now - mod.watchdog_pending_since) > timeout_threshold + if gone or hub_state in ("unprovisioned", "deprovisioning"): + mod.state = "idle" + mod.base_uri = None + mod.written_keys = [] + counters.inc("deprovisions_watchdog") + elif timed_out: + mod.state = "deprovisioning" + to_deprovision_explicitly.append(mid) + for mid in to_deprovision_explicitly: + try: + pool.submit(_task_deprovision, mid, port, state_map, state_lock, counters) + except RuntimeError: + with state_lock: + if state_map[mid].state == "deprovisioning": + state_map[mid].state = "idle" + + stop_event.wait(timeout=0.05) + + +# --------------------------------------------------------------------------- +# Stats display +# --------------------------------------------------------------------------- + +def _stats_thread( + state_map: dict[str, ModuleState], + state_lock: threading.Lock, + counters: Counters, + stop_event: threading.Event, + interval_s: float = 5.0, +) -> None: + is_tty = sys.stdout.isatty() + prev_lines = 0 + t0 = time.monotonic() + prev_snap: Optional[dict] = None + prev_t = t0 + + while not stop_event.is_set(): + stop_event.wait(timeout=interval_s) + now = time.monotonic() + elapsed = now - t0 + dt = now - prev_t + + snap = counters.snapshot() + with state_lock: + states: dict[str, int] = {} + for m in state_map.values(): + states[m.state] = states.get(m.state, 0) + 1 + + def rate(key: str) -> float: + if prev_snap is None or dt <= 0: + return 0.0 + return (snap[key] - prev_snap[key]) / dt * 60.0 + + lines = [ + f"[{time.strftime('%H:%M:%S')}] elapsed={elapsed:.0f}s", + f" modules: idle={states.get('idle', 0)} " + f"provisioning={states.get('provisioning', 0)} " + f"active={states.get('active', 0)} " + f"watchdog-pending={states.get('watchdog-pending', 0)} " + f"deprovisioning={states.get('deprovisioning', 0)}", + f" totals: provisions={snap['provisions']} " + f"deprov-explicit={snap['deprovisions_explicit']} " + f"deprov-watchdog={snap['deprovisions_watchdog']} " + f"rejected={snap['provision_rejected']} " + f"errors={snap['errors']}", + f" data: writes={snap['writes']} reads={snap['reads']} " + f"rounds={snap['activity_rounds']}", + f" rates/min: provisions={rate('provisions'):.1f} " + f"deprov={rate('deprovisions_explicit') + rate('deprovisions_watchdog'):.1f} " + f"writes={rate('writes'):.1f} reads={rate('reads'):.1f}", + ] + + if is_tty and prev_lines > 0: + # Move cursor up to overwrite previous block + sys.stdout.write(f"\033[{prev_lines}A\033[J") + + sys.stdout.write("\n".join(lines) + "\n") + sys.stdout.flush() + prev_lines = len(lines) + prev_snap = snap + prev_t = now + + +# --------------------------------------------------------------------------- +# Shutdown +# --------------------------------------------------------------------------- + +def _wait_for_hub_idle(port: int, hub_proc: subprocess.Popen, timeout_s: float = 120.0) -> None: + """Wait until the hub reports no transitioning instances (dehydration done).""" + _STABLE = {"provisioned", "hibernated", "crashed", "unprovisioned"} + print(f"[shutdown] waiting for hub dehydration (up to {timeout_s:.0f}s)...") + deadline = time.monotonic() + timeout_s + while time.monotonic() < deadline: + if hub_proc.poll() is not None: + print("[shutdown] hub process has exited") + return + modules = _hub_status(port, timeout_s=5.0) + if modules is None: + # Hub not responding. If it has exited, we're done. If it's still alive + # it may be saturated with S3 uploads - keep waiting rather than assuming done. + if hub_proc.poll() is not None: + print("[shutdown] hub process has exited") + return + time.sleep(1.0) + continue + transitioning = [m for m in modules if m.get("state") not in _STABLE] + remaining = len(modules) + if not transitioning: + if remaining: + print(f"[shutdown] hub idle ({remaining} instances in stable state)") + else: + print("[shutdown] hub idle (no instances remaining)") + return + print(f"[shutdown] {len(transitioning)} instance(s) still dehydrating...", end="\r") + time.sleep(1.0) + print(f"\n[shutdown] WARNING: hub did not become idle within {timeout_s:.0f}s") + + +def _shutdown_deprovision_all( + port: int, + state_map: dict[str, ModuleState], + state_lock: threading.Lock, + counters: Counters, + workers: int, + timeout_s: float = 60.0, +) -> None: + with state_lock: + to_deprovision = [ + mid for mid, m in state_map.items() + if m.state in ("active", "watchdog-pending", "provisioning") + ] + for mid in to_deprovision: + state_map[mid].state = "deprovisioning" + + if not to_deprovision: + return + + print(f"\n[shutdown] deprovisioning {len(to_deprovision)} active modules...") + pool = ThreadPoolExecutor(max_workers=min(workers, len(to_deprovision))) + futures: list[Future] = [ + pool.submit(_task_deprovision, mid, port, state_map, state_lock, counters) + for mid in to_deprovision + ] + pool.shutdown(wait=False) + + done_set, not_done_set = futures_wait(futures, timeout=timeout_s) + if not_done_set: + print(f"[shutdown] WARNING: {len(not_done_set)} deprovision tasks did not complete within {timeout_s}s") + else: + print(f"[shutdown] all modules deprovisioned") + + +# --------------------------------------------------------------------------- +# Main +# --------------------------------------------------------------------------- + +def main() -> None: + parser = argparse.ArgumentParser(description=__doc__, + formatter_class=argparse.RawDescriptionHelpFormatter) + parser.add_argument("--data-dir", default="E:/Dev/hub-loadtest", + help="Hub --data-dir (default: E:/Dev/hub-loadtest)") + parser.add_argument("--port", type=int, default=8558, + help="Hub HTTP port (default: 8558)") + parser.add_argument("--workers", type=int, default=50, + help="Thread pool size for HTTP calls (default: 50)") + parser.add_argument("--active-modules", type=int, default=100, + help="Target number of concurrently provisioned modules (default: 100)") + parser.add_argument("--module-count", type=int, default=1000, + help="Total predefined module name pool size (default: 1000)") + parser.add_argument("--idle-timeout", type=int, default=90, + help="Hub watchdog inactivity timeout in seconds (default: 90)") + parser.add_argument("--explicit-deprovision-rate", type=float, default=0.3, + help="Fraction of modules explicitly deprovisioned (default: 0.3)") + parser.add_argument("--zenserver-dir", + help="Directory containing zenserver executable (auto-detected by default)") + parser.add_argument("--s3", action="store_true", + help="Use local MinIO for hub de/hydration instead of filesystem") + parser.add_argument("--minio-port", type=int, default=9000, + help="MinIO S3 API port (default: 9000)") + parser.add_argument("--minio-console-port", type=int, default=9001, + help="MinIO web console port (default: 9001)") + parser.add_argument("--s3-bucket", default="zen-load-test", + help="S3 bucket name for MinIO hydration (default: zen-load-test)") + parser.add_argument("--no-browser", action="store_true", + help="Skip opening browser tabs") + args = parser.parse_args() + + if args.active_modules > args.module_count: + sys.exit( + f"--active-modules ({args.active_modules}) must not exceed " + f"--module-count ({args.module_count})" + ) + + _init_blobs() + + data_dir = Path(args.data_dir) + hub_log = data_dir / "hub.log" + + zenserver_exe = _find_zenserver(args.zenserver_dir) + print(f"[setup] zenserver: {zenserver_exe}") + + module_names = [f"load-test-module-{i:04d}" for i in range(args.module_count)] + state_map: dict[str, ModuleState] = {mid: ModuleState() for mid in module_names} + state_lock = threading.Lock() + activity_queue: queue.PriorityQueue = queue.PriorityQueue() + counters = Counters() + stop_event = threading.Event() + + minio_proc: Optional[subprocess.Popen] = None + hub_proc: Optional[subprocess.Popen] = None + hub_log_handle = None + + hub_extra_args: list[str] = [] + hub_extra_env: Optional[dict[str, str]] = None + + try: + if args.s3: + minio_exe = _find_minio(zenserver_exe) + minio_proc = _start_minio(minio_exe, data_dir, args.minio_port, args.minio_console_port) + _wait_for_minio(args.minio_port) + _create_minio_bucket(args.minio_port, args.s3_bucket) + if not args.no_browser: + webbrowser.open(f"http://localhost:{args.minio_console_port}") + config_json = { + "type": "s3", + "settings": { + "uri": f"s3://{args.s3_bucket}", + "endpoint": f"http://localhost:{args.minio_port}", + "path-style": True, + "region": "us-east-1", + }, + } + data_dir.mkdir(parents=True, exist_ok=True) + config_path = data_dir / "hydration_config.json" + config_path.write_text(json.dumps(config_json), encoding="ascii") + hub_extra_args = [f"--hub-hydration-target-config={config_path}"] + hub_extra_env = { + "AWS_ACCESS_KEY_ID": _MINIO_USER, + "AWS_SECRET_ACCESS_KEY": _MINIO_PASS, + } + + hub_proc, hub_log_handle = _start_hub( + zenserver_exe, data_dir, args.port, + hub_log, args.idle_timeout, + hub_extra_args, hub_extra_env, + ) + _wait_for_hub(hub_proc, args.port) + if not args.no_browser: + webbrowser.open(f"http://localhost:{args.port}/dashboard") + + print(f"[load-test] starting: pool={args.module_count} modules, " + f"target={args.active_modules} active, " + f"idle-timeout={args.idle_timeout}s, " + f"explicit-deprovision={args.explicit_deprovision_rate:.0%}") + print("[load-test] press Ctrl-C to stop") + + with ThreadPoolExecutor(max_workers=args.workers) as pool: + stats_t = threading.Thread( + target=_stats_thread, + args=(state_map, state_lock, counters, stop_event), + daemon=True, + ) + stats_t.start() + + orch_t = threading.Thread( + target=_orchestrate, + args=( + args.port, state_map, state_lock, + activity_queue, counters, pool, + stop_event, args.active_modules, + args.explicit_deprovision_rate, args.idle_timeout, + ), + daemon=True, + ) + orch_t.start() + + try: + while not stop_event.is_set(): + time.sleep(0.5) + if hub_proc.poll() is not None: + print(f"\n[hub] process exited unexpectedly (rc={hub_proc.returncode})") + break + except KeyboardInterrupt: + print("\n[load-test] Ctrl-C received, shutting down...") + + stop_event.set() + orch_t.join(timeout=3.0) + stats_t.join(timeout=3.0) + + _shutdown_deprovision_all( + args.port, state_map, state_lock, counters, + args.workers, timeout_s=60.0, + ) + _wait_for_hub_idle(args.port, hub_proc, timeout_s=max(120.0, args.active_modules * 1.0)) + + _stop_process(hub_proc, "hub", timeout_s=120.0) + if hub_log_handle is not None: + hub_log_handle.close() + hub_log_handle = None + if minio_proc is not None: + _stop_process(minio_proc, "minio") + minio_proc = None + + finally: + # Safety net: only reached if an exception occurred before normal shutdown + if hub_proc is not None and hub_proc.poll() is None: + _stop_process(hub_proc, "hub") + if hub_log_handle is not None: + hub_log_handle.close() + if minio_proc is not None: + _stop_process(minio_proc, "minio") + + +if __name__ == "__main__": + main() diff --git a/scripts/test_scripts/hub_provision_perf_test.py b/scripts/test_scripts/hub_provision_perf_test.py new file mode 100644 index 000000000..5b264ad62 --- /dev/null +++ b/scripts/test_scripts/hub_provision_perf_test.py @@ -0,0 +1,501 @@ +#!/usr/bin/env python3 +"""Hub provisioning performance test. + +Floods a zenserver hub with concurrent provision requests until the hub rejects +further provisioning (HTTP 409), then deprovisions all instances and exits. +A WPR ETW trace runs for the entire test and produces a .etl file for analysis +in WPA or PerfView (CPU sampling + context-switch events for lock contention). + +Optional --s3: starts a local MinIO server and configures the hub to use it +as the de/hydration backend instead of the default local filesystem. + +Requirements: + pip install boto3 (only needed with --s3) +WPR (wpr.exe) must be available (ships with Windows). Running as Administrator +is required for WPR to collect ETW traces. +""" + +from __future__ import annotations + +import argparse +import json +import os +import subprocess +import sys +import time +import urllib.error +import urllib.request +import uuid +import webbrowser +from concurrent.futures import FIRST_COMPLETED, Future, ThreadPoolExecutor, as_completed, wait +from pathlib import Path +from typing import Optional + +_EXE_SUFFIX = ".exe" if sys.platform == "win32" else "" +_STABLE_STATES = {"provisioned", "hibernated", "crashed", "unprovisioned"} +_MINIO_USER = "minioadmin" +_MINIO_PASS = "minioadmin" + + +# --------------------------------------------------------------------------- +# Executable discovery +# --------------------------------------------------------------------------- + +def _find_zenserver(override: Optional[str]) -> Path: + if override: + p = Path(override) / f"zenserver{_EXE_SUFFIX}" + if not p.exists(): + sys.exit(f"zenserver not found at {p}") + return p + + script_dir = Path(__file__).resolve().parent + repo_root = script_dir.parent.parent + candidates = [ + repo_root / "build" / "windows" / "x64" / "release" / f"zenserver{_EXE_SUFFIX}", + repo_root / "build" / "linux" / "x86_64" / "release" / f"zenserver{_EXE_SUFFIX}", + repo_root / "build" / "macosx" / "x86_64" / "release" / f"zenserver{_EXE_SUFFIX}", + ] + for c in candidates: + if c.exists(): + return c + + matches = repo_root.glob(f"build/**/release/zenserver{_EXE_SUFFIX}") + for match in sorted(matches, key=lambda p: p.stat().st_mtime, reverse=True): + return match + + sys.exit( + "zenserver executable not found in build/. " + "Run: xmake config -y -m release -a x64 && xmake -y\n" + "Or pass --zenserver-dir <dir>." + ) + + +def _find_minio(zenserver_path: Path) -> Path: + p = zenserver_path.parent / f"minio{_EXE_SUFFIX}" + if not p.exists(): + sys.exit( + f"minio executable not found at {p}. " + "Build with: xmake config -y -m release -a x64 && xmake -y" + ) + return p + + +# --------------------------------------------------------------------------- +# WPR (Windows Performance Recorder) +# --------------------------------------------------------------------------- + +def _is_elevated() -> bool: + if sys.platform != "win32": + return False + try: + import ctypes + return bool(ctypes.windll.shell32.IsUserAnAdmin()) + except Exception: + return False + + +def _wpr_start(trace_file: Path) -> bool: + if sys.platform != "win32": + print("[wpr] WPR is Windows-only; skipping ETW trace.") + return False + if not _is_elevated(): + print("[wpr] skipping ETW trace - re-run from an elevated (Administrator) prompt to collect traces.") + return False + result = subprocess.run( + ["wpr.exe", "-start", "CPU", "-filemode"], + capture_output=True, text=True + ) + if result.returncode != 0: + print(f"[wpr] WARNING: wpr -start failed (code {result.returncode}):\n{result.stderr.strip()}") + return False + print(f"[wpr] ETW trace started (CPU profile, file mode)") + return True + + +def _wpr_stop(trace_file: Path) -> None: + if sys.platform != "win32": + return + result = subprocess.run( + ["wpr.exe", "-stop", str(trace_file)], + capture_output=True, text=True + ) + if result.returncode != 0: + print(f"[wpr] WARNING: wpr -stop failed (code {result.returncode}):\n{result.stderr.strip()}") + else: + print(f"[wpr] ETW trace saved: {trace_file}") + + +# --------------------------------------------------------------------------- +# MinIO +# --------------------------------------------------------------------------- + +def _start_minio(minio_exe: Path, data_dir: Path, port: int, console_port: int) -> subprocess.Popen: + minio_data = data_dir / "minio" + minio_data.mkdir(parents=True, exist_ok=True) + env = os.environ.copy() + env["MINIO_ROOT_USER"] = _MINIO_USER + env["MINIO_ROOT_PASSWORD"] = _MINIO_PASS + proc = subprocess.Popen( + [str(minio_exe), "server", str(minio_data), + "--address", f":{port}", + "--console-address", f":{console_port}", + "--quiet"], + env=env, + stdout=subprocess.DEVNULL, + stderr=subprocess.DEVNULL, + ) + print(f"[minio] started (pid {proc.pid}) on port {port}, console on port {console_port}") + return proc + + +def _wait_for_minio(port: int, timeout_s: float = 30.0) -> None: + deadline = time.monotonic() + timeout_s + url = f"http://localhost:{port}/minio/health/live" + while time.monotonic() < deadline: + try: + with urllib.request.urlopen(url, timeout=1): + print("[minio] ready") + return + except Exception: + time.sleep(0.1) + sys.exit(f"[minio] timed out waiting for readiness after {timeout_s}s") + + +def _create_minio_bucket(port: int, bucket: str) -> None: + try: + import boto3 + import botocore.config + import botocore.exceptions + except ImportError: + sys.exit( + "[minio] boto3 is required for --s3.\n" + "Install it with: pip install boto3" + ) + s3 = boto3.client( + "s3", + endpoint_url=f"http://localhost:{port}", + aws_access_key_id=_MINIO_USER, + aws_secret_access_key=_MINIO_PASS, + region_name="us-east-1", + config=botocore.config.Config(signature_version="s3v4"), + ) + try: + s3.create_bucket(Bucket=bucket) + print(f"[minio] created bucket '{bucket}'") + except botocore.exceptions.ClientError as e: + if e.response["Error"]["Code"] in ("BucketAlreadyOwnedByYou", "BucketAlreadyExists"): + print(f"[minio] bucket '{bucket}' already exists") + else: + raise + + +# --------------------------------------------------------------------------- +# Hub lifecycle +# --------------------------------------------------------------------------- + +def _start_hub( + zenserver_exe: Path, + data_dir: Path, + port: int, + log_file: Path, + extra_args: list[str], + extra_env: Optional[dict[str, str]], +) -> tuple[subprocess.Popen, object]: + data_dir.mkdir(parents=True, exist_ok=True) + cmd = [ + str(zenserver_exe), + "hub", + f"--data-dir={data_dir}", + f"--port={port}", + "--hub-instance-http-threads=8", + "--hub-instance-corelimit=4", + "--hub-provision-disk-limit-percent=99", + "--hub-provision-memory-limit-percent=80", + "--hub-instance-limit=100", + ] + extra_args + + env = os.environ.copy() + if extra_env: + env.update(extra_env) + + log_handle = log_file.open("wb") + try: + proc = subprocess.Popen( + cmd, env=env, stdout=log_handle, stderr=subprocess.STDOUT + ) + except Exception: + log_handle.close() + raise + print(f"[hub] started (pid {proc.pid}), log: {log_file}") + return proc, log_handle + + +def _wait_for_hub(proc: subprocess.Popen, port: int, timeout_s: float = 100.0) -> None: + deadline = time.monotonic() + timeout_s + req = urllib.request.Request(f"http://localhost:{port}/hub/status", + headers={"Accept": "application/json"}) + while time.monotonic() < deadline: + if proc.poll() is not None: + sys.exit(f"[hub] process exited unexpectedly (rc={proc.returncode}) - " + f"is another zenserver already running on port {port}?") + try: + with urllib.request.urlopen(req, timeout=2): + print("[hub] ready") + return + except Exception: + time.sleep(0.2) + sys.exit(f"[hub] timed out waiting for readiness after {timeout_s}s") + + +def _stop_process(proc: subprocess.Popen, name: str, timeout_s: float = 10.0) -> None: + if proc.poll() is not None: + return + proc.terminate() + try: + proc.wait(timeout=timeout_s) + except subprocess.TimeoutExpired: + print(f"[{name}] did not exit after {timeout_s}s, killing") + proc.kill() + proc.wait() + + +# --------------------------------------------------------------------------- +# Hub HTTP helpers +# --------------------------------------------------------------------------- + +def _hub_url(port: int, path: str) -> str: + return f"http://localhost:{port}{path}" + + +def _provision_one(port: int) -> tuple[str, int]: + module_id = str(uuid.uuid4()) + url = _hub_url(port, f"/hub/modules/{module_id}/provision") + req = urllib.request.Request(url, data=b"{}", method="POST", + headers={"Content-Type": "application/json", + "Accept": "application/json"}) + try: + with urllib.request.urlopen(req, timeout=30) as resp: + return module_id, resp.status + except urllib.error.HTTPError as e: + return module_id, e.code + except Exception: + return module_id, 0 + + +def _deprovision_one(port: int, module_id: str, retries: int = 5) -> int: + url = _hub_url(port, f"/hub/modules/{module_id}/deprovision") + req = urllib.request.Request(url, data=b"{}", method="POST", + headers={"Content-Type": "application/json", + "Accept": "application/json"}) + for attempt in range(retries + 1): + try: + with urllib.request.urlopen(req, timeout=30) as resp: + return resp.status + except urllib.error.HTTPError as e: + if e.code == 409 and attempt < retries: + time.sleep(0.2) + continue + return e.code + except Exception: + return 0 + + +def _hub_status(port: int, timeout_s: float = 5.0) -> Optional[list[dict]]: + try: + req = urllib.request.Request(_hub_url(port, "/hub/status"), + headers={"Accept": "application/json"}) + with urllib.request.urlopen(req, timeout=timeout_s) as resp: + data = json.loads(resp.read()) + return data.get("modules", []) + except Exception: + return None + + +# --------------------------------------------------------------------------- +# Test phases +# --------------------------------------------------------------------------- + +def _flood_provision(port: int, workers: int) -> tuple[list[str], float, float]: + stopped = False + provisioned_ids: list[str] = [] + time_to_rejection: Optional[float] = None + t0 = time.monotonic() + + with ThreadPoolExecutor(max_workers=workers) as pool: + pending: set[Future] = {pool.submit(_provision_one, port) for _ in range(workers)} + + while pending: + done, pending = wait(pending, return_when=FIRST_COMPLETED) + + for f in done: + module_id, status = f.result() + if status in (200, 202): + provisioned_ids.append(module_id) + elif status == 409: + if not stopped: + time_to_rejection = time.monotonic() - t0 + stopped = True + print(f"\n[flood] hub rejected provisioning after " + f"{len(provisioned_ids)} instances " + f"({time_to_rejection:.2f}s)") + elif status == 0: + if not stopped: + stopped = True + print(f"\n[flood] hub unreachable - stopping flood " + f"({len(provisioned_ids)} instances so far)") + else: + print(f"[flood] unexpected status {status} for {module_id}") + + if not stopped: + pending.add(pool.submit(_provision_one, port)) + + wall_clock = time.monotonic() - t0 + return provisioned_ids, time_to_rejection or wall_clock, wall_clock + + +def _wait_stable(port: int, timeout_s: float = 20.0) -> None: + print("[hub] waiting for all instances to reach stable state...") + deadline = time.monotonic() + timeout_s + status_timeout_s = 5.0 + while time.monotonic() < deadline: + modules = _hub_status(port, timeout_s=status_timeout_s) + if modules is None: + time.sleep(0.5) + continue + transitioning = [m for m in modules if m.get("state") not in _STABLE_STATES] + elapsed = time.monotonic() - (deadline - timeout_s) + print(f"[hub] {elapsed:.1f}s: {len(modules) - len(transitioning)}/{len(modules)} stable", end="\r") + if not transitioning: + print(f"\n[hub] all {len(modules)} instances in stable state") + return + time.sleep(0.5) + print(f"\n[hub] WARNING: timed out waiting for stable states after {timeout_s}s") + + +def _deprovision_all(port: int, module_ids: list[str], workers: int) -> None: + raw_status = _hub_status(port, timeout_s=60.0) + if raw_status is None: + print("[deprovision] WARNING: could not reach hub to enumerate extra modules - " + "only deprovisioning tracked instances") + extra_ids = {m["moduleId"] for m in (raw_status or []) + if m.get("state") not in ("unprovisioned",)} - set(module_ids) + all_ids = list(module_ids) + list(extra_ids) + + print(f"[deprovision] deprovisioning {len(all_ids)} instances...") + t0 = time.monotonic() + errors = 0 + with ThreadPoolExecutor(max_workers=workers) as pool: + futures = {pool.submit(_deprovision_one, port, mid): mid for mid in all_ids} + for f in as_completed(futures): + status = f.result() + if status not in (200, 202, 409): + errors += 1 + print(f"[deprovision] module {futures[f]}: unexpected status {status}") + elapsed = time.monotonic() - t0 + print(f"[deprovision] done in {elapsed:.2f}s ({errors} errors)") + + +# --------------------------------------------------------------------------- +# Main +# --------------------------------------------------------------------------- + +def main() -> None: + parser = argparse.ArgumentParser(description=__doc__, + formatter_class=argparse.RawDescriptionHelpFormatter) + parser.add_argument("--data-dir", default="E:/Dev/hub-perftest", + help="Hub --data-dir (default: E:/Dev/hub-perftest)") + parser.add_argument("--port", type=int, default=8558, + help="Hub HTTP port (default: 8558)") + parser.add_argument("--workers", type=int, default=20, + help="Concurrent provisioning threads (default: 20)") + parser.add_argument("--trace-file", default="hub_perf_trace.etl", + help="WPR output .etl path (default: hub_perf_trace.etl)") + parser.add_argument("--zenserver-dir", + help="Directory containing zenserver executable (auto-detected by default)") + parser.add_argument("--s3", action="store_true", + help="Use local MinIO for hub de/hydration instead of filesystem") + parser.add_argument("--minio-port", type=int, default=9000, + help="MinIO S3 API port when --s3 is used (default: 9000)") + parser.add_argument("--minio-console-port", type=int, default=9001, + help="MinIO web console port when --s3 is used (default: 9001)") + parser.add_argument("--s3-bucket", default="zen-hydration-test", + help="S3 bucket name for MinIO hydration (default: zen-hydration-test)") + args = parser.parse_args() + + data_dir = Path(args.data_dir) + trace_file = Path(args.trace_file).resolve() + hub_log = data_dir / "hub.log" + + zenserver_exe = _find_zenserver(args.zenserver_dir) + print(f"[setup] zenserver: {zenserver_exe}") + + minio_proc: Optional[subprocess.Popen] = None + hub_proc: Optional[subprocess.Popen] = None + hub_log_handle = None + wpr_started = False + + hub_extra_args: list[str] = [] + hub_extra_env: Optional[dict[str, str]] = None + + try: + if args.s3: + minio_exe = _find_minio(zenserver_exe) + minio_proc = _start_minio(minio_exe, data_dir, args.minio_port, args.minio_console_port) + _wait_for_minio(args.minio_port) + _create_minio_bucket(args.minio_port, args.s3_bucket) + webbrowser.open(f"http://localhost:{args.minio_console_port}") + config_json = { + "type": "s3", + "settings": { + "uri": f"s3://{args.s3_bucket}", + "endpoint": f"http://localhost:{args.minio_port}", + "path-style": True, + "region": "us-east-1", + }, + } + data_dir.mkdir(parents=True, exist_ok=True) + config_path = data_dir / "hydration_config.json" + config_path.write_text(json.dumps(config_json), encoding="ascii") + hub_extra_args = [ + f"--hub-hydration-target-config={config_path}", + ] + hub_extra_env = { + "AWS_ACCESS_KEY_ID": _MINIO_USER, + "AWS_SECRET_ACCESS_KEY": _MINIO_PASS, + } + + wpr_started = _wpr_start(trace_file) + + hub_proc, hub_log_handle = _start_hub( + zenserver_exe, data_dir, args.port, + hub_log, hub_extra_args, hub_extra_env + ) + _wait_for_hub(hub_proc, args.port) + webbrowser.open(f"http://localhost:{args.port}/dashboard") + + provisioned_ids, time_to_rejection, wall_clock = _flood_provision(args.port, args.workers) + + print(f"\n[results] provisioned : {len(provisioned_ids)}") + print(f"[results] time to 409 : {time_to_rejection:.3f}s") + print(f"[results] wall clock : {wall_clock:.3f}s") + if time_to_rejection > 0: + print(f"[results] rate : {len(provisioned_ids) / time_to_rejection:.1f} provisions/s") + + _wait_stable(args.port, timeout_s=120.0) + _deprovision_all(args.port, provisioned_ids, args.workers) + dehydration_timeout_s = max(60.0, len(provisioned_ids) * 0.5) + _wait_stable(args.port, timeout_s=dehydration_timeout_s) + + finally: + if wpr_started: + _wpr_stop(trace_file) + if hub_proc is not None: + _stop_process(hub_proc, "hub") + if hub_log_handle is not None: + hub_log_handle.close() + if minio_proc is not None: + _stop_process(minio_proc, "minio") + + +if __name__ == "__main__": + main() |