mirror of
https://github.com/OpenRouterTeam/spawn.git
synced 2026-05-02 13:50:26 +00:00
- Add trap 'rm -f "${ENV_TEMP}"' EXIT after mktemp creation
- Scripts with DOTENV_TEMP get combined trap for both files
- Remove manual rm calls that are now redundant
- Prevents temp file leaks on early script exit (errors, signals)
- Affects 67 agent scripts across all providers
Impact: Prevents /tmp pollution in production deployments
Score: 90 (Impact: 9, Confidence: 10, Risk: 1)
68 lines
1.9 KiB
Bash
Executable file
68 lines
1.9 KiB
Bash
Executable file
#!/bin/bash
|
|
set -eo pipefail
|
|
|
|
# Source common functions - try local file first, fall back to remote
|
|
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" 2>/dev/null && pwd)"
|
|
# shellcheck source=lambda/lib/common.sh
|
|
if [[ -f "${SCRIPT_DIR}/lib/common.sh" ]]; then
|
|
source "${SCRIPT_DIR}/lib/common.sh"
|
|
else
|
|
eval "$(curl -fsSL https://raw.githubusercontent.com/OpenRouterTeam/spawn/main/lambda/lib/common.sh)"
|
|
fi
|
|
|
|
log_info "Codex CLI on Lambda Cloud"
|
|
echo ""
|
|
|
|
# 1. Ensure Lambda API key is configured
|
|
ensure_lambda_token
|
|
|
|
# 2. Generate + register SSH key
|
|
ensure_ssh_key
|
|
|
|
# 3. Get instance name and create server
|
|
SERVER_NAME=$(get_server_name)
|
|
create_server "${SERVER_NAME}"
|
|
|
|
# 4. Wait for SSH and cloud-init
|
|
verify_server_connectivity "${LAMBDA_SERVER_IP}"
|
|
wait_for_cloud_init "${LAMBDA_SERVER_IP}"
|
|
|
|
# 5. Install Codex CLI
|
|
log_warn "Installing Codex CLI..."
|
|
run_server "${LAMBDA_SERVER_IP}" "npm install -g @openai/codex"
|
|
log_info "Codex CLI installed"
|
|
|
|
# 6. Get OpenRouter API key
|
|
echo ""
|
|
if [[ -n "${OPENROUTER_API_KEY:-}" ]]; then
|
|
log_info "Using OpenRouter API key from environment"
|
|
else
|
|
OPENROUTER_API_KEY=$(get_openrouter_api_key_oauth 5180)
|
|
fi
|
|
|
|
# 7. Inject environment variables into ~/.zshrc
|
|
log_warn "Setting up environment variables..."
|
|
|
|
ENV_TEMP=$(mktemp)
|
|
trap 'rm -f "${ENV_TEMP}"' EXIT
|
|
cat > "${ENV_TEMP}" << EOF
|
|
|
|
# [spawn:env]
|
|
export OPENROUTER_API_KEY="${OPENROUTER_API_KEY}"
|
|
export OPENAI_API_KEY="${OPENROUTER_API_KEY}"
|
|
export OPENAI_BASE_URL="https://openrouter.ai/api/v1"
|
|
EOF
|
|
|
|
upload_file "${LAMBDA_SERVER_IP}" "${ENV_TEMP}" "/tmp/env_config"
|
|
run_server "${LAMBDA_SERVER_IP}" "cat /tmp/env_config >> ~/.zshrc && rm /tmp/env_config"
|
|
|
|
echo ""
|
|
log_info "Lambda Cloud instance setup completed successfully!"
|
|
log_info "Instance: ${SERVER_NAME} (IP: ${LAMBDA_SERVER_IP})"
|
|
echo ""
|
|
|
|
# 8. Start Codex interactively
|
|
log_warn "Starting Codex..."
|
|
sleep 1
|
|
clear
|
|
interactive_session "${LAMBDA_SERVER_IP}" "source ~/.zshrc && codex"
|