#!/bin/bash set -euo pipefail # ========================== 全局配置与工具函数 ========================== # 颜色定义 RED='\033[0;31m' GREEN='\033[0;32m' YELLOW='\033[1;33m' BLUE='\033[0;34m' NC='\033[0m' # 打印函数 info() { echo -e "${YELLOW}[INFO]${NC} $1"; } success() { echo -e "${GREEN}[SUCCESS]${NC} $1"; } error() { echo -e "${RED}[ERROR]${NC} $1"; exit 1; } step() { echo -e "\n${BLUE}===== $1 =====${NC}"; } # 固定路径(可根据实际调整) KUBEADM_CONF="/opt/k8s-install-conf/kubeadm-conf.yaml" # Master1的kubeadm配置文件 INIT_RESULT_FILE="/opt/k8s-install-conf/kubeadm-init-result.txt" # 初始化结果保存文件 # 关键修改1:Calico v3.25官方YAML地址(替换原v3.26) CALICO_YAML_OFFICIAL="https://docs.projectcalico.org/v3.25/manifests/calico.yaml" # 本地临时Calico YAML路径(用于镜像替换) LOCAL_CALICO_YAML="/tmp/calico-v3.25-modified.yaml" # 关键修改2:华为云SWR镜像前缀(替换原docker.io/calico) SWR_CALICO_PREFIX="swr.cn-north-4.myhuaweicloud.com/ddn-k8s/docker.io/calico" REMOTE_USER="root" # 远程节点登录用户 # ========================== 参数解析(核心:指定Master/Node IP) ========================== # 初始化参数 MASTER_IPS=() # 3个Master节点IP(顺序:Master1, Master2, Master3) NODE_IPS=() # Node节点IP列表(支持多个,用逗号分隔) parse_args() { while [[ $# -gt 0 ]]; do case "$1" in --master-ips) # 解析3个Master IP(格式:"192.168.61.10,192.168.61.11,192.168.61.12") IFS=',' read -ra IPS <<< "$2" if [[ ${#IPS[@]} -ne 3 ]]; then error "--master-ips必须包含3个IP(逗号分隔),当前提供${#IPS[@]}个" fi MASTER_IPS=("${IPS[@]}") shift 2 ;; --node-ips) # 解析Node IP列表(格式:"192.168.61.20,192.168.61.21") IFS=',' read -ra IPS <<< "$2" NODE_IPS=("${IPS[@]}") shift 2 ;; --help) echo "用法:$0 --master-ips [--node-ips ]" echo "功能:Master1初始化→远程Master/Node加入→Calico(v3.25+华为云SWR)部署" echo "参数说明:" echo " --master-ips 必选,3个Master节点IP(顺序:Master1(初始化节点), Master2, Master3)" echo " --node-ips 可选,Node节点IP列表(逗号分隔,如不指定则仅部署Master集群)" echo "示例:" echo " $0 --master-ips 192.168.61.10,192.168.61.11,192.168.61.12 --node-ips 192.168.61.20,192.168.61.21" exit 0 ;; *) error "未知参数:$1(执行$0 --help查看用法)" ;; esac done # 校验必选参数 if [[ ${#MASTER_IPS[@]} -eq 0 ]]; then error "缺少必选参数--master-ips(需指定3个Master节点IP)" fi # 提取Master1 IP(初始化节点) export MASTER1_IP="${MASTER_IPS[0]}" export MASTER2_IP="${MASTER_IPS[1]}" export MASTER3_IP="${MASTER_IPS[2]}" info "已识别节点:" info "Master1(初始化):$MASTER1_IP" info "Master2:$MASTER2_IP" info "Master3:$MASTER3_IP" [[ ${#NODE_IPS[@]} -gt 0 ]] && info "Node节点:${NODE_IPS[*]}" || info "未指定Node节点,仅部署Master集群" } # ========================== 步骤1:Master1初始化(核心) ========================== init_master1() { step "Step 1:Master1节点初始化($MASTER1_IP)" # 检查Master1的kubeadm配置文件是否存在 if ! ssh "$REMOTE_USER@$MASTER1_IP" "test -f $KUBEADM_CONF"; then error "Master1的kubeadm配置文件不存在:$MASTER1_IP:$KUBEADM_CONF(请先执行配置分发脚本)" fi # 执行kubeadm init(带证书上传,支持其他Master加入) info "执行初始化命令:kubeadm init --upload-certs --config $KUBEADM_CONF" info "初始化结果将保存到:$INIT_RESULT_FILE" # 远程执行初始化,同时保存输出到本地文件 ssh "$REMOTE_USER@$MASTER1_IP" "kubeadm init --upload-certs --config $KUBEADM_CONF 2>&1" | tee "$INIT_RESULT_FILE" local init_exit_code=${PIPESTATUS[0]} # 获取远程命令退出码 if [[ $init_exit_code -ne 0 ]]; then error "Master1初始化失败!查看详情:cat $INIT_RESULT_FILE" fi success "Master1初始化完成,结果已保存到:$INIT_RESULT_FILE" info "初始化输出预览(前20行):" head -n 20 "$INIT_RESULT_FILE" } # ========================== 步骤2:Master1配置kubeconfig ========================== config_master1_kubeconfig() { step "Step 2:配置Master1的kubeconfig" # 远程执行kubeconfig配置命令 local kube_cmds=( "mkdir -p \$HOME/.kube" "cp -i /etc/kubernetes/admin.conf \$HOME/.kube/config" "chown \$(id -u):\$(id -g) \$HOME/.kube/config" "echo 'export KUBECONFIG=/etc/kubernetes/admin.conf' >> ~/.bash_profile" "source ~/.bash_profile" ) for cmd in "${kube_cmds[@]}"; do info "在Master1执行:$cmd" ssh "$REMOTE_USER@$MASTER1_IP" "$cmd" || error "执行命令失败:$cmd" done # 验证kubectl是否可用 info "验证kubectl:在Master1执行 kubectl get nodes" ssh "$REMOTE_USER@$MASTER1_IP" "kubectl get nodes" || error "kubectl配置失败,无法获取节点信息" success "Master1 kubeconfig配置完成" } # ========================== 步骤3:提取join命令(从初始化结果中) ========================== extract_join_commands() { step "Step 3:从初始化结果中提取join命令" # 检查结果文件是否存在 [[ -f "$INIT_RESULT_FILE" ]] || error "初始化结果文件不存在:$INIT_RESULT_FILE" # 1. 提取Master节点join命令(含--control-plane和--certificate-key) info "提取Master节点join命令" local master_join_cmd master_join_cmd=$(grep -A 3 "You can now join any number of control-plane nodes by running the following command on each as root" "$INIT_RESULT_FILE" | grep -v "You can now" | tr -d '\n' | sed 's/ */ /g') [[ -z "$master_join_cmd" ]] && error "未从初始化结果中提取到Master join命令" # 补充--config参数(其他Master也需用自己的配置文件) master_join_cmd="$master_join_cmd --config $KUBEADM_CONF" export MASTER_JOIN_CMD="$master_join_cmd" success "Master join命令提取完成:" echo -e "${YELLOW}${MASTER_JOIN_CMD}${NC}" # 2. 提取Node节点join命令 info "提取Node节点join命令" local node_join_cmd node_join_cmd=$(grep -A 2 "Then you can join any number of worker nodes by running the following on each as root" "$INIT_RESULT_FILE" | grep -v "Then you can" | tr -d '\n' | sed 's/ */ /g') [[ -z "$node_join_cmd" ]] && error "未从初始化结果中提取到Node join命令" export NODE_JOIN_CMD="$node_join_cmd" success "Node join命令提取完成:" echo -e "${YELLOW}${NODE_JOIN_CMD}${NC}" } # ========================== 步骤4:远程其他Master节点加入 ========================== join_other_masters() { step "Step 4:远程Master2($MASTER2_IP)和Master3($MASTER3_IP)加入集群" # 定义要加入的Master节点列表(排除Master1) local other_masters=("$MASTER2_IP" "$MASTER3_IP") for master_ip in "${other_masters[@]}"; do info "处理Master节点:$master_ip" # 检查该Master的kubeadm配置文件是否存在 if ! ssh "$REMOTE_USER@$master_ip" "test -f $KUBEADM_CONF"; then error "Master $master_ip的kubeconfig配置文件不存在:$KUBEADM_CONF" fi # 远程执行join命令 info "在$master_ip执行join命令:$MASTER_JOIN_CMD" ssh "$REMOTE_USER@$master_ip" "$MASTER_JOIN_CMD" || error "Master $master_ip加入失败" # 配置该Master的kubeconfig(否则无法本地执行kubectl) info "配置$master_ip的kubeconfig" ssh "$REMOTE_USER@$master_ip" "mkdir -p \$HOME/.kube && cp -i /etc/kubernetes/admin.conf \$HOME/.kube/config && chown \$(id -u):\$(id -g) \$HOME/.kube/config" success "Master $master_ip加入集群并配置完成" done # 验证所有Master节点状态 info "验证所有Master节点状态(从Master1执行)" ssh "$REMOTE_USER@$MASTER1_IP" "kubectl get nodes | grep master" || error "Master节点状态验证失败" success "所有Master节点加入完成" } # ========================== 步骤5:远程Node节点加入(可选) ========================== join_nodes() { if [[ ${#NODE_IPS[@]} -eq 0 ]]; then info "未指定Node节点,跳过Node加入步骤" return fi step "Step 5:远程Node节点加入集群(共${#NODE_IPS[@]}个)" for node_ip in "${NODE_IPS[@]}"; do info "处理Node节点:$node_ip" # 检查Node节点是否已安装基础组件(kubeadm、kubelet、containerd) if ! ssh "$REMOTE_USER@$node_ip" "command -v kubeadm &>/dev/null && command -v kubelet &>/dev/null && systemctl is-active --quiet containerd"; then error "Node $node_ip未安装基础组件(需先安装kubeadm、kubelet、containerd)" fi # 远程执行Node join命令 info "在$node_ip执行join命令:$NODE_JOIN_CMD" ssh "$REMOTE_USER@$node_ip" "$NODE_JOIN_CMD" || error "Node $node_ip加入失败" success "Node $node_ip加入集群完成" done # 验证所有节点状态 info "验证所有节点状态(从Master1执行)" ssh "$REMOTE_USER@$MASTER1_IP" "kubectl get nodes" || error "节点状态验证失败" success "所有Node节点加入完成" } # ========================== 步骤6:部署Calico网络插件(v3.25+华为云SWR) ========================== deploy_calico() { step "Step 6:部署Calico v3.25(华为云SWR镜像)" # 子步骤1:下载Calico v3.25官方YAML到本地 info "下载Calico v3.25官方YAML:$CALICO_YAML_OFFICIAL" if command -v wget &>/dev/null; then wget -q -O "$LOCAL_CALICO_YAML" "$CALICO_YAML_OFFICIAL" || error "wget下载Calico YAML失败" elif command -v curl &>/dev/null; then curl -s -o "$LOCAL_CALICO_YAML" "$CALICO_YAML_OFFICIAL" || error "curl下载Calico YAML失败" else error "请先安装wget或curl" fi [[ -f "$LOCAL_CALICO_YAML" ]] || error "Calico YAML下载失败,文件不存在" success "Calico YAML下载完成(本地路径:$LOCAL_CALICO_YAML)" # 子步骤2:替换YAML中的所有镜像地址为华为云SWR(核心修改) info "替换镜像地址为华为云SWR:$SWR_CALICO_PREFIX" # 替换3个核心镜像(cni、node、kube-controllers,均为v3.25.0版本) sed -i "s#docker.io/calico/cni:v3.25.0#${SWR_CALICO_PREFIX}/cni:v3.25.0#g" "$LOCAL_CALICO_YAML" sed -i "s#docker.io/calico/node:v3.25.0#${SWR_CALICO_PREFIX}/node:v3.25.0#g" "$LOCAL_CALICO_YAML" sed -i "s#docker.io/calico/kube-controllers:v3.25.0#${SWR_CALICO_PREFIX}/kube-controllers:v3.25.0#g" "$LOCAL_CALICO_YAML" # 验证替换结果 info "验证镜像替换结果(查看前10行含镜像的内容)" grep -n "image:" "$LOCAL_CALICO_YAML" | head -n 10 || error "Calico YAML中未找到镜像配置,替换失败" success "Calico镜像地址替换完成" # 子步骤3:将修改后的YAML上传到Master1节点 local remote_calico_yaml="/opt/k8s-install-conf/calico-v3.25-swr.yaml" info "上传修改后的YAML到Master1:$remote_calico_yaml" scp -o StrictHostKeyChecking=no "$LOCAL_CALICO_YAML" "$REMOTE_USER@$MASTER1_IP:$remote_calico_yaml" || error "YAML上传失败" success "YAML上传完成" # 子步骤4:在Master1执行Calico部署 info "执行部署命令:kubectl apply -f $remote_calico_yaml" ssh "$REMOTE_USER@$MASTER1_IP" "kubectl apply -f $remote_calico_yaml" || error "Calico部署失败" # 子步骤5:等待Calico Pod启动(最多等待5分钟) info "等待Calico Pod启动(最多5分钟)..." local wait_time=0 while true; do # 检查calico-system命名空间下所有Pod是否Running local pod_status pod_status=$(ssh "$REMOTE_USER@$MASTER1_IP" "kubectl get pods -n calico-system -o jsonpath='{.items[*].status.phase}' 2>/dev/null") if [[ "$pod_status" =~ ^(Running\ )*Running$ && -n "$pod_status" ]]; then break fi if [[ $wait_time -ge 300 ]]; then # 5分钟=300秒 error "Calico Pod启动超时(5分钟),请手动检查:kubectl get pods -n calico-system" fi sleep 10 wait_time=$((wait_time + 10)) info "已等待${wait_time}秒,Calico Pod状态:$pod_status" done # 子步骤6:验证Calico状态 info "验证Calico状态(从Master1执行)" ssh "$REMOTE_USER@$MASTER1_IP" "kubectl get pods -n calico-system" || error "Calico状态验证失败" # 额外验证网络插件是否就绪 ssh "$REMOTE_USER@$MASTER1_IP" "kubectl get nodes -o wide | grep -E 'STATUS|Ready'" || error "节点网络未就绪" success "Calico v3.25(华为云SWR)部署完成,集群网络已就绪" # 子步骤7:清理本地临时YAML文件 info "清理本地临时文件:$LOCAL_CALICO_YAML" rm -f "$LOCAL_CALICO_YAML" || info "本地文件清理失败,可手动删除" } # ========================== 主流程:串联所有步骤 ========================== main() { # 1. 解析参数 parse_args "$@" # 2. 前置检查:确保SSH免密登录正常 step "前置检查:SSH免密登录验证" local all_ips=("${MASTER_IPS[@]}" "${NODE_IPS[@]}") for ip in "${all_ips[@]}"; do info "验证SSH免密登录:$REMOTE_USER@$ip" if ! ssh -o StrictHostKeyChecking=no "$REMOTE_USER@$ip" "echo 'SSH连通正常' >/dev/null"; then error "SSH免密登录失败:$ip(请先执行 ssh-copy-id $REMOTE_USER@$ip 配置免密)" fi done success "所有节点SSH免密登录验证通过" # 3. 执行核心步骤 init_master1 # Master1初始化 config_master1_kubeconfig # Master1配置kubeconfig extract_join_commands # 提取join命令 join_other_masters # 其他Master加入 join_nodes # Node加入(可选) deploy_calico # 部署Calico(v3.25+SWR) # 4. 集群部署完成总结 step "集群部署完成总结" echo -e "${GREEN}========================================" echo -e "K8s集群部署全流程完成!" echo -e "关键信息:" echo -e " 1. Calico版本:v3.25(镜像源:华为云SWR)" echo -e " 2. 集群节点状态:ssh $REMOTE_USER@$MASTER1_IP 'kubectl get nodes'" echo -e " 3. Calico状态:ssh $REMOTE_USER@$MASTER1_IP 'kubectl get pods -n calico-system'" echo -e " 4. 初始化结果:$INIT_RESULT_FILE" echo -e "========================================" } # 启动主流程 main "$@" chmod +x k8s-cluster-deploy.sh # 部署3个Master + 2个Node(替换为你的IP) #bash k8s-cluster-deploy.sh --master-ips 192.168.61.10,192.168.61.11,192.168.61.12 --node-ips 192.168.61.20,192.168.61.21