我无法删除/更新集群。我得到:
I0107 19:54:02.618454 8397 request_logger.go:45] AWS request: autoscaling/DescribeAutoScalingGroups
I0107 19:54:02.812764 8397 request_logger.go:45] AWS request: ec2/DescribeNatGateways
W0107 19:54:03.032646 8397 executor.go:130] error running task "ElasticIP/us-east-1a.my.domain" (9m56s remaining to succeed): error finding AssociatedNatGatewayRouteTable: error listing NatGateway %!q(*string=0xc42169eb08): NatGatewayNotFound: NAT gateway nat-083300682d9a0fa74 was not found
status code: 400, request id: 8408a79d-1f8f-4886-83d9-ae0a26c1cc47
I0107 19:54:03.032738 8397 executor.go:103] Tasks: 98 done / 101 total; 1 can run
I0107 19:54:03.032828 8397 executor.go:178] Executing task "ElasticIP/us-east-1a.my.domain": *awstasks.ElasticIP {"Name":"us-east-1a.my.domain","Lifecycle":"Sync","ID":null,"PublicIP":null,"TagOnSubnet":null,"Tags":{"KubernetesCluster":"my.domain","Name":"us-east-1a.my.domain","kubernetes.io/cluster/my.domain":"owned"},"AssociatedNatGatewayRouteTable":{"Name":"private-us-east-1a.my.domain","Lifecycle":"Sync","ID":"rtb-089bd4ffc062a3b15","VPC":{"Name":"my.domain","Lifecycle":"Sync","ID":"vpc-0b638e55c11fc9021","CIDR":"172.10.0.0/16","EnableDNSHostnames":null,"EnableDNSSupport":true,"Shared":true,"Tags":null},"Shared":false,"Tags":{"KubernetesCluster":"my.domain","Name":"private-us-east-1a.my.domain","kubernetes.io/cluster/my.domain":"owned","kubernetes.io/kops/role":"private-us-east-1a"}}}
I0107 19:54:03.033039 8397 natgateway.go:205] trying to match NatGateway via RouteTable rtb-089bd4ffc062a3b15
I0107 19:54:03.033304 8397 request_logger.go:45] AWS request: ec2/DescribeRouteTables
I0107 19:54:03.741980 8397 request_logger.go:45] AWS request: ec2/DescribeNatGateways
W0107 19:54:03.981744 8397 executor.go:130] error running task "ElasticIP/us-east-1a.my.domain" (9m55s remaining to succeed): error finding AssociatedNatGatewayRouteTable: error listing NatGateway %!q(*string=0xc4217e8da8): NatGatewayNotFound: NAT gateway nat-083300682d9a0fa74 was not found
status code: 400, request id: 3be6843a-38e2-4584-b2cd-b29f6a132d2d
I0107 19:54:03.981881 8397 executor.go:145] No progress made, sleeping before retrying 1 failed task(s)
I0107 19:54:13.982261 8397 executor.go:103] Tasks: 98 done / 101 total; 1 can run
我更改了kubectl版本以对其他集群执行一些任务,然后又恢复到最新状态,我一直在测试新集群的删除,创建,更新,没有任何问题...直到现在,我已经可以使用该集群了。要修改并花钱,请确保我可以删除kops IAM,但是我将其用于同一帐户的其他环境。
至少,有一个文件可以编辑正在查看AWS的kops,以便删除该对象吗?我在config / spec S3文件中找不到。
由于这个原因,我有一个无法使用的已部署群集,请确保我可以拒绝kops权限并删除该群集,以便kops无法重新创建它,但是我也有其他群集。
kops版本:1.10.0版(git-8b52ea6d1)
答案 0 :(得分:0)
我先删除了存储桶,然后手动删除了所有资源。
对于将来的读者,请在导出集群配置的存储桶中启用版本控制。
答案 1 :(得分:0)
几分钟前,我们遇到了同样的问题。我们可以通过搜索指向各自的NatGateway的VPC RouteTable条目来修复它(状态为Blackhole)。删除这些群集之后,我们终于可以删除群集,而没有任何其他问题。
this issue comment向我们指明了正确的方向。
答案 2 :(得分:0)
只需删除集群死亡的主节点。我在测试KOPS时遇到了类似的问题,导致支付了少量款项。当我删除一个子节点时,立即创建了一个新节点,这是可以理解的。所以我删除了主节点,集群死了。