mysql-deployment-756f9d8cdf-8kzdw microk8s kubectl
注意11分钟的年龄。
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal SandboxChanged 11m kubelet Pod sandbox changed, it will be killed and re-created.
Normal Pulled 11m kubelet Container image "mysql:5.7" already present on machine
Normal Created 11m kubelet Created container mysql-container
Normal Started 11m kubelet Started container mysql-container
microk8s get pods -o wide注意到41h,IP地址也在大约11分钟前更改。
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
mysql-deployment-756f9d8cdf-8kzdw 1/1 Running 3 41h 10.1.167.149 john-trx40-designare <none> <none>
mysql-deployment-756f9d8cdf-8kzdw microk8s kubectl日志
报告了一些
2020-12-08T02:10:10.264100Z 32 [Note] Aborted connection 32 to db: 'jjg_script_db' user: 'root' host: '10.1.167.159' (Got an error reading communication packets)
其他豆荚报告dhcp查找失败,然后崩溃并重新创建..。
我觉得IP租约快用完了,但我宁愿看日志,也不愿猜测。这样说是因为sql的年龄在不断增加。同样的sql映像和数据在码头组合上也能维持几个月的运行。频率与交通无关。
sudo microk8s检查会生成大量的日志文件,查看每个人,但可能因为日志数量众多而忽略了关键事件,并且不确定应该在哪里查看。
在哪里查找导致SandboxChanged的原因/触发器的日志?如果我的猜测是正确的,那就是IP租约的问题,我在哪里可以找到microk8s kubernetes的IP分配日志?
我的设置是一个ubuntu18.04主机,很少安装,除了码头,对接-合成,git,可视化演播室代码和microk8s kubernetes。
一切都恢复了,但这段插曲令人难以忍受,不知道该往哪里去看,让我发疯了。
PjoterS要求的额外信息
入口
apiVersion: networking.k8s.io/v1beta1
kind: Ingress
metadata:
name: ingress-myservice-jjg
annotations:
# use the shared ingress-nginx
kubernetes.io/ingress.class: "nginx"
nginx.ingress.kubernetes.io/affinity: "cookie"
nginx.ingress.kubernetes.io/affinity-mode: "balanced" # "persistent"
# added next 2 lines for secured https after I got a certificate
certmanager.k8s.io/cluster-issuer: "letsencrypt-issuer"
nginx.ingress.kubernetes.io/ssl-redirect: "true"
spec:
# added tls lines for secured https after I got a certificate
tls:
- hosts:
- ancient-script.org
- www.ancient-script.org
secretName: ancient-script-org-crt-secret
rules:
- host: ancient-script.org
http:
paths:
- path: /
backend:
serviceName: express-service
servicePort: 3000
- host: www.ancient-script.org
http:
paths:
- path: /
backend:
serviceName: express-service
servicePort: 3000
部署
apiVersion: v1
kind: Service
metadata:
name: mysql-service
spec:
ports:
- protocol: TCP # default is TCP
port: 3306 # incoming port from within kubernetes
targetPort: 3306 # default, port on the pod
selector:
app: mysql-pod
clusterIP: None
---
apiVersion: apps/v1 # for versions before 1.9.0 use apps/v1beta2
kind: Deployment
metadata:
name: mysql-deployment
spec:
selector:
matchLabels:
app: mysql-pod
strategy:
type: Recreate
template:
metadata:
labels:
app: mysql-pod
spec:
containers:
- image: mysql:5.7
name: mysql-container
args:
- --sql-mode=STRICT_TRANS_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION
env:
- name: MYSQL_ROOT_PASSWORD
value: 'todochange'
- name: MYSQL_DATABASE
value: ancient_script_db
ports:
- containerPort: 3306
name: mysql
volumeMounts:
- name: mysql-persistent-storage
mountPath: /var/lib/mysql
volumes:
- name: mysql-persistent-storage
persistentVolumeClaim:
claimName: mysql-pv-claim
在IP更改的同时从MySQL登录
2020-12-10T19:39:43.795268Z 44 [Note] Aborted connection 44 to db: 'ancient_script_db' user: 'root' host: '10.1.167.141' (Got an error reading communication packets)
2020-12-10T19:39:43.796587Z 43 [Note] Aborted connection 43 to db: 'ancient_script_db' user: 'root' host: '10.1.167.141' (Got an error reading communication packets)
2020-12-10T19:39:43.796761Z 41 [Note] Aborted connection 41 to db: 'ancient_script_db' user: 'root' host: '10.1.167.141' (Got an error reading communication packets)
2020-12-10T19:39:43.796831Z 38 [Note] Aborted connection 38 to db: 'ancient_script_db' user: 'root' host: '10.1.167.141' (Got an error reading communication packets)
2020-12-10T19:39:43.796889Z 42 [Note] Aborted connection 42 to db: 'ancient_script_db' user: 'root' host: '10.1.167.141' (Got an error reading communication packets)
来自另一个事件的,它们每天发生两次:
注意来自microk8s的错误消息,当试图从另一个使用MySQL的荚获取日志时,当microk9s最终允许日志时,日志将显示dhcp问题。
john@john-trx40-designare:~/Documents/GitHub/help-me-transcribe$ k describe pod express-deployment-64947b66b9-84vzc
Name: express-deployment-64947b66b9-84vzc
Namespace: default
Priority: 0
Node: john-trx40-designare/99.153.71.9
Start Time: Fri, 11 Dec 2020 12:27:13 -0600
Labels: app=express-pod
pod-template-hash=64947b66b9
Annotations: cni.projectcalico.org/podIP: 10.1.167.153/32
cni.projectcalico.org/podIPs: 10.1.167.153/32
Status: Running
IP: 10.1.167.153
IPs:
IP: 10.1.167.153
Controlled By: ReplicaSet/express-deployment-64947b66b9
Containers:
express:
Container ID: containerd://e9d178313638d3a8985caef57ae6e45f2b37b5ae08032e0eeba01c30a12676ce
Image: localhost:32000/express-server:20201211d
Image ID: localhost:32000/express-server@sha256:95fcc8679727820cff428f657ea7c32811681c2782e3692fbad0041ffcd3d935
Port: <none>
Host Port: <none>
State: Running
Started: Fri, 11 Dec 2020 12:27:15 -0600
Ready: True
Restart Count: 0
Environment:
mySQL_connection_limit: 200
mySQL_host: mysql-service
mySQL_port: 3306
mySQL_user: root
ROOT_PATH_user_files: /user_files
ROOT_PATH_crop_images: /crop_images
ROOT_PATH_bulk_input_AI_transcriptions: /temp/FEB_2020_AI_transcription.txt #is this used?
ROOT_PATH_CURRICULUM: /temp/curriculum/ #is this used?
ROOT_PATH_transcribed_words: /transcription_db/crops_sets/transcribed_words/ #is this used?
ROOT_PATH_train_test: /transcription_db/train_test #is this used?
IMAGE_SERVICE_ADDRESS: image-service
TRANSCRIBE_SERVICE_ADDRESS: transcription-service
LOGIN_COOKIE_NAME: ancient_script_signed_login_token
Mounts:
/crop_images from name-ci (rw)
/user_files from name-uf (rw)
/var/run/secrets/kubernetes.io/serviceaccount from default-token-qnqqd (ro)
Conditions:
Type Status
Initialized True
Ready True
ContainersReady True
PodScheduled True
Volumes:
name-uf:
Type: HostPath (bare host directory volume)
Path: /mnt/disk2/Documents/help_me_transcribe/production/pv/user_files
HostPathType:
name-ci:
Type: HostPath (bare host directory volume)
Path: /mnt/disk2/Documents/help_me_transcribe/production/pv/crop_images
HostPathType:
default-token-qnqqd:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-qnqqd
Optional: false
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 28m default-scheduler Successfully assigned default/express-deployment-64947b66b9-84vzc to john-trx40-designare
Normal Pulling 28m kubelet Pulling image "localhost:32000/express-server:20201211d"
Normal Pulled 28m kubelet Successfully pulled image "localhost:32000/express-server:20201211d" in 446.977029ms
Normal Created 28m kubelet Created container express
Normal Started 28m kubelet Started container express
john@john-trx40-designare:~/Documents/GitHub/help-me-transcribe$ k logs express-deployment-64947b66b9-84vzc
Error from server (NotFound): the server could not find the requested resource ( pods/log express-deployment-64947b66b9-84vzc)
john@john-trx40-designare:~/Documents/GitHub/help-me-transcribe$ k logs express-deployment-64947b66b9-84vzc
process.env.mySQL_host = mysql-service
process.env.mySQL_port = 3306
process.env.mySQL_user = root
initializing socketApi.js
FOCUS
TODO, need to process all user files not just john_grabner
Scanning all files to make sure present in database ... this will take some time
Listening on port 3000 when launched native or from docker networks, maybe remapped to host in docker-compose
Unhandled Rejection at: Promise Promise {
<rejected> { Error: getaddrinfo ENOTFOUND mysql-service mysql-service:3306
at errnoException (dns.js:55:10)
at GetAddrInfoReqWrap.onlookup [as oncomplete] (dns.js:97:26)
--------------------
at Protocol._enqueue (/app/src/help-me-transcribe-server/node_modules/mysql/lib/protocol/Protocol.js:145:48)
at Protocol.handshake (/app/src/help-me-transcribe-server/node_modules/mysql/lib/protocol/Protocol.js:52:23)
at PoolConnection.connect (/app/src/help-me-transcribe-server/node_modules/mysql/lib/Connection.js:130:18)
at Pool.getConnection (/app/src/help-me-transcribe-server/node_modules/mysql/lib/Pool.js:48:16)
at Promise (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:47:22)
at new Promise (<anonymous>)
at getConnection (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:42:16)
at query (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:77:12)
at query (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:74:16)
at Object.query_one_or_null (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:101:12)
code: 'ENOTFOUND',
errno: 'ENOTFOUND',
syscall: 'getaddrinfo',
hostname: 'mysql-service',
host: 'mysql-service',
port: 3306,
fatal: true } } reason: { Error: getaddrinfo ENOTFOUND mysql-service mysql-service:3306
at errnoException (dns.js:55:10)
at GetAddrInfoReqWrap.onlookup [as oncomplete] (dns.js:97:26)
--------------------
at Protocol._enqueue (/app/src/help-me-transcribe-server/node_modules/mysql/lib/protocol/Protocol.js:145:48)
at Protocol.handshake (/app/src/help-me-transcribe-server/node_modules/mysql/lib/protocol/Protocol.js:52:23)
at PoolConnection.connect (/app/src/help-me-transcribe-server/node_modules/mysql/lib/Connection.js:130:18)
at Pool.getConnection (/app/src/help-me-transcribe-server/node_modules/mysql/lib/Pool.js:48:16)
at Promise (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:47:22)
at new Promise (<anonymous>)
at getConnection (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:42:16)
at query (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:77:12)
at query (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:74:16)
at Object.query_one_or_null (/app/src/help-me-transcribe-server/help-me-data-access/sql.js:101:12)
code: 'ENOTFOUND',
errno: 'ENOTFOUND',
syscall: 'getaddrinfo',
hostname: 'mysql-service',
host: 'mysql-service',
port: 3306,
fatal: true }
as per request, more info on mysql config
# microk8s kubectl get pv --sort-by=.spec.capacity.storage --namespace=production
apiVersion: v1
kind: PersistentVolume
metadata:
# namespace: production
name: mysql-pv-volume
labels:
type: local
spec:
storageClassName: manual
capacity:
storage: 100Gi
accessModes:
- ReadWriteOnce
hostPath:
path: "/Disk2/Documents/help_me_transcribe/production/pv/mysql-pv-volume"
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
# namespace: production
name: mysql-pv-claim
spec:
storageClassName: manual
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 100Gi
# microk8s kubectl get services --namespace=production # List all services in the namespace
apiVersion: v1
kind: Service
metadata:
# namespace: production
name: mysql-service
spec:
#type: xxxxxxx # default, ClusterIP: Exposes the Service on a cluster-internal IP
# NodePort: Exposes the Service on each Node's IP at a static port (the NodePort, 30000-32767)
ports:
- protocol: TCP # default is TCP
port: 3306 # incoming port from within kubernetes
targetPort: 3306 # default, port on the pod
#nodePort: 33306
selector:
app: mysql-pod
clusterIP: None
---
# microk8s kubectl get pods --namespace=production
# microk8s kubectl get pods -o wide --namespace=production # List all pods in the current namespace, with more details
# notice the IP address 10.1.167.xxx ... use this for "MySQL Workbench"
# microk8s kubectl describe pods my-pod --namespace=production
apiVersion: apps/v1 # for versions before 1.9.0 use apps/v1beta2
kind: Deployment
metadata:
# namespace: production
name: mysql-deployment
spec:
selector:
matchLabels:
app: mysql-pod
strategy:
type: Recreate
template:
metadata:
labels:
app: mysql-pod
spec:
containers:
- image: mysql:5.7
name: mysql-container
args:
- --sql-mode=STRICT_TRANS_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION
env:
- name: MYSQL_DATABASE
value: ancient_script_db
ports:
- containerPort: 3306
name: mysql
volumeMounts:
- name: mysql-persistent-storage
mountPath: /var/lib/mysql
volumes:
- name: mysql-persistent-storage
# hostPath:
# path: "/Disk2/Documents/help_me_transcribe/production/pv/mysql-pv-volume"
persistentVolumeClaim:
claimName: mysql-pv-claim
2021年1月5日
microk8s中肯定有东西在吃"microk8s kubectl事件“,因为这个规范报告了”默认名称空间中找不到任何资源“。幸运的是,刚刚捕获了一个看起来像是taintManager在做什么的事件。
microk8s kubectl get events
LAST SEEN TYPE REASON OBJECT MESSAGE
49m Normal Starting node/john-trx40-designare Starting kube-proxy.
49m Normal Starting node/john-trx40-designare Starting kubelet.
49m Warning InvalidDiskCapacity node/john-trx40-designare invalid capacity 0 on image filesystem
49m Normal NodeHasSufficientMemory node/john-trx40-designare Node john-trx40-designare status is now: NodeHasSufficientMemory
49m Normal NodeHasNoDiskPressure node/john-trx40-designare Node john-trx40-designare status is now: NodeHasNoDiskPressure
49m Normal NodeHasSufficientPID node/john-trx40-designare Node john-trx40-designare status is now: NodeHasSufficientPID
49m Normal NodeNotReady node/john-trx40-designare Node john-trx40-designare status is now: NodeNotReady
49m Normal TaintManagerEviction pod/image-deployment-78c4c9fd7f-5vllb Cancelling deletion of Pod default/image-deployment-78c4c9fd7f-5vllb
49m Normal TaintManagerEviction pod/mysql-deployment-756f9d8cdf-lbfrg Cancelling deletion of Pod default/mysql-deployment-756f9d8cdf-lbfrg
49m Normal TaintManagerEviction pod/express-deployment-6dbb578fbb-dmjqs Cancelling deletion of Pod default/express-deployment-6dbb578fbb-dmjqs
49m Normal TaintManagerEviction pod/transcription-deployment-84fddcdff8-7sd9d Cancelling deletion of Pod default/transcription-deployment-84fddcdff8-7sd9d
47m Normal NodeAllocatableEnforced node/john-trx40-designare Updated Node Allocatable limit across pods
47m Normal NodeReady node/john-trx40-designare Node john-trx40-designare status is now: NodeReady
47m Normal SandboxChanged pod/transcription-deployment-84fddcdff8-7sd9d Pod sandbox changed, it will be killed and re-created.
47m Normal SandboxChanged pod/image-deployment-78c4c9fd7f-5vllb Pod sandbox changed, it will be killed and re-created.
47m Normal Pulled pod/transcription-deployment-84fddcdff8-7sd9d Container image "localhost:32000/py_transcribe_service:20201226d" already present on machine
47m Normal Created pod/transcription-deployment-84fddcdff8-7sd9d Created container transcription-container
47m Normal Started pod/transcription-deployment-84fddcdff8-7sd9d Started container transcription-container
47m Normal SandboxChanged pod/express-deployment-6dbb578fbb-dmjqs Pod sandbox changed, it will be killed and re-created.
47m Normal SandboxChanged pod/mysql-deployment-756f9d8cdf-lbfrg Pod sandbox changed, it will be killed and re-created.
47m Normal CREATE ingress/ingress-myservice-jjg Ingress default/ingress-myservice-jjg
47m Normal UPDATE ingress/ingress-myservice-jjg Ingress default/ingress-myservice-jjg
47m Normal Pulled pod/image-deployment-78c4c9fd7f-5vllb Container image "localhost:32000/py_image_service:2020126a" already present on machine
47m Normal Created pod/image-deployment-78c4c9fd7f-5vllb Created container image-container
47m Normal Pulled pod/mysql-deployment-756f9d8cdf-lbfrg Container image "mysql:5.7" already present on machine
47m Normal Created pod/mysql-deployment-756f9d8cdf-lbfrg Created container mysql-container
47m Normal Started pod/image-deployment-78c4c9fd7f-5vllb Started container image-container
47m Normal Started pod/mysql-deployment-756f9d8cdf-lbfrg Started container mysql-container
47m Normal Pulled pod/express-deployment-6dbb578fbb-dmjqs Container image "localhost:32000/express-server:20201226b" already present on machine
47m Normal Created pod/express-deployment-6dbb578fbb-dmjqs Created container express
47m Normal Started pod/express-deployment-6dbb578fbb-dmjqs Started container express
下一次更新:
在重新启动节点并获得新的IP地址之前,inspection-report/snap.microk8s.daemon-controller-manager似乎包含了一系列问题。
有人知道这意味着什么吗?
Jan 05 09:01:27 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:27.587942 1704 cronjob_controller.go:123] Failed to extract job list: Get "https://127.0.0.1:16443/apis/batch/v1/jobs?limit=500": dial tcp 127.0.0.1:16443: connect: connection refused
Jan 05 09:01:27 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:27.854454 1704 leaderelection.go:325] error retrieving resource lock kube-system/kube-controller-manager: Get "https://127.0.0.1:16443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=10s": dial tcp 127.0.0.1:16443: connect: connection refused
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.286465 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.286510 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.286594 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.286889 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.286909 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.287010 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.288178 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.289698 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.289797 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.289898 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.293707 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.294074 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.294088 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.294092 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.294191 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.294221 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.294322 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:30 john-trx40-designare microk8s.daemon-controller-manager[1704]: E0105 09:01:30.294869 1704 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server is currently unable to handle the request
Jan 05 09:01:35 john-trx40-designare microk8s.daemon-controller-manager[1704]: I0105 09:01:35.877899 1704 request.go:655] Throttling request took 1.048035134s, request: GET:https://127.0.0.1:16443/apis/discovery.k8s.io/v1beta1?timeout=32s
Jan 05 09:01:37 john-trx40-designare microk8s.daemon-controller-manager[1704]: W0105 09:01:37.179011 1704 garbagecollector.go:703] failed to discover some groups: map[metrics.k8s.io/v1beta1:the server is currently unable to handle the request]
Jan 05 09:01:38 john-trx40-designare microk8s.daemon-controller-manager[1704]: I0105 09:01:38.242607 1704 node_lifecycle_controller.go:1195] Controller detected that all Nodes are not-Ready. Entering master disruption mode.
Jan 05 09:01:38 john-trx40-designare microk8s.daemon-controller-manager[1704]: I0105 09:01:38.255936 1704 event.go:291] "Event occurred" object="kube-system/dashboard-metrics-scraper-6c4568dc68-pzj4s" kind="Pod" apiVersion="" type="Normal" reason="TaintManagerEviction" message="Cancelling deletion of Pod kube-system/dashboard-metrics-scraper-6c4568dc68-pzj4s"
Jan 05 09:01:38 john-trx40-designare microk8s.daemon-controller-manager[1704]: I0105 09:01:38.255960 1704 event.go:291] "Event occurred" object="kube-system/coredns-86f78bb79c-f5plg" kind="Pod" apiVersion="" type="Normal" reason="TaintManagerEviction" message="Cancelling deletion of Pod kube-system/coredns-86f78bb79c-f5plg"
Jan 05 09:01:38 john-trx40-designare microk8s.daemon-controller-manager[1704]: I0105 09:01:38.255974 1704 event.go:291] "Event occurred" object="kube-system/calico-kube-controllers-847c8c99d-hkgn6" kind="Pod" apiVersion="" type="Normal" reason="TaintManagerEviction"
发布于 2021-12-09 11:54:14
https://github.com/ubuntu/microk8s/issues/2241提供了一个解决方案。
/var/snap/microk8s/2695/args$ nano 加法线
--bind-address=0.0.0.0
https://stackoverflow.com/questions/65207734
复制相似问题