-
Notifications
You must be signed in to change notification settings - Fork 441
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] need to handle the failure during allocate multiple IPs for a single pod, or it will exhausted the whole IP Pool. #4210
Comments
can some expert help fix this issue as I have no authority to do it, |
please attach the err log in the kube-ovn-controller pod about the nat gw pod |
701 I0619 12:07:04.058569 6 ipam.go:60] allocate v4 192.168.1.10, v6 , mac for kube-system/vpc-nat-gw-gw1-vpc-1-0 from subnet ovn-vpc-external-network |
err: please attatch |
the root cause should be identified as above, we need to handle the exception gracefully.
|
per the log above,it looks exist another problem(as you mentioned ), the controller shouldn't allocate the "10.0.1.254" for ovn-default subnet |
where is your 10.0.1.0/24 subnet ??? |
could you help take a deep look at the problem? it should be easy to reproduce with my configuration above. many thanks |
anyway, I will reproduce it and upload all the log file later, thanks |
I have reproduced it with a new VPC named "vpc-3" with related log files |
could you refer to my reply above : #4210 (comment) |
it looks like the problem is obviously, can we help fix it if possible? |
"if you use vpc3 subnet, I think you should use 10.0.3.254." yes, I'm using 10.0.3.254 for vpc3, please refer to vpc3(rather than vpc1) related configuration/debug info in the tar ball thanks |
hi @zhangzujian, it seems IPAM has a problem ? |
Kube-OVN Version
v.1.12.17 and master
Kubernetes Version
v1.27
Operation-system/Kernel Version
"Ubuntu 20.04.6 LTS" / 5.4.0-186-generic
Description
it looks like no handling for the failure of IP allocation during create the VPC GW pod, and the whole External IP Pool get exhausted by this problem.
after done some research, I found out the root cause looks like below:
an vpc-gw pod info:
Steps To Reproduce
create and delete vpc nat gate multiple times
Current Behavior
the external IP CIDR get exhausted by this problem
Expected Behavior
nice handling for the IP allocating/releasing to avoid such a problem.
The text was updated successfully, but these errors were encountered: