-
Notifications
You must be signed in to change notification settings - Fork 811
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Volume attachment limits for p4d.24xlarge are too low? #2301
Comments
Output of
|
Hi @j-vizcaino, thank you for opening this issue and providing great reproduction steps! Let me look into this. We will prioritize a fix in the driver or correct the docs. In the meantime, you can rely on our Additional Node DaemonSets feature to automate overriding the volume attachment limit for Pasting the relevant AWS docs wording below for posterity:
|
/priority important-soon |
/assign |
/kind bug
What happened?
csinode
forp4d.24xlarge
reports 6 allocatable EBS volumes but can support more.According to the AWS docs these instance types should support up to 11 EBS volumes.
As our
p4d.24xlarge
instances include 4 EFA/ENI devices, this brings down the number to 7. Taking into account the root EBS volume, this brings us to 6.BUT, those instances, with EFA, support at least 8 EBS (+1 for root) volumes (see below)
How to reproduce it (as minimally and precisely as possible)?
pd4.24xlarge
instance, with EFA enabledcsinode
resource (or look for theebs-csi-node
pod log line): the allocatable volume count is 6ebs-csi-node
daemonset and force the number of allocatable EBS volumes to 11, adding--volume-attachment-limits=11
Anything else we need to know?:
It's unclear if the issue is related to how the available EBS volumes computation is performed in the ebs-csi code, or if it's an AWS issue, with the EC2 metadata endpoint not reporting numbers correctly, but it's clear that those instances support more EBS attachments than what the driver reports.
Environment
kubectl version
):v1.29.12-eks-2d5f260
v1.38.1-eksbuild.2
(EKS addon)The text was updated successfully, but these errors were encountered: