weixin_39943799
weixin_39943799
2020-12-01 17:26

Test 2653 (NFS) fails on ipv6 lane, VM stuck on scheduling

/kind bug Please see suggested possible solution at the end.

What happened: Using PR #3221 (ipv5 migration fix) , and PR #3171 (both arent merged yet right now)

testing ipv6 lane , test_id 2653 VM fails to schedule

describing the virt-launcher pod gives this event


Mounting command: systemd-run
Mounting arguments: --description=Kubernetes transient mount for /var/lib/kubelet/pods/7dccb5f8-a4ae-4227-bade-4ed1343df2f2/volumes/kubernetes.io~nfs/test-nfssfsn9f6rj5r44svhp6ljnr2nhr8q5xx2vdwwxsbtx4gwzcj8 --scope -- mount -t nfs fd00:10:244:0:1::e:/ /var/lib/kubelet/pods/7dccb5f8-a4ae-4227-bade-4ed1343df2f2/volumes/kubernetes.io~nfs/test-nfssfsn9f6rj5r44svhp6ljnr2nhr8q5xx2vdwwxsbtx4gwzcj8
Output: Running scope as unit: run-rf4a661c741b741c8a10cb6b9cd4524f8.scope
mount: /var/lib/kubelet/pods/7dccb5f8-a4ae-4227-bade-4ed1343df2f2/volumes/kubernetes.io~nfs/test-nfssfsn9f6rj5r44svhp6ljnr2nhr8q5xx2vdwwxsbtx4gwzcj8: bad option; for several filesystems (e.g. nfs, cifs) you might need a /sbin/mount.<type> helper program.
  Warning  FailedMount  17s  kubelet, kind-1.17.0-worker  MountVolume.SetUp failed for volume "test-nfssfsn9f6rj5r44svhp6ljnr2nhr8q5xx2vdwwxsbtx4gwzcj8" : mount failed: exit status 32
</type>

test fails because VM doesnt get Running state


• Failure [229.328 seconds]
[rfe_id:393][crit:high][vendor:cnv-qe.com][level:system] VM Live Migration
/root/go/src/kubevirt.io/kubevirt/tests/migration_test.go:62
  Starting a VirtualMachineInstance 
  /root/go/src/kubevirt.io/kubevirt/tests/migration_test.go:347
    with an Fedora shared NFS PVC, cloud init and service account
    /root/go/src/kubevirt.io/kubevirt/tests/migration_test.go:847
      [test_id:2653]  should be migrated successfully, using guest agent on VM [It]
      /root/go/src/kubevirt.io/kubevirt/tests/migration_test.go:876

      Timed out after 180.000s.
      Timed out waiting for VMI testvmi5p754bzlhss9v696lsh2b7xjrd99rl8xdpx8nkz9d52zxxlz to enter [Running] phase(s)
      Expected
          <v1.virtualmachineinstancephase>: Scheduling
      to be an element of
          : [["Running"]]

      /root/go/src/kubevirt.io/kubevirt/tests/utils.go:2507
------------------------------
</v1.virtualmachineinstancephase>

What you expected to happen: test should pass, or at least the VM should be running

How to reproduce it (as minimally and precisely as possible): FUNC_TEST_ARGS=" -ginkgo.v --ginkgo.focus=test_id:2653" make functest make sure you have the PRs mentioned above (assuming 3171 is needed)

Note: i had this problem on real machine once solution was to install nfs-utils thanks to Yaniv L

该提问来源于开源项目:kubevirt/kubevirt

  • 点赞
  • 写回答
  • 关注问题
  • 收藏
  • 复制链接分享
  • 邀请回答

6条回答

  • weixin_39573822 weixin_39573822 5月前

    Is the test working on a kind setup with IPv4?

    点赞 评论 复制链接分享
  • weixin_39943799 weixin_39943799 5月前

    Other kinds we have is sriov, where migration isnt supported and just after this issue was opened, Quique added new kind provider which is IPv4 but i believe the root cause is written above and the solution need to install nfs-utils so i doubt it would work on ipv4.

    Also kind must have unique UUID which we fixed only on ipv6 atm, and need to fix for the other kinds as well

    点赞 评论 复制链接分享
  • weixin_39890633 weixin_39890633 5月前

    Issues go stale after 90d of inactivity. Mark the issue as fresh with /remove-lifecycle stale. Stale issues rot after an additional 30d of inactivity and eventually close.

    If this issue is safe to close now please do so with /close.

    /lifecycle stale

    点赞 评论 复制链接分享
  • weixin_39890633 weixin_39890633 5月前

    Stale issues rot after 30d of inactivity. Mark the issue as fresh with /remove-lifecycle rotten. Rotten issues close after an additional 30d of inactivity.

    If this issue is safe to close now please do so with /close.

    /lifecycle rotten

    点赞 评论 复制链接分享
  • weixin_39890633 weixin_39890633 5月前

    Rotten issues close after 30d of inactivity. Reopen the issue with /reopen. Mark the issue as fresh with /remove-lifecycle rotten.

    /close

    点赞 评论 复制链接分享
  • weixin_39890633 weixin_39890633 5月前

    -bot: Closing this issue.

    In response to [this](https://github.com/kubevirt/kubevirt/issues/3322#issuecomment-710767558): >Rotten issues close after 30d of inactivity. >Reopen the issue with `/reopen`. >Mark the issue as fresh with `/remove-lifecycle rotten`. > >/close Instructions for interacting with me using PR comments are available [here](https://git.k8s.io/community/contributors/guide/pull-requests.md). If you have questions or suggestions related to my behavior, please file an issue against the [kubernetes/test-infra](https://github.com/kubernetes/test-infra/issues/new?title=Prow%20issue:) repository.
    点赞 评论 复制链接分享

相关推荐