[dpdk-dev] [Dpdk-ovs] problem in binding interfaces of virtio-pci on the VM

Bruce Richardson bruce.richardson at intel.com
Fri Feb 27 11:06:32 CET 2015


On Thu, Feb 26, 2015 at 10:46:58PM +0530, Srinivasreddy R wrote:
> hi Bruce ,
> Thank you for your response .
> I am accessing my VM via  " vncviewer " . so ssh doesn't come into picture .
> Is there any way to find the root cause of my problem . does dpdk stores
> any logs while binding interfaces to igb_uio.
> i have seen my /var/log/messages . but could not find any clue.
> 
> the movement i gave the below command my vm got struck and not responding
> untill i forcefully kill the qemu and relaunch .
> ./dpdk_nic_bind.py --bind=igb_uio 00:04.0 00:05.0
> 

Does VNC not also connect using a network port? What is the output of 
./dpdk_nic_bind.py --status before you run this command?

/Bruce

> 
> 
> thanks,
> srinivas.
> 
> 
> 
> On Thu, Feb 26, 2015 at 10:30 PM, Bruce Richardson <
> bruce.richardson at intel.com> wrote:
> 
> > On Thu, Feb 26, 2015 at 10:08:59PM +0530, Srinivasreddy R wrote:
> > > hi Mike,
> > > Thanks for our detailed explanation of your example . usually i do
> > similar
> > > to u and i am aware of working with dpdk applications .
> > > my problem is :
> > > 1. i have written a code for  host to guest communication .[taken form
> > > usvhost which is developed in ovdk vswitch] .
> > > 2. i launched VM with two  interfaces .
> > > 3. i am able to send and receive traffic to and from guest to host on
> > these
> > > interfaces .
> > > 4. when i  try to bind these interfaces to igb_uio  to run dpdk
> > application
> > > . i am not able to access my instance . it got struck and not responding
> > .
> > > i need to hard reboot the vm.
> >
> > Are you sure you are not trying to access the vm via one of the interfaces
> > now bount to igb_uio? If you bind the interface you use for ssh to igb_uio,
> > you won't be able to ssh to that vm any more.
> >
> > /Bruce
> >
> > >
> > > My Question is  :
> > > surely i might done something wrong in code . as my VM is not able to
> > > access any more when i try to bind interfaces to igb_uio  . not able to
> > > debug the issue .
> > > someone please help me in figuring the issue . i dont find anything in
> > > /var/log/messages after relaunching the instance .
> > >
> > >
> > > thanks,
> > > srinivas.
> > >
> > >
> > >
> > > On Thu, Feb 26, 2015 at 8:42 PM, Polehn, Mike A <mike.a.polehn at intel.com
> > >
> > > wrote:
> > >
> > > > In this example, the control network 00:03.0, remains unbound to UIO
> > > > driver but remains attached
> > > >  to Linux device driver (ssh access with putty) and just the target
> > > > interfaces are bound.
> > > > Below, it shows all 3 interfaces bound to the uio driver, which are not
> > > > usable until a task uses the UIO driver.
> > > >
> > > > [root at F21vm l3fwd-vf]# lspci -nn
> > > > 00:00.0 Host bridge [0600]: Intel Corporation 440FX - 82441FX PMC
> > [Natoma]
> > > > [8086:1237] (rev 02)
> > > > 00:01.0 ISA bridge [0601]: Intel Corporation 82371SB PIIX3 ISA
> > > > [Natoma/Triton II] [8086:7000]
> > > > 00:01.1 IDE interface [0101]: Intel Corporation 82371SB PIIX3 IDE
> > > > [Natoma/Triton II] [8086:7010]
> > > > 00:01.3 Bridge [0680]: Intel Corporation 82371AB/EB/MB PIIX4 ACPI
> > > > [8086:7113] (rev 03)
> > > > 00:02.0 VGA compatible controller [0300]: Cirrus Logic GD 5446
> > [1013:00b8]
> > > > 00:03.0 Ethernet controller [0200]: Red Hat, Inc Virtio network device
> > > > [1af4:1000]
> > > > 00:04.0 Ethernet controller [0200]: Intel Corporation XL710/X710
> > Virtual
> > > > Function [8086:154c] (rev 01)
> > > > 00:05.0 Ethernet controller [0200]: Intel Corporation XL710/X710
> > Virtual
> > > > Function [8086:154c] (rev 01)
> > > >
> > > > [root at F21vm l3fwd-vf]# /usr/src/dpdk/tools/dpdk_nic_bind.py
> > > > --bind=igb_uio 00:04.0
> > > > [root at F21vm l3fwd-vf]# /usr/src/dpdk/tools/dpdk_nic_bind.py
> > > > --bind=igb_uio 00:05.0
> > > > [root at F21vm l3fwd-vf]# /usr/src/dpdk/tools/dpdk_nic_bind.py --status
> > > >
> > > > Network devices using DPDK-compatible driver
> > > > ============================================
> > > > 0000:00:04.0 'XL710/X710 Virtual Function' drv=igb_uio unused=i40evf
> > > > 0000:00:05.0 'XL710/X710 Virtual Function' drv=igb_uio unused=i40evf
> > > >
> > > > Network devices using kernel driver
> > > > ===================================
> > > > 0000:00:03.0 'Virtio network device' if= drv=virtio-pci
> > > > unused=virtio_pci,igb_uio
> > > >
> > > > Other network devices
> > > > =====================
> > > > <none>
> > > >
> > > > -----Original Message-----
> > > > From: Dpdk-ovs [mailto:dpdk-ovs-bounces at lists.01.org] On Behalf Of
> > > > Srinivasreddy R
> > > > Sent: Thursday, February 26, 2015 6:11 AM
> > > > To: dev at dpdk.org; dpdk-ovs at lists.01.org
> > > > Subject: [Dpdk-ovs] problem in binding interfaces of virtio-pci on the
> > VM
> > > >
> > > > hi ,
> > > > I have written sample program for usvhost  supported by ovdk.
> > > >
> > > > i have initialized VM using the below command .
> > > > On the VM :
> > > >
> > > > I am able to see two interfaces . and working fine with traffic in
> > > > rawsocket mode .
> > > > my problem is when i bind the interfaces to pmd driver[ ibg_uio ] my
> > > > virtual machine is getting hanged . and  i am not able to access it
> > further
> > > > .
> > > > now my question is . what may be the reason for the behavior . and how
> > can
> > > > in debug the root cause .
> > > > please help in finding out the problem .
> > > >
> > > >
> > > >
> > > >  ./tools/dpdk_nic_bind.py --status
> > > >
> > > > Network devices using DPDK-compatible driver
> > > > ============================================
> > > > <none>
> > > >
> > > > Network devices using kernel driver
> > > > ===================================
> > > > 0000:00:03.0 '82540EM Gigabit Ethernet Controller' if=ens3 drv=e1000
> > > > unused=igb_uio *Active*
> > > > 0000:00:04.0 'Virtio network device' if= drv=virtio-pci unused=igb_uio
> > > > 0000:00:05.0 'Virtio network device' if= drv=virtio-pci unused=igb_uio
> > > >
> > > > Other network devices
> > > > =====================
> > > > <none>
> > > >
> > > >
> > > > ./dpdk_nic_bind.py --bind=igb_uio 00:04.0 00:05.0
> > > >
> > > >
> > > >
> > > > ./x86_64-softmmu/qemu-system-x86_64 -cpu host -boot c  -hda
> > > > /home/utils/images/vm1.img  -m 2048M -smp 3 --enable-kvm -name 'VM1'
> > > > -nographic -vnc :1 -pidfile /tmp/vm1.pid -drive
> > > > file=fat:rw:/tmp/qemu_share,snapshot=off -monitor
> > > > unix:/tmp/vm1monitor,server,nowait  -net none -no-reboot -mem-path
> > > > /dev/hugepages -mem-prealloc -netdev
> > > > type=tap,id=net1,script=no,downscript=no,ifname=usvhost1,vhost=on
> > -device
> > > >
> > virtio-net-pci,netdev=net1,mac=00:16:3e:00:03:03,csum=off,gso=off,guest_tso4=off,guest_tso6=off,guest_ecn=off
> > > > -netdev
> > type=tap,id=net2,script=no,downscript=no,ifname=usvhost2,vhost=on
> > > > -device
> > > >
> > > >
> > virtio-net-pci,netdev=net2,mac=00:16:3e:00:03:04,csum=off,gso=off,guest_tso4=off,guest_tso6=off,guest_ecn=off
> > > >
> > > >
> > > >
> > > >
> > > > ----------
> > > > thanks
> > > > srinivas.
> > > > _______________________________________________
> > > > Dpdk-ovs mailing list
> > > > Dpdk-ovs at lists.01.org
> > > > https://lists.01.org/mailman/listinfo/dpdk-ovs
> > > >
> > >
> > >
> > >
> > > --
> > > thanks
> > > srinivas.
> >
> 
> 
> 
> -- 
> thanks
> srinivas.


More information about the dev mailing list