Age | Commit message (Collapse) | Author |
|
|
|
Bug 6845, 6821
Need to avoid using /sys/class/net/ethX/device/irq (or /sys/bus/pci/.../irq)
because these don't handle MSI or multiqueue. This also resolves issues
with Vmware/ESX.
|
|
Bug 6845
Warn (and ignore) attempt to assign IRQ directly on multiqueuej
NIC.
|
|
Bug 6784
No point in trying to force affinity if device is offline.
|
|
Adapt to irq naming convention in 2.6.37 kernel for vmxnet3 driver.
|
|
Bug 6784
Disabled device has no IRQ, so don't change it.
|
|
The initial CPU selection function needs to take hyperthreading
into account.
|
|
(cherry picked from commit a943568e64bca73bb2951e968d0cc752d72874ab)
|
|
The problem was due to incorrect initialization of the $q and $cpu
variables. Their initializations were reversed.
|
|
Need to handle virtual devices with no IRQ, and older processors
without concept of multi-core.
|
|
Replace old script with new cleaner script that handles both
IRQ affinity and Receive Packet Steering. Instead of two scripts
(one for mask and one for auto), do it all with one script.
Receive Packet Steering is supported in two ways.
If 'auto' is used, then both threads on HT system will be
used for receive processing.
If explicit mask is given, then two masks can be used to set
both IRQ cpus and RPS cpus.
|