#!/bin/sh # SPDX-License-Identifier: GPL-2.0-only export KSELFTESTS_SKIP=4 log() { echo >/dev/stderr $* } pe_ok() { local dev="$1" local path="/sys/bus/pci/devices/$dev/eeh_pe_state" # if a driver doesn't support the error handling callbacks then the # device is recovered by removing and re-probing it. This causes the # sysfs directory to disappear so read the PE state once and squash # any potential error messages local eeh_state="$(cat $path 2>/dev/null)" if [ -z "$eeh_state" ]; then return 1; fi local fw_state="$(echo $eeh_state | cut -d' ' -f1)" local sw_state="$(echo $eeh_state | cut -d' ' -f2)" # If EEH_PE_ISOLATED or EEH_PE_RECOVERING are set then the PE is in an # error state or being recovered. Either way, not ok. if [ "$((sw_state & 0x3))" -ne 0 ] ; then return 1 fi # A functioning PE should have the EEH_STATE_MMIO_ACTIVE and # EEH_STATE_DMA_ACTIVE flags set. For some goddamn stupid reason # the platform backends set these when the PE is in reset. The # RECOVERING check above should stop any false positives though. if [ "$((fw_state & 0x18))" -ne "$((0x18))" ] ; then return 1 fi return 0; } eeh_supported() { test -e /proc/powerpc/eeh && \ grep -q 'EEH Subsystem is enabled' /proc/powerpc/eeh } eeh_test_prep() { if ! eeh_supported ; then echo "EEH not supported on this system, skipping" exit $KSELFTESTS_SKIP; fi if [ ! -e "/sys/kernel/debug/powerpc/eeh_dev_check" ] && \ [ ! -e "/sys/kernel/debug/powerpc/eeh_dev_break" ] ; then log "debugfs EEH testing files are missing. Is debugfs mounted?" exit $KSELFTESTS_SKIP; fi # Bump the max freeze count to something absurd so we don't # trip over it while breaking things. echo 5000 > /sys/kernel/debug/powerpc/eeh_max_freezes } eeh_can_break() { # skip bridges since we can't recover them (yet...) if [ -e "/sys/bus/pci/devices/$dev/pci_bus" ] ; then log "$dev, Skipped: bridge" return 1; fi # The ahci driver doesn't support error recovery. If the ahci device # happens to be hosting the root filesystem, and then we go and break # it the system will generally go down. We should probably fix that # at some point if [ "ahci" = "$(basename $(realpath /sys/bus/pci/devices/$dev/driver))" ] ; then log "$dev, Skipped: ahci doesn't support recovery" return 1; fi # Don't inject errosr into an already-frozen PE. This happens with # PEs that contain multiple PCI devices (e.g. multi-function cards) # and injecting new errors during the recovery process will probably # result in the recovery failing and the device being marked as # failed. if ! pe_ok $dev ; then log "$dev, Skipped: Bad initial PE state" return 1; fi return 0 } eeh_one_dev() { local dev="$1" # Using this function from the command line is sometimes useful for # testing so check that the argument is a well-formed sysfs device # name. if ! test -e /sys/bus/pci/devices/$dev/ ; then log "Error: '$dev' must be a sysfs device name (DDDD:BB:DD.F)" return 1; fi # Break it echo $dev >/sys/kernel/debug/powerpc/eeh_dev_break # Force an EEH device check. If the kernel has already # noticed the EEH (due to a driver poll or whatever), this # is a no-op. echo $dev >/sys/kernel/debug/powerpc/eeh_dev_check # Default to a 60s timeout when waiting for a device to recover. This # is an arbitrary default which can be overridden by setting the # EEH_MAX_WAIT environmental variable when required. # The current record holder for longest recovery time is: # "Adaptec Series 8 12G SAS/PCIe 3" at 39 seconds max_wait=${EEH_MAX_WAIT:=60} for i in `seq 0 ${max_wait}` ; do if pe_ok $dev ; then break; fi log "$dev, waited $i/${max_wait}" sleep 1 done if ! pe_ok $dev ; then log "$dev, Failed to recover!" return 1; fi log "$dev, Recovered after $i seconds" return 0; } eeh_has_driver() { test -e /sys/bus/pci/devices/$1/driver; return $? } eeh_can_recover() { # we'll get an IO error if the device's current driver doesn't support # error recovery echo $1 > '/sys/kernel/debug/powerpc/eeh_dev_can_recover' 2>/dev/null return $? } eeh_find_all_pfs() { devices="" # SR-IOV on pseries requires hypervisor support, so check for that is_pseries="" if grep -q pSeries /proc/cpuinfo ; then if [ ! -f /proc/device-tree/rtas/ibm,open-sriov-allow-unfreeze ] || [ ! -f /proc/device-tree/rtas/ibm,open-sriov-map-pe-number ] ; then return 1; fi is_pseries="true" fi for dev in `ls -1 /sys/bus/pci/devices/` ; do sysfs="/sys/bus/pci/devices/$dev" if [ ! -e "$sysfs/sriov_numvfs" ] ; then continue fi # skip unsupported PFs on pseries if [ -z "$is_pseries" ] && [ ! -f "$sysfs/of_node/ibm,is-open-sriov-pf" ] && [ ! -f "$sysfs/of_node/ibm,open-sriov-vf-bar-info" ] ; then continue; fi # no driver, no vfs if ! eeh_has_driver $dev ; then continue fi devices="$devices $dev" done if [ -z "$devices" ] ; then return 1; fi echo $devices return 0; } # attempts to enable one VF on each PF so we can do VF specific tests. # stdout: list of enabled VFs, one per line # return code: 0 if vfs are found, 1 otherwise eeh_enable_vfs() { pf_list="$(eeh_find_all_pfs)" vfs=0 for dev in $pf_list ; do pf_sysfs="/sys/bus/pci/devices/$dev" # make sure we have a single VF echo 0 > "$pf_sysfs/sriov_numvfs" echo 1 > "$pf_sysfs/sriov_numvfs" if [ "$?" != 0 ] ; then log "Unable to enable VFs on $pf, skipping" continue; fi vf="$(basename $(realpath "$pf_sysfs/virtfn0"))" if [ $? != 0 ] ; then log "unable to find enabled vf on $pf" echo 0 > "$pf_sysfs/sriov_numvfs" continue; fi if ! eeh_can_break $vf ; then log "skipping " echo 0 > "$pf_sysfs/sriov_numvfs" continue; fi vfs="$((vfs + 1))" echo $vf done test "$vfs" != 0 return $? } eeh_disable_vfs() { pf_list="$(eeh_find_all_pfs)" if [ -z "$pf_list" ] ; then return 1; fi for dev in $pf_list ; do echo 0 > "/sys/bus/pci/devices/$dev/sriov_numvfs" done return 0; }