/****************************************************************************** * * This file is provided under a dual BSD/GPLv2 license. When using or * redistributing this file, you may do so under either license. * * GPL LICENSE SUMMARY * * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. * * This program is free software; you can redistribute it and/or modify * it under the terms of version 2 of the GNU General Public License as * published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110, * USA * * The full GNU General Public License is included in this distribution * in the file called COPYING. * * Contact Information: * Intel Linux Wireless * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 * * BSD LICENSE * * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in * the documentation and/or other materials provided with the * distribution. * * Neither the name Intel Corporation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * *****************************************************************************/ #include #include #include "mvm.h" #include "iwl-eeprom-parse.h" #include "fw-api-scan.h" #define IWL_PLCP_QUIET_THRESH 1 #define IWL_ACTIVE_QUIET_TIME 10 struct iwl_mvm_scan_params { u32 max_out_time; u32 suspend_time; bool passive_fragmented; struct _dwell { u16 passive; u16 active; } dwell[IEEE80211_NUM_BANDS]; }; static inline __le16 iwl_mvm_scan_rx_chain(struct iwl_mvm *mvm) { u16 rx_chain; u8 rx_ant; if (mvm->scan_rx_ant != ANT_NONE) rx_ant = mvm->scan_rx_ant; else rx_ant = mvm->fw->valid_rx_ant; rx_chain = rx_ant << PHY_RX_CHAIN_VALID_POS; rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_MIMO_SEL_POS; rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_SEL_POS; rx_chain |= 0x1 << PHY_RX_CHAIN_DRIVER_FORCE_POS; return cpu_to_le16(rx_chain); } static inline __le32 iwl_mvm_scan_rxon_flags(struct cfg80211_scan_request *req) { if (req->channels[0]->band == IEEE80211_BAND_2GHZ) return cpu_to_le32(PHY_BAND_24); else return cpu_to_le32(PHY_BAND_5); } static inline __le32 iwl_mvm_scan_rate_n_flags(struct iwl_mvm *mvm, enum ieee80211_band band, bool no_cck) { u32 tx_ant; mvm->scan_last_antenna_idx = iwl_mvm_next_antenna(mvm, mvm->fw->valid_tx_ant, mvm->scan_last_antenna_idx); tx_ant = BIT(mvm->scan_last_antenna_idx) << RATE_MCS_ANT_POS; if (band == IEEE80211_BAND_2GHZ && !no_cck) return cpu_to_le32(IWL_RATE_1M_PLCP | RATE_MCS_CCK_MSK | tx_ant); else return cpu_to_le32(IWL_RATE_6M_PLCP | tx_ant); } /* * We insert the SSIDs in an inverted order, because the FW will * invert it back. The most prioritized SSID, which is first in the * request list, is not copied here, but inserted directly to the probe * request. */ static void iwl_mvm_scan_fill_ssids(struct iwl_scan_cmd *cmd, struct cfg80211_scan_request *req, int first) { int fw_idx, req_idx; for (req_idx = req->n_ssids - 1, fw_idx = 0; req_idx >= first; req_idx--, fw_idx++) { cmd->direct_scan[fw_idx].id = WLAN_EID_SSID; cmd->direct_scan[fw_idx].len = req->ssids[req_idx].ssid_len; memcpy(cmd->direct_scan[fw_idx].ssid, req->ssids[req_idx].ssid, req->ssids[req_idx].ssid_len); } } /* * If req->n_ssids > 0, it means we should do an active scan. * In case of active scan w/o directed scan, we receive a zero-length SSID * just to notify that this scan is active and not passive. * In order to notify the FW of the number of SSIDs we wish to scan (including * the zero-length one), we need to set the corresponding bits in chan->type, * one for each SSID, and set the active bit (first). If the first SSID is * already included in the probe template, so we need to set only * req->n_ssids - 1 bits in addition to the first bit. */ static u16 iwl_mvm_get_active_dwell(enum ieee80211_band band, int n_ssids) { if (band == IEEE80211_BAND_2GHZ) return 30 + 3 * (n_ssids + 1); return 20 + 2 * (n_ssids + 1); } static u16 iwl_mvm_get_passive_dwell(enum ieee80211_band band) { return band == IEEE80211_BAND_2GHZ ? 100 + 20 : 100 + 10; } static void iwl_mvm_scan_fill_channels(struct iwl_scan_cmd *cmd, struct cfg80211_scan_request *req, bool basic_ssid, struct iwl_mvm_scan_params *params) { struct iwl_scan_channel *chan = (struct iwl_scan_channel *) (cmd->data + le16_to_cpu(cmd->tx_cmd.len)); int i; int type = BIT(req->n_ssids) - 1; enum ieee80211_band band = req->channels[0]->band; if (!basic_ssid) type |= BIT(req->n_ssids); for (i = 0; i < cmd->channel_count; i++) { chan->channel = cpu_to_le16(req->channels[i]->hw_value); chan->type = cpu_to_le32(type); if (req->channels[i]->flags & IEEE80211_CHAN_NO_IR) chan->type &= cpu_to_le32(~SCAN_CHANNEL_TYPE_ACTIVE); chan->active_dwell = cpu_to_le16(params->dwell[band].active); chan->passive_dwell = cpu_to_le16(params->dwell[band].passive); chan->iteration_count = cpu_to_le16(1); chan++; } } /* * Fill in probe request with the following parameters: * TA is our vif HW address, which mac80211 ensures we have. * Packet is broadcasted, so this is both SA and DA. * The probe request IE is made out of two: first comes the most prioritized * SSID if a directed scan is requested. Second comes whatever extra * information was given to us as the scan request IE. */ static u16 iwl_mvm_fill_probe_req(struct ieee80211_mgmt *frame, const u8 *ta, int n_ssids, const u8 *ssid, int ssid_len, const u8 *ie, int ie_len, int left) { int len = 0; u8 *pos = NULL; /* Make sure there is enough space for the probe request, * two mandatory IEs and the data */ left -= 24; if (left < 0) return 0; frame->frame_control = cpu_to_le16(IEEE80211_STYPE_PROBE_REQ); eth_broadcast_addr(frame->da); memcpy(frame->sa, ta, ETH_ALEN); eth_broadcast_addr(frame->bssid); frame->seq_ctrl = 0; len += 24; /* for passive scans, no need to fill anything */ if (n_ssids == 0) return (u16)len; /* points to the payload of the request */ pos = &frame->u.probe_req.variable[0]; /* fill in our SSID IE */ left -= ssid_len + 2; if (left < 0) return 0; *pos++ = WLAN_EID_SSID; *pos++ = ssid_len; if (ssid && ssid_len) { /* ssid_len may be == 0 even if ssid is valid */ memcpy(pos, ssid, ssid_len); pos += ssid_len; } len += ssid_len + 2; if (WARN_ON(left < ie_len)) return len; if (ie && ie_len) { memcpy(pos, ie, ie_len); len += ie_len; } return (u16)len; } static void iwl_mvm_scan_condition_iterator(void *data, u8 *mac, struct ieee80211_vif *vif) { struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); bool *global_bound = data; if (mvmvif->phy_ctxt && mvmvif->phy_ctxt->id < MAX_PHYS) *global_bound = true; } static void iwl_mvm_scan_calc_params(struct iwl_mvm *mvm, struct ieee80211_vif *vif, int n_ssids, struct iwl_mvm_scan_params *params) { bool global_bound = false; enum ieee80211_band band; ieee80211_iterate_active_interfaces_atomic(mvm->hw, IEEE80211_IFACE_ITER_NORMAL, iwl_mvm_scan_condition_iterator, &global_bound); /* * Under low latency traffic passive scan is fragmented meaning * that dwell on a particular channel will be fragmented. Each fragment * dwell time is 20ms and fragments period is 105ms. Skipping to next * channel will be delayed by the same period - 105ms. So suspend_time * parameter describing both fragments and channels skipping periods is * set to 105ms. This value is chosen so that overall passive scan * duration will not be too long. Max_out_time in this case is set to * 70ms, so for active scanning operating channel will be left for 70ms * while for passive still for 20ms (fragment dwell). */ if (global_bound) { if (!iwl_mvm_low_latency(mvm)) { params->suspend_time = ieee80211_tu_to_usec(100); params->max_out_time = ieee80211_tu_to_usec(600); } else { params->suspend_time = ieee80211_tu_to_usec(105); /* P2P doesn't support fragmented passive scan, so * configure max_out_time to be at least longest dwell * time for passive scan. */ if (vif->type == NL80211_IFTYPE_STATION && !vif->p2p) { params->max_out_time = ieee80211_tu_to_usec(70); params->passive_fragmented = true; } else { u32 passive_dwell; /* * Use band G so that passive channel dwell time * will be assigned with maximum value. */ band = IEEE80211_BAND_2GHZ; passive_dwell = iwl_mvm_get_passive_dwell(band); params->max_out_time = ieee80211_tu_to_usec(passive_dwell); } } } for (band = IEEE80211_BAND_2GHZ; band < IEEE80211_NUM_BANDS; band++) { if (params->passive_fragmented) params->dwell[band].passive = 20; else params->dwell[band].passive = iwl_mvm_get_passive_dwell(band); params->dwell[band].active = iwl_mvm_get_active_dwell(band, n_ssids); } } int iwl_mvm_scan_request(struct iwl_mvm *mvm, struct ieee80211_vif *vif, struct cfg80211_scan_request *req) { struct iwl_host_cmd hcmd = { .id = SCAN_REQUEST_CMD, .len = { 0, }, .data = { mvm->scan_cmd, }, .flags = CMD_SYNC, .dataflags = { IWL_HCMD_DFL_NOCOPY, }, }; struct iwl_scan_cmd *cmd = mvm->scan_cmd; int ret; u32 status; int ssid_len = 0; u8 *ssid = NULL; bool basic_ssid = !(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_NO_BASIC_SSID); struct iwl_mvm_scan_params params = {}; lockdep_assert_held(&mvm->mutex); BUG_ON(mvm->scan_cmd == NULL); IWL_DEBUG_SCAN(mvm, "Handling mac80211 scan request\n"); mvm->scan_status = IWL_MVM_SCAN_OS; memset(cmd, 0, sizeof(struct iwl_scan_cmd) + mvm->fw->ucode_capa.max_probe_length + (MAX_NUM_SCAN_CHANNELS * sizeof(struct iwl_scan_channel))); cmd->channel_count = (u8)req->n_channels; cmd->quiet_time = cpu_to_le16(IWL_ACTIVE_QUIET_TIME); cmd->quiet_plcp_th = cpu_to_le16(IWL_PLCP_QUIET_THRESH); cmd->rxchain_sel_flags = iwl_mvm_scan_rx_chain(mvm); iwl_mvm_scan_calc_params(mvm, vif, req->n_ssids, ¶ms); cmd->max_out_time = cpu_to_le32(params.max_out_time); cmd->suspend_time = cpu_to_le32(params.suspend_time); if (params.passive_fragmented) cmd->scan_flags |= SCAN_FLAGS_FRAGMENTED_SCAN; cmd->rxon_flags = iwl_mvm_scan_rxon_flags(req); cmd->filter_flags = cpu_to_le32(MAC_FILTER_ACCEPT_GRP | MAC_FILTER_IN_BEACON); if (vif->type == NL80211_IFTYPE_P2P_DEVICE) cmd->type = cpu_to_le32(SCAN_TYPE_DISCOVERY_FORCED); else cmd->type = cpu_to_le32(SCAN_TYPE_FORCED); cmd->repeats = cpu_to_le32(1); /* * If the user asked for passive scan, don't change to active scan if * you see any activity on the channel - remain passive. */ if (req->n_ssids > 0) { cmd->passive2active = cpu_to_le16(1); cmd->scan_flags |= SCAN_FLAGS_PASSIVE2ACTIVE; if (basic_ssid) { ssid = req->ssids[0].ssid; ssid_len = req->ssids[0].ssid_len; } } else { cmd->passive2active = 0; cmd->scan_flags &= ~SCAN_FLAGS_PASSIVE2ACTIVE; } iwl_mvm_scan_fill_ssids(cmd, req, basic_ssid ? 1 : 0); cmd->tx_cmd.tx_flags = cpu_to_le32(TX_CMD_FLG_SEQ_CTL | TX_CMD_FLG_BT_DIS); cmd->tx_cmd.sta_id = mvm->aux_sta.sta_id; cmd->tx_cmd.life_time = cpu_to_le32(TX_CMD_LIFE_TIME_INFINITE); cmd->tx_cmd.rate_n_flags = iwl_mvm_scan_rate_n_flags(mvm, req->channels[0]->band, req->no_cck); cmd->tx_cmd.len = cpu_to_le16(iwl_mvm_fill_probe_req( (struct ieee80211_mgmt *)cmd->data, vif->addr, req->n_ssids, ssid, ssid_len, req->ie, req->ie_len, mvm->fw->ucode_capa.max_probe_length)); iwl_mvm_scan_fill_channels(cmd, req, basic_ssid, ¶ms); cmd->len = cpu_to_le16(sizeof(struct iwl_scan_cmd) + le16_to_cpu(cmd->tx_cmd.len) + (cmd->channel_count * sizeof(struct iwl_scan_channel))); hcmd.len[0] = le16_to_cpu(cmd->len); status = SCAN_RESPONSE_OK; ret = iwl_mvm_send_cmd_status(mvm, &hcmd, &status); if (!ret && status == SCAN_RESPONSE_OK) { IWL_DEBUG_SCAN(mvm, "Scan request was sent successfully\n"); } else { /* * If the scan failed, it usually means that the FW was unable * to allocate the time events. Warn on it, but maybe we * should try to send the command again with different params. */ IWL_ERR(mvm, "Scan failed! status 0x%x ret %d\n", status, ret); mvm->scan_status = IWL_MVM_SCAN_NONE; ret = -EIO; } return ret; } int iwl_mvm_rx_scan_response(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb, struct iwl_device_cmd *cmd) { struct iwl_rx_packet *pkt = rxb_addr(rxb); struct iwl_cmd_response *resp = (void *)pkt->data; IWL_DEBUG_SCAN(mvm, "Scan response received. status 0x%x\n", le32_to_cpu(resp->status)); return 0; } int iwl_mvm_rx_scan_complete(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb, struct iwl_device_cmd *cmd) { struct iwl_rx_packet *pkt = rxb_addr(rxb); struct iwl_scan_complete_notif *notif = (void *)pkt->data; lockdep_assert_held(&mvm->mutex); IWL_DEBUG_SCAN(mvm, "Scan complete: status=0x%x scanned channels=%d\n", notif->status, notif->scanned_channels); if (mvm->scan_status == IWL_MVM_SCAN_OS) mvm->scan_status = IWL_MVM_SCAN_NONE; ieee80211_scan_completed(mvm->hw, notif->status != SCAN_COMP_STATUS_OK); iwl_mvm_unref(mvm, IWL_MVM_REF_SCAN); return 0; } int iwl_mvm_rx_sched_scan_results(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb, struct iwl_device_cmd *cmd) { struct iwl_rx_packet *pkt = rxb_addr(rxb); struct iwl_sched_scan_results *notif = (void *)pkt->data; if (notif->client_bitmap & SCAN_CLIENT_SCHED_SCAN) { IWL_DEBUG_SCAN(mvm, "Scheduled scan results\n"); ieee80211_sched_scan_results(mvm->hw); } return 0; } static bool iwl_mvm_scan_abort_notif(struct iwl_notif_wait_data *notif_wait, struct iwl_rx_packet *pkt, void *data) { struct iwl_mvm *mvm = container_of(notif_wait, struct iwl_mvm, notif_wait); struct iwl_scan_complete_notif *notif; u32 *resp; switch (pkt->hdr.cmd) { case SCAN_ABORT_CMD: resp = (void *)pkt->data; if (*resp == CAN_ABORT_STATUS) { IWL_DEBUG_SCAN(mvm, "Scan can be aborted, wait until completion\n"); return false; } /* * If scan cannot be aborted, it means that we had a * SCAN_COMPLETE_NOTIFICATION in the pipe and it called * ieee80211_scan_completed already. */ IWL_DEBUG_SCAN(mvm, "Scan cannot be aborted, exit now: %d\n", *resp); return true; case SCAN_COMPLETE_NOTIFICATION: notif = (void *)pkt->data; IWL_DEBUG_SCAN(mvm, "Scan aborted: status 0x%x\n", notif->status); return true; default: WARN_ON(1); return false; }; } int iwl_mvm_cancel_scan(struct iwl_mvm *mvm) { struct iwl_notification_wait wait_scan_abort; static const u8 scan_abort_notif[] = { SCAN_ABORT_CMD, SCAN_COMPLETE_NOTIFICATION }; int ret; if (mvm->scan_status == IWL_MVM_SCAN_NONE) return 0; if (iwl_mvm_is_radio_killed(mvm)) { ieee80211_scan_completed(mvm->hw, true); iwl_mvm_unref(mvm, IWL_MVM_REF_SCAN); mvm->scan_status = IWL_MVM_SCAN_NONE; return 0; } iwl_init_notification_wait(&mvm->notif_wait, &wait_scan_abort, scan_abort_notif, ARRAY_SIZE(scan_abort_notif), iwl_mvm_scan_abort_notif, NULL); ret = iwl_mvm_send_cmd_pdu(mvm, SCAN_ABORT_CMD, CMD_SYNC, 0, NULL); if (ret) { IWL_ERR(mvm, "Couldn't send SCAN_ABORT_CMD: %d\n", ret); /* mac80211's state will be cleaned in the nic_restart flow */ goto out_remove_notif; } return iwl_wait_notification(&mvm->notif_wait, &wait_scan_abort, HZ); out_remove_notif: iwl_remove_notification(&mvm->notif_wait, &wait_scan_abort); return ret; } int iwl_mvm_rx_scan_offload_complete_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb, struct iwl_device_cmd *cmd) { struct iwl_rx_packet *pkt = rxb_addr(rxb); struct iwl_scan_offload_complete *scan_notif = (void *)pkt->data; /* scan status must be locked for proper checking */ lockdep_assert_held(&mvm->mutex); IWL_DEBUG_SCAN(mvm, "Scheduled scan completed, status %s\n", scan_notif->status == IWL_SCAN_OFFLOAD_COMPLETED ? "completed" : "aborted"); /* only call mac80211 completion if the stop was initiated by FW */ if (mvm->scan_status == IWL_MVM_SCAN_SCHED) { mvm->scan_status = IWL_MVM_SCAN_NONE; ieee80211_sched_scan_stopped(mvm->hw); } return 0; } static void iwl_scan_offload_build_tx_cmd(struct iwl_mvm *mvm, struct ieee80211_vif *vif, struct ieee80211_sched_scan_ies *ies, enum ieee80211_band band, struct iwl_tx_cmd *cmd, u8 *data) { u16 cmd_len; cmd->tx_flags = cpu_to_le32(TX_CMD_FLG_SEQ_CTL); cmd->life_time = cpu_to_le32(TX_CMD_LIFE_TIME_INFINITE); cmd->sta_id = mvm->aux_sta.sta_id; cmd->rate_n_flags = iwl_mvm_scan_rate_n_flags(mvm, band, false); cmd_len = iwl_mvm_fill_probe_req((struct ieee80211_mgmt *)data, vif->addr, 1, NULL, 0, ies->ie[band], ies->len[band], SCAN_OFFLOAD_PROBE_REQ_SIZE); cmd->len = cpu_to_le16(cmd_len); } static void iwl_build_scan_cmd(struct iwl_mvm *mvm, struct ieee80211_vif *vif, struct cfg80211_sched_scan_request *req, struct iwl_scan_offload_cmd *scan, struct iwl_mvm_scan_params *params) { scan->channel_count = mvm->nvm_data->bands[IEEE80211_BAND_2GHZ].n_channels + mvm->nvm_data->bands[IEEE80211_BAND_5GHZ].n_channels; scan->quiet_time = cpu_to_le16(IWL_ACTIVE_QUIET_TIME); scan->quiet_plcp_th = cpu_to_le16(IWL_PLCP_QUIET_THRESH); scan->good_CRC_th = IWL_GOOD_CRC_TH_DEFAULT; scan->rx_chain = iwl_mvm_scan_rx_chain(mvm); scan->max_out_time = cpu_to_le32(params->max_out_time); scan->suspend_time = cpu_to_le32(params->suspend_time); scan->filter_flags |= cpu_to_le32(MAC_FILTER_ACCEPT_GRP | MAC_FILTER_IN_BEACON); scan->scan_type = cpu_to_le32(SCAN_TYPE_BACKGROUND); scan->rep_count = cpu_to_le32(1); if (params->passive_fragmented) scan->scan_flags |= SCAN_FLAGS_FRAGMENTED_SCAN; } static int iwl_ssid_exist(u8 *ssid, u8 ssid_len, struct iwl_ssid_ie *ssid_list) { int i; for (i = 0; i < PROBE_OPTION_MAX; i++) { if (!ssid_list[i].len) break; if (ssid_list[i].len == ssid_len && !memcmp(ssid_list->ssid, ssid, ssid_len)) return i; } return -1; } static void iwl_scan_offload_build_ssid(struct cfg80211_sched_scan_request *req, struct iwl_scan_offload_cmd *scan, u32 *ssid_bitmap) { int i, j; int index; /* * copy SSIDs from match list. * iwl_config_sched_scan_profiles() uses the order of these ssids to * config match list. */ for (i = 0; i < req->n_match_sets && i < PROBE_OPTION_MAX; i++) { /* skip empty SSID matchsets */ if (!req->match_sets[i].ssid.ssid_len) continue; scan->direct_scan[i].id = WLAN_EID_SSID; scan->direct_scan[i].len = req->match_sets[i].ssid.ssid_len; memcpy(scan->direct_scan[i].ssid, req->match_sets[i].ssid.ssid, scan->direct_scan[i].len); } /* add SSIDs from scan SSID list */ *ssid_bitmap = 0; for (j = 0; j < req->n_ssids && i < PROBE_OPTION_MAX; j++) { index = iwl_ssid_exist(req->ssids[j].ssid, req->ssids[j].ssid_len, scan->direct_scan); if (index < 0) { if (!req->ssids[j].ssid_len) continue; scan->direct_scan[i].id = WLAN_EID_SSID; scan->direct_scan[i].len = req->ssids[j].ssid_len; memcpy(scan->direct_scan[i].ssid, req->ssids[j].ssid, scan->direct_scan[i].len); *ssid_bitmap |= BIT(i + 1); i++; } else { *ssid_bitmap |= BIT(index + 1); } } } static void iwl_build_channel_cfg(struct iwl_mvm *mvm, struct cfg80211_sched_scan_request *req, struct iwl_scan_channel_cfg *channels, enum ieee80211_band band, int *head, int *tail, u32 ssid_bitmap, struct iwl_mvm_scan_params *params) { struct ieee80211_supported_band *s_band; int n_channels = req->n_channels; int i, j, index = 0; bool partial; /* * We have to configure all supported channels, even if we don't want to * scan on them, but we have to send channels in the order that we want * to scan. So add requested channels to head of the list and others to * the end. */ s_band = &mvm->nvm_data->bands[band]; for (i = 0; i < s_band->n_channels && *head <= *tail; i++) { partial = false; for (j = 0; j < n_channels; j++) if (s_band->channels[i].center_freq == req->channels[j]->center_freq) { index = *head; (*head)++; /* * Channels that came with the request will be * in partial scan . */ partial = true; break; } if (!partial) { index = *tail; (*tail)--; } channels->channel_number[index] = cpu_to_le16(ieee80211_frequency_to_channel( s_band->channels[i].center_freq)); channels->dwell_time[index][0] = params->dwell[band].active; channels->dwell_time[index][1] = params->dwell[band].passive; channels->iter_count[index] = cpu_to_le16(1); channels->iter_interval[index] = 0; if (!(s_band->channels[i].flags & IEEE80211_CHAN_NO_IR)) channels->type[index] |= cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_ACTIVE); channels->type[index] |= cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_FULL); if (partial) channels->type[index] |= cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_PARTIAL); if (s_band->channels[i].flags & IEEE80211_CHAN_NO_HT40) channels->type[index] |= cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_NARROW); /* scan for all SSIDs from req->ssids */ channels->type[index] |= cpu_to_le32(ssid_bitmap); } } int iwl_mvm_config_sched_scan(struct iwl_mvm *mvm, struct ieee80211_vif *vif, struct cfg80211_sched_scan_request *req, struct ieee80211_sched_scan_ies *ies) { int band_2ghz = mvm->nvm_data->bands[IEEE80211_BAND_2GHZ].n_channels; int band_5ghz = mvm->nvm_data->bands[IEEE80211_BAND_5GHZ].n_channels; int head = 0; int tail = band_2ghz + band_5ghz; u32 ssid_bitmap; int cmd_len; int ret; struct iwl_scan_offload_cfg *scan_cfg; struct iwl_host_cmd cmd = { .id = SCAN_OFFLOAD_CONFIG_CMD, .flags = CMD_SYNC, }; struct iwl_mvm_scan_params params = {}; lockdep_assert_held(&mvm->mutex); cmd_len = sizeof(struct iwl_scan_offload_cfg) + 2 * SCAN_OFFLOAD_PROBE_REQ_SIZE; scan_cfg = kzalloc(cmd_len, GFP_KERNEL); if (!scan_cfg) return -ENOMEM; iwl_mvm_scan_calc_params(mvm, vif, req->n_ssids, ¶ms); iwl_build_scan_cmd(mvm, vif, req, &scan_cfg->scan_cmd, ¶ms); scan_cfg->scan_cmd.len = cpu_to_le16(cmd_len); iwl_scan_offload_build_ssid(req, &scan_cfg->scan_cmd, &ssid_bitmap); /* build tx frames for supported bands */ if (band_2ghz) { iwl_scan_offload_build_tx_cmd(mvm, vif, ies, IEEE80211_BAND_2GHZ, &scan_cfg->scan_cmd.tx_cmd[0], scan_cfg->data); iwl_build_channel_cfg(mvm, req, &scan_cfg->channel_cfg, IEEE80211_BAND_2GHZ, &head, &tail, ssid_bitmap, ¶ms); } if (band_5ghz) { iwl_scan_offload_build_tx_cmd(mvm, vif, ies, IEEE80211_BAND_5GHZ, &scan_cfg->scan_cmd.tx_cmd[1], scan_cfg->data + SCAN_OFFLOAD_PROBE_REQ_SIZE); iwl_build_channel_cfg(mvm, req, &scan_cfg->channel_cfg, IEEE80211_BAND_5GHZ, &head, &tail, ssid_bitmap, ¶ms); } cmd.data[0] = scan_cfg; cmd.len[0] = cmd_len; cmd.dataflags[0] = IWL_HCMD_DFL_NOCOPY; IWL_DEBUG_SCAN(mvm, "Sending scheduled scan config\n"); ret = iwl_mvm_send_cmd(mvm, &cmd); kfree(scan_cfg); return ret; } int iwl_mvm_config_sched_scan_profiles(struct iwl_mvm *mvm, struct cfg80211_sched_scan_request *req) { struct iwl_scan_offload_profile *profile; struct iwl_scan_offload_profile_cfg *profile_cfg; struct iwl_scan_offload_blacklist *blacklist; struct iwl_host_cmd cmd = { .id = SCAN_OFFLOAD_UPDATE_PROFILES_CMD, .flags = CMD_SYNC, .len[1] = sizeof(*profile_cfg), .dataflags[0] = IWL_HCMD_DFL_NOCOPY, .dataflags[1] = IWL_HCMD_DFL_NOCOPY, }; int blacklist_len; int i; int ret; if (WARN_ON(req->n_match_sets > IWL_SCAN_MAX_PROFILES)) return -EIO; if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_SHORT_BL) blacklist_len = IWL_SCAN_SHORT_BLACKLIST_LEN; else blacklist_len = IWL_SCAN_MAX_BLACKLIST_LEN; blacklist = kzalloc(sizeof(*blacklist) * blacklist_len, GFP_KERNEL); if (!blacklist) return -ENOMEM; profile_cfg = kzalloc(sizeof(*profile_cfg), GFP_KERNEL); if (!profile_cfg) { ret = -ENOMEM; goto free_blacklist; } cmd.data[0] = blacklist; cmd.len[0] = sizeof(*blacklist) * blacklist_len; cmd.data[1] = profile_cfg; /* No blacklist configuration */ profile_cfg->num_profiles = req->n_match_sets; profile_cfg->active_clients = SCAN_CLIENT_SCHED_SCAN; profile_cfg->pass_match = SCAN_CLIENT_SCHED_SCAN; profile_cfg->match_notify = SCAN_CLIENT_SCHED_SCAN; if (!req->n_match_sets || !req->match_sets[0].ssid.ssid_len) profile_cfg->any_beacon_notify = SCAN_CLIENT_SCHED_SCAN; for (i = 0; i < req->n_match_sets; i++) { profile = &profile_cfg->profiles[i]; profile->ssid_index = i; /* Support any cipher and auth algorithm */ profile->unicast_cipher = 0xff; profile->auth_alg = 0xff; profile->network_type = IWL_NETWORK_TYPE_ANY; profile->band_selection = IWL_SCAN_OFFLOAD_SELECT_ANY; profile->client_bitmap = SCAN_CLIENT_SCHED_SCAN; } IWL_DEBUG_SCAN(mvm, "Sending scheduled scan profile config\n"); ret = iwl_mvm_send_cmd(mvm, &cmd); kfree(profile_cfg); free_blacklist: kfree(blacklist); return ret; } int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm, struct cfg80211_sched_scan_request *req) { struct iwl_scan_offload_req scan_req = { .watchdog = IWL_SCHED_SCAN_WATCHDOG, .schedule_line[0].iterations = IWL_FAST_SCHED_SCAN_ITERATIONS, .schedule_line[0].delay = req->interval / 1000, .schedule_line[0].full_scan_mul = 1, .schedule_line[1].iterations = 0xff, .schedule_line[1].delay = req->interval / 1000, .schedule_line[1].full_scan_mul = IWL_FULL_SCAN_MULTIPLIER, }; if (req->n_match_sets && req->match_sets[0].ssid.ssid_len) { IWL_DEBUG_SCAN(mvm, "Sending scheduled scan with filtering, filter len %d\n", req->n_match_sets); } else { IWL_DEBUG_SCAN(mvm, "Sending Scheduled scan without filtering\n"); scan_req.flags |= cpu_to_le16(IWL_SCAN_OFFLOAD_FLAG_PASS_ALL); } return iwl_mvm_send_cmd_pdu(mvm, SCAN_OFFLOAD_REQUEST_CMD, CMD_SYNC, sizeof(scan_req), &scan_req); } static int iwl_mvm_send_sched_scan_abort(struct iwl_mvm *mvm) { int ret; struct iwl_host_cmd cmd = { .id = SCAN_OFFLOAD_ABORT_CMD, .flags = CMD_SYNC, }; u32 status; /* Exit instantly with error when device is not ready * to receive scan abort command or it does not perform * scheduled scan currently */ if (mvm->scan_status != IWL_MVM_SCAN_SCHED) return -EIO; ret = iwl_mvm_send_cmd_status(mvm, &cmd, &status); if (ret) return ret; if (status != CAN_ABORT_STATUS) { /* * The scan abort will return 1 for success or * 2 for "failure". A failure condition can be * due to simply not being in an active scan which * can occur if we send the scan abort before the * microcode has notified us that a scan is completed. */ IWL_DEBUG_SCAN(mvm, "SCAN OFFLOAD ABORT ret %d.\n", status); ret = -ENOENT; } return ret; } int iwl_mvm_sched_scan_stop(struct iwl_mvm *mvm) { int ret; struct iwl_notification_wait wait_scan_done; static const u8 scan_done_notif[] = { SCAN_OFFLOAD_COMPLETE, }; lockdep_assert_held(&mvm->mutex); if (mvm->scan_status != IWL_MVM_SCAN_SCHED) { IWL_DEBUG_SCAN(mvm, "No offloaded scan to stop\n"); return 0; } iwl_init_notification_wait(&mvm->notif_wait, &wait_scan_done, scan_done_notif, ARRAY_SIZE(scan_done_notif), NULL, NULL); ret = iwl_mvm_send_sched_scan_abort(mvm); if (ret) { IWL_DEBUG_SCAN(mvm, "Send stop offload scan failed %d\n", ret); iwl_remove_notification(&mvm->notif_wait, &wait_scan_done); return ret; } IWL_DEBUG_SCAN(mvm, "Successfully sent stop offload scan\n"); ret = iwl_wait_notification(&mvm->notif_wait, &wait_scan_done, 1 * HZ); if (ret) return ret; /* * Clear the scan status so the next scan requests will succeed. This * also ensures the Rx handler doesn't do anything, as the scan was * stopped from above. */ mvm->scan_status = IWL_MVM_SCAN_NONE; return 0; }