671 lines
15 KiB
C
671 lines
15 KiB
C
// SPDX-License-Identifier: BSD-3-Clause-Clear
|
|
/*
|
|
* Copyright (c) 2018-2019 The Linux Foundation. All rights reserved.
|
|
* Copyright (c) 2021-2022 Qualcomm Innovation Center, Inc. All rights reserved.
|
|
*/
|
|
|
|
#include "core.h"
|
|
#include "peer.h"
|
|
#include "debug.h"
|
|
|
|
static struct ath11k_peer *ath11k_peer_find_list_by_id(struct ath11k_base *ab,
|
|
int peer_id)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
|
|
lockdep_assert_held(&ab->base_lock);
|
|
|
|
list_for_each_entry(peer, &ab->peers, list) {
|
|
if (peer->peer_id != peer_id)
|
|
continue;
|
|
|
|
return peer;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct ath11k_peer *ath11k_peer_find(struct ath11k_base *ab, int vdev_id,
|
|
const u8 *addr)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
|
|
lockdep_assert_held(&ab->base_lock);
|
|
|
|
list_for_each_entry(peer, &ab->peers, list) {
|
|
if (peer->vdev_id != vdev_id)
|
|
continue;
|
|
if (!ether_addr_equal(peer->addr, addr))
|
|
continue;
|
|
|
|
return peer;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct ath11k_peer *ath11k_peer_find_by_addr(struct ath11k_base *ab,
|
|
const u8 *addr)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
|
|
lockdep_assert_held(&ab->base_lock);
|
|
|
|
if (!ab->rhead_peer_addr)
|
|
return NULL;
|
|
|
|
peer = rhashtable_lookup_fast(ab->rhead_peer_addr, addr,
|
|
ab->rhash_peer_addr_param);
|
|
|
|
return peer;
|
|
}
|
|
|
|
struct ath11k_peer *ath11k_peer_find_by_id(struct ath11k_base *ab,
|
|
int peer_id)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
|
|
lockdep_assert_held(&ab->base_lock);
|
|
|
|
if (!ab->rhead_peer_id)
|
|
return NULL;
|
|
|
|
peer = rhashtable_lookup_fast(ab->rhead_peer_id, &peer_id,
|
|
ab->rhash_peer_id_param);
|
|
|
|
return peer;
|
|
}
|
|
|
|
struct ath11k_peer *ath11k_peer_find_by_vdev_id(struct ath11k_base *ab,
|
|
int vdev_id)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
|
|
spin_lock_bh(&ab->base_lock);
|
|
|
|
list_for_each_entry(peer, &ab->peers, list) {
|
|
if (vdev_id == peer->vdev_id) {
|
|
spin_unlock_bh(&ab->base_lock);
|
|
return peer;
|
|
}
|
|
}
|
|
spin_unlock_bh(&ab->base_lock);
|
|
return NULL;
|
|
}
|
|
|
|
void ath11k_peer_unmap_event(struct ath11k_base *ab, u16 peer_id)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
|
|
spin_lock_bh(&ab->base_lock);
|
|
|
|
peer = ath11k_peer_find_list_by_id(ab, peer_id);
|
|
if (!peer) {
|
|
ath11k_warn(ab, "peer-unmap-event: unknown peer id %d\n",
|
|
peer_id);
|
|
goto exit;
|
|
}
|
|
|
|
ath11k_dbg(ab, ATH11K_DBG_DP_HTT, "htt peer unmap vdev %d peer %pM id %d\n",
|
|
peer->vdev_id, peer->addr, peer_id);
|
|
|
|
list_del(&peer->list);
|
|
kfree(peer);
|
|
wake_up(&ab->peer_mapping_wq);
|
|
|
|
exit:
|
|
spin_unlock_bh(&ab->base_lock);
|
|
}
|
|
|
|
void ath11k_peer_map_event(struct ath11k_base *ab, u8 vdev_id, u16 peer_id,
|
|
u8 *mac_addr, u16 ast_hash, u16 hw_peer_id)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
|
|
spin_lock_bh(&ab->base_lock);
|
|
peer = ath11k_peer_find(ab, vdev_id, mac_addr);
|
|
if (!peer) {
|
|
peer = kzalloc(sizeof(*peer), GFP_ATOMIC);
|
|
if (!peer)
|
|
goto exit;
|
|
|
|
peer->vdev_id = vdev_id;
|
|
peer->peer_id = peer_id;
|
|
peer->ast_hash = ast_hash;
|
|
peer->hw_peer_id = hw_peer_id;
|
|
ether_addr_copy(peer->addr, mac_addr);
|
|
list_add(&peer->list, &ab->peers);
|
|
wake_up(&ab->peer_mapping_wq);
|
|
}
|
|
|
|
ath11k_dbg(ab, ATH11K_DBG_DP_HTT, "htt peer map vdev %d peer %pM id %d\n",
|
|
vdev_id, mac_addr, peer_id);
|
|
|
|
exit:
|
|
spin_unlock_bh(&ab->base_lock);
|
|
}
|
|
|
|
static int ath11k_wait_for_peer_common(struct ath11k_base *ab, int vdev_id,
|
|
const u8 *addr, bool expect_mapped)
|
|
{
|
|
int ret;
|
|
|
|
ret = wait_event_timeout(ab->peer_mapping_wq, ({
|
|
bool mapped;
|
|
|
|
spin_lock_bh(&ab->base_lock);
|
|
mapped = !!ath11k_peer_find(ab, vdev_id, addr);
|
|
spin_unlock_bh(&ab->base_lock);
|
|
|
|
(mapped == expect_mapped ||
|
|
test_bit(ATH11K_FLAG_CRASH_FLUSH, &ab->dev_flags));
|
|
}), 3 * HZ);
|
|
|
|
if (ret <= 0)
|
|
return -ETIMEDOUT;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int ath11k_peer_rhash_insert(struct ath11k_base *ab,
|
|
struct rhashtable *rtbl,
|
|
struct rhash_head *rhead,
|
|
struct rhashtable_params *params,
|
|
void *key)
|
|
{
|
|
struct ath11k_peer *tmp;
|
|
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
tmp = rhashtable_lookup_get_insert_fast(rtbl, rhead, *params);
|
|
|
|
if (!tmp)
|
|
return 0;
|
|
else if (IS_ERR(tmp))
|
|
return PTR_ERR(tmp);
|
|
else
|
|
return -EEXIST;
|
|
}
|
|
|
|
static inline int ath11k_peer_rhash_remove(struct ath11k_base *ab,
|
|
struct rhashtable *rtbl,
|
|
struct rhash_head *rhead,
|
|
struct rhashtable_params *params)
|
|
{
|
|
int ret;
|
|
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
ret = rhashtable_remove_fast(rtbl, rhead, *params);
|
|
if (ret && ret != -ENOENT)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath11k_peer_rhash_add(struct ath11k_base *ab, struct ath11k_peer *peer)
|
|
{
|
|
int ret;
|
|
|
|
lockdep_assert_held(&ab->base_lock);
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
if (!ab->rhead_peer_id || !ab->rhead_peer_addr)
|
|
return -EPERM;
|
|
|
|
ret = ath11k_peer_rhash_insert(ab, ab->rhead_peer_id, &peer->rhash_id,
|
|
&ab->rhash_peer_id_param, &peer->peer_id);
|
|
if (ret) {
|
|
ath11k_warn(ab, "failed to add peer %pM with id %d in rhash_id ret %d\n",
|
|
peer->addr, peer->peer_id, ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = ath11k_peer_rhash_insert(ab, ab->rhead_peer_addr, &peer->rhash_addr,
|
|
&ab->rhash_peer_addr_param, &peer->addr);
|
|
if (ret) {
|
|
ath11k_warn(ab, "failed to add peer %pM with id %d in rhash_addr ret %d\n",
|
|
peer->addr, peer->peer_id, ret);
|
|
goto err_clean;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_clean:
|
|
ath11k_peer_rhash_remove(ab, ab->rhead_peer_id, &peer->rhash_id,
|
|
&ab->rhash_peer_id_param);
|
|
return ret;
|
|
}
|
|
|
|
void ath11k_peer_cleanup(struct ath11k *ar, u32 vdev_id)
|
|
{
|
|
struct ath11k_peer *peer, *tmp;
|
|
struct ath11k_base *ab = ar->ab;
|
|
|
|
lockdep_assert_held(&ar->conf_mutex);
|
|
|
|
mutex_lock(&ab->tbl_mtx_lock);
|
|
spin_lock_bh(&ab->base_lock);
|
|
list_for_each_entry_safe(peer, tmp, &ab->peers, list) {
|
|
if (peer->vdev_id != vdev_id)
|
|
continue;
|
|
|
|
ath11k_warn(ab, "removing stale peer %pM from vdev_id %d\n",
|
|
peer->addr, vdev_id);
|
|
|
|
ath11k_peer_rhash_delete(ab, peer);
|
|
list_del(&peer->list);
|
|
kfree(peer);
|
|
ar->num_peers--;
|
|
}
|
|
|
|
spin_unlock_bh(&ab->base_lock);
|
|
mutex_unlock(&ab->tbl_mtx_lock);
|
|
}
|
|
|
|
static int ath11k_wait_for_peer_deleted(struct ath11k *ar, int vdev_id, const u8 *addr)
|
|
{
|
|
return ath11k_wait_for_peer_common(ar->ab, vdev_id, addr, false);
|
|
}
|
|
|
|
int ath11k_wait_for_peer_delete_done(struct ath11k *ar, u32 vdev_id,
|
|
const u8 *addr)
|
|
{
|
|
int ret;
|
|
unsigned long time_left;
|
|
|
|
ret = ath11k_wait_for_peer_deleted(ar, vdev_id, addr);
|
|
if (ret) {
|
|
ath11k_warn(ar->ab, "failed wait for peer deleted");
|
|
return ret;
|
|
}
|
|
|
|
time_left = wait_for_completion_timeout(&ar->peer_delete_done,
|
|
3 * HZ);
|
|
if (time_left == 0) {
|
|
ath11k_warn(ar->ab, "Timeout in receiving peer delete response\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __ath11k_peer_delete(struct ath11k *ar, u32 vdev_id, const u8 *addr)
|
|
{
|
|
int ret;
|
|
struct ath11k_peer *peer;
|
|
struct ath11k_base *ab = ar->ab;
|
|
|
|
lockdep_assert_held(&ar->conf_mutex);
|
|
|
|
mutex_lock(&ab->tbl_mtx_lock);
|
|
spin_lock_bh(&ab->base_lock);
|
|
|
|
peer = ath11k_peer_find_by_addr(ab, addr);
|
|
/* Check if the found peer is what we want to remove.
|
|
* While the sta is transitioning to another band we may
|
|
* have 2 peer with the same addr assigned to different
|
|
* vdev_id. Make sure we are deleting the correct peer.
|
|
*/
|
|
if (peer && peer->vdev_id == vdev_id)
|
|
ath11k_peer_rhash_delete(ab, peer);
|
|
|
|
/* Fallback to peer list search if the correct peer can't be found.
|
|
* Skip the deletion of the peer from the rhash since it has already
|
|
* been deleted in peer add.
|
|
*/
|
|
if (!peer)
|
|
peer = ath11k_peer_find(ab, vdev_id, addr);
|
|
|
|
if (!peer) {
|
|
spin_unlock_bh(&ab->base_lock);
|
|
mutex_unlock(&ab->tbl_mtx_lock);
|
|
|
|
ath11k_warn(ab,
|
|
"failed to find peer vdev_id %d addr %pM in delete\n",
|
|
vdev_id, addr);
|
|
return -EINVAL;
|
|
}
|
|
|
|
spin_unlock_bh(&ab->base_lock);
|
|
mutex_unlock(&ab->tbl_mtx_lock);
|
|
|
|
reinit_completion(&ar->peer_delete_done);
|
|
|
|
ret = ath11k_wmi_send_peer_delete_cmd(ar, addr, vdev_id);
|
|
if (ret) {
|
|
ath11k_warn(ab,
|
|
"failed to delete peer vdev_id %d addr %pM ret %d\n",
|
|
vdev_id, addr, ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = ath11k_wait_for_peer_delete_done(ar, vdev_id, addr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ath11k_peer_delete(struct ath11k *ar, u32 vdev_id, u8 *addr)
|
|
{
|
|
int ret;
|
|
|
|
lockdep_assert_held(&ar->conf_mutex);
|
|
|
|
ret = __ath11k_peer_delete(ar, vdev_id, addr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ar->num_peers--;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath11k_wait_for_peer_created(struct ath11k *ar, int vdev_id, const u8 *addr)
|
|
{
|
|
return ath11k_wait_for_peer_common(ar->ab, vdev_id, addr, true);
|
|
}
|
|
|
|
int ath11k_peer_create(struct ath11k *ar, struct ath11k_vif *arvif,
|
|
struct ieee80211_sta *sta, struct peer_create_params *param)
|
|
{
|
|
struct ath11k_peer *peer;
|
|
struct ath11k_sta *arsta;
|
|
int ret, fbret;
|
|
|
|
lockdep_assert_held(&ar->conf_mutex);
|
|
|
|
if (ar->num_peers > (ar->max_num_peers - 1)) {
|
|
ath11k_warn(ar->ab,
|
|
"failed to create peer due to insufficient peer entry resource in firmware\n");
|
|
return -ENOBUFS;
|
|
}
|
|
|
|
mutex_lock(&ar->ab->tbl_mtx_lock);
|
|
spin_lock_bh(&ar->ab->base_lock);
|
|
peer = ath11k_peer_find_by_addr(ar->ab, param->peer_addr);
|
|
if (peer) {
|
|
if (peer->vdev_id == param->vdev_id) {
|
|
spin_unlock_bh(&ar->ab->base_lock);
|
|
mutex_unlock(&ar->ab->tbl_mtx_lock);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Assume sta is transitioning to another band.
|
|
* Remove here the peer from rhash.
|
|
*/
|
|
ath11k_peer_rhash_delete(ar->ab, peer);
|
|
}
|
|
spin_unlock_bh(&ar->ab->base_lock);
|
|
mutex_unlock(&ar->ab->tbl_mtx_lock);
|
|
|
|
ret = ath11k_wmi_send_peer_create_cmd(ar, param);
|
|
if (ret) {
|
|
ath11k_warn(ar->ab,
|
|
"failed to send peer create vdev_id %d ret %d\n",
|
|
param->vdev_id, ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = ath11k_wait_for_peer_created(ar, param->vdev_id,
|
|
param->peer_addr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
mutex_lock(&ar->ab->tbl_mtx_lock);
|
|
spin_lock_bh(&ar->ab->base_lock);
|
|
|
|
peer = ath11k_peer_find(ar->ab, param->vdev_id, param->peer_addr);
|
|
if (!peer) {
|
|
spin_unlock_bh(&ar->ab->base_lock);
|
|
mutex_unlock(&ar->ab->tbl_mtx_lock);
|
|
ath11k_warn(ar->ab, "failed to find peer %pM on vdev %i after creation\n",
|
|
param->peer_addr, param->vdev_id);
|
|
|
|
ret = -ENOENT;
|
|
goto cleanup;
|
|
}
|
|
|
|
ret = ath11k_peer_rhash_add(ar->ab, peer);
|
|
if (ret) {
|
|
spin_unlock_bh(&ar->ab->base_lock);
|
|
mutex_unlock(&ar->ab->tbl_mtx_lock);
|
|
goto cleanup;
|
|
}
|
|
|
|
peer->pdev_idx = ar->pdev_idx;
|
|
peer->sta = sta;
|
|
|
|
if (arvif->vif->type == NL80211_IFTYPE_STATION) {
|
|
arvif->ast_hash = peer->ast_hash;
|
|
arvif->ast_idx = peer->hw_peer_id;
|
|
}
|
|
|
|
peer->sec_type = HAL_ENCRYPT_TYPE_OPEN;
|
|
peer->sec_type_grp = HAL_ENCRYPT_TYPE_OPEN;
|
|
|
|
if (sta) {
|
|
arsta = (struct ath11k_sta *)sta->drv_priv;
|
|
arsta->tcl_metadata |= FIELD_PREP(HTT_TCL_META_DATA_TYPE, 0) |
|
|
FIELD_PREP(HTT_TCL_META_DATA_PEER_ID,
|
|
peer->peer_id);
|
|
|
|
/* set HTT extension valid bit to 0 by default */
|
|
arsta->tcl_metadata &= ~HTT_TCL_META_DATA_VALID_HTT;
|
|
}
|
|
|
|
ar->num_peers++;
|
|
|
|
spin_unlock_bh(&ar->ab->base_lock);
|
|
mutex_unlock(&ar->ab->tbl_mtx_lock);
|
|
|
|
return 0;
|
|
|
|
cleanup:
|
|
fbret = __ath11k_peer_delete(ar, param->vdev_id, param->peer_addr);
|
|
if (fbret)
|
|
ath11k_warn(ar->ab, "failed peer %pM delete vdev_id %d fallback ret %d\n",
|
|
param->peer_addr, param->vdev_id, fbret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int ath11k_peer_rhash_delete(struct ath11k_base *ab, struct ath11k_peer *peer)
|
|
{
|
|
int ret;
|
|
|
|
lockdep_assert_held(&ab->base_lock);
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
if (!ab->rhead_peer_id || !ab->rhead_peer_addr)
|
|
return -EPERM;
|
|
|
|
ret = ath11k_peer_rhash_remove(ab, ab->rhead_peer_addr, &peer->rhash_addr,
|
|
&ab->rhash_peer_addr_param);
|
|
if (ret) {
|
|
ath11k_warn(ab, "failed to remove peer %pM id %d in rhash_addr ret %d\n",
|
|
peer->addr, peer->peer_id, ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = ath11k_peer_rhash_remove(ab, ab->rhead_peer_id, &peer->rhash_id,
|
|
&ab->rhash_peer_id_param);
|
|
if (ret) {
|
|
ath11k_warn(ab, "failed to remove peer %pM id %d in rhash_id ret %d\n",
|
|
peer->addr, peer->peer_id, ret);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath11k_peer_rhash_id_tbl_init(struct ath11k_base *ab)
|
|
{
|
|
struct rhashtable_params *param;
|
|
struct rhashtable *rhash_id_tbl;
|
|
int ret;
|
|
size_t size;
|
|
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
if (ab->rhead_peer_id)
|
|
return 0;
|
|
|
|
size = sizeof(*ab->rhead_peer_id);
|
|
rhash_id_tbl = kzalloc(size, GFP_KERNEL);
|
|
if (!rhash_id_tbl) {
|
|
ath11k_warn(ab, "failed to init rhash id table due to no mem (size %zu)\n",
|
|
size);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
param = &ab->rhash_peer_id_param;
|
|
|
|
param->key_offset = offsetof(struct ath11k_peer, peer_id);
|
|
param->head_offset = offsetof(struct ath11k_peer, rhash_id);
|
|
param->key_len = sizeof_field(struct ath11k_peer, peer_id);
|
|
param->automatic_shrinking = true;
|
|
param->nelem_hint = ab->num_radios * TARGET_NUM_PEERS_PDEV(ab);
|
|
|
|
ret = rhashtable_init(rhash_id_tbl, param);
|
|
if (ret) {
|
|
ath11k_warn(ab, "failed to init peer id rhash table %d\n", ret);
|
|
goto err_free;
|
|
}
|
|
|
|
spin_lock_bh(&ab->base_lock);
|
|
|
|
if (!ab->rhead_peer_id) {
|
|
ab->rhead_peer_id = rhash_id_tbl;
|
|
} else {
|
|
spin_unlock_bh(&ab->base_lock);
|
|
goto cleanup_tbl;
|
|
}
|
|
|
|
spin_unlock_bh(&ab->base_lock);
|
|
|
|
return 0;
|
|
|
|
cleanup_tbl:
|
|
rhashtable_destroy(rhash_id_tbl);
|
|
err_free:
|
|
kfree(rhash_id_tbl);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ath11k_peer_rhash_addr_tbl_init(struct ath11k_base *ab)
|
|
{
|
|
struct rhashtable_params *param;
|
|
struct rhashtable *rhash_addr_tbl;
|
|
int ret;
|
|
size_t size;
|
|
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
if (ab->rhead_peer_addr)
|
|
return 0;
|
|
|
|
size = sizeof(*ab->rhead_peer_addr);
|
|
rhash_addr_tbl = kzalloc(size, GFP_KERNEL);
|
|
if (!rhash_addr_tbl) {
|
|
ath11k_warn(ab, "failed to init rhash addr table due to no mem (size %zu)\n",
|
|
size);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
param = &ab->rhash_peer_addr_param;
|
|
|
|
param->key_offset = offsetof(struct ath11k_peer, addr);
|
|
param->head_offset = offsetof(struct ath11k_peer, rhash_addr);
|
|
param->key_len = sizeof_field(struct ath11k_peer, addr);
|
|
param->automatic_shrinking = true;
|
|
param->nelem_hint = ab->num_radios * TARGET_NUM_PEERS_PDEV(ab);
|
|
|
|
ret = rhashtable_init(rhash_addr_tbl, param);
|
|
if (ret) {
|
|
ath11k_warn(ab, "failed to init peer addr rhash table %d\n", ret);
|
|
goto err_free;
|
|
}
|
|
|
|
spin_lock_bh(&ab->base_lock);
|
|
|
|
if (!ab->rhead_peer_addr) {
|
|
ab->rhead_peer_addr = rhash_addr_tbl;
|
|
} else {
|
|
spin_unlock_bh(&ab->base_lock);
|
|
goto cleanup_tbl;
|
|
}
|
|
|
|
spin_unlock_bh(&ab->base_lock);
|
|
|
|
return 0;
|
|
|
|
cleanup_tbl:
|
|
rhashtable_destroy(rhash_addr_tbl);
|
|
err_free:
|
|
kfree(rhash_addr_tbl);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline void ath11k_peer_rhash_id_tbl_destroy(struct ath11k_base *ab)
|
|
{
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
if (!ab->rhead_peer_id)
|
|
return;
|
|
|
|
rhashtable_destroy(ab->rhead_peer_id);
|
|
kfree(ab->rhead_peer_id);
|
|
ab->rhead_peer_id = NULL;
|
|
}
|
|
|
|
static inline void ath11k_peer_rhash_addr_tbl_destroy(struct ath11k_base *ab)
|
|
{
|
|
lockdep_assert_held(&ab->tbl_mtx_lock);
|
|
|
|
if (!ab->rhead_peer_addr)
|
|
return;
|
|
|
|
rhashtable_destroy(ab->rhead_peer_addr);
|
|
kfree(ab->rhead_peer_addr);
|
|
ab->rhead_peer_addr = NULL;
|
|
}
|
|
|
|
int ath11k_peer_rhash_tbl_init(struct ath11k_base *ab)
|
|
{
|
|
int ret;
|
|
|
|
mutex_lock(&ab->tbl_mtx_lock);
|
|
|
|
ret = ath11k_peer_rhash_id_tbl_init(ab);
|
|
if (ret)
|
|
goto out;
|
|
|
|
ret = ath11k_peer_rhash_addr_tbl_init(ab);
|
|
if (ret)
|
|
goto cleanup_tbl;
|
|
|
|
mutex_unlock(&ab->tbl_mtx_lock);
|
|
|
|
return 0;
|
|
|
|
cleanup_tbl:
|
|
ath11k_peer_rhash_id_tbl_destroy(ab);
|
|
out:
|
|
mutex_unlock(&ab->tbl_mtx_lock);
|
|
return ret;
|
|
}
|
|
|
|
void ath11k_peer_rhash_tbl_destroy(struct ath11k_base *ab)
|
|
{
|
|
mutex_lock(&ab->tbl_mtx_lock);
|
|
|
|
ath11k_peer_rhash_addr_tbl_destroy(ab);
|
|
ath11k_peer_rhash_id_tbl_destroy(ab);
|
|
|
|
mutex_unlock(&ab->tbl_mtx_lock);
|
|
}
|