Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/net/mac802154/util.c
26282 views
1
// SPDX-License-Identifier: GPL-2.0-only
2
/*
3
*
4
* Authors:
5
* Alexander Aring <[email protected]>
6
*
7
* Based on: net/mac80211/util.c
8
*/
9
10
#include "ieee802154_i.h"
11
#include "driver-ops.h"
12
13
/* privid for wpan_phys to determine whether they belong to us or not */
14
const void *const mac802154_wpan_phy_privid = &mac802154_wpan_phy_privid;
15
16
/**
17
* ieee802154_wake_queue - wake ieee802154 queue
18
* @hw: main hardware object
19
*
20
* Tranceivers usually have either one transmit framebuffer or one framebuffer
21
* for both transmitting and receiving. Hence, the core currently only handles
22
* one frame at a time for each phy, which means we had to stop the queue to
23
* avoid new skb to come during the transmission. The queue then needs to be
24
* woken up after the operation.
25
*/
26
static void ieee802154_wake_queue(struct ieee802154_hw *hw)
27
{
28
struct ieee802154_local *local = hw_to_local(hw);
29
struct ieee802154_sub_if_data *sdata;
30
31
rcu_read_lock();
32
clear_bit(WPAN_PHY_FLAG_STATE_QUEUE_STOPPED, &local->phy->flags);
33
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
34
if (!sdata->dev)
35
continue;
36
37
netif_wake_queue(sdata->dev);
38
}
39
rcu_read_unlock();
40
}
41
42
/**
43
* ieee802154_stop_queue - stop ieee802154 queue
44
* @hw: main hardware object
45
*
46
* Tranceivers usually have either one transmit framebuffer or one framebuffer
47
* for both transmitting and receiving. Hence, the core currently only handles
48
* one frame at a time for each phy, which means we need to tell upper layers to
49
* stop giving us new skbs while we are busy with the transmitted one. The queue
50
* must then be stopped before transmitting.
51
*/
52
static void ieee802154_stop_queue(struct ieee802154_hw *hw)
53
{
54
struct ieee802154_local *local = hw_to_local(hw);
55
struct ieee802154_sub_if_data *sdata;
56
57
rcu_read_lock();
58
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
59
if (!sdata->dev)
60
continue;
61
62
netif_stop_queue(sdata->dev);
63
}
64
rcu_read_unlock();
65
}
66
67
void ieee802154_hold_queue(struct ieee802154_local *local)
68
{
69
unsigned long flags;
70
71
spin_lock_irqsave(&local->phy->queue_lock, flags);
72
if (!atomic_fetch_inc(&local->phy->hold_txs))
73
ieee802154_stop_queue(&local->hw);
74
spin_unlock_irqrestore(&local->phy->queue_lock, flags);
75
}
76
77
void ieee802154_release_queue(struct ieee802154_local *local)
78
{
79
unsigned long flags;
80
81
spin_lock_irqsave(&local->phy->queue_lock, flags);
82
if (atomic_dec_and_test(&local->phy->hold_txs))
83
ieee802154_wake_queue(&local->hw);
84
spin_unlock_irqrestore(&local->phy->queue_lock, flags);
85
}
86
87
void ieee802154_disable_queue(struct ieee802154_local *local)
88
{
89
struct ieee802154_sub_if_data *sdata;
90
91
rcu_read_lock();
92
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
93
if (!sdata->dev)
94
continue;
95
96
netif_tx_disable(sdata->dev);
97
}
98
rcu_read_unlock();
99
}
100
101
enum hrtimer_restart ieee802154_xmit_ifs_timer(struct hrtimer *timer)
102
{
103
struct ieee802154_local *local =
104
container_of(timer, struct ieee802154_local, ifs_timer);
105
106
ieee802154_release_queue(local);
107
108
return HRTIMER_NORESTART;
109
}
110
111
void ieee802154_xmit_complete(struct ieee802154_hw *hw, struct sk_buff *skb,
112
bool ifs_handling)
113
{
114
struct ieee802154_local *local = hw_to_local(hw);
115
116
local->tx_result = IEEE802154_SUCCESS;
117
118
if (ifs_handling) {
119
u8 max_sifs_size;
120
121
/* If transceiver sets CRC on his own we need to use lifs
122
* threshold len above 16 otherwise 18, because it's not
123
* part of skb->len.
124
*/
125
if (hw->flags & IEEE802154_HW_TX_OMIT_CKSUM)
126
max_sifs_size = IEEE802154_MAX_SIFS_FRAME_SIZE -
127
IEEE802154_FCS_LEN;
128
else
129
max_sifs_size = IEEE802154_MAX_SIFS_FRAME_SIZE;
130
131
if (skb->len > max_sifs_size)
132
hrtimer_start(&local->ifs_timer,
133
hw->phy->lifs_period * NSEC_PER_USEC,
134
HRTIMER_MODE_REL);
135
else
136
hrtimer_start(&local->ifs_timer,
137
hw->phy->sifs_period * NSEC_PER_USEC,
138
HRTIMER_MODE_REL);
139
} else {
140
ieee802154_release_queue(local);
141
}
142
143
dev_consume_skb_any(skb);
144
if (atomic_dec_and_test(&hw->phy->ongoing_txs))
145
wake_up(&hw->phy->sync_txq);
146
}
147
EXPORT_SYMBOL(ieee802154_xmit_complete);
148
149
void ieee802154_xmit_error(struct ieee802154_hw *hw, struct sk_buff *skb,
150
int reason)
151
{
152
struct ieee802154_local *local = hw_to_local(hw);
153
154
local->tx_result = reason;
155
ieee802154_release_queue(local);
156
dev_kfree_skb_any(skb);
157
if (atomic_dec_and_test(&hw->phy->ongoing_txs))
158
wake_up(&hw->phy->sync_txq);
159
}
160
EXPORT_SYMBOL(ieee802154_xmit_error);
161
162
void ieee802154_xmit_hw_error(struct ieee802154_hw *hw, struct sk_buff *skb)
163
{
164
ieee802154_xmit_error(hw, skb, IEEE802154_SYSTEM_ERROR);
165
}
166
EXPORT_SYMBOL(ieee802154_xmit_hw_error);
167
168
void ieee802154_stop_device(struct ieee802154_local *local)
169
{
170
flush_workqueue(local->workqueue);
171
hrtimer_cancel(&local->ifs_timer);
172
drv_stop(local);
173
}
174
175