2 Copyright (C) 2004 - 2009 Ivo van Doorn <IvDoorn@gmail.com>
3 <http://rt2x00.serialmonkey.com>
5 This program is free software; you can redistribute it and/or modify
6 it under the terms of the GNU General Public License as published by
7 the Free Software Foundation; either version 2 of the License, or
8 (at your option) any later version.
10 This program is distributed in the hope that it will be useful,
11 but WITHOUT ANY WARRANTY; without even the implied warranty of
12 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 GNU General Public License for more details.
15 You should have received a copy of the GNU General Public License
16 along with this program; if not, write to the
17 Free Software Foundation, Inc.,
18 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
23 Abstract: rt2x00 generic pci device routines.
26 #include <linux/dma-mapping.h>
27 #include <linux/kernel.h>
28 #include <linux/module.h>
29 #include <linux/pci.h>
30 #include <linux/slab.h>
33 #include "rt2x00pci.h"
38 int rt2x00pci_regbusy_read(struct rt2x00_dev
*rt2x00dev
,
39 const unsigned int offset
,
40 const struct rt2x00_field32 field
,
45 if (!test_bit(DEVICE_STATE_PRESENT
, &rt2x00dev
->flags
))
48 for (i
= 0; i
< REGISTER_BUSY_COUNT
; i
++) {
49 rt2x00pci_register_read(rt2x00dev
, offset
, reg
);
50 if (!rt2x00_get_field32(*reg
, field
))
52 udelay(REGISTER_BUSY_DELAY
);
55 ERROR(rt2x00dev
, "Indirect register access failed: "
56 "offset=0x%.08x, value=0x%.08x\n", offset
, *reg
);
61 EXPORT_SYMBOL_GPL(rt2x00pci_regbusy_read
);
66 int rt2x00pci_write_tx_data(struct queue_entry
*entry
,
67 struct txentry_desc
*txdesc
)
69 struct rt2x00_dev
*rt2x00dev
= entry
->queue
->rt2x00dev
;
70 struct queue_entry_priv_pci
*entry_priv
= entry
->priv_data
;
71 struct skb_frame_desc
*skbdesc
;
74 * This should not happen, we already checked the entry
75 * was ours. When the hardware disagrees there has been
78 if (unlikely(rt2x00dev
->ops
->lib
->get_entry_state(entry
))) {
80 "Corrupt queue %d, accessing entry which is not ours.\n"
81 "Please file bug report to %s.\n",
82 entry
->queue
->qid
, DRV_PROJECT
);
87 * Fill in skb descriptor
89 skbdesc
= get_skb_frame_desc(entry
->skb
);
90 skbdesc
->desc
= entry_priv
->desc
;
91 skbdesc
->desc_len
= entry
->queue
->desc_size
;
95 EXPORT_SYMBOL_GPL(rt2x00pci_write_tx_data
);
98 * TX/RX data handlers.
100 void rt2x00pci_rxdone(struct rt2x00_dev
*rt2x00dev
)
102 struct data_queue
*queue
= rt2x00dev
->rx
;
103 struct queue_entry
*entry
;
104 struct queue_entry_priv_pci
*entry_priv
;
105 struct skb_frame_desc
*skbdesc
;
108 entry
= rt2x00queue_get_entry(queue
, Q_INDEX
);
109 entry_priv
= entry
->priv_data
;
111 if (rt2x00dev
->ops
->lib
->get_entry_state(entry
))
115 * Fill in desc fields of the skb descriptor
117 skbdesc
= get_skb_frame_desc(entry
->skb
);
118 skbdesc
->desc
= entry_priv
->desc
;
119 skbdesc
->desc_len
= entry
->queue
->desc_size
;
122 * Send the frame to rt2x00lib for further processing.
124 rt2x00lib_rxdone(rt2x00dev
, entry
);
127 EXPORT_SYMBOL_GPL(rt2x00pci_rxdone
);
130 * Device initialization handlers.
132 static int rt2x00pci_alloc_queue_dma(struct rt2x00_dev
*rt2x00dev
,
133 struct data_queue
*queue
)
135 struct queue_entry_priv_pci
*entry_priv
;
141 * Allocate DMA memory for descriptor and buffer.
143 addr
= dma_alloc_coherent(rt2x00dev
->dev
,
144 queue
->limit
* queue
->desc_size
,
145 &dma
, GFP_KERNEL
| GFP_DMA
);
149 memset(addr
, 0, queue
->limit
* queue
->desc_size
);
152 * Initialize all queue entries to contain valid addresses.
154 for (i
= 0; i
< queue
->limit
; i
++) {
155 entry_priv
= queue
->entries
[i
].priv_data
;
156 entry_priv
->desc
= addr
+ i
* queue
->desc_size
;
157 entry_priv
->desc_dma
= dma
+ i
* queue
->desc_size
;
163 static void rt2x00pci_free_queue_dma(struct rt2x00_dev
*rt2x00dev
,
164 struct data_queue
*queue
)
166 struct queue_entry_priv_pci
*entry_priv
=
167 queue
->entries
[0].priv_data
;
169 if (entry_priv
->desc
)
170 dma_free_coherent(rt2x00dev
->dev
,
171 queue
->limit
* queue
->desc_size
,
172 entry_priv
->desc
, entry_priv
->desc_dma
);
173 entry_priv
->desc
= NULL
;
176 int rt2x00pci_initialize(struct rt2x00_dev
*rt2x00dev
)
178 struct data_queue
*queue
;
184 queue_for_each(rt2x00dev
, queue
) {
185 status
= rt2x00pci_alloc_queue_dma(rt2x00dev
, queue
);
191 * Register interrupt handler.
193 status
= request_irq(rt2x00dev
->irq
, rt2x00dev
->ops
->lib
->irq_handler
,
194 IRQF_SHARED
, rt2x00dev
->name
, rt2x00dev
);
196 ERROR(rt2x00dev
, "IRQ %d allocation failed (error %d).\n",
197 rt2x00dev
->irq
, status
);
204 queue_for_each(rt2x00dev
, queue
)
205 rt2x00pci_free_queue_dma(rt2x00dev
, queue
);
209 EXPORT_SYMBOL_GPL(rt2x00pci_initialize
);
211 void rt2x00pci_uninitialize(struct rt2x00_dev
*rt2x00dev
)
213 struct data_queue
*queue
;
218 free_irq(to_pci_dev(rt2x00dev
->dev
)->irq
, rt2x00dev
);
223 queue_for_each(rt2x00dev
, queue
)
224 rt2x00pci_free_queue_dma(rt2x00dev
, queue
);
226 EXPORT_SYMBOL_GPL(rt2x00pci_uninitialize
);
229 * PCI driver handlers.
231 static void rt2x00pci_free_reg(struct rt2x00_dev
*rt2x00dev
)
233 kfree(rt2x00dev
->rf
);
234 rt2x00dev
->rf
= NULL
;
236 kfree(rt2x00dev
->eeprom
);
237 rt2x00dev
->eeprom
= NULL
;
239 if (rt2x00dev
->csr
.base
) {
240 iounmap(rt2x00dev
->csr
.base
);
241 rt2x00dev
->csr
.base
= NULL
;
245 static int rt2x00pci_alloc_reg(struct rt2x00_dev
*rt2x00dev
)
247 struct pci_dev
*pci_dev
= to_pci_dev(rt2x00dev
->dev
);
249 rt2x00dev
->csr
.base
= pci_ioremap_bar(pci_dev
, 0);
250 if (!rt2x00dev
->csr
.base
)
253 rt2x00dev
->eeprom
= kzalloc(rt2x00dev
->ops
->eeprom_size
, GFP_KERNEL
);
254 if (!rt2x00dev
->eeprom
)
257 rt2x00dev
->rf
= kzalloc(rt2x00dev
->ops
->rf_size
, GFP_KERNEL
);
264 ERROR_PROBE("Failed to allocate registers.\n");
266 rt2x00pci_free_reg(rt2x00dev
);
271 int rt2x00pci_probe(struct pci_dev
*pci_dev
, const struct pci_device_id
*id
)
273 struct rt2x00_ops
*ops
= (struct rt2x00_ops
*)id
->driver_data
;
274 struct ieee80211_hw
*hw
;
275 struct rt2x00_dev
*rt2x00dev
;
278 retval
= pci_request_regions(pci_dev
, pci_name(pci_dev
));
280 ERROR_PROBE("PCI request regions failed.\n");
284 retval
= pci_enable_device(pci_dev
);
286 ERROR_PROBE("Enable device failed.\n");
287 goto exit_release_regions
;
290 pci_set_master(pci_dev
);
292 if (pci_set_mwi(pci_dev
))
293 ERROR_PROBE("MWI not available.\n");
295 if (dma_set_mask(&pci_dev
->dev
, DMA_BIT_MASK(32))) {
296 ERROR_PROBE("PCI DMA not supported.\n");
298 goto exit_disable_device
;
301 hw
= ieee80211_alloc_hw(sizeof(struct rt2x00_dev
), ops
->hw
);
303 ERROR_PROBE("Failed to allocate hardware.\n");
305 goto exit_disable_device
;
308 pci_set_drvdata(pci_dev
, hw
);
310 rt2x00dev
= hw
->priv
;
311 rt2x00dev
->dev
= &pci_dev
->dev
;
312 rt2x00dev
->ops
= ops
;
314 rt2x00dev
->irq
= pci_dev
->irq
;
315 rt2x00dev
->name
= pci_name(pci_dev
);
317 rt2x00_set_chip_intf(rt2x00dev
, RT2X00_CHIP_INTF_PCI
);
319 retval
= rt2x00pci_alloc_reg(rt2x00dev
);
321 goto exit_free_device
;
323 retval
= rt2x00lib_probe_dev(rt2x00dev
);
330 rt2x00pci_free_reg(rt2x00dev
);
333 ieee80211_free_hw(hw
);
336 if (retval
!= -EBUSY
)
337 pci_disable_device(pci_dev
);
339 exit_release_regions
:
340 pci_release_regions(pci_dev
);
342 pci_set_drvdata(pci_dev
, NULL
);
346 EXPORT_SYMBOL_GPL(rt2x00pci_probe
);
348 void rt2x00pci_remove(struct pci_dev
*pci_dev
)
350 struct ieee80211_hw
*hw
= pci_get_drvdata(pci_dev
);
351 struct rt2x00_dev
*rt2x00dev
= hw
->priv
;
354 * Free all allocated data.
356 rt2x00lib_remove_dev(rt2x00dev
);
357 rt2x00pci_free_reg(rt2x00dev
);
358 ieee80211_free_hw(hw
);
361 * Free the PCI device data.
363 pci_set_drvdata(pci_dev
, NULL
);
364 pci_disable_device(pci_dev
);
365 pci_release_regions(pci_dev
);
367 EXPORT_SYMBOL_GPL(rt2x00pci_remove
);
370 int rt2x00pci_suspend(struct pci_dev
*pci_dev
, pm_message_t state
)
372 struct ieee80211_hw
*hw
= pci_get_drvdata(pci_dev
);
373 struct rt2x00_dev
*rt2x00dev
= hw
->priv
;
376 retval
= rt2x00lib_suspend(rt2x00dev
, state
);
380 pci_save_state(pci_dev
);
381 pci_disable_device(pci_dev
);
382 return pci_set_power_state(pci_dev
, pci_choose_state(pci_dev
, state
));
384 EXPORT_SYMBOL_GPL(rt2x00pci_suspend
);
386 int rt2x00pci_resume(struct pci_dev
*pci_dev
)
388 struct ieee80211_hw
*hw
= pci_get_drvdata(pci_dev
);
389 struct rt2x00_dev
*rt2x00dev
= hw
->priv
;
391 if (pci_set_power_state(pci_dev
, PCI_D0
) ||
392 pci_enable_device(pci_dev
) ||
393 pci_restore_state(pci_dev
)) {
394 ERROR(rt2x00dev
, "Failed to resume device.\n");
398 return rt2x00lib_resume(rt2x00dev
);
400 EXPORT_SYMBOL_GPL(rt2x00pci_resume
);
401 #endif /* CONFIG_PM */
404 * rt2x00pci module information.
406 MODULE_AUTHOR(DRV_PROJECT
);
407 MODULE_VERSION(DRV_VERSION
);
408 MODULE_DESCRIPTION("rt2x00 pci library");
409 MODULE_LICENSE("GPL");