1 // SPDX-License-Identifier: GPL-2.0+
3 * Copyright (c) 1996, 2003 VIA Networking Technologies, Inc.
6 * Purpose: driver entry for initial, open, close, tx and rx.
14 * vt6655_probe - module initial (insmod) driver entry
15 * vt6655_remove - module remove entry
16 * device_free_info - device structure resource free function
17 * device_print_info - print out resource
18 * device_rx_srv - rx service function
19 * device_alloc_rx_buf - rx buffer pre-allocated function
20 * device_free_rx_buf - free rx buffer function
21 * device_free_tx_buf - free tx buffer function
22 * device_init_rd0_ring - initial rd dma0 ring
23 * device_init_rd1_ring - initial rd dma1 ring
24 * device_init_td0_ring - initial tx dma0 ring buffer
25 * device_init_td1_ring - initial tx dma1 ring buffer
26 * device_init_registers - initial MAC & BBP & RF internal registers.
27 * device_init_rings - initial tx/rx ring buffer
28 * device_free_rings - free all allocated ring buffer
29 * device_tx_srv - tx interrupt service function
34 #include <linux/file.h>
44 #include <linux/delay.h>
45 #include <linux/kthread.h>
46 #include <linux/slab.h>
48 /*--------------------- Static Definitions -------------------------*/
50 * Define module options
52 MODULE_AUTHOR("VIA Networking Technologies, Inc., <lyndonchen@vntek.com.tw>");
53 MODULE_LICENSE("GPL");
54 MODULE_DESCRIPTION("VIA Networking Solomon-A/B/G Wireless LAN Adapter Driver");
56 #define DEVICE_PARAM(N, D)
58 #define RX_DESC_MIN0 16
59 #define RX_DESC_MAX0 128
60 #define RX_DESC_DEF0 32
61 DEVICE_PARAM(RxDescriptors0, "Number of receive descriptors0");
63 #define RX_DESC_MIN1 16
64 #define RX_DESC_MAX1 128
65 #define RX_DESC_DEF1 32
66 DEVICE_PARAM(RxDescriptors1, "Number of receive descriptors1");
68 #define TX_DESC_MIN0 16
69 #define TX_DESC_MAX0 128
70 #define TX_DESC_DEF0 32
71 DEVICE_PARAM(TxDescriptors0, "Number of transmit descriptors0");
73 #define TX_DESC_MIN1 16
74 #define TX_DESC_MAX1 128
75 #define TX_DESC_DEF1 64
76 DEVICE_PARAM(TxDescriptors1, "Number of transmit descriptors1");
78 #define INT_WORKS_DEF 20
79 #define INT_WORKS_MIN 10
80 #define INT_WORKS_MAX 64
82 DEVICE_PARAM(int_works, "Number of packets per interrupt services");
84 #define RTS_THRESH_DEF 2347
86 #define FRAG_THRESH_DEF 2346
88 #define SHORT_RETRY_MIN 0
89 #define SHORT_RETRY_MAX 31
90 #define SHORT_RETRY_DEF 8
92 DEVICE_PARAM(ShortRetryLimit, "Short frame retry limits");
94 #define LONG_RETRY_MIN 0
95 #define LONG_RETRY_MAX 15
96 #define LONG_RETRY_DEF 4
98 DEVICE_PARAM(LongRetryLimit, "long frame retry limits");
100 /* BasebandType[] baseband type selected
101 * 0: indicate 802.11a type
102 * 1: indicate 802.11b type
103 * 2: indicate 802.11g type
105 #define BBP_TYPE_MIN 0
106 #define BBP_TYPE_MAX 2
107 #define BBP_TYPE_DEF 2
109 DEVICE_PARAM(BasebandType, "baseband type");
112 * Static vars definitions
114 static const struct pci_device_id vt6655_pci_id_table[] = {
115 { PCI_VDEVICE(VIA, 0x3253) },
119 /*--------------------- Static Functions --------------------------*/
121 static int vt6655_probe(struct pci_dev *pcid, const struct pci_device_id *ent);
122 static void device_free_info(struct vnt_private *priv);
123 static void device_print_info(struct vnt_private *priv);
125 static int device_init_rd0_ring(struct vnt_private *priv);
126 static int device_init_rd1_ring(struct vnt_private *priv);
127 static int device_init_td0_ring(struct vnt_private *priv);
128 static int device_init_td1_ring(struct vnt_private *priv);
130 static int device_rx_srv(struct vnt_private *priv, unsigned int idx);
131 static int device_tx_srv(struct vnt_private *priv, unsigned int idx);
132 static bool device_alloc_rx_buf(struct vnt_private *, struct vnt_rx_desc *);
133 static void device_free_rx_buf(struct vnt_private *priv,
134 struct vnt_rx_desc *rd);
135 static void device_init_registers(struct vnt_private *priv);
136 static void device_free_tx_buf(struct vnt_private *, struct vnt_tx_desc *);
137 static void device_free_td0_ring(struct vnt_private *priv);
138 static void device_free_td1_ring(struct vnt_private *priv);
139 static void device_free_rd0_ring(struct vnt_private *priv);
140 static void device_free_rd1_ring(struct vnt_private *priv);
141 static void device_free_rings(struct vnt_private *priv);
143 /*--------------------- Export Variables --------------------------*/
145 /*--------------------- Export Functions --------------------------*/
147 static void vt6655_remove(struct pci_dev *pcid)
149 struct vnt_private *priv = pci_get_drvdata(pcid);
153 device_free_info(priv);
156 static void device_get_options(struct vnt_private *priv)
158 struct vnt_options *opts = &priv->opts;
160 opts->rx_descs0 = RX_DESC_DEF0;
161 opts->rx_descs1 = RX_DESC_DEF1;
162 opts->tx_descs[0] = TX_DESC_DEF0;
163 opts->tx_descs[1] = TX_DESC_DEF1;
164 opts->int_works = INT_WORKS_DEF;
166 opts->short_retry = SHORT_RETRY_DEF;
167 opts->long_retry = LONG_RETRY_DEF;
168 opts->bbp_type = BBP_TYPE_DEF;
172 device_set_options(struct vnt_private *priv)
174 priv->byShortRetryLimit = priv->opts.short_retry;
175 priv->byLongRetryLimit = priv->opts.long_retry;
176 priv->byBBType = priv->opts.bbp_type;
177 priv->byPacketType = priv->byBBType;
178 priv->byAutoFBCtrl = AUTO_FB_0;
179 priv->bUpdateBBVGA = true;
180 priv->preamble_type = 0;
182 pr_debug(" byShortRetryLimit= %d\n", (int)priv->byShortRetryLimit);
183 pr_debug(" byLongRetryLimit= %d\n", (int)priv->byLongRetryLimit);
184 pr_debug(" preamble_type= %d\n", (int)priv->preamble_type);
185 pr_debug(" byShortPreamble= %d\n", (int)priv->byShortPreamble);
186 pr_debug(" byBBType= %d\n", (int)priv->byBBType);
190 * Initialisation of MAC & BBP registers
193 static void device_init_registers(struct vnt_private *priv)
197 unsigned char byValue;
198 unsigned char byCCKPwrdBm = 0;
199 unsigned char byOFDMPwrdBm = 0;
202 bb_software_reset(priv);
204 /* Do MACbSoftwareReset in MACvInitialize */
205 MACbSoftwareReset(priv);
209 /* Only used in 11g type, sync with ERP IE */
210 priv->bProtectMode = false;
212 priv->bNonERPPresent = false;
213 priv->bBarkerPreambleMd = false;
214 priv->wCurrentRate = RATE_1M;
215 priv->byTopOFDMBasicRate = RATE_24M;
216 priv->byTopCCKBasicRate = RATE_1M;
219 MACvInitialize(priv);
222 priv->local_id = ioread8(priv->port_offset + MAC_REG_LOCALID);
224 spin_lock_irqsave(&priv->lock, flags);
226 SROMvReadAllContents(priv->port_offset, priv->abyEEPROM);
228 spin_unlock_irqrestore(&priv->lock, flags);
230 /* Get Channel range */
231 priv->byMinChannel = 1;
232 priv->byMaxChannel = CB_MAX_CHANNEL;
235 byValue = SROMbyReadEmbedded(priv->port_offset, EEP_OFS_ANTENNA);
236 if (byValue & EEP_ANTINV)
237 priv->bTxRxAntInv = true;
239 priv->bTxRxAntInv = false;
241 byValue &= (EEP_ANTENNA_AUX | EEP_ANTENNA_MAIN);
242 /* if not set default is All */
244 byValue = (EEP_ANTENNA_AUX | EEP_ANTENNA_MAIN);
246 if (byValue == (EEP_ANTENNA_AUX | EEP_ANTENNA_MAIN)) {
247 priv->byAntennaCount = 2;
248 priv->byTxAntennaMode = ANT_B;
249 priv->dwTxAntennaSel = 1;
250 priv->dwRxAntennaSel = 1;
252 if (priv->bTxRxAntInv)
253 priv->byRxAntennaMode = ANT_A;
255 priv->byRxAntennaMode = ANT_B;
257 priv->byAntennaCount = 1;
258 priv->dwTxAntennaSel = 0;
259 priv->dwRxAntennaSel = 0;
261 if (byValue & EEP_ANTENNA_AUX) {
262 priv->byTxAntennaMode = ANT_A;
264 if (priv->bTxRxAntInv)
265 priv->byRxAntennaMode = ANT_B;
267 priv->byRxAntennaMode = ANT_A;
269 priv->byTxAntennaMode = ANT_B;
271 if (priv->bTxRxAntInv)
272 priv->byRxAntennaMode = ANT_A;
274 priv->byRxAntennaMode = ANT_B;
278 /* Set initial antenna mode */
279 bb_set_tx_antenna_mode(priv, priv->byTxAntennaMode);
280 bb_set_rx_antenna_mode(priv, priv->byRxAntennaMode);
282 /* zonetype initial */
283 priv->byOriginalZonetype = priv->abyEEPROM[EEP_OFS_ZONETYPE];
285 if (!priv->bZoneRegExist)
286 priv->byZoneType = priv->abyEEPROM[EEP_OFS_ZONETYPE];
288 pr_debug("priv->byZoneType = %x\n", priv->byZoneType);
293 /* Get Desire Power Value */
294 priv->byCurPwr = 0xFF;
295 priv->byCCKPwr = SROMbyReadEmbedded(priv->port_offset, EEP_OFS_PWR_CCK);
296 priv->byOFDMPwrG = SROMbyReadEmbedded(priv->port_offset,
299 /* Load power Table */
300 for (ii = 0; ii < CB_MAX_CHANNEL_24G; ii++) {
301 priv->abyCCKPwrTbl[ii + 1] =
302 SROMbyReadEmbedded(priv->port_offset,
303 (unsigned char)(ii + EEP_OFS_CCK_PWR_TBL));
304 if (priv->abyCCKPwrTbl[ii + 1] == 0)
305 priv->abyCCKPwrTbl[ii + 1] = priv->byCCKPwr;
307 priv->abyOFDMPwrTbl[ii + 1] =
308 SROMbyReadEmbedded(priv->port_offset,
309 (unsigned char)(ii + EEP_OFS_OFDM_PWR_TBL));
310 if (priv->abyOFDMPwrTbl[ii + 1] == 0)
311 priv->abyOFDMPwrTbl[ii + 1] = priv->byOFDMPwrG;
313 priv->abyCCKDefaultPwr[ii + 1] = byCCKPwrdBm;
314 priv->abyOFDMDefaultPwr[ii + 1] = byOFDMPwrdBm;
317 /* recover 12,13 ,14channel for EUROPE by 11 channel */
318 for (ii = 11; ii < 14; ii++) {
319 priv->abyCCKPwrTbl[ii] = priv->abyCCKPwrTbl[10];
320 priv->abyOFDMPwrTbl[ii] = priv->abyOFDMPwrTbl[10];
323 /* Load OFDM A Power Table */
324 for (ii = 0; ii < CB_MAX_CHANNEL_5G; ii++) {
325 priv->abyOFDMPwrTbl[ii + CB_MAX_CHANNEL_24G + 1] =
326 SROMbyReadEmbedded(priv->port_offset,
327 (unsigned char)(ii + EEP_OFS_OFDMA_PWR_TBL));
329 priv->abyOFDMDefaultPwr[ii + CB_MAX_CHANNEL_24G + 1] =
330 SROMbyReadEmbedded(priv->port_offset,
331 (unsigned char)(ii + EEP_OFS_OFDMA_PWR_dBm));
334 if (priv->local_id > REV_ID_VT3253_B1) {
335 MACvSelectPage1(priv->port_offset);
337 iowrite8(MSRCTL1_TXPWR | MSRCTL1_CSAPAREN, priv->port_offset + MAC_REG_MSRCTL + 1);
339 MACvSelectPage0(priv->port_offset);
342 /* use relative tx timeout and 802.11i D4 */
343 MACvWordRegBitsOn(priv->port_offset,
344 MAC_REG_CFG, (CFG_TKIPOPT | CFG_NOTXTIMEOUT));
346 /* set performance parameter by registry */
347 MACvSetShortRetryLimit(priv, priv->byShortRetryLimit);
348 MACvSetLongRetryLimit(priv, priv->byLongRetryLimit);
350 /* reset TSF counter */
351 iowrite8(TFTCTL_TSFCNTRST, priv->port_offset + MAC_REG_TFTCTL);
352 /* enable TSF counter */
353 iowrite8(TFTCTL_TSFCNTREN, priv->port_offset + MAC_REG_TFTCTL);
355 /* initialize BBP registers */
356 bb_vt3253_init(priv);
358 if (priv->bUpdateBBVGA) {
359 priv->byBBVGACurrent = priv->abyBBVGA[0];
360 priv->byBBVGANew = priv->byBBVGACurrent;
361 bb_set_vga_gain_offset(priv, priv->abyBBVGA[0]);
364 bb_set_rx_antenna_mode(priv, priv->byRxAntennaMode);
365 bb_set_tx_antenna_mode(priv, priv->byTxAntennaMode);
367 /* Set BB and packet type at the same time. */
368 /* Set Short Slot Time, xIFS, and RSPINF. */
369 priv->wCurrentRate = RATE_54M;
371 priv->radio_off = false;
373 priv->byRadioCtl = SROMbyReadEmbedded(priv->port_offset,
375 priv->hw_radio_off = false;
377 if (priv->byRadioCtl & EEP_RADIOCTL_ENABLE) {
379 priv->byGPIO = ioread8(priv->port_offset + MAC_REG_GPIOCTL1);
381 if (((priv->byGPIO & GPIO0_DATA) &&
382 !(priv->byRadioCtl & EEP_RADIOCTL_INV)) ||
383 (!(priv->byGPIO & GPIO0_DATA) &&
384 (priv->byRadioCtl & EEP_RADIOCTL_INV)))
385 priv->hw_radio_off = true;
388 if (priv->hw_radio_off || priv->bRadioControlOff)
389 CARDbRadioPowerOff(priv);
391 /* get Permanent network address */
392 SROMvReadEtherAddress(priv->port_offset, priv->abyCurrentNetAddr);
393 pr_debug("Network address = %pM\n", priv->abyCurrentNetAddr);
395 /* reset Tx pointer */
396 CARDvSafeResetRx(priv);
397 /* reset Rx pointer */
398 CARDvSafeResetTx(priv);
400 if (priv->local_id <= REV_ID_VT3253_A1)
401 MACvRegBitsOn(priv->port_offset, MAC_REG_RCR, RCR_WPAERR);
404 MACvReceive0(priv->port_offset);
405 MACvReceive1(priv->port_offset);
407 /* start the adapter */
408 iowrite8(HOSTCR_MACEN | HOSTCR_RXON | HOSTCR_TXON, priv->port_offset + MAC_REG_HOSTCR);
411 static void device_print_info(struct vnt_private *priv)
413 dev_info(&priv->pcid->dev, "MAC=%pM IO=0x%lx Mem=0x%lx IRQ=%d\n",
414 priv->abyCurrentNetAddr, (unsigned long)priv->ioaddr,
415 (unsigned long)priv->port_offset, priv->pcid->irq);
418 static void device_free_info(struct vnt_private *priv)
424 ieee80211_unregister_hw(priv->hw);
426 if (priv->port_offset)
427 iounmap(priv->port_offset);
430 pci_release_regions(priv->pcid);
433 ieee80211_free_hw(priv->hw);
436 static bool device_init_rings(struct vnt_private *priv)
440 /*allocate all RD/TD rings a single pool*/
441 vir_pool = dma_alloc_coherent(&priv->pcid->dev,
442 priv->opts.rx_descs0 * sizeof(struct vnt_rx_desc) +
443 priv->opts.rx_descs1 * sizeof(struct vnt_rx_desc) +
444 priv->opts.tx_descs[0] * sizeof(struct vnt_tx_desc) +
445 priv->opts.tx_descs[1] * sizeof(struct vnt_tx_desc),
446 &priv->pool_dma, GFP_ATOMIC);
448 dev_err(&priv->pcid->dev, "allocate desc dma memory failed\n");
452 priv->aRD0Ring = vir_pool;
453 priv->aRD1Ring = vir_pool +
454 priv->opts.rx_descs0 * sizeof(struct vnt_rx_desc);
456 priv->rd0_pool_dma = priv->pool_dma;
457 priv->rd1_pool_dma = priv->rd0_pool_dma +
458 priv->opts.rx_descs0 * sizeof(struct vnt_rx_desc);
460 priv->tx0_bufs = dma_alloc_coherent(&priv->pcid->dev,
461 priv->opts.tx_descs[0] * PKT_BUF_SZ +
462 priv->opts.tx_descs[1] * PKT_BUF_SZ +
465 &priv->tx_bufs_dma0, GFP_ATOMIC);
466 if (!priv->tx0_bufs) {
467 dev_err(&priv->pcid->dev, "allocate buf dma memory failed\n");
469 dma_free_coherent(&priv->pcid->dev,
470 priv->opts.rx_descs0 * sizeof(struct vnt_rx_desc) +
471 priv->opts.rx_descs1 * sizeof(struct vnt_rx_desc) +
472 priv->opts.tx_descs[0] * sizeof(struct vnt_tx_desc) +
473 priv->opts.tx_descs[1] * sizeof(struct vnt_tx_desc),
474 vir_pool, priv->pool_dma);
478 priv->td0_pool_dma = priv->rd1_pool_dma +
479 priv->opts.rx_descs1 * sizeof(struct vnt_rx_desc);
481 priv->td1_pool_dma = priv->td0_pool_dma +
482 priv->opts.tx_descs[0] * sizeof(struct vnt_tx_desc);
484 /* vir_pool: pvoid type */
485 priv->apTD0Rings = vir_pool
486 + priv->opts.rx_descs0 * sizeof(struct vnt_rx_desc)
487 + priv->opts.rx_descs1 * sizeof(struct vnt_rx_desc);
489 priv->apTD1Rings = vir_pool
490 + priv->opts.rx_descs0 * sizeof(struct vnt_rx_desc)
491 + priv->opts.rx_descs1 * sizeof(struct vnt_rx_desc)
492 + priv->opts.tx_descs[0] * sizeof(struct vnt_tx_desc);
494 priv->tx1_bufs = priv->tx0_bufs +
495 priv->opts.tx_descs[0] * PKT_BUF_SZ;
497 priv->tx_beacon_bufs = priv->tx1_bufs +
498 priv->opts.tx_descs[1] * PKT_BUF_SZ;
500 priv->pbyTmpBuff = priv->tx_beacon_bufs +
503 priv->tx_bufs_dma1 = priv->tx_bufs_dma0 +
504 priv->opts.tx_descs[0] * PKT_BUF_SZ;
506 priv->tx_beacon_dma = priv->tx_bufs_dma1 +
507 priv->opts.tx_descs[1] * PKT_BUF_SZ;
512 static void device_free_rings(struct vnt_private *priv)
514 dma_free_coherent(&priv->pcid->dev,
515 priv->opts.rx_descs0 * sizeof(struct vnt_rx_desc) +
516 priv->opts.rx_descs1 * sizeof(struct vnt_rx_desc) +
517 priv->opts.tx_descs[0] * sizeof(struct vnt_tx_desc) +
518 priv->opts.tx_descs[1] * sizeof(struct vnt_tx_desc),
519 priv->aRD0Ring, priv->pool_dma);
522 dma_free_coherent(&priv->pcid->dev,
523 priv->opts.tx_descs[0] * PKT_BUF_SZ +
524 priv->opts.tx_descs[1] * PKT_BUF_SZ +
527 priv->tx0_bufs, priv->tx_bufs_dma0);
530 static int device_init_rd0_ring(struct vnt_private *priv)
533 dma_addr_t curr = priv->rd0_pool_dma;
534 struct vnt_rx_desc *desc;
537 /* Init the RD0 ring entries */
538 for (i = 0; i < priv->opts.rx_descs0;
539 i ++, curr += sizeof(struct vnt_rx_desc)) {
540 desc = &priv->aRD0Ring[i];
541 desc->rd_info = kzalloc(sizeof(*desc->rd_info), GFP_KERNEL);
542 if (!desc->rd_info) {
547 if (!device_alloc_rx_buf(priv, desc)) {
548 dev_err(&priv->pcid->dev, "can not alloc rx bufs\n");
553 desc->next = &priv->aRD0Ring[(i + 1) % priv->opts.rx_descs0];
554 desc->next_desc = cpu_to_le32(curr + sizeof(struct vnt_rx_desc));
558 priv->aRD0Ring[i - 1].next_desc = cpu_to_le32(priv->rd0_pool_dma);
559 priv->pCurrRD[0] = &priv->aRD0Ring[0];
564 kfree(desc->rd_info);
568 desc = &priv->aRD0Ring[i];
569 device_free_rx_buf(priv, desc);
570 kfree(desc->rd_info);
576 static int device_init_rd1_ring(struct vnt_private *priv)
579 dma_addr_t curr = priv->rd1_pool_dma;
580 struct vnt_rx_desc *desc;
583 /* Init the RD1 ring entries */
584 for (i = 0; i < priv->opts.rx_descs1;
585 i ++, curr += sizeof(struct vnt_rx_desc)) {
586 desc = &priv->aRD1Ring[i];
587 desc->rd_info = kzalloc(sizeof(*desc->rd_info), GFP_KERNEL);
588 if (!desc->rd_info) {
593 if (!device_alloc_rx_buf(priv, desc)) {
594 dev_err(&priv->pcid->dev, "can not alloc rx bufs\n");
599 desc->next = &priv->aRD1Ring[(i + 1) % priv->opts.rx_descs1];
600 desc->next_desc = cpu_to_le32(curr + sizeof(struct vnt_rx_desc));
604 priv->aRD1Ring[i - 1].next_desc = cpu_to_le32(priv->rd1_pool_dma);
605 priv->pCurrRD[1] = &priv->aRD1Ring[0];
610 kfree(desc->rd_info);
614 desc = &priv->aRD1Ring[i];
615 device_free_rx_buf(priv, desc);
616 kfree(desc->rd_info);
622 static void device_free_rd0_ring(struct vnt_private *priv)
626 for (i = 0; i < priv->opts.rx_descs0; i++) {
627 struct vnt_rx_desc *desc = &priv->aRD0Ring[i];
629 device_free_rx_buf(priv, desc);
630 kfree(desc->rd_info);
634 static void device_free_rd1_ring(struct vnt_private *priv)
638 for (i = 0; i < priv->opts.rx_descs1; i++) {
639 struct vnt_rx_desc *desc = &priv->aRD1Ring[i];
641 device_free_rx_buf(priv, desc);
642 kfree(desc->rd_info);
646 static int device_init_td0_ring(struct vnt_private *priv)
650 struct vnt_tx_desc *desc;
653 curr = priv->td0_pool_dma;
654 for (i = 0; i < priv->opts.tx_descs[0];
655 i++, curr += sizeof(struct vnt_tx_desc)) {
656 desc = &priv->apTD0Rings[i];
657 desc->td_info = kzalloc(sizeof(*desc->td_info), GFP_KERNEL);
658 if (!desc->td_info) {
663 desc->td_info->buf = priv->tx0_bufs + i * PKT_BUF_SZ;
664 desc->td_info->buf_dma = priv->tx_bufs_dma0 + i * PKT_BUF_SZ;
666 desc->next = &(priv->apTD0Rings[(i + 1) % priv->opts.tx_descs[0]]);
667 desc->next_desc = cpu_to_le32(curr +
668 sizeof(struct vnt_tx_desc));
672 priv->apTD0Rings[i - 1].next_desc = cpu_to_le32(priv->td0_pool_dma);
673 priv->apTailTD[0] = priv->apCurrTD[0] = &priv->apTD0Rings[0];
679 desc = &priv->apTD0Rings[i];
680 kfree(desc->td_info);
686 static int device_init_td1_ring(struct vnt_private *priv)
690 struct vnt_tx_desc *desc;
693 /* Init the TD ring entries */
694 curr = priv->td1_pool_dma;
695 for (i = 0; i < priv->opts.tx_descs[1];
696 i++, curr += sizeof(struct vnt_tx_desc)) {
697 desc = &priv->apTD1Rings[i];
698 desc->td_info = kzalloc(sizeof(*desc->td_info), GFP_KERNEL);
699 if (!desc->td_info) {
704 desc->td_info->buf = priv->tx1_bufs + i * PKT_BUF_SZ;
705 desc->td_info->buf_dma = priv->tx_bufs_dma1 + i * PKT_BUF_SZ;
707 desc->next = &(priv->apTD1Rings[(i + 1) % priv->opts.tx_descs[1]]);
708 desc->next_desc = cpu_to_le32(curr + sizeof(struct vnt_tx_desc));
712 priv->apTD1Rings[i - 1].next_desc = cpu_to_le32(priv->td1_pool_dma);
713 priv->apTailTD[1] = priv->apCurrTD[1] = &priv->apTD1Rings[0];
719 desc = &priv->apTD1Rings[i];
720 kfree(desc->td_info);
726 static void device_free_td0_ring(struct vnt_private *priv)
730 for (i = 0; i < priv->opts.tx_descs[0]; i++) {
731 struct vnt_tx_desc *desc = &priv->apTD0Rings[i];
732 struct vnt_td_info *td_info = desc->td_info;
734 dev_kfree_skb(td_info->skb);
735 kfree(desc->td_info);
739 static void device_free_td1_ring(struct vnt_private *priv)
743 for (i = 0; i < priv->opts.tx_descs[1]; i++) {
744 struct vnt_tx_desc *desc = &priv->apTD1Rings[i];
745 struct vnt_td_info *td_info = desc->td_info;
747 dev_kfree_skb(td_info->skb);
748 kfree(desc->td_info);
752 /*-----------------------------------------------------------------*/
754 static int device_rx_srv(struct vnt_private *priv, unsigned int idx)
756 struct vnt_rx_desc *rd;
759 for (rd = priv->pCurrRD[idx];
760 rd->rd0.owner == OWNED_BY_HOST;
765 if (!rd->rd_info->skb)
768 if (vnt_receive_frame(priv, rd)) {
769 if (!device_alloc_rx_buf(priv, rd)) {
770 dev_err(&priv->pcid->dev,
771 "can not allocate rx buf\n");
775 rd->rd0.owner = OWNED_BY_NIC;
778 priv->pCurrRD[idx] = rd;
783 static bool device_alloc_rx_buf(struct vnt_private *priv,
784 struct vnt_rx_desc *rd)
786 struct vnt_rd_info *rd_info = rd->rd_info;
788 rd_info->skb = dev_alloc_skb((int)priv->rx_buf_sz);
793 dma_map_single(&priv->pcid->dev,
794 skb_put(rd_info->skb, skb_tailroom(rd_info->skb)),
795 priv->rx_buf_sz, DMA_FROM_DEVICE);
796 if (dma_mapping_error(&priv->pcid->dev, rd_info->skb_dma)) {
797 dev_kfree_skb(rd_info->skb);
802 *((unsigned int *)&rd->rd0) = 0; /* FIX cast */
804 rd->rd0.res_count = cpu_to_le16(priv->rx_buf_sz);
805 rd->rd0.owner = OWNED_BY_NIC;
806 rd->rd1.req_count = cpu_to_le16(priv->rx_buf_sz);
807 rd->buff_addr = cpu_to_le32(rd_info->skb_dma);
812 static void device_free_rx_buf(struct vnt_private *priv,
813 struct vnt_rx_desc *rd)
815 struct vnt_rd_info *rd_info = rd->rd_info;
817 dma_unmap_single(&priv->pcid->dev, rd_info->skb_dma,
818 priv->rx_buf_sz, DMA_FROM_DEVICE);
819 dev_kfree_skb(rd_info->skb);
822 static const u8 fallback_rate0[5][5] = {
823 {RATE_18M, RATE_18M, RATE_12M, RATE_12M, RATE_12M},
824 {RATE_24M, RATE_24M, RATE_18M, RATE_12M, RATE_12M},
825 {RATE_36M, RATE_36M, RATE_24M, RATE_18M, RATE_18M},
826 {RATE_48M, RATE_48M, RATE_36M, RATE_24M, RATE_24M},
827 {RATE_54M, RATE_54M, RATE_48M, RATE_36M, RATE_36M}
830 static const u8 fallback_rate1[5][5] = {
831 {RATE_18M, RATE_18M, RATE_12M, RATE_6M, RATE_6M},
832 {RATE_24M, RATE_24M, RATE_18M, RATE_6M, RATE_6M},
833 {RATE_36M, RATE_36M, RATE_24M, RATE_12M, RATE_12M},
834 {RATE_48M, RATE_48M, RATE_24M, RATE_12M, RATE_12M},
835 {RATE_54M, RATE_54M, RATE_36M, RATE_18M, RATE_18M}
838 static int vnt_int_report_rate(struct vnt_private *priv,
839 struct vnt_td_info *context, u8 tsr0, u8 tsr1)
841 struct vnt_tx_fifo_head *fifo_head;
842 struct ieee80211_tx_info *info;
843 struct ieee80211_rate *rate;
845 u8 tx_retry = (tsr0 & TSR0_NCR);
854 fifo_head = (struct vnt_tx_fifo_head *)context->buf;
855 fb_option = (le16_to_cpu(fifo_head->fifo_ctl) &
856 (FIFOCTL_AUTO_FB_0 | FIFOCTL_AUTO_FB_1));
858 info = IEEE80211_SKB_CB(context->skb);
859 idx = info->control.rates[0].idx;
861 if (fb_option && !(tsr1 & TSR1_TERR)) {
865 rate = ieee80211_get_tx_rate(priv->hw, info);
866 tx_rate = rate->hw_value - RATE_18M;
871 if (fb_option & FIFOCTL_AUTO_FB_0)
872 tx_rate = fallback_rate0[tx_rate][retry];
873 else if (fb_option & FIFOCTL_AUTO_FB_1)
874 tx_rate = fallback_rate1[tx_rate][retry];
876 if (info->band == NL80211_BAND_5GHZ)
877 idx = tx_rate - RATE_6M;
882 ieee80211_tx_info_clear_status(info);
884 info->status.rates[0].count = tx_retry;
886 if (!(tsr1 & TSR1_TERR)) {
887 info->status.rates[0].idx = idx;
889 if (info->flags & IEEE80211_TX_CTL_NO_ACK)
890 info->flags |= IEEE80211_TX_STAT_NOACK_TRANSMITTED;
892 info->flags |= IEEE80211_TX_STAT_ACK;
898 static int device_tx_srv(struct vnt_private *priv, unsigned int idx)
900 struct vnt_tx_desc *desc;
902 unsigned char byTsr0;
903 unsigned char byTsr1;
905 for (desc = priv->apTailTD[idx]; priv->iTDUsed[idx] > 0; desc = desc->next) {
906 if (desc->td0.owner == OWNED_BY_NIC)
911 byTsr0 = desc->td0.tsr0;
912 byTsr1 = desc->td0.tsr1;
914 /* Only the status of first TD in the chain is correct */
915 if (desc->td1.tcr & TCR_STP) {
916 if ((desc->td_info->flags & TD_FLAGS_NETIF_SKB) != 0) {
917 if (!(byTsr1 & TSR1_TERR)) {
919 pr_debug(" Tx[%d] OK but has error. tsr1[%02X] tsr0[%02X]\n",
924 pr_debug(" Tx[%d] dropped & tsr1[%02X] tsr0[%02X]\n",
925 (int)idx, byTsr1, byTsr0);
929 if (byTsr1 & TSR1_TERR) {
930 if ((desc->td_info->flags & TD_FLAGS_PRIV_SKB) != 0) {
931 pr_debug(" Tx[%d] fail has error. tsr1[%02X] tsr0[%02X]\n",
932 (int)idx, byTsr1, byTsr0);
936 vnt_int_report_rate(priv, desc->td_info, byTsr0, byTsr1);
938 device_free_tx_buf(priv, desc);
939 priv->iTDUsed[idx]--;
943 priv->apTailTD[idx] = desc;
948 static void device_error(struct vnt_private *priv, unsigned short status)
950 if (status & ISR_FETALERR) {
951 dev_err(&priv->pcid->dev, "Hardware fatal error\n");
958 static void device_free_tx_buf(struct vnt_private *priv,
959 struct vnt_tx_desc *desc)
961 struct vnt_td_info *td_info = desc->td_info;
962 struct sk_buff *skb = td_info->skb;
965 ieee80211_tx_status_irqsafe(priv->hw, skb);
971 static void vnt_check_bb_vga(struct vnt_private *priv)
976 if (!priv->bUpdateBBVGA)
979 if (priv->hw->conf.flags & IEEE80211_CONF_OFFCHANNEL)
982 if (!(priv->vif->bss_conf.assoc && priv->current_rssi))
985 RFvRSSITodBm(priv, (u8)priv->current_rssi, &dbm);
987 for (i = 0; i < BB_VGA_LEVEL; i++) {
988 if (dbm < priv->dbm_threshold[i]) {
989 priv->byBBVGANew = priv->abyBBVGA[i];
994 if (priv->byBBVGANew == priv->byBBVGACurrent) {
995 priv->uBBVGADiffCount = 1;
999 priv->uBBVGADiffCount++;
1001 if (priv->uBBVGADiffCount == 1) {
1002 /* first VGA diff gain */
1003 bb_set_vga_gain_offset(priv, priv->byBBVGANew);
1005 dev_dbg(&priv->pcid->dev,
1006 "First RSSI[%d] NewGain[%d] OldGain[%d] Count[%d]\n",
1007 (int)dbm, priv->byBBVGANew,
1008 priv->byBBVGACurrent,
1009 (int)priv->uBBVGADiffCount);
1012 if (priv->uBBVGADiffCount >= BB_VGA_CHANGE_THRESHOLD) {
1013 dev_dbg(&priv->pcid->dev,
1014 "RSSI[%d] NewGain[%d] OldGain[%d] Count[%d]\n",
1015 (int)dbm, priv->byBBVGANew,
1016 priv->byBBVGACurrent,
1017 (int)priv->uBBVGADiffCount);
1019 bb_set_vga_gain_offset(priv, priv->byBBVGANew);
1023 static void vnt_interrupt_process(struct vnt_private *priv)
1025 struct ieee80211_low_level_stats *low_stats = &priv->low_stats;
1029 unsigned long flags;
1031 isr = ioread32(priv->port_offset + MAC_REG_ISR);
1036 if (isr == 0xffffffff) {
1037 pr_debug("isr = 0xffff\n");
1041 spin_lock_irqsave(&priv->lock, flags);
1043 /* Read low level stats */
1044 mib_counter = ioread32(priv->port_offset + MAC_REG_MIBCNTR);
1046 low_stats->dot11RTSSuccessCount += mib_counter & 0xff;
1047 low_stats->dot11RTSFailureCount += (mib_counter >> 8) & 0xff;
1048 low_stats->dot11ACKFailureCount += (mib_counter >> 16) & 0xff;
1049 low_stats->dot11FCSErrorCount += (mib_counter >> 24) & 0xff;
1053 * Must do this after doing rx/tx, cause ISR bit is slow
1054 * than RD/TD write back
1055 * update ISR counter
1057 while (isr && priv->vif) {
1058 MACvWriteISR(priv->port_offset, isr);
1060 if (isr & ISR_FETALERR) {
1061 pr_debug(" ISR_FETALERR\n");
1062 iowrite8(0, priv->port_offset + MAC_REG_SOFTPWRCTL);
1063 VNSvOutPortW(priv->port_offset +
1064 MAC_REG_SOFTPWRCTL, SOFTPWRCTL_SWPECTI);
1065 device_error(priv, isr);
1068 if (isr & ISR_TBTT) {
1069 if (priv->op_mode != NL80211_IFTYPE_ADHOC)
1070 vnt_check_bb_vga(priv);
1072 priv->bBeaconSent = false;
1073 if (priv->bEnablePSMode)
1074 PSbIsNextTBTTWakeUp((void *)priv);
1076 if ((priv->op_mode == NL80211_IFTYPE_AP ||
1077 priv->op_mode == NL80211_IFTYPE_ADHOC) &&
1078 priv->vif->bss_conf.enable_beacon)
1079 MACvOneShotTimer1MicroSec(priv,
1080 (priv->vif->bss_conf.beacon_int -
1081 MAKE_BEACON_RESERVED) << 10);
1083 /* TODO: adhoc PS mode */
1086 if (isr & ISR_BNTX) {
1087 if (priv->op_mode == NL80211_IFTYPE_ADHOC) {
1088 priv->bIsBeaconBufReadySet = false;
1089 priv->cbBeaconBufReadySetCnt = 0;
1092 priv->bBeaconSent = true;
1095 if (isr & ISR_RXDMA0)
1096 max_count += device_rx_srv(priv, TYPE_RXDMA0);
1098 if (isr & ISR_RXDMA1)
1099 max_count += device_rx_srv(priv, TYPE_RXDMA1);
1101 if (isr & ISR_TXDMA0)
1102 max_count += device_tx_srv(priv, TYPE_TXDMA0);
1104 if (isr & ISR_AC0DMA)
1105 max_count += device_tx_srv(priv, TYPE_AC0DMA);
1107 if (isr & ISR_SOFTTIMER1) {
1108 if (priv->vif->bss_conf.enable_beacon)
1109 vnt_beacon_make(priv, priv->vif);
1112 /* If both buffers available wake the queue */
1113 if (AVAIL_TD(priv, TYPE_TXDMA0) &&
1114 AVAIL_TD(priv, TYPE_AC0DMA) &&
1115 ieee80211_queue_stopped(priv->hw, 0))
1116 ieee80211_wake_queues(priv->hw);
1118 isr = ioread32(priv->port_offset + MAC_REG_ISR);
1120 MACvReceive0(priv->port_offset);
1121 MACvReceive1(priv->port_offset);
1123 if (max_count > priv->opts.int_works)
1127 spin_unlock_irqrestore(&priv->lock, flags);
1130 static void vnt_interrupt_work(struct work_struct *work)
1132 struct vnt_private *priv =
1133 container_of(work, struct vnt_private, interrupt_work);
1136 vnt_interrupt_process(priv);
1138 MACvIntEnable(priv->port_offset, IMR_MASK_VALUE);
1141 static irqreturn_t vnt_interrupt(int irq, void *arg)
1143 struct vnt_private *priv = arg;
1145 schedule_work(&priv->interrupt_work);
1147 MACvIntDisable(priv->port_offset);
1152 static int vnt_tx_packet(struct vnt_private *priv, struct sk_buff *skb)
1154 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
1155 struct vnt_tx_desc *head_td;
1157 unsigned long flags;
1159 spin_lock_irqsave(&priv->lock, flags);
1161 if (ieee80211_is_data(hdr->frame_control))
1162 dma_idx = TYPE_AC0DMA;
1164 dma_idx = TYPE_TXDMA0;
1166 if (AVAIL_TD(priv, dma_idx) < 1) {
1167 spin_unlock_irqrestore(&priv->lock, flags);
1168 ieee80211_stop_queues(priv->hw);
1172 head_td = priv->apCurrTD[dma_idx];
1174 head_td->td1.tcr = 0;
1176 head_td->td_info->skb = skb;
1178 if (dma_idx == TYPE_AC0DMA)
1179 head_td->td_info->flags = TD_FLAGS_NETIF_SKB;
1181 priv->apCurrTD[dma_idx] = head_td->next;
1183 spin_unlock_irqrestore(&priv->lock, flags);
1185 vnt_generate_fifo_header(priv, dma_idx, head_td, skb);
1187 spin_lock_irqsave(&priv->lock, flags);
1189 priv->bPWBitOn = false;
1191 /* Set TSR1 & ReqCount in TxDescHead */
1192 head_td->td1.tcr |= (TCR_STP | TCR_EDP | EDMSDU);
1193 head_td->td1.req_count = cpu_to_le16(head_td->td_info->req_count);
1195 head_td->buff_addr = cpu_to_le32(head_td->td_info->buf_dma);
1197 /* Poll Transmit the adapter */
1199 head_td->td0.owner = OWNED_BY_NIC;
1200 wmb(); /* second memory barrier */
1202 if (head_td->td_info->flags & TD_FLAGS_NETIF_SKB)
1203 MACvTransmitAC0(priv->port_offset);
1205 MACvTransmit0(priv->port_offset);
1207 priv->iTDUsed[dma_idx]++;
1209 spin_unlock_irqrestore(&priv->lock, flags);
1214 static void vnt_tx_80211(struct ieee80211_hw *hw,
1215 struct ieee80211_tx_control *control,
1216 struct sk_buff *skb)
1218 struct vnt_private *priv = hw->priv;
1220 if (vnt_tx_packet(priv, skb))
1221 ieee80211_free_txskb(hw, skb);
1224 static int vnt_start(struct ieee80211_hw *hw)
1226 struct vnt_private *priv = hw->priv;
1229 priv->rx_buf_sz = PKT_BUF_SZ;
1230 if (!device_init_rings(priv))
1233 ret = request_irq(priv->pcid->irq, vnt_interrupt,
1234 IRQF_SHARED, "vt6655", priv);
1236 dev_dbg(&priv->pcid->dev, "failed to start irq\n");
1237 goto err_free_rings;
1240 dev_dbg(&priv->pcid->dev, "call device init rd0 ring\n");
1241 ret = device_init_rd0_ring(priv);
1244 ret = device_init_rd1_ring(priv);
1246 goto err_free_rd0_ring;
1247 ret = device_init_td0_ring(priv);
1249 goto err_free_rd1_ring;
1250 ret = device_init_td1_ring(priv);
1252 goto err_free_td0_ring;
1254 device_init_registers(priv);
1256 dev_dbg(&priv->pcid->dev, "call MACvIntEnable\n");
1257 MACvIntEnable(priv->port_offset, IMR_MASK_VALUE);
1259 ieee80211_wake_queues(hw);
1264 device_free_td0_ring(priv);
1266 device_free_rd1_ring(priv);
1268 device_free_rd0_ring(priv);
1270 free_irq(priv->pcid->irq, priv);
1272 device_free_rings(priv);
1276 static void vnt_stop(struct ieee80211_hw *hw)
1278 struct vnt_private *priv = hw->priv;
1280 ieee80211_stop_queues(hw);
1282 cancel_work_sync(&priv->interrupt_work);
1285 MACbSoftwareReset(priv);
1286 CARDbRadioPowerOff(priv);
1288 device_free_td0_ring(priv);
1289 device_free_td1_ring(priv);
1290 device_free_rd0_ring(priv);
1291 device_free_rd1_ring(priv);
1292 device_free_rings(priv);
1294 free_irq(priv->pcid->irq, priv);
1297 static int vnt_add_interface(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
1299 struct vnt_private *priv = hw->priv;
1303 switch (vif->type) {
1304 case NL80211_IFTYPE_STATION:
1306 case NL80211_IFTYPE_ADHOC:
1307 MACvRegBitsOff(priv->port_offset, MAC_REG_RCR, RCR_UNICAST);
1309 MACvRegBitsOn(priv->port_offset, MAC_REG_HOSTCR, HOSTCR_ADHOC);
1312 case NL80211_IFTYPE_AP:
1313 MACvRegBitsOff(priv->port_offset, MAC_REG_RCR, RCR_UNICAST);
1315 MACvRegBitsOn(priv->port_offset, MAC_REG_HOSTCR, HOSTCR_AP);
1322 priv->op_mode = vif->type;
1327 static void vnt_remove_interface(struct ieee80211_hw *hw,
1328 struct ieee80211_vif *vif)
1330 struct vnt_private *priv = hw->priv;
1332 switch (vif->type) {
1333 case NL80211_IFTYPE_STATION:
1335 case NL80211_IFTYPE_ADHOC:
1336 MACvRegBitsOff(priv->port_offset, MAC_REG_TCR, TCR_AUTOBCNTX);
1337 MACvRegBitsOff(priv->port_offset,
1338 MAC_REG_TFTCTL, TFTCTL_TSFCNTREN);
1339 MACvRegBitsOff(priv->port_offset, MAC_REG_HOSTCR, HOSTCR_ADHOC);
1341 case NL80211_IFTYPE_AP:
1342 MACvRegBitsOff(priv->port_offset, MAC_REG_TCR, TCR_AUTOBCNTX);
1343 MACvRegBitsOff(priv->port_offset,
1344 MAC_REG_TFTCTL, TFTCTL_TSFCNTREN);
1345 MACvRegBitsOff(priv->port_offset, MAC_REG_HOSTCR, HOSTCR_AP);
1351 priv->op_mode = NL80211_IFTYPE_UNSPECIFIED;
1354 static int vnt_config(struct ieee80211_hw *hw, u32 changed)
1356 struct vnt_private *priv = hw->priv;
1357 struct ieee80211_conf *conf = &hw->conf;
1360 if (changed & IEEE80211_CONF_CHANGE_PS) {
1361 if (conf->flags & IEEE80211_CONF_PS)
1362 PSvEnablePowerSaving(priv, conf->listen_interval);
1364 PSvDisablePowerSaving(priv);
1367 if ((changed & IEEE80211_CONF_CHANGE_CHANNEL) ||
1368 (conf->flags & IEEE80211_CONF_OFFCHANNEL)) {
1369 set_channel(priv, conf->chandef.chan);
1371 if (conf->chandef.chan->band == NL80211_BAND_5GHZ)
1372 bb_type = BB_TYPE_11A;
1374 bb_type = BB_TYPE_11G;
1376 if (priv->byBBType != bb_type) {
1377 priv->byBBType = bb_type;
1379 CARDbSetPhyParameter(priv, priv->byBBType);
1383 if (changed & IEEE80211_CONF_CHANGE_POWER) {
1384 if (priv->byBBType == BB_TYPE_11B)
1385 priv->wCurrentRate = RATE_1M;
1387 priv->wCurrentRate = RATE_54M;
1389 RFbSetPower(priv, priv->wCurrentRate,
1390 conf->chandef.chan->hw_value);
1396 static void vnt_bss_info_changed(struct ieee80211_hw *hw,
1397 struct ieee80211_vif *vif,
1398 struct ieee80211_bss_conf *conf, u32 changed)
1400 struct vnt_private *priv = hw->priv;
1402 priv->current_aid = conf->aid;
1404 if (changed & BSS_CHANGED_BSSID && conf->bssid) {
1405 unsigned long flags;
1407 spin_lock_irqsave(&priv->lock, flags);
1409 MACvWriteBSSIDAddress(priv->port_offset, conf->bssid);
1411 spin_unlock_irqrestore(&priv->lock, flags);
1414 if (changed & BSS_CHANGED_BASIC_RATES) {
1415 priv->basic_rates = conf->basic_rates;
1417 CARDvUpdateBasicTopRate(priv);
1419 dev_dbg(&priv->pcid->dev,
1420 "basic rates %x\n", conf->basic_rates);
1423 if (changed & BSS_CHANGED_ERP_PREAMBLE) {
1424 if (conf->use_short_preamble) {
1425 MACvEnableBarkerPreambleMd(priv->port_offset);
1426 priv->preamble_type = true;
1428 MACvDisableBarkerPreambleMd(priv->port_offset);
1429 priv->preamble_type = false;
1433 if (changed & BSS_CHANGED_ERP_CTS_PROT) {
1434 if (conf->use_cts_prot)
1435 MACvEnableProtectMD(priv->port_offset);
1437 MACvDisableProtectMD(priv->port_offset);
1440 if (changed & BSS_CHANGED_ERP_SLOT) {
1441 if (conf->use_short_slot)
1442 priv->short_slot_time = true;
1444 priv->short_slot_time = false;
1446 CARDbSetPhyParameter(priv, priv->byBBType);
1447 bb_set_vga_gain_offset(priv, priv->abyBBVGA[0]);
1450 if (changed & BSS_CHANGED_TXPOWER)
1451 RFbSetPower(priv, priv->wCurrentRate,
1452 conf->chandef.chan->hw_value);
1454 if (changed & BSS_CHANGED_BEACON_ENABLED) {
1455 dev_dbg(&priv->pcid->dev,
1456 "Beacon enable %d\n", conf->enable_beacon);
1458 if (conf->enable_beacon) {
1459 vnt_beacon_enable(priv, vif, conf);
1461 MACvRegBitsOn(priv->port_offset, MAC_REG_TCR,
1464 MACvRegBitsOff(priv->port_offset, MAC_REG_TCR,
1469 if (changed & (BSS_CHANGED_ASSOC | BSS_CHANGED_BEACON_INFO) &&
1470 priv->op_mode != NL80211_IFTYPE_AP) {
1471 if (conf->assoc && conf->beacon_rate) {
1472 CARDbUpdateTSF(priv, conf->beacon_rate->hw_value,
1475 CARDbSetBeaconPeriod(priv, conf->beacon_int);
1477 CARDvSetFirstNextTBTT(priv, conf->beacon_int);
1479 iowrite8(TFTCTL_TSFCNTRST, priv->port_offset + MAC_REG_TFTCTL);
1480 iowrite8(TFTCTL_TSFCNTREN, priv->port_offset + MAC_REG_TFTCTL);
1485 static u64 vnt_prepare_multicast(struct ieee80211_hw *hw,
1486 struct netdev_hw_addr_list *mc_list)
1488 struct vnt_private *priv = hw->priv;
1489 struct netdev_hw_addr *ha;
1493 netdev_hw_addr_list_for_each(ha, mc_list) {
1494 bit_nr = ether_crc(ETH_ALEN, ha->addr) >> 26;
1496 mc_filter |= 1ULL << (bit_nr & 0x3f);
1499 priv->mc_list_count = mc_list->count;
1504 static void vnt_configure(struct ieee80211_hw *hw,
1505 unsigned int changed_flags,
1506 unsigned int *total_flags, u64 multicast)
1508 struct vnt_private *priv = hw->priv;
1511 *total_flags &= FIF_ALLMULTI | FIF_OTHER_BSS | FIF_BCN_PRBRESP_PROMISC;
1513 rx_mode = ioread8(priv->port_offset + MAC_REG_RCR);
1515 dev_dbg(&priv->pcid->dev, "rx mode in = %x\n", rx_mode);
1517 if (changed_flags & FIF_ALLMULTI) {
1518 if (*total_flags & FIF_ALLMULTI) {
1519 unsigned long flags;
1521 spin_lock_irqsave(&priv->lock, flags);
1523 if (priv->mc_list_count > 2) {
1524 MACvSelectPage1(priv->port_offset);
1526 VNSvOutPortD(priv->port_offset +
1527 MAC_REG_MAR0, 0xffffffff);
1528 VNSvOutPortD(priv->port_offset +
1529 MAC_REG_MAR0 + 4, 0xffffffff);
1531 MACvSelectPage0(priv->port_offset);
1533 MACvSelectPage1(priv->port_offset);
1535 VNSvOutPortD(priv->port_offset +
1536 MAC_REG_MAR0, (u32)multicast);
1537 VNSvOutPortD(priv->port_offset +
1539 (u32)(multicast >> 32));
1541 MACvSelectPage0(priv->port_offset);
1544 spin_unlock_irqrestore(&priv->lock, flags);
1546 rx_mode |= RCR_MULTICAST | RCR_BROADCAST;
1548 rx_mode &= ~(RCR_MULTICAST | RCR_BROADCAST);
1552 if (changed_flags & (FIF_OTHER_BSS | FIF_BCN_PRBRESP_PROMISC)) {
1553 rx_mode |= RCR_MULTICAST | RCR_BROADCAST;
1555 if (*total_flags & (FIF_OTHER_BSS | FIF_BCN_PRBRESP_PROMISC))
1556 rx_mode &= ~RCR_BSSID;
1558 rx_mode |= RCR_BSSID;
1561 iowrite8(rx_mode, priv->port_offset + MAC_REG_RCR);
1563 dev_dbg(&priv->pcid->dev, "rx mode out= %x\n", rx_mode);
1566 static int vnt_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
1567 struct ieee80211_vif *vif, struct ieee80211_sta *sta,
1568 struct ieee80211_key_conf *key)
1570 struct vnt_private *priv = hw->priv;
1574 if (vnt_set_keys(hw, sta, vif, key))
1578 if (test_bit(key->hw_key_idx, &priv->key_entry_inuse))
1579 clear_bit(key->hw_key_idx, &priv->key_entry_inuse);
1588 static int vnt_get_stats(struct ieee80211_hw *hw,
1589 struct ieee80211_low_level_stats *stats)
1591 struct vnt_private *priv = hw->priv;
1593 memcpy(stats, &priv->low_stats, sizeof(*stats));
1598 static u64 vnt_get_tsf(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
1600 struct vnt_private *priv = hw->priv;
1603 tsf = vt6655_get_current_tsf(priv);
1608 static void vnt_set_tsf(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
1611 struct vnt_private *priv = hw->priv;
1613 CARDvUpdateNextTBTT(priv, tsf, vif->bss_conf.beacon_int);
1616 static void vnt_reset_tsf(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
1618 struct vnt_private *priv = hw->priv;
1620 /* reset TSF counter */
1621 iowrite8(TFTCTL_TSFCNTRST, priv->port_offset + MAC_REG_TFTCTL);
1624 static const struct ieee80211_ops vnt_mac_ops = {
1628 .add_interface = vnt_add_interface,
1629 .remove_interface = vnt_remove_interface,
1630 .config = vnt_config,
1631 .bss_info_changed = vnt_bss_info_changed,
1632 .prepare_multicast = vnt_prepare_multicast,
1633 .configure_filter = vnt_configure,
1634 .set_key = vnt_set_key,
1635 .get_stats = vnt_get_stats,
1636 .get_tsf = vnt_get_tsf,
1637 .set_tsf = vnt_set_tsf,
1638 .reset_tsf = vnt_reset_tsf,
1641 static int vnt_init(struct vnt_private *priv)
1643 SET_IEEE80211_PERM_ADDR(priv->hw, priv->abyCurrentNetAddr);
1645 vnt_init_bands(priv);
1647 if (ieee80211_register_hw(priv->hw))
1650 priv->mac_hw = true;
1652 CARDbRadioPowerOff(priv);
1658 vt6655_probe(struct pci_dev *pcid, const struct pci_device_id *ent)
1660 struct vnt_private *priv;
1661 struct ieee80211_hw *hw;
1662 struct wiphy *wiphy;
1665 dev_notice(&pcid->dev,
1666 "%s Ver. %s\n", DEVICE_FULL_DRV_NAM, DEVICE_VERSION);
1668 dev_notice(&pcid->dev,
1669 "Copyright (c) 2003 VIA Networking Technologies, Inc.\n");
1671 hw = ieee80211_alloc_hw(sizeof(*priv), &vnt_mac_ops);
1673 dev_err(&pcid->dev, "could not register ieee80211_hw\n");
1680 spin_lock_init(&priv->lock);
1684 SET_IEEE80211_DEV(priv->hw, &pcid->dev);
1686 if (pci_enable_device(pcid)) {
1687 device_free_info(priv);
1692 "Before get pci_info memaddr is %x\n", priv->memaddr);
1694 pci_set_master(pcid);
1696 priv->memaddr = pci_resource_start(pcid, 0);
1697 priv->ioaddr = pci_resource_start(pcid, 1);
1698 priv->port_offset = ioremap(priv->memaddr & PCI_BASE_ADDRESS_MEM_MASK,
1700 if (!priv->port_offset) {
1701 dev_err(&pcid->dev, ": Failed to IO remapping ..\n");
1702 device_free_info(priv);
1706 rc = pci_request_regions(pcid, DEVICE_NAME);
1708 dev_err(&pcid->dev, ": Failed to find PCI device\n");
1709 device_free_info(priv);
1713 if (dma_set_mask(&pcid->dev, DMA_BIT_MASK(32))) {
1714 dev_err(&pcid->dev, ": Failed to set dma 32 bit mask\n");
1715 device_free_info(priv);
1719 INIT_WORK(&priv->interrupt_work, vnt_interrupt_work);
1722 if (!MACbSoftwareReset(priv)) {
1723 dev_err(&pcid->dev, ": Failed to access MAC hardware..\n");
1724 device_free_info(priv);
1727 /* initial to reload eeprom */
1728 MACvInitialize(priv);
1729 MACvReadEtherAddress(priv->port_offset, priv->abyCurrentNetAddr);
1732 priv->byRFType = SROMbyReadEmbedded(priv->port_offset, EEP_OFS_RFTYPE);
1733 priv->byRFType &= RF_MASK;
1735 dev_dbg(&pcid->dev, "RF Type = %x\n", priv->byRFType);
1737 device_get_options(priv);
1738 device_set_options(priv);
1740 wiphy = priv->hw->wiphy;
1742 wiphy->frag_threshold = FRAG_THRESH_DEF;
1743 wiphy->rts_threshold = RTS_THRESH_DEF;
1744 wiphy->interface_modes = BIT(NL80211_IFTYPE_STATION) |
1745 BIT(NL80211_IFTYPE_ADHOC) | BIT(NL80211_IFTYPE_AP);
1747 ieee80211_hw_set(priv->hw, TIMING_BEACON_ONLY);
1748 ieee80211_hw_set(priv->hw, SIGNAL_DBM);
1749 ieee80211_hw_set(priv->hw, RX_INCLUDES_FCS);
1750 ieee80211_hw_set(priv->hw, REPORTS_TX_ACK_STATUS);
1751 ieee80211_hw_set(priv->hw, SUPPORTS_PS);
1753 priv->hw->max_signal = 100;
1755 if (vnt_init(priv)) {
1756 device_free_info(priv);
1760 device_print_info(priv);
1761 pci_set_drvdata(pcid, priv);
1766 /*------------------------------------------------------------------*/
1768 static int __maybe_unused vt6655_suspend(struct device *dev_d)
1770 struct vnt_private *priv = dev_get_drvdata(dev_d);
1771 unsigned long flags;
1773 spin_lock_irqsave(&priv->lock, flags);
1777 spin_unlock_irqrestore(&priv->lock, flags);
1782 static int __maybe_unused vt6655_resume(struct device *dev_d)
1784 device_wakeup_disable(dev_d);
1789 MODULE_DEVICE_TABLE(pci, vt6655_pci_id_table);
1791 static SIMPLE_DEV_PM_OPS(vt6655_pm_ops, vt6655_suspend, vt6655_resume);
1793 static struct pci_driver device_driver = {
1794 .name = DEVICE_NAME,
1795 .id_table = vt6655_pci_id_table,
1796 .probe = vt6655_probe,
1797 .remove = vt6655_remove,
1798 .driver.pm = &vt6655_pm_ops,
1801 module_pci_driver(device_driver);