Loading drivers/atm/he.c +16 −131 Original line number Diff line number Diff line Loading @@ -780,59 +780,18 @@ he_init_group(struct he_dev *he_dev, int group) { int i; /* small buffer pool */ he_dev->rbps_pool = pci_pool_create("rbps", he_dev->pci_dev, CONFIG_RBPS_BUFSIZE, 8, 0); if (he_dev->rbps_pool == NULL) { hprintk("unable to create rbps pages\n"); return -ENOMEM; } he_dev->rbps_base = pci_alloc_consistent(he_dev->pci_dev, CONFIG_RBPS_SIZE * sizeof(struct he_rbp), &he_dev->rbps_phys); if (he_dev->rbps_base == NULL) { hprintk("failed to alloc rbps_base\n"); goto out_destroy_rbps_pool; } memset(he_dev->rbps_base, 0, CONFIG_RBPS_SIZE * sizeof(struct he_rbp)); he_dev->rbps_virt = kmalloc(CONFIG_RBPS_SIZE * sizeof(struct he_virt), GFP_KERNEL); if (he_dev->rbps_virt == NULL) { hprintk("failed to alloc rbps_virt\n"); goto out_free_rbps_base; } for (i = 0; i < CONFIG_RBPS_SIZE; ++i) { dma_addr_t dma_handle; void *cpuaddr; cpuaddr = pci_pool_alloc(he_dev->rbps_pool, GFP_KERNEL|GFP_DMA, &dma_handle); if (cpuaddr == NULL) goto out_free_rbps_virt; he_dev->rbps_virt[i].virt = cpuaddr; he_dev->rbps_base[i].status = RBP_LOANED | RBP_SMALLBUF | (i << RBP_INDEX_OFF); he_dev->rbps_base[i].phys = dma_handle; } he_dev->rbps_tail = &he_dev->rbps_base[CONFIG_RBPS_SIZE - 1]; he_writel(he_dev, he_dev->rbps_phys, G0_RBPS_S + (group * 32)); he_writel(he_dev, RBPS_MASK(he_dev->rbps_tail), G0_RBPS_T + (group * 32)); he_writel(he_dev, CONFIG_RBPS_BUFSIZE/4, he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32)); he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32)); he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32)); he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0), G0_RBPS_BS + (group * 32)); he_writel(he_dev, RBP_THRESH(CONFIG_RBPS_THRESH) | RBP_QSIZE(CONFIG_RBPS_SIZE - 1) | RBP_INT_ENB, G0_RBPS_QI + (group * 32)); /* large buffer pool */ he_dev->rbpl_pool = pci_pool_create("rbpl", he_dev->pci_dev, CONFIG_RBPL_BUFSIZE, 8, 0); if (he_dev->rbpl_pool == NULL) { hprintk("unable to create rbpl pool\n"); goto out_free_rbps_virt; return -ENOMEM; } he_dev->rbpl_base = pci_alloc_consistent(he_dev->pci_dev, Loading Loading @@ -934,19 +893,6 @@ he_init_group(struct he_dev *he_dev, int group) out_destroy_rbpl_pool: pci_pool_destroy(he_dev->rbpl_pool); i = CONFIG_RBPS_SIZE; out_free_rbps_virt: while (i--) pci_pool_free(he_dev->rbps_pool, he_dev->rbps_virt[i].virt, he_dev->rbps_base[i].phys); kfree(he_dev->rbps_virt); out_free_rbps_base: pci_free_consistent(he_dev->pci_dev, CONFIG_RBPS_SIZE * sizeof(struct he_rbp), he_dev->rbps_base, he_dev->rbps_phys); out_destroy_rbps_pool: pci_pool_destroy(he_dev->rbps_pool); return -ENOMEM; } Loading Loading @@ -1634,22 +1580,6 @@ he_stop(struct he_dev *he_dev) if (he_dev->rbpl_pool) pci_pool_destroy(he_dev->rbpl_pool); if (he_dev->rbps_base) { int i; for (i = 0; i < CONFIG_RBPS_SIZE; ++i) { void *cpuaddr = he_dev->rbps_virt[i].virt; dma_addr_t dma_handle = he_dev->rbps_base[i].phys; pci_pool_free(he_dev->rbps_pool, cpuaddr, dma_handle); } pci_free_consistent(he_dev->pci_dev, CONFIG_RBPS_SIZE * sizeof(struct he_rbp), he_dev->rbps_base, he_dev->rbps_phys); } if (he_dev->rbps_pool) pci_pool_destroy(he_dev->rbps_pool); if (he_dev->rbrq_base) pci_free_consistent(he_dev->pci_dev, CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq), he_dev->rbrq_base, he_dev->rbrq_phys); Loading Loading @@ -1740,9 +1670,6 @@ he_service_rbrq(struct he_dev *he_dev, int group) RBRQ_CON_CLOSED(he_dev->rbrq_head) ? " CON_CLOSED" : "", RBRQ_HBUF_ERR(he_dev->rbrq_head) ? " HBUF_ERR" : ""); if (RBRQ_ADDR(he_dev->rbrq_head) & RBP_SMALLBUF) rbp = &he_dev->rbps_base[RBP_INDEX(RBRQ_ADDR(he_dev->rbrq_head))]; else rbp = &he_dev->rbpl_base[RBP_INDEX(RBRQ_ADDR(he_dev->rbrq_head))]; buf_len = RBRQ_BUFLEN(he_dev->rbrq_head) * 4; Loading Loading @@ -1819,15 +1746,9 @@ he_service_rbrq(struct he_dev *he_dev, int group) __net_timestamp(skb); for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) { if (iov->iov_base & RBP_SMALLBUF) memcpy(skb_put(skb, iov->iov_len), he_dev->rbps_virt[RBP_INDEX(iov->iov_base)].virt, iov->iov_len); else for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) memcpy(skb_put(skb, iov->iov_len), he_dev->rbpl_virt[RBP_INDEX(iov->iov_base)].virt, iov->iov_len); } switch (vcc->qos.aal) { case ATM_AAL0: Loading Loading @@ -1867,13 +1788,8 @@ he_service_rbrq(struct he_dev *he_dev, int group) return_host_buffers: ++pdus_assembled; for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) { if (iov->iov_base & RBP_SMALLBUF) rbp = &he_dev->rbps_base[RBP_INDEX(iov->iov_base)]; else for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) { rbp = &he_dev->rbpl_base[RBP_INDEX(iov->iov_base)]; rbp->status &= ~RBP_LOANED; } Loading Loading @@ -1978,7 +1894,6 @@ he_service_tbrq(struct he_dev *he_dev, int group) } } static void he_service_rbpl(struct he_dev *he_dev, int group) { Loading Loading @@ -2006,33 +1921,6 @@ he_service_rbpl(struct he_dev *he_dev, int group) he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail), G0_RBPL_T); } static void he_service_rbps(struct he_dev *he_dev, int group) { struct he_rbp *newtail; struct he_rbp *rbps_head; int moved = 0; rbps_head = (struct he_rbp *) ((unsigned long)he_dev->rbps_base | RBPS_MASK(he_readl(he_dev, G0_RBPS_S))); for (;;) { newtail = (struct he_rbp *) ((unsigned long)he_dev->rbps_base | RBPS_MASK(he_dev->rbps_tail+1)); /* table 3.42 -- rbps_tail should never be set to rbps_head */ if ((newtail == rbps_head) || (newtail->status & RBP_LOANED)) break; newtail->status |= RBP_LOANED; he_dev->rbps_tail = newtail; ++moved; } if (moved) he_writel(he_dev, RBPS_MASK(he_dev->rbps_tail), G0_RBPS_T); } static void he_tasklet(unsigned long data) { Loading @@ -2055,10 +1943,8 @@ he_tasklet(unsigned long data) HPRINTK("rbrq%d threshold\n", group); /* fall through */ case ITYPE_RBRQ_TIMER: if (he_service_rbrq(he_dev, group)) { if (he_service_rbrq(he_dev, group)) he_service_rbpl(he_dev, group); he_service_rbps(he_dev, group); } break; case ITYPE_TBRQ_THRESH: HPRINTK("tbrq%d threshold\n", group); Loading @@ -2070,7 +1956,7 @@ he_tasklet(unsigned long data) he_service_rbpl(he_dev, group); break; case ITYPE_RBPS_THRESH: he_service_rbps(he_dev, group); /* shouldn't happen unless small buffers enabled */ break; case ITYPE_PHY: HPRINTK("phy interrupt\n"); Loading Loading @@ -2098,7 +1984,6 @@ he_tasklet(unsigned long data) he_service_rbrq(he_dev, 0); he_service_rbpl(he_dev, 0); he_service_rbps(he_dev, 0); he_service_tbrq(he_dev, 0); break; default: Loading Loading @@ -2406,8 +2291,8 @@ he_open(struct atm_vcc *vcc) goto open_failed; } rsr1 = RSR1_GROUP(0); rsr4 = RSR4_GROUP(0); rsr1 = RSR1_GROUP(0) | RSR1_RBPL_ONLY; rsr4 = RSR4_GROUP(0) | RSR4_RBPL_ONLY; rsr0 = vcc->qos.rxtp.traffic_class == ATM_UBR ? (RSR0_EPD_ENABLE|RSR0_PPD_ENABLE) : 0; Loading drivers/atm/he.h +0 −17 Original line number Diff line number Diff line Loading @@ -67,11 +67,6 @@ #define CONFIG_RBPL_BUFSIZE 4096 #define RBPL_MASK(x) (((unsigned long)(x))&((CONFIG_RBPL_SIZE<<3)-1)) #define CONFIG_RBPS_SIZE 1024 #define CONFIG_RBPS_THRESH 64 #define CONFIG_RBPS_BUFSIZE 128 #define RBPS_MASK(x) (((unsigned long)(x))&((CONFIG_RBPS_SIZE<<3)-1)) /* 5.1.3 initialize connection memory */ #define CONFIG_RSRA 0x00000 Loading Loading @@ -225,14 +220,8 @@ struct he_virt { void *virt; }; #define RBPL_ALIGNMENT CONFIG_RBPL_SIZE #define RBPS_ALIGNMENT CONFIG_RBPS_SIZE #ifdef notyet struct he_group { u32 rpbs_size, rpbs_qsize; struct he_rbp rbps_ba; u32 rpbl_size, rpbl_qsize; struct he_rpb_entry *rbpl_ba; }; Loading Loading @@ -303,12 +292,6 @@ struct he_dev { struct he_virt *rbpl_virt; int rbpl_peak; struct pci_pool *rbps_pool; dma_addr_t rbps_phys; struct he_rbp *rbps_base, *rbps_tail; struct he_virt *rbps_virt; int rbps_peak; dma_addr_t tbrq_phys; struct he_tbrq *tbrq_base, *tbrq_head; int tbrq_peak; Loading Loading
drivers/atm/he.c +16 −131 Original line number Diff line number Diff line Loading @@ -780,59 +780,18 @@ he_init_group(struct he_dev *he_dev, int group) { int i; /* small buffer pool */ he_dev->rbps_pool = pci_pool_create("rbps", he_dev->pci_dev, CONFIG_RBPS_BUFSIZE, 8, 0); if (he_dev->rbps_pool == NULL) { hprintk("unable to create rbps pages\n"); return -ENOMEM; } he_dev->rbps_base = pci_alloc_consistent(he_dev->pci_dev, CONFIG_RBPS_SIZE * sizeof(struct he_rbp), &he_dev->rbps_phys); if (he_dev->rbps_base == NULL) { hprintk("failed to alloc rbps_base\n"); goto out_destroy_rbps_pool; } memset(he_dev->rbps_base, 0, CONFIG_RBPS_SIZE * sizeof(struct he_rbp)); he_dev->rbps_virt = kmalloc(CONFIG_RBPS_SIZE * sizeof(struct he_virt), GFP_KERNEL); if (he_dev->rbps_virt == NULL) { hprintk("failed to alloc rbps_virt\n"); goto out_free_rbps_base; } for (i = 0; i < CONFIG_RBPS_SIZE; ++i) { dma_addr_t dma_handle; void *cpuaddr; cpuaddr = pci_pool_alloc(he_dev->rbps_pool, GFP_KERNEL|GFP_DMA, &dma_handle); if (cpuaddr == NULL) goto out_free_rbps_virt; he_dev->rbps_virt[i].virt = cpuaddr; he_dev->rbps_base[i].status = RBP_LOANED | RBP_SMALLBUF | (i << RBP_INDEX_OFF); he_dev->rbps_base[i].phys = dma_handle; } he_dev->rbps_tail = &he_dev->rbps_base[CONFIG_RBPS_SIZE - 1]; he_writel(he_dev, he_dev->rbps_phys, G0_RBPS_S + (group * 32)); he_writel(he_dev, RBPS_MASK(he_dev->rbps_tail), G0_RBPS_T + (group * 32)); he_writel(he_dev, CONFIG_RBPS_BUFSIZE/4, he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32)); he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32)); he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32)); he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0), G0_RBPS_BS + (group * 32)); he_writel(he_dev, RBP_THRESH(CONFIG_RBPS_THRESH) | RBP_QSIZE(CONFIG_RBPS_SIZE - 1) | RBP_INT_ENB, G0_RBPS_QI + (group * 32)); /* large buffer pool */ he_dev->rbpl_pool = pci_pool_create("rbpl", he_dev->pci_dev, CONFIG_RBPL_BUFSIZE, 8, 0); if (he_dev->rbpl_pool == NULL) { hprintk("unable to create rbpl pool\n"); goto out_free_rbps_virt; return -ENOMEM; } he_dev->rbpl_base = pci_alloc_consistent(he_dev->pci_dev, Loading Loading @@ -934,19 +893,6 @@ he_init_group(struct he_dev *he_dev, int group) out_destroy_rbpl_pool: pci_pool_destroy(he_dev->rbpl_pool); i = CONFIG_RBPS_SIZE; out_free_rbps_virt: while (i--) pci_pool_free(he_dev->rbps_pool, he_dev->rbps_virt[i].virt, he_dev->rbps_base[i].phys); kfree(he_dev->rbps_virt); out_free_rbps_base: pci_free_consistent(he_dev->pci_dev, CONFIG_RBPS_SIZE * sizeof(struct he_rbp), he_dev->rbps_base, he_dev->rbps_phys); out_destroy_rbps_pool: pci_pool_destroy(he_dev->rbps_pool); return -ENOMEM; } Loading Loading @@ -1634,22 +1580,6 @@ he_stop(struct he_dev *he_dev) if (he_dev->rbpl_pool) pci_pool_destroy(he_dev->rbpl_pool); if (he_dev->rbps_base) { int i; for (i = 0; i < CONFIG_RBPS_SIZE; ++i) { void *cpuaddr = he_dev->rbps_virt[i].virt; dma_addr_t dma_handle = he_dev->rbps_base[i].phys; pci_pool_free(he_dev->rbps_pool, cpuaddr, dma_handle); } pci_free_consistent(he_dev->pci_dev, CONFIG_RBPS_SIZE * sizeof(struct he_rbp), he_dev->rbps_base, he_dev->rbps_phys); } if (he_dev->rbps_pool) pci_pool_destroy(he_dev->rbps_pool); if (he_dev->rbrq_base) pci_free_consistent(he_dev->pci_dev, CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq), he_dev->rbrq_base, he_dev->rbrq_phys); Loading Loading @@ -1740,9 +1670,6 @@ he_service_rbrq(struct he_dev *he_dev, int group) RBRQ_CON_CLOSED(he_dev->rbrq_head) ? " CON_CLOSED" : "", RBRQ_HBUF_ERR(he_dev->rbrq_head) ? " HBUF_ERR" : ""); if (RBRQ_ADDR(he_dev->rbrq_head) & RBP_SMALLBUF) rbp = &he_dev->rbps_base[RBP_INDEX(RBRQ_ADDR(he_dev->rbrq_head))]; else rbp = &he_dev->rbpl_base[RBP_INDEX(RBRQ_ADDR(he_dev->rbrq_head))]; buf_len = RBRQ_BUFLEN(he_dev->rbrq_head) * 4; Loading Loading @@ -1819,15 +1746,9 @@ he_service_rbrq(struct he_dev *he_dev, int group) __net_timestamp(skb); for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) { if (iov->iov_base & RBP_SMALLBUF) memcpy(skb_put(skb, iov->iov_len), he_dev->rbps_virt[RBP_INDEX(iov->iov_base)].virt, iov->iov_len); else for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) memcpy(skb_put(skb, iov->iov_len), he_dev->rbpl_virt[RBP_INDEX(iov->iov_base)].virt, iov->iov_len); } switch (vcc->qos.aal) { case ATM_AAL0: Loading Loading @@ -1867,13 +1788,8 @@ he_service_rbrq(struct he_dev *he_dev, int group) return_host_buffers: ++pdus_assembled; for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) { if (iov->iov_base & RBP_SMALLBUF) rbp = &he_dev->rbps_base[RBP_INDEX(iov->iov_base)]; else for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) { rbp = &he_dev->rbpl_base[RBP_INDEX(iov->iov_base)]; rbp->status &= ~RBP_LOANED; } Loading Loading @@ -1978,7 +1894,6 @@ he_service_tbrq(struct he_dev *he_dev, int group) } } static void he_service_rbpl(struct he_dev *he_dev, int group) { Loading Loading @@ -2006,33 +1921,6 @@ he_service_rbpl(struct he_dev *he_dev, int group) he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail), G0_RBPL_T); } static void he_service_rbps(struct he_dev *he_dev, int group) { struct he_rbp *newtail; struct he_rbp *rbps_head; int moved = 0; rbps_head = (struct he_rbp *) ((unsigned long)he_dev->rbps_base | RBPS_MASK(he_readl(he_dev, G0_RBPS_S))); for (;;) { newtail = (struct he_rbp *) ((unsigned long)he_dev->rbps_base | RBPS_MASK(he_dev->rbps_tail+1)); /* table 3.42 -- rbps_tail should never be set to rbps_head */ if ((newtail == rbps_head) || (newtail->status & RBP_LOANED)) break; newtail->status |= RBP_LOANED; he_dev->rbps_tail = newtail; ++moved; } if (moved) he_writel(he_dev, RBPS_MASK(he_dev->rbps_tail), G0_RBPS_T); } static void he_tasklet(unsigned long data) { Loading @@ -2055,10 +1943,8 @@ he_tasklet(unsigned long data) HPRINTK("rbrq%d threshold\n", group); /* fall through */ case ITYPE_RBRQ_TIMER: if (he_service_rbrq(he_dev, group)) { if (he_service_rbrq(he_dev, group)) he_service_rbpl(he_dev, group); he_service_rbps(he_dev, group); } break; case ITYPE_TBRQ_THRESH: HPRINTK("tbrq%d threshold\n", group); Loading @@ -2070,7 +1956,7 @@ he_tasklet(unsigned long data) he_service_rbpl(he_dev, group); break; case ITYPE_RBPS_THRESH: he_service_rbps(he_dev, group); /* shouldn't happen unless small buffers enabled */ break; case ITYPE_PHY: HPRINTK("phy interrupt\n"); Loading Loading @@ -2098,7 +1984,6 @@ he_tasklet(unsigned long data) he_service_rbrq(he_dev, 0); he_service_rbpl(he_dev, 0); he_service_rbps(he_dev, 0); he_service_tbrq(he_dev, 0); break; default: Loading Loading @@ -2406,8 +2291,8 @@ he_open(struct atm_vcc *vcc) goto open_failed; } rsr1 = RSR1_GROUP(0); rsr4 = RSR4_GROUP(0); rsr1 = RSR1_GROUP(0) | RSR1_RBPL_ONLY; rsr4 = RSR4_GROUP(0) | RSR4_RBPL_ONLY; rsr0 = vcc->qos.rxtp.traffic_class == ATM_UBR ? (RSR0_EPD_ENABLE|RSR0_PPD_ENABLE) : 0; Loading
drivers/atm/he.h +0 −17 Original line number Diff line number Diff line Loading @@ -67,11 +67,6 @@ #define CONFIG_RBPL_BUFSIZE 4096 #define RBPL_MASK(x) (((unsigned long)(x))&((CONFIG_RBPL_SIZE<<3)-1)) #define CONFIG_RBPS_SIZE 1024 #define CONFIG_RBPS_THRESH 64 #define CONFIG_RBPS_BUFSIZE 128 #define RBPS_MASK(x) (((unsigned long)(x))&((CONFIG_RBPS_SIZE<<3)-1)) /* 5.1.3 initialize connection memory */ #define CONFIG_RSRA 0x00000 Loading Loading @@ -225,14 +220,8 @@ struct he_virt { void *virt; }; #define RBPL_ALIGNMENT CONFIG_RBPL_SIZE #define RBPS_ALIGNMENT CONFIG_RBPS_SIZE #ifdef notyet struct he_group { u32 rpbs_size, rpbs_qsize; struct he_rbp rbps_ba; u32 rpbl_size, rpbl_qsize; struct he_rpb_entry *rbpl_ba; }; Loading Loading @@ -303,12 +292,6 @@ struct he_dev { struct he_virt *rbpl_virt; int rbpl_peak; struct pci_pool *rbps_pool; dma_addr_t rbps_phys; struct he_rbp *rbps_base, *rbps_tail; struct he_virt *rbps_virt; int rbps_peak; dma_addr_t tbrq_phys; struct he_tbrq *tbrq_base, *tbrq_head; int tbrq_peak; Loading