|  | /* | 
|  | *  IBM eServer eHCA Infiniband device driver for Linux on POWER | 
|  | * | 
|  | *  internal queue handling | 
|  | * | 
|  | *  Authors: Waleri Fomin <fomin@de.ibm.com> | 
|  | *           Reinhard Ernst <rernst@de.ibm.com> | 
|  | *           Christoph Raisch <raisch@de.ibm.com> | 
|  | * | 
|  | *  Copyright (c) 2005 IBM Corporation | 
|  | * | 
|  | *  This source code is distributed under a dual license of GPL v2.0 and OpenIB | 
|  | *  BSD. | 
|  | * | 
|  | * OpenIB BSD License | 
|  | * | 
|  | * Redistribution and use in source and binary forms, with or without | 
|  | * modification, are permitted provided that the following conditions are met: | 
|  | * | 
|  | * Redistributions of source code must retain the above copyright notice, this | 
|  | * list of conditions and the following disclaimer. | 
|  | * | 
|  | * Redistributions in binary form must reproduce the above copyright notice, | 
|  | * this list of conditions and the following disclaimer in the documentation | 
|  | * and/or other materials | 
|  | * provided with the distribution. | 
|  | * | 
|  | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | 
|  | * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | 
|  | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | 
|  | * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE | 
|  | * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR | 
|  | * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF | 
|  | * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR | 
|  | * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER | 
|  | * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) | 
|  | * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE | 
|  | * POSSIBILITY OF SUCH DAMAGE. | 
|  | */ | 
|  |  | 
|  | #include "ehca_tools.h" | 
|  | #include "ipz_pt_fn.h" | 
|  | #include "ehca_classes.h" | 
|  |  | 
|  | #define PAGES_PER_KPAGE (PAGE_SIZE >> EHCA_PAGESHIFT) | 
|  |  | 
|  | struct kmem_cache *small_qp_cache; | 
|  |  | 
|  | void *ipz_qpageit_get_inc(struct ipz_queue *queue) | 
|  | { | 
|  | void *ret = ipz_qeit_get(queue); | 
|  | queue->current_q_offset += queue->pagesize; | 
|  | if (queue->current_q_offset > queue->queue_length) { | 
|  | queue->current_q_offset -= queue->pagesize; | 
|  | ret = NULL; | 
|  | } | 
|  | if (((u64)ret) % queue->pagesize) { | 
|  | ehca_gen_err("ERROR!! not at PAGE-Boundary"); | 
|  | return NULL; | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | void *ipz_qeit_eq_get_inc(struct ipz_queue *queue) | 
|  | { | 
|  | void *ret = ipz_qeit_get(queue); | 
|  | u64 last_entry_in_q = queue->queue_length - queue->qe_size; | 
|  |  | 
|  | queue->current_q_offset += queue->qe_size; | 
|  | if (queue->current_q_offset > last_entry_in_q) { | 
|  | queue->current_q_offset = 0; | 
|  | queue->toggle_state = (~queue->toggle_state) & 1; | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | int ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset) | 
|  | { | 
|  | int i; | 
|  | for (i = 0; i < queue->queue_length / queue->pagesize; i++) { | 
|  | u64 page = (u64)virt_to_abs(queue->queue_pages[i]); | 
|  | if (addr >= page && addr < page + queue->pagesize) { | 
|  | *q_offset = addr - page + i * queue->pagesize; | 
|  | return 0; | 
|  | } | 
|  | } | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | #if PAGE_SHIFT < EHCA_PAGESHIFT | 
|  | #error Kernel pages must be at least as large than eHCA pages (4K) ! | 
|  | #endif | 
|  |  | 
|  | /* | 
|  | * allocate pages for queue: | 
|  | * outer loop allocates whole kernel pages (page aligned) and | 
|  | * inner loop divides a kernel page into smaller hca queue pages | 
|  | */ | 
|  | static int alloc_queue_pages(struct ipz_queue *queue, const u32 nr_of_pages) | 
|  | { | 
|  | int k, f = 0; | 
|  | u8 *kpage; | 
|  |  | 
|  | while (f < nr_of_pages) { | 
|  | kpage = (u8 *)get_zeroed_page(GFP_KERNEL); | 
|  | if (!kpage) | 
|  | goto out; | 
|  |  | 
|  | for (k = 0; k < PAGES_PER_KPAGE && f < nr_of_pages; k++) { | 
|  | queue->queue_pages[f] = (struct ipz_page *)kpage; | 
|  | kpage += EHCA_PAGESIZE; | 
|  | f++; | 
|  | } | 
|  | } | 
|  | return 1; | 
|  |  | 
|  | out: | 
|  | for (f = 0; f < nr_of_pages && queue->queue_pages[f]; | 
|  | f += PAGES_PER_KPAGE) | 
|  | free_page((unsigned long)(queue->queue_pages)[f]); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) | 
|  | { | 
|  | int order = ilog2(queue->pagesize) - 9; | 
|  | struct ipz_small_queue_page *page; | 
|  | unsigned long bit; | 
|  |  | 
|  | mutex_lock(&pd->lock); | 
|  |  | 
|  | if (!list_empty(&pd->free[order])) | 
|  | page = list_entry(pd->free[order].next, | 
|  | struct ipz_small_queue_page, list); | 
|  | else { | 
|  | page = kmem_cache_zalloc(small_qp_cache, GFP_KERNEL); | 
|  | if (!page) | 
|  | goto out; | 
|  |  | 
|  | page->page = get_zeroed_page(GFP_KERNEL); | 
|  | if (!page->page) { | 
|  | kmem_cache_free(small_qp_cache, page); | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | list_add(&page->list, &pd->free[order]); | 
|  | } | 
|  |  | 
|  | bit = find_first_zero_bit(page->bitmap, IPZ_SPAGE_PER_KPAGE >> order); | 
|  | __set_bit(bit, page->bitmap); | 
|  | page->fill++; | 
|  |  | 
|  | if (page->fill == IPZ_SPAGE_PER_KPAGE >> order) | 
|  | list_move(&page->list, &pd->full[order]); | 
|  |  | 
|  | mutex_unlock(&pd->lock); | 
|  |  | 
|  | queue->queue_pages[0] = (void *)(page->page | (bit << (order + 9))); | 
|  | queue->small_page = page; | 
|  | queue->offset = bit << (order + 9); | 
|  | return 1; | 
|  |  | 
|  | out: | 
|  | ehca_err(pd->ib_pd.device, "failed to allocate small queue page"); | 
|  | mutex_unlock(&pd->lock); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) | 
|  | { | 
|  | int order = ilog2(queue->pagesize) - 9; | 
|  | struct ipz_small_queue_page *page = queue->small_page; | 
|  | unsigned long bit; | 
|  | int free_page = 0; | 
|  |  | 
|  | bit = ((unsigned long)queue->queue_pages[0] & ~PAGE_MASK) | 
|  | >> (order + 9); | 
|  |  | 
|  | mutex_lock(&pd->lock); | 
|  |  | 
|  | __clear_bit(bit, page->bitmap); | 
|  | page->fill--; | 
|  |  | 
|  | if (page->fill == 0) { | 
|  | list_del(&page->list); | 
|  | free_page = 1; | 
|  | } | 
|  |  | 
|  | if (page->fill == (IPZ_SPAGE_PER_KPAGE >> order) - 1) | 
|  | /* the page was full until we freed the chunk */ | 
|  | list_move_tail(&page->list, &pd->free[order]); | 
|  |  | 
|  | mutex_unlock(&pd->lock); | 
|  |  | 
|  | if (free_page) { | 
|  | free_page(page->page); | 
|  | kmem_cache_free(small_qp_cache, page); | 
|  | } | 
|  | } | 
|  |  | 
|  | int ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue, | 
|  | const u32 nr_of_pages, const u32 pagesize, | 
|  | const u32 qe_size, const u32 nr_of_sg, | 
|  | int is_small) | 
|  | { | 
|  | if (pagesize > PAGE_SIZE) { | 
|  | ehca_gen_err("FATAL ERROR: pagesize=%x " | 
|  | "is greater than kernel page size", pagesize); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* init queue fields */ | 
|  | queue->queue_length = nr_of_pages * pagesize; | 
|  | queue->pagesize = pagesize; | 
|  | queue->qe_size = qe_size; | 
|  | queue->act_nr_of_sg = nr_of_sg; | 
|  | queue->current_q_offset = 0; | 
|  | queue->toggle_state = 1; | 
|  | queue->small_page = NULL; | 
|  |  | 
|  | /* allocate queue page pointers */ | 
|  | queue->queue_pages = vmalloc(nr_of_pages * sizeof(void *)); | 
|  | if (!queue->queue_pages) { | 
|  | ehca_gen_err("Couldn't allocate queue page list"); | 
|  | return 0; | 
|  | } | 
|  | memset(queue->queue_pages, 0, nr_of_pages * sizeof(void *)); | 
|  |  | 
|  | /* allocate actual queue pages */ | 
|  | if (is_small) { | 
|  | if (!alloc_small_queue_page(queue, pd)) | 
|  | goto ipz_queue_ctor_exit0; | 
|  | } else | 
|  | if (!alloc_queue_pages(queue, nr_of_pages)) | 
|  | goto ipz_queue_ctor_exit0; | 
|  |  | 
|  | return 1; | 
|  |  | 
|  | ipz_queue_ctor_exit0: | 
|  | ehca_gen_err("Couldn't alloc pages queue=%p " | 
|  | "nr_of_pages=%x",  queue, nr_of_pages); | 
|  | vfree(queue->queue_pages); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue) | 
|  | { | 
|  | int i, nr_pages; | 
|  |  | 
|  | if (!queue || !queue->queue_pages) { | 
|  | ehca_gen_dbg("queue or queue_pages is NULL"); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (queue->small_page) | 
|  | free_small_queue_page(queue, pd); | 
|  | else { | 
|  | nr_pages = queue->queue_length / queue->pagesize; | 
|  | for (i = 0; i < nr_pages; i += PAGES_PER_KPAGE) | 
|  | free_page((unsigned long)queue->queue_pages[i]); | 
|  | } | 
|  |  | 
|  | vfree(queue->queue_pages); | 
|  |  | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | int ehca_init_small_qp_cache(void) | 
|  | { | 
|  | small_qp_cache = kmem_cache_create("ehca_cache_small_qp", | 
|  | sizeof(struct ipz_small_queue_page), | 
|  | 0, SLAB_HWCACHE_ALIGN, NULL); | 
|  | if (!small_qp_cache) | 
|  | return -ENOMEM; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | void ehca_cleanup_small_qp_cache(void) | 
|  | { | 
|  | kmem_cache_destroy(small_qp_cache); | 
|  | } |