123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495 |
- /*
- * Copyright (c) 2005, 2006, 2007, 2008 Mellanox Technologies. All rights reserved.
- * Copyright (c) 2006, 2007 Cisco Systems, Inc. All rights reserved.
- *
- * This software is available to you under a choice of one of two
- * licenses. You may choose to be licensed under the terms of the GNU
- * General Public License (GPL) Version 2, available from the file
- * COPYING in the main directory of this source tree, or the
- * OpenIB.org BSD license below:
- *
- * Redistribution and use in source and binary forms, with or
- * without modification, are permitted provided that the following
- * conditions are met:
- *
- * - Redistributions of source code must retain the above
- * copyright notice, this list of conditions and the following
- * disclaimer.
- *
- * - Redistributions in binary form must reproduce the above
- * copyright notice, this list of conditions and the following
- * disclaimer in the documentation and/or other materials
- * provided with the distribution.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
- * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
- * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
- * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
- * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
- * SOFTWARE.
- */
- #include <linux/errno.h>
- #include <linux/mm.h>
- #include <linux/scatterlist.h>
- #include <linux/slab.h>
- #include <linux/mlx4/cmd.h>
- #include "mlx4.h"
- #include "icm.h"
- #include "fw.h"
- /*
- * We allocate in as big chunks as we can, up to a maximum of 256 KB
- * per chunk. Note that the chunks are not necessarily in contiguous
- * physical memory.
- */
- enum {
- MLX4_ICM_ALLOC_SIZE = 1 << 18,
- MLX4_TABLE_CHUNK_SIZE = 1 << 18,
- };
- static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk)
- {
- int i;
- if (chunk->nsg > 0)
- pci_unmap_sg(dev->persist->pdev, chunk->sg, chunk->npages,
- PCI_DMA_BIDIRECTIONAL);
- for (i = 0; i < chunk->npages; ++i)
- __free_pages(sg_page(&chunk->sg[i]),
- get_order(chunk->sg[i].length));
- }
- static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk)
- {
- int i;
- for (i = 0; i < chunk->npages; ++i)
- dma_free_coherent(&dev->persist->pdev->dev,
- chunk->buf[i].size,
- chunk->buf[i].addr,
- chunk->buf[i].dma_addr);
- }
- void mlx4_free_icm(struct mlx4_dev *dev, struct mlx4_icm *icm, int coherent)
- {
- struct mlx4_icm_chunk *chunk, *tmp;
- if (!icm)
- return;
- list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) {
- if (coherent)
- mlx4_free_icm_coherent(dev, chunk);
- else
- mlx4_free_icm_pages(dev, chunk);
- kfree(chunk);
- }
- kfree(icm);
- }
- static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order,
- gfp_t gfp_mask, int node)
- {
- struct page *page;
- page = alloc_pages_node(node, gfp_mask, order);
- if (!page) {
- page = alloc_pages(gfp_mask, order);
- if (!page)
- return -ENOMEM;
- }
- sg_set_page(mem, page, PAGE_SIZE << order, 0);
- return 0;
- }
- static int mlx4_alloc_icm_coherent(struct device *dev, struct mlx4_icm_buf *buf,
- int order, gfp_t gfp_mask)
- {
- buf->addr = dma_alloc_coherent(dev, PAGE_SIZE << order,
- &buf->dma_addr, gfp_mask);
- if (!buf->addr)
- return -ENOMEM;
- if (offset_in_page(buf->addr)) {
- dma_free_coherent(dev, PAGE_SIZE << order, buf->addr,
- buf->dma_addr);
- return -ENOMEM;
- }
- buf->size = PAGE_SIZE << order;
- return 0;
- }
- struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages,
- gfp_t gfp_mask, int coherent)
- {
- struct mlx4_icm *icm;
- struct mlx4_icm_chunk *chunk = NULL;
- int cur_order;
- gfp_t mask;
- int ret;
- /* We use sg_set_buf for coherent allocs, which assumes low memory */
- BUG_ON(coherent && (gfp_mask & __GFP_HIGHMEM));
- icm = kmalloc_node(sizeof(*icm),
- gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN),
- dev->numa_node);
- if (!icm) {
- icm = kmalloc(sizeof(*icm),
- gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN));
- if (!icm)
- return NULL;
- }
- icm->refcount = 0;
- INIT_LIST_HEAD(&icm->chunk_list);
- cur_order = get_order(MLX4_ICM_ALLOC_SIZE);
- while (npages > 0) {
- if (!chunk) {
- chunk = kzalloc_node(sizeof(*chunk),
- gfp_mask & ~(__GFP_HIGHMEM |
- __GFP_NOWARN),
- dev->numa_node);
- if (!chunk) {
- chunk = kzalloc(sizeof(*chunk),
- gfp_mask & ~(__GFP_HIGHMEM |
- __GFP_NOWARN));
- if (!chunk)
- goto fail;
- }
- chunk->coherent = coherent;
- if (!coherent)
- sg_init_table(chunk->sg, MLX4_ICM_CHUNK_LEN);
- list_add_tail(&chunk->list, &icm->chunk_list);
- }
- while (1 << cur_order > npages)
- --cur_order;
- mask = gfp_mask;
- if (cur_order)
- mask &= ~__GFP_DIRECT_RECLAIM;
- if (coherent)
- ret = mlx4_alloc_icm_coherent(&dev->persist->pdev->dev,
- &chunk->buf[chunk->npages],
- cur_order, mask);
- else
- ret = mlx4_alloc_icm_pages(&chunk->sg[chunk->npages],
- cur_order, mask,
- dev->numa_node);
- if (ret) {
- if (--cur_order < 0)
- goto fail;
- else
- continue;
- }
- ++chunk->npages;
- if (coherent)
- ++chunk->nsg;
- else if (chunk->npages == MLX4_ICM_CHUNK_LEN) {
- chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->sg,
- chunk->npages,
- PCI_DMA_BIDIRECTIONAL);
- if (chunk->nsg <= 0)
- goto fail;
- }
- if (chunk->npages == MLX4_ICM_CHUNK_LEN)
- chunk = NULL;
- npages -= 1 << cur_order;
- }
- if (!coherent && chunk) {
- chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->sg,
- chunk->npages,
- PCI_DMA_BIDIRECTIONAL);
- if (chunk->nsg <= 0)
- goto fail;
- }
- return icm;
- fail:
- mlx4_free_icm(dev, icm, coherent);
- return NULL;
- }
- static int mlx4_MAP_ICM(struct mlx4_dev *dev, struct mlx4_icm *icm, u64 virt)
- {
- return mlx4_map_cmd(dev, MLX4_CMD_MAP_ICM, icm, virt);
- }
- static int mlx4_UNMAP_ICM(struct mlx4_dev *dev, u64 virt, u32 page_count)
- {
- return mlx4_cmd(dev, virt, page_count, 0, MLX4_CMD_UNMAP_ICM,
- MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
- }
- int mlx4_MAP_ICM_AUX(struct mlx4_dev *dev, struct mlx4_icm *icm)
- {
- return mlx4_map_cmd(dev, MLX4_CMD_MAP_ICM_AUX, icm, -1);
- }
- int mlx4_UNMAP_ICM_AUX(struct mlx4_dev *dev)
- {
- return mlx4_cmd(dev, 0, 0, 0, MLX4_CMD_UNMAP_ICM_AUX,
- MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
- }
- int mlx4_table_get(struct mlx4_dev *dev, struct mlx4_icm_table *table, u32 obj)
- {
- u32 i = (obj & (table->num_obj - 1)) /
- (MLX4_TABLE_CHUNK_SIZE / table->obj_size);
- int ret = 0;
- mutex_lock(&table->mutex);
- if (table->icm[i]) {
- ++table->icm[i]->refcount;
- goto out;
- }
- table->icm[i] = mlx4_alloc_icm(dev, MLX4_TABLE_CHUNK_SIZE >> PAGE_SHIFT,
- (table->lowmem ? GFP_KERNEL : GFP_HIGHUSER) |
- __GFP_NOWARN, table->coherent);
- if (!table->icm[i]) {
- ret = -ENOMEM;
- goto out;
- }
- if (mlx4_MAP_ICM(dev, table->icm[i], table->virt +
- (u64) i * MLX4_TABLE_CHUNK_SIZE)) {
- mlx4_free_icm(dev, table->icm[i], table->coherent);
- table->icm[i] = NULL;
- ret = -ENOMEM;
- goto out;
- }
- ++table->icm[i]->refcount;
- out:
- mutex_unlock(&table->mutex);
- return ret;
- }
- void mlx4_table_put(struct mlx4_dev *dev, struct mlx4_icm_table *table, u32 obj)
- {
- u32 i;
- u64 offset;
- i = (obj & (table->num_obj - 1)) / (MLX4_TABLE_CHUNK_SIZE / table->obj_size);
- mutex_lock(&table->mutex);
- if (--table->icm[i]->refcount == 0) {
- offset = (u64) i * MLX4_TABLE_CHUNK_SIZE;
- mlx4_UNMAP_ICM(dev, table->virt + offset,
- MLX4_TABLE_CHUNK_SIZE / MLX4_ICM_PAGE_SIZE);
- mlx4_free_icm(dev, table->icm[i], table->coherent);
- table->icm[i] = NULL;
- }
- mutex_unlock(&table->mutex);
- }
- void *mlx4_table_find(struct mlx4_icm_table *table, u32 obj,
- dma_addr_t *dma_handle)
- {
- int offset, dma_offset, i;
- u64 idx;
- struct mlx4_icm_chunk *chunk;
- struct mlx4_icm *icm;
- void *addr = NULL;
- if (!table->lowmem)
- return NULL;
- mutex_lock(&table->mutex);
- idx = (u64) (obj & (table->num_obj - 1)) * table->obj_size;
- icm = table->icm[idx / MLX4_TABLE_CHUNK_SIZE];
- dma_offset = offset = idx % MLX4_TABLE_CHUNK_SIZE;
- if (!icm)
- goto out;
- list_for_each_entry(chunk, &icm->chunk_list, list) {
- for (i = 0; i < chunk->npages; ++i) {
- dma_addr_t dma_addr;
- size_t len;
- if (table->coherent) {
- len = chunk->buf[i].size;
- dma_addr = chunk->buf[i].dma_addr;
- addr = chunk->buf[i].addr;
- } else {
- struct page *page;
- len = sg_dma_len(&chunk->sg[i]);
- dma_addr = sg_dma_address(&chunk->sg[i]);
- /* XXX: we should never do this for highmem
- * allocation. This function either needs
- * to be split, or the kernel virtual address
- * return needs to be made optional.
- */
- page = sg_page(&chunk->sg[i]);
- addr = lowmem_page_address(page);
- }
- if (dma_handle && dma_offset >= 0) {
- if (len > dma_offset)
- *dma_handle = dma_addr + dma_offset;
- dma_offset -= len;
- }
- /*
- * DMA mapping can merge pages but not split them,
- * so if we found the page, dma_handle has already
- * been assigned to.
- */
- if (len > offset)
- goto out;
- offset -= len;
- }
- }
- addr = NULL;
- out:
- mutex_unlock(&table->mutex);
- return addr ? addr + offset : NULL;
- }
- int mlx4_table_get_range(struct mlx4_dev *dev, struct mlx4_icm_table *table,
- u32 start, u32 end)
- {
- int inc = MLX4_TABLE_CHUNK_SIZE / table->obj_size;
- int err;
- u32 i;
- for (i = start; i <= end; i += inc) {
- err = mlx4_table_get(dev, table, i);
- if (err)
- goto fail;
- }
- return 0;
- fail:
- while (i > start) {
- i -= inc;
- mlx4_table_put(dev, table, i);
- }
- return err;
- }
- void mlx4_table_put_range(struct mlx4_dev *dev, struct mlx4_icm_table *table,
- u32 start, u32 end)
- {
- u32 i;
- for (i = start; i <= end; i += MLX4_TABLE_CHUNK_SIZE / table->obj_size)
- mlx4_table_put(dev, table, i);
- }
- int mlx4_init_icm_table(struct mlx4_dev *dev, struct mlx4_icm_table *table,
- u64 virt, int obj_size, u32 nobj, int reserved,
- int use_lowmem, int use_coherent)
- {
- int obj_per_chunk;
- int num_icm;
- unsigned chunk_size;
- int i;
- u64 size;
- obj_per_chunk = MLX4_TABLE_CHUNK_SIZE / obj_size;
- if (WARN_ON(!obj_per_chunk))
- return -EINVAL;
- num_icm = (nobj + obj_per_chunk - 1) / obj_per_chunk;
- table->icm = kvcalloc(num_icm, sizeof(*table->icm), GFP_KERNEL);
- if (!table->icm)
- return -ENOMEM;
- table->virt = virt;
- table->num_icm = num_icm;
- table->num_obj = nobj;
- table->obj_size = obj_size;
- table->lowmem = use_lowmem;
- table->coherent = use_coherent;
- mutex_init(&table->mutex);
- size = (u64) nobj * obj_size;
- for (i = 0; i * MLX4_TABLE_CHUNK_SIZE < reserved * obj_size; ++i) {
- chunk_size = MLX4_TABLE_CHUNK_SIZE;
- if ((i + 1) * MLX4_TABLE_CHUNK_SIZE > size)
- chunk_size = PAGE_ALIGN(size -
- i * MLX4_TABLE_CHUNK_SIZE);
- table->icm[i] = mlx4_alloc_icm(dev, chunk_size >> PAGE_SHIFT,
- (use_lowmem ? GFP_KERNEL : GFP_HIGHUSER) |
- __GFP_NOWARN, use_coherent);
- if (!table->icm[i])
- goto err;
- if (mlx4_MAP_ICM(dev, table->icm[i], virt + i * MLX4_TABLE_CHUNK_SIZE)) {
- mlx4_free_icm(dev, table->icm[i], use_coherent);
- table->icm[i] = NULL;
- goto err;
- }
- /*
- * Add a reference to this ICM chunk so that it never
- * gets freed (since it contains reserved firmware objects).
- */
- ++table->icm[i]->refcount;
- }
- return 0;
- err:
- for (i = 0; i < num_icm; ++i)
- if (table->icm[i]) {
- mlx4_UNMAP_ICM(dev, virt + i * MLX4_TABLE_CHUNK_SIZE,
- MLX4_TABLE_CHUNK_SIZE / MLX4_ICM_PAGE_SIZE);
- mlx4_free_icm(dev, table->icm[i], use_coherent);
- }
- kvfree(table->icm);
- return -ENOMEM;
- }
- void mlx4_cleanup_icm_table(struct mlx4_dev *dev, struct mlx4_icm_table *table)
- {
- int i;
- for (i = 0; i < table->num_icm; ++i)
- if (table->icm[i]) {
- mlx4_UNMAP_ICM(dev, table->virt + i * MLX4_TABLE_CHUNK_SIZE,
- MLX4_TABLE_CHUNK_SIZE / MLX4_ICM_PAGE_SIZE);
- mlx4_free_icm(dev, table->icm[i], table->coherent);
- }
- kvfree(table->icm);
- }
|