|  | /* | 
|  | * Copyright(c) 2004 - 2006 Intel Corporation. All rights reserved. | 
|  | * Portions based on net/core/datagram.c and copyrighted by their authors. | 
|  | * | 
|  | * This program is free software; you can redistribute it and/or modify it | 
|  | * under the terms of the GNU General Public License as published by the Free | 
|  | * Software Foundation; either version 2 of the License, or (at your option) | 
|  | * any later version. | 
|  | * | 
|  | * This program is distributed in the hope that it will be useful, but WITHOUT | 
|  | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 
|  | * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for | 
|  | * more details. | 
|  | * | 
|  | * You should have received a copy of the GNU General Public License along with | 
|  | * this program; if not, write to the Free Software Foundation, Inc., 59 | 
|  | * Temple Place - Suite 330, Boston, MA  02111-1307, USA. | 
|  | * | 
|  | * The full GNU General Public License is included in this distribution in the | 
|  | * file called COPYING. | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * This code allows the net stack to make use of a DMA engine for | 
|  | * skb to iovec copies. | 
|  | */ | 
|  |  | 
|  | #include <linux/dmaengine.h> | 
|  | #include <linux/pagemap.h> | 
|  | #include <linux/slab.h> | 
|  | #include <net/tcp.h> /* for memcpy_toiovec */ | 
|  | #include <asm/io.h> | 
|  | #include <asm/uaccess.h> | 
|  |  | 
|  | static int num_pages_spanned(struct iovec *iov) | 
|  | { | 
|  | return | 
|  | ((PAGE_ALIGN((unsigned long)iov->iov_base + iov->iov_len) - | 
|  | ((unsigned long)iov->iov_base & PAGE_MASK)) >> PAGE_SHIFT); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Pin down all the iovec pages needed for len bytes. | 
|  | * Return a struct dma_pinned_list to keep track of pages pinned down. | 
|  | * | 
|  | * We are allocating a single chunk of memory, and then carving it up into | 
|  | * 3 sections, the latter 2 whose size depends on the number of iovecs and the | 
|  | * total number of pages, respectively. | 
|  | */ | 
|  | struct dma_pinned_list *dma_pin_iovec_pages(struct iovec *iov, size_t len) | 
|  | { | 
|  | struct dma_pinned_list *local_list; | 
|  | struct page **pages; | 
|  | int i; | 
|  | int ret; | 
|  | int nr_iovecs = 0; | 
|  | int iovec_len_used = 0; | 
|  | int iovec_pages_used = 0; | 
|  |  | 
|  | /* don't pin down non-user-based iovecs */ | 
|  | if (segment_eq(get_fs(), KERNEL_DS)) | 
|  | return NULL; | 
|  |  | 
|  | /* determine how many iovecs/pages there are, up front */ | 
|  | do { | 
|  | iovec_len_used += iov[nr_iovecs].iov_len; | 
|  | iovec_pages_used += num_pages_spanned(&iov[nr_iovecs]); | 
|  | nr_iovecs++; | 
|  | } while (iovec_len_used < len); | 
|  |  | 
|  | /* single kmalloc for pinned list, page_list[], and the page arrays */ | 
|  | local_list = kmalloc(sizeof(*local_list) | 
|  | + (nr_iovecs * sizeof (struct dma_page_list)) | 
|  | + (iovec_pages_used * sizeof (struct page*)), GFP_KERNEL); | 
|  | if (!local_list) | 
|  | goto out; | 
|  |  | 
|  | /* list of pages starts right after the page list array */ | 
|  | pages = (struct page **) &local_list->page_list[nr_iovecs]; | 
|  |  | 
|  | local_list->nr_iovecs = 0; | 
|  |  | 
|  | for (i = 0; i < nr_iovecs; i++) { | 
|  | struct dma_page_list *page_list = &local_list->page_list[i]; | 
|  |  | 
|  | len -= iov[i].iov_len; | 
|  |  | 
|  | if (!access_ok(VERIFY_WRITE, iov[i].iov_base, iov[i].iov_len)) | 
|  | goto unpin; | 
|  |  | 
|  | page_list->nr_pages = num_pages_spanned(&iov[i]); | 
|  | page_list->base_address = iov[i].iov_base; | 
|  |  | 
|  | page_list->pages = pages; | 
|  | pages += page_list->nr_pages; | 
|  |  | 
|  | /* pin pages down */ | 
|  | down_read(¤t->mm->mmap_sem); | 
|  | ret = get_user_pages( | 
|  | current, | 
|  | current->mm, | 
|  | (unsigned long) iov[i].iov_base, | 
|  | page_list->nr_pages, | 
|  | 1,	/* write */ | 
|  | 0,	/* force */ | 
|  | page_list->pages, | 
|  | NULL); | 
|  | up_read(¤t->mm->mmap_sem); | 
|  |  | 
|  | if (ret != page_list->nr_pages) | 
|  | goto unpin; | 
|  |  | 
|  | local_list->nr_iovecs = i + 1; | 
|  | } | 
|  |  | 
|  | return local_list; | 
|  |  | 
|  | unpin: | 
|  | dma_unpin_iovec_pages(local_list); | 
|  | out: | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | void dma_unpin_iovec_pages(struct dma_pinned_list *pinned_list) | 
|  | { | 
|  | int i, j; | 
|  |  | 
|  | if (!pinned_list) | 
|  | return; | 
|  |  | 
|  | for (i = 0; i < pinned_list->nr_iovecs; i++) { | 
|  | struct dma_page_list *page_list = &pinned_list->page_list[i]; | 
|  | for (j = 0; j < page_list->nr_pages; j++) { | 
|  | set_page_dirty_lock(page_list->pages[j]); | 
|  | page_cache_release(page_list->pages[j]); | 
|  | } | 
|  | } | 
|  |  | 
|  | kfree(pinned_list); | 
|  | } | 
|  |  | 
|  |  | 
|  | /* | 
|  | * We have already pinned down the pages we will be using in the iovecs. | 
|  | * Each entry in iov array has corresponding entry in pinned_list->page_list. | 
|  | * Using array indexing to keep iov[] and page_list[] in sync. | 
|  | * Initial elements in iov array's iov->iov_len will be 0 if already copied into | 
|  | *   by another call. | 
|  | * iov array length remaining guaranteed to be bigger than len. | 
|  | */ | 
|  | dma_cookie_t dma_memcpy_to_iovec(struct dma_chan *chan, struct iovec *iov, | 
|  | struct dma_pinned_list *pinned_list, unsigned char *kdata, size_t len) | 
|  | { | 
|  | int iov_byte_offset; | 
|  | int copy; | 
|  | dma_cookie_t dma_cookie = 0; | 
|  | int iovec_idx; | 
|  | int page_idx; | 
|  |  | 
|  | if (!chan) | 
|  | return memcpy_toiovec(iov, kdata, len); | 
|  |  | 
|  | iovec_idx = 0; | 
|  | while (iovec_idx < pinned_list->nr_iovecs) { | 
|  | struct dma_page_list *page_list; | 
|  |  | 
|  | /* skip already used-up iovecs */ | 
|  | while (!iov[iovec_idx].iov_len) | 
|  | iovec_idx++; | 
|  |  | 
|  | page_list = &pinned_list->page_list[iovec_idx]; | 
|  |  | 
|  | iov_byte_offset = ((unsigned long)iov[iovec_idx].iov_base & ~PAGE_MASK); | 
|  | page_idx = (((unsigned long)iov[iovec_idx].iov_base & PAGE_MASK) | 
|  | - ((unsigned long)page_list->base_address & PAGE_MASK)) >> PAGE_SHIFT; | 
|  |  | 
|  | /* break up copies to not cross page boundary */ | 
|  | while (iov[iovec_idx].iov_len) { | 
|  | copy = min_t(int, PAGE_SIZE - iov_byte_offset, len); | 
|  | copy = min_t(int, copy, iov[iovec_idx].iov_len); | 
|  |  | 
|  | dma_cookie = dma_async_memcpy_buf_to_pg(chan, | 
|  | page_list->pages[page_idx], | 
|  | iov_byte_offset, | 
|  | kdata, | 
|  | copy); | 
|  | /* poll for a descriptor slot */ | 
|  | if (unlikely(dma_cookie < 0)) { | 
|  | dma_async_issue_pending(chan); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | len -= copy; | 
|  | iov[iovec_idx].iov_len -= copy; | 
|  | iov[iovec_idx].iov_base += copy; | 
|  |  | 
|  | if (!len) | 
|  | return dma_cookie; | 
|  |  | 
|  | kdata += copy; | 
|  | iov_byte_offset = 0; | 
|  | page_idx++; | 
|  | } | 
|  | iovec_idx++; | 
|  | } | 
|  |  | 
|  | /* really bad if we ever run out of iovecs */ | 
|  | BUG(); | 
|  | return -EFAULT; | 
|  | } | 
|  |  | 
|  | dma_cookie_t dma_memcpy_pg_to_iovec(struct dma_chan *chan, struct iovec *iov, | 
|  | struct dma_pinned_list *pinned_list, struct page *page, | 
|  | unsigned int offset, size_t len) | 
|  | { | 
|  | int iov_byte_offset; | 
|  | int copy; | 
|  | dma_cookie_t dma_cookie = 0; | 
|  | int iovec_idx; | 
|  | int page_idx; | 
|  | int err; | 
|  |  | 
|  | /* this needs as-yet-unimplemented buf-to-buff, so punt. */ | 
|  | /* TODO: use dma for this */ | 
|  | if (!chan || !pinned_list) { | 
|  | u8 *vaddr = kmap(page); | 
|  | err = memcpy_toiovec(iov, vaddr + offset, len); | 
|  | kunmap(page); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | iovec_idx = 0; | 
|  | while (iovec_idx < pinned_list->nr_iovecs) { | 
|  | struct dma_page_list *page_list; | 
|  |  | 
|  | /* skip already used-up iovecs */ | 
|  | while (!iov[iovec_idx].iov_len) | 
|  | iovec_idx++; | 
|  |  | 
|  | page_list = &pinned_list->page_list[iovec_idx]; | 
|  |  | 
|  | iov_byte_offset = ((unsigned long)iov[iovec_idx].iov_base & ~PAGE_MASK); | 
|  | page_idx = (((unsigned long)iov[iovec_idx].iov_base & PAGE_MASK) | 
|  | - ((unsigned long)page_list->base_address & PAGE_MASK)) >> PAGE_SHIFT; | 
|  |  | 
|  | /* break up copies to not cross page boundary */ | 
|  | while (iov[iovec_idx].iov_len) { | 
|  | copy = min_t(int, PAGE_SIZE - iov_byte_offset, len); | 
|  | copy = min_t(int, copy, iov[iovec_idx].iov_len); | 
|  |  | 
|  | dma_cookie = dma_async_memcpy_pg_to_pg(chan, | 
|  | page_list->pages[page_idx], | 
|  | iov_byte_offset, | 
|  | page, | 
|  | offset, | 
|  | copy); | 
|  | /* poll for a descriptor slot */ | 
|  | if (unlikely(dma_cookie < 0)) { | 
|  | dma_async_issue_pending(chan); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | len -= copy; | 
|  | iov[iovec_idx].iov_len -= copy; | 
|  | iov[iovec_idx].iov_base += copy; | 
|  |  | 
|  | if (!len) | 
|  | return dma_cookie; | 
|  |  | 
|  | offset += copy; | 
|  | iov_byte_offset = 0; | 
|  | page_idx++; | 
|  | } | 
|  | iovec_idx++; | 
|  | } | 
|  |  | 
|  | /* really bad if we ever run out of iovecs */ | 
|  | BUG(); | 
|  | return -EFAULT; | 
|  | } |