summaryrefslogtreecommitdiffstats
path: root/driver/kmem.c
diff options
context:
space:
mode:
authorSuren A. Chilingaryan <csa@dside.dyndns.org>2011-02-13 03:07:11 +0100
committerSuren A. Chilingaryan <csa@dside.dyndns.org>2011-02-13 03:07:11 +0100
commitb72e8567723c30f0e5a83675743e9380c48e5d55 (patch)
treef0dc052508543771fca08c358836507770864249 /driver/kmem.c
downloadipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.tar.gz
ipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.tar.bz2
ipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.tar.xz
ipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.zip
Initial import
Diffstat (limited to 'driver/kmem.c')
-rw-r--r--driver/kmem.c321
1 files changed, 321 insertions, 0 deletions
diff --git a/driver/kmem.c b/driver/kmem.c
new file mode 100644
index 0000000..737b74d
--- /dev/null
+++ b/driver/kmem.c
@@ -0,0 +1,321 @@
+/**
+ *
+ * @file kmem.c
+ * @brief This file contains all functions dealing with kernel memory.
+ * @author Guillermo Marcus
+ * @date 2009-04-05
+ *
+ */
+#include <linux/version.h>
+#include <linux/string.h>
+#include <linux/types.h>
+#include <linux/list.h>
+#include <linux/interrupt.h>
+#include <linux/pci.h>
+#include <linux/cdev.h>
+#include <linux/wait.h>
+#include <linux/mm.h>
+#include <linux/pagemap.h>
+
+#include "config.h" /* compile-time configuration */
+#include "compat.h" /* compatibility definitions for older linux */
+#include "pciDriver.h" /* external interface for the driver */
+#include "common.h" /* internal definitions for all parts */
+#include "kmem.h" /* prototypes for kernel memory */
+#include "sysfs.h" /* prototypes for sysfs */
+
+/**
+ *
+ * Allocates new kernel memory including the corresponding management structure, makes
+ * it available via sysfs if possible.
+ *
+ */
+int pcidriver_kmem_alloc(pcidriver_privdata_t *privdata, kmem_handle_t *kmem_handle)
+{
+ pcidriver_kmem_entry_t *kmem_entry;
+ void *retptr;
+
+ /* First, allocate zeroed memory for the kmem_entry */
+ if ((kmem_entry = kcalloc(1, sizeof(pcidriver_kmem_entry_t), GFP_KERNEL)) == NULL)
+ goto kmem_alloc_entry_fail;
+
+ /* Initialize the kmem_entry */
+ kmem_entry->id = atomic_inc_return(&privdata->kmem_count) - 1;
+ kmem_entry->size = kmem_handle->size;
+ kmem_handle->handle_id = kmem_entry->id;
+
+ /* Initialize sysfs if possible */
+ if (pcidriver_sysfs_initialize_kmem(privdata, kmem_entry->id, &(kmem_entry->sysfs_attr)) != 0)
+ goto kmem_alloc_mem_fail;
+
+ /* ...and allocate the DMA memory */
+ /* note this is a memory pair, referencing the same area: the cpu address (cpua)
+ * and the PCI bus address (pa). The CPU and PCI addresses may not be the same.
+ * The CPU sees only CPU addresses, while the device sees only PCI addresses.
+ * CPU address is used for the mmap (internal to the driver), and
+ * PCI address is the address passed to the DMA Controller in the device.
+ */
+ retptr = pci_alloc_consistent( privdata->pdev, kmem_handle->size, &(kmem_entry->dma_handle) );
+ if (retptr == NULL)
+ goto kmem_alloc_mem_fail;
+ kmem_entry->cpua = (unsigned long)retptr;
+ kmem_handle->pa = (unsigned long)(kmem_entry->dma_handle);
+
+ set_pages_reserved_compat(kmem_entry->cpua, kmem_entry->size);
+
+ /* Add the kmem_entry to the list of the device */
+ spin_lock( &(privdata->kmemlist_lock) );
+ list_add_tail( &(kmem_entry->list), &(privdata->kmem_list) );
+ spin_unlock( &(privdata->kmemlist_lock) );
+
+ return 0;
+
+kmem_alloc_mem_fail:
+ kfree(kmem_entry);
+kmem_alloc_entry_fail:
+ return -ENOMEM;
+}
+
+/**
+ *
+ * Called via sysfs, frees kernel memory and the corresponding management structure
+ *
+ */
+int pcidriver_kmem_free( pcidriver_privdata_t *privdata, kmem_handle_t *kmem_handle )
+{
+ pcidriver_kmem_entry_t *kmem_entry;
+
+ /* Find the associated kmem_entry for this buffer */
+ if ((kmem_entry = pcidriver_kmem_find_entry(privdata, kmem_handle)) == NULL)
+ return -EINVAL; /* kmem_handle is not valid */
+
+ return pcidriver_kmem_free_entry(privdata, kmem_entry);
+}
+
+/**
+ *
+ * Called when cleaning up, frees all kernel memory and their corresponding management structure
+ *
+ */
+int pcidriver_kmem_free_all(pcidriver_privdata_t *privdata)
+{
+ struct list_head *ptr, *next;
+ pcidriver_kmem_entry_t *kmem_entry;
+
+ /* iterate safely over the entries and delete them */
+ list_for_each_safe(ptr, next, &(privdata->kmem_list)) {
+ kmem_entry = list_entry(ptr, pcidriver_kmem_entry_t, list);
+ pcidriver_kmem_free_entry(privdata, kmem_entry); /* spin lock inside! */
+ }
+
+ return 0;
+}
+
+/**
+ *
+ * Synchronize memory to/from the device (or in both directions).
+ *
+ */
+int pcidriver_kmem_sync( pcidriver_privdata_t *privdata, kmem_sync_t *kmem_sync )
+{
+ pcidriver_kmem_entry_t *kmem_entry;
+
+ /* Find the associated kmem_entry for this buffer */
+ if ((kmem_entry = pcidriver_kmem_find_entry(privdata, &(kmem_sync->handle))) == NULL)
+ return -EINVAL; /* kmem_handle is not valid */
+
+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,11)
+ switch (kmem_sync->dir) {
+ case PCIDRIVER_DMA_TODEVICE:
+ pci_dma_sync_single_for_device( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_TODEVICE );
+ break;
+ case PCIDRIVER_DMA_FROMDEVICE:
+ pci_dma_sync_single_for_cpu( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_FROMDEVICE );
+ break;
+ case PCIDRIVER_DMA_BIDIRECTIONAL:
+ pci_dma_sync_single_for_device( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_BIDIRECTIONAL );
+ pci_dma_sync_single_for_cpu( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_BIDIRECTIONAL );
+ break;
+ default:
+ return -EINVAL; /* wrong direction parameter */
+ }
+#else
+ switch (kmem_sync->dir) {
+ case PCIDRIVER_DMA_TODEVICE:
+ pci_dma_sync_single( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_TODEVICE );
+ break;
+ case PCIDRIVER_DMA_FROMDEVICE:
+ pci_dma_sync_single( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_FROMDEVICE );
+ break;
+ case PCIDRIVER_DMA_BIDIRECTIONAL:
+ pci_dma_sync_single( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_BIDIRECTIONAL );
+ break;
+ default:
+ return -EINVAL; /* wrong direction parameter */
+ }
+#endif
+
+ return 0; /* success */
+}
+
+/**
+ *
+ * Free the given kmem_entry and its memory.
+ *
+ */
+int pcidriver_kmem_free_entry(pcidriver_privdata_t *privdata, pcidriver_kmem_entry_t *kmem_entry)
+{
+ pcidriver_sysfs_remove(privdata, &(kmem_entry->sysfs_attr));
+
+ /* Go over the pages of the kmem buffer, and mark them as not reserved */
+#if 0
+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,15)
+ /*
+ * This code is DISABLED.
+ * Apparently, it is not needed to unreserve them. Doing so here
+ * hangs the machine. Why?
+ *
+ * Uhm.. see links:
+ *
+ * http://lwn.net/Articles/161204/
+ * http://lists.openfabrics.org/pipermail/general/2007-March/034101.html
+ *
+ * I insist, this should be enabled, but doing so hangs the machine.
+ * Literature supports the point, and there is even a similar problem (see link)
+ * But this is not the case. It seems right to me. but obviously is not.
+ *
+ * Anyway, this goes away in kernel >=2.6.15.
+ */
+ unsigned long start = __pa(kmem_entry->cpua) >> PAGE_SHIFT;
+ unsigned long end = __pa(kmem_entry->cpua + kmem_entry->size) >> PAGE_SHIFT;
+ unsigned long i;
+ for(i=start;i<end;i++) {
+ struct page *kpage = pfn_to_page(i);
+ ClearPageReserved(kpage);
+ }
+#endif
+#endif
+
+ /* Release DMA memory */
+ pci_free_consistent( privdata->pdev, kmem_entry->size, (void *)(kmem_entry->cpua), kmem_entry->dma_handle );
+
+ /* Remove the kmem list entry */
+ spin_lock( &(privdata->kmemlist_lock) );
+ list_del( &(kmem_entry->list) );
+ spin_unlock( &(privdata->kmemlist_lock) );
+
+ /* Release kmem_entry memory */
+ kfree(kmem_entry);
+
+ return 0;
+}
+
+/**
+ *
+ * Find the corresponding kmem_entry for the given kmem_handle.
+ *
+ */
+pcidriver_kmem_entry_t *pcidriver_kmem_find_entry(pcidriver_privdata_t *privdata, kmem_handle_t *kmem_handle)
+{
+ struct list_head *ptr;
+ pcidriver_kmem_entry_t *entry, *result = NULL;
+
+ /* should I implement it better using the handle_id? */
+
+ spin_lock(&(privdata->kmemlist_lock));
+ list_for_each(ptr, &(privdata->kmem_list)) {
+ entry = list_entry(ptr, pcidriver_kmem_entry_t, list);
+
+ if (entry->dma_handle == kmem_handle->pa) {
+ result = entry;
+ break;
+ }
+ }
+
+ spin_unlock(&(privdata->kmemlist_lock));
+ return result;
+}
+
+/**
+ *
+ * find the corresponding kmem_entry for the given id.
+ *
+ */
+pcidriver_kmem_entry_t *pcidriver_kmem_find_entry_id(pcidriver_privdata_t *privdata, int id)
+{
+ struct list_head *ptr;
+ pcidriver_kmem_entry_t *entry, *result = NULL;
+
+ spin_lock(&(privdata->kmemlist_lock));
+ list_for_each(ptr, &(privdata->kmem_list)) {
+ entry = list_entry(ptr, pcidriver_kmem_entry_t, list);
+
+ if (entry->id == id) {
+ result = entry;
+ break;
+ }
+ }
+
+ spin_unlock(&(privdata->kmemlist_lock));
+ return result;
+}
+
+/**
+ *
+ * mmap() kernel memory to userspace.
+ *
+ */
+int pcidriver_mmap_kmem(pcidriver_privdata_t *privdata, struct vm_area_struct *vma)
+{
+ unsigned long vma_size;
+ pcidriver_kmem_entry_t *kmem_entry;
+ int ret;
+
+ mod_info_dbg("Entering mmap_kmem\n");
+
+ /* FIXME: Is this really right? Always just the latest one? Can't we identify one? */
+ /* Get latest entry on the kmem_list */
+ spin_lock(&(privdata->kmemlist_lock));
+ if (list_empty(&(privdata->kmem_list))) {
+ spin_unlock(&(privdata->kmemlist_lock));
+ mod_info("Trying to mmap a kernel memory buffer without creating it first!\n");
+ return -EFAULT;
+ }
+ kmem_entry = list_entry(privdata->kmem_list.prev, pcidriver_kmem_entry_t, list);
+ spin_unlock(&(privdata->kmemlist_lock));
+
+ mod_info_dbg("Got kmem_entry with id: %d\n", kmem_entry->id);
+
+ /* Check sizes */
+ vma_size = (vma->vm_end - vma->vm_start);
+ if ((vma_size != kmem_entry->size) &&
+ ((kmem_entry->size < PAGE_SIZE) && (vma_size != PAGE_SIZE))) {
+ mod_info("kem_entry size(%lu) and vma size do not match(%lu)\n", kmem_entry->size, vma_size);
+ return -EINVAL;
+ }
+
+ vma->vm_flags |= (VM_RESERVED);
+
+#ifdef pgprot_noncached
+ // This is coherent memory, so it must not be cached.
+ vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
+#endif
+
+ mod_info_dbg("Mapping address %08lx / PFN %08lx\n",
+ virt_to_phys((void*)kmem_entry->cpua),
+ page_to_pfn(virt_to_page((void*)kmem_entry->cpua)));
+
+ ret = remap_pfn_range_cpua_compat(
+ vma,
+ vma->vm_start,
+ kmem_entry->cpua,
+ kmem_entry->size,
+ vma->vm_page_prot );
+
+ if (ret) {
+ mod_info("kmem remap failed: %d (%lx)\n", ret,kmem_entry->cpua);
+ return -EAGAIN;
+ }
+
+ return ret;
+}