2 * Copyright (C) 2012 Red Hat
4 * This file is subject to the terms and conditions of the GNU General Public
5 * License v2. See the file COPYING in the main directory of this archive for
11 #include <linux/shmem_fs.h>
13 struct udl_gem_object *udl_gem_alloc_object(struct drm_device *dev,
16 struct udl_gem_object *obj;
18 obj = kzalloc(sizeof(*obj), GFP_KERNEL);
22 if (drm_gem_object_init(dev, &obj->base, size) != 0) {
31 udl_gem_create(struct drm_file *file,
32 struct drm_device *dev,
36 struct udl_gem_object *obj;
40 size = roundup(size, PAGE_SIZE);
42 obj = udl_gem_alloc_object(dev, size);
46 ret = drm_gem_handle_create(file, &obj->base, &handle);
48 drm_gem_object_release(&obj->base);
53 drm_gem_object_unreference(&obj->base);
58 int udl_dumb_create(struct drm_file *file,
59 struct drm_device *dev,
60 struct drm_mode_create_dumb *args)
62 args->pitch = args->width * ((args->bpp + 1) / 8);
63 args->size = args->pitch * args->height;
64 return udl_gem_create(file, dev,
65 args->size, &args->handle);
68 int udl_dumb_destroy(struct drm_file *file, struct drm_device *dev,
71 return drm_gem_handle_delete(file, handle);
74 int udl_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
76 struct udl_gem_object *obj = to_udl_bo(vma->vm_private_data);
78 unsigned int page_offset;
81 page_offset = ((unsigned long)vmf->virtual_address - vma->vm_start) >>
85 return VM_FAULT_SIGBUS;
87 page = obj->pages[page_offset];
88 ret = vm_insert_page(vma, (unsigned long)vmf->virtual_address, page);
94 return VM_FAULT_NOPAGE;
98 return VM_FAULT_SIGBUS;
102 int udl_gem_init_object(struct drm_gem_object *obj)
109 static int udl_gem_get_pages(struct udl_gem_object *obj, gfp_t gfpmask)
114 struct address_space *mapping;
119 page_count = obj->base.size / PAGE_SIZE;
120 BUG_ON(obj->pages != NULL);
121 obj->pages = drm_malloc_ab(page_count, sizeof(struct page *));
122 if (obj->pages == NULL)
125 inode = obj->base.filp->f_path.dentry->d_inode;
126 mapping = inode->i_mapping;
127 gfpmask |= mapping_gfp_mask(mapping);
129 for (i = 0; i < page_count; i++) {
130 page = shmem_read_mapping_page_gfp(mapping, i, gfpmask);
133 obj->pages[i] = page;
139 page_cache_release(obj->pages[i]);
140 drm_free_large(obj->pages);
142 return PTR_ERR(page);
145 static void udl_gem_put_pages(struct udl_gem_object *obj)
147 int page_count = obj->base.size / PAGE_SIZE;
150 for (i = 0; i < page_count; i++)
151 page_cache_release(obj->pages[i]);
153 drm_free_large(obj->pages);
157 int udl_gem_vmap(struct udl_gem_object *obj)
159 int page_count = obj->base.size / PAGE_SIZE;
162 ret = udl_gem_get_pages(obj, GFP_KERNEL);
166 obj->vmapping = vmap(obj->pages, page_count, 0, PAGE_KERNEL);
172 void udl_gem_vunmap(struct udl_gem_object *obj)
175 vunmap(obj->vmapping);
177 udl_gem_put_pages(obj);
180 void udl_gem_free_object(struct drm_gem_object *gem_obj)
182 struct udl_gem_object *obj = to_udl_bo(gem_obj);
188 udl_gem_put_pages(obj);
190 if (gem_obj->map_list.map)
191 drm_gem_free_mmap_offset(gem_obj);
194 /* the dumb interface doesn't work with the GEM straight MMAP
195 interface, it expects to do MMAP on the drm fd, like normal */
196 int udl_gem_mmap(struct drm_file *file, struct drm_device *dev,
197 uint32_t handle, uint64_t *offset)
199 struct udl_gem_object *gobj;
200 struct drm_gem_object *obj;
203 mutex_lock(&dev->struct_mutex);
204 obj = drm_gem_object_lookup(dev, file, handle);
209 gobj = to_udl_bo(obj);
211 ret = udl_gem_get_pages(gobj, GFP_KERNEL);
214 if (!gobj->base.map_list.map) {
215 ret = drm_gem_create_mmap_offset(obj);
220 *offset = (u64)gobj->base.map_list.hash.key << PAGE_SHIFT;
223 drm_gem_object_unreference(&gobj->base);
225 mutex_unlock(&dev->struct_mutex);