aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorRob Clark2013-12-13 11:48:30 -0600
committerRob Clark2013-12-13 14:48:10 -0600
commit068ea68b3f7ebd5efcfcc2f6ae417651423c8382 (patch)
treee094e44e8d01c33ab8e5d74aa489308f267d3eea /freedreno/freedreno_device.c
parent1489811a805fb6b5b19d61fa99b9b962cc06bd22 (diff)
downloadexternal-libgbm-068ea68b3f7ebd5efcfcc2f6ae417651423c8382.tar.gz
external-libgbm-068ea68b3f7ebd5efcfcc2f6ae417651423c8382.tar.xz
external-libgbm-068ea68b3f7ebd5efcfcc2f6ae417651423c8382.zip
freedreno: add bo cache
Workloads which create many transient buffers cause significant CPU overhead in buffer allocation, zeroing, cache maint, and mmap setup. By caching and re-using existing buffers, the CPU overhead drops significantly. See: http://bloggingthemonkey.blogspot.com/2013/09/freedreno-update-moar-fps.html A simple time based policy is used for purging the cache. Once the kernel supports it, we could use madvise style API to handle memory pressure scenarios a bit better. Signed-off-by: Rob Clark <robclark@freedesktop.org>
Diffstat (limited to 'freedreno/freedreno_device.c')
-rw-r--r--freedreno/freedreno_device.c60
1 files changed, 56 insertions, 4 deletions
diff --git a/freedreno/freedreno_device.c b/freedreno/freedreno_device.c
index 1e3d9df2..6486983d 100644
--- a/freedreno/freedreno_device.c
+++ b/freedreno/freedreno_device.c
@@ -39,6 +39,44 @@ static void * dev_table;
39struct fd_device * kgsl_device_new(int fd); 39struct fd_device * kgsl_device_new(int fd);
40struct fd_device * msm_device_new(int fd); 40struct fd_device * msm_device_new(int fd);
41 41
42static void
43add_bucket(struct fd_device *dev, int size)
44{
45 unsigned int i = dev->num_buckets;
46
47 assert(i < ARRAY_SIZE(dev->cache_bucket));
48
49 list_inithead(&dev->cache_bucket[i].list);
50 dev->cache_bucket[i].size = size;
51 dev->num_buckets++;
52}
53
54static void
55init_cache_buckets(struct fd_device *dev)
56{
57 unsigned long size, cache_max_size = 64 * 1024 * 1024;
58
59 /* OK, so power of two buckets was too wasteful of memory.
60 * Give 3 other sizes between each power of two, to hopefully
61 * cover things accurately enough. (The alternative is
62 * probably to just go for exact matching of sizes, and assume
63 * that for things like composited window resize the tiled
64 * width/height alignment and rounding of sizes to pages will
65 * get us useful cache hit rates anyway)
66 */
67 add_bucket(dev, 4096);
68 add_bucket(dev, 4096 * 2);
69 add_bucket(dev, 4096 * 3);
70
71 /* Initialize the linked lists for BO reuse cache. */
72 for (size = 4 * 4096; size <= cache_max_size; size *= 2) {
73 add_bucket(dev, size);
74 add_bucket(dev, size + size * 1 / 4);
75 add_bucket(dev, size + size * 2 / 4);
76 add_bucket(dev, size + size * 3 / 4);
77 }
78}
79
42static struct fd_device * fd_device_new_impl(int fd) 80static struct fd_device * fd_device_new_impl(int fd)
43{ 81{
44 struct fd_device *dev; 82 struct fd_device *dev;
@@ -69,6 +107,7 @@ static struct fd_device * fd_device_new_impl(int fd)
69 dev->fd = fd; 107 dev->fd = fd;
70 dev->handle_table = drmHashCreate(); 108 dev->handle_table = drmHashCreate();
71 dev->name_table = drmHashCreate(); 109 dev->name_table = drmHashCreate();
110 init_cache_buckets(dev);
72 111
73 return dev; 112 return dev;
74} 113}
@@ -102,14 +141,27 @@ struct fd_device * fd_device_ref(struct fd_device *dev)
102 return dev; 141 return dev;
103} 142}
104 143
144static void fd_device_del_impl(struct fd_device *dev)
145{
146 fd_cleanup_bo_cache(dev, 0);
147 drmHashDestroy(dev->handle_table);
148 drmHashDestroy(dev->name_table);
149 drmHashDelete(dev_table, dev->fd);
150 dev->funcs->destroy(dev);
151}
152
153void fd_device_del_locked(struct fd_device *dev)
154{
155 if (!atomic_dec_and_test(&dev->refcnt))
156 return;
157 fd_device_del_impl(dev);
158}
159
105void fd_device_del(struct fd_device *dev) 160void fd_device_del(struct fd_device *dev)
106{ 161{
107 if (!atomic_dec_and_test(&dev->refcnt)) 162 if (!atomic_dec_and_test(&dev->refcnt))
108 return; 163 return;
109 pthread_mutex_lock(&table_lock); 164 pthread_mutex_lock(&table_lock);
110 drmHashDestroy(dev->handle_table); 165 fd_device_del_impl(dev);
111 drmHashDestroy(dev->name_table);
112 drmHashDelete(dev_table, dev->fd);
113 pthread_mutex_unlock(&table_lock); 166 pthread_mutex_unlock(&table_lock);
114 dev->funcs->destroy(dev);
115} 167}