aboutsummaryrefslogtreecommitdiffstats
path: root/lib
diff options
context:
space:
mode:
authorLinus Torvalds2017-05-03 12:05:38 -0500
committerLinus Torvalds2017-05-03 12:05:38 -0500
commite5021876c91dc3894b2174cca8fa797f8e29e7b9 (patch)
treecf6cc6591a8421e0f75cfcfbc10312421bd8e9f1 /lib
parent46f0537b1ecf672052007c97f102a7e6bf0791e4 (diff)
parente265eb3a30543a237b2ebc4e0422ac82e55b07e4 (diff)
downloadkernel-e5021876c91dc3894b2174cca8fa797f8e29e7b9.tar.gz
kernel-e5021876c91dc3894b2174cca8fa797f8e29e7b9.tar.xz
kernel-e5021876c91dc3894b2174cca8fa797f8e29e7b9.zip
Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/shli/md
Pull MD updates from Shaohua Li: - Add Partial Parity Log (ppl) feature found in Intel IMSM raid array by Artur Paszkiewicz. This feature is another way to close RAID5 writehole. The Linux implementation is also available for normal RAID5 array if specific superblock bit is set. - A number of md-cluser fixes and enabling md-cluster array resize from Guoqing Jiang - A bunch of patches from Ming Lei and Neil Brown to rewrite MD bio handling related code. Now MD doesn't directly access bio bvec, bi_phys_segments and uses modern bio API for bio split. - Improve RAID5 IO pattern to improve performance for hard disk based RAID5/6 from me. - Several patches from Song Liu to speed up raid5-cache recovery and allow raid5 cache feature disabling in runtime. - Fix a performance regression in raid1 resync from Xiao Ni. - Other cleanup and fixes from various people. * 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/shli/md: (84 commits) md/raid10: skip spare disk as 'first' disk md/raid1: Use a new variable to count flighting sync requests md: clear WantReplacement once disk is removed md/raid1/10: remove unused queue md: handle read-only member devices better. md/raid10: wait up frozen array in handle_write_completed uapi: fix linux/raid/md_p.h userspace compilation error md-cluster: Fix a memleak in an error handling path md: support disabling of create-on-open semantics. md: allow creation of mdNNN arrays via md_mod/parameters/new_array raid5-ppl: use a single mempool for ppl_io_unit and header_page md/raid0: fix up bio splitting. md/linear: improve bio splitting. md/raid5: make chunk_aligned_read() split bios more cleanly. md/raid10: simplify handle_read_error() md/raid10: simplify the splitting of requests. md/raid1: factor out flush_bio_list() md/raid1: simplify handle_read_error(). Revert "block: introduce bio_copy_data_partial" md/raid1: simplify alloc_behind_master_bio() ...
Diffstat (limited to 'lib')
-rw-r--r--lib/percpu-refcount.c17
1 files changed, 17 insertions, 0 deletions
diff --git a/lib/percpu-refcount.c b/lib/percpu-refcount.c
index 9ac959ef4cae..fe03c6d52761 100644
--- a/lib/percpu-refcount.c
+++ b/lib/percpu-refcount.c
@@ -260,6 +260,22 @@ void percpu_ref_switch_to_atomic(struct percpu_ref *ref,
260 260
261 spin_unlock_irqrestore(&percpu_ref_switch_lock, flags); 261 spin_unlock_irqrestore(&percpu_ref_switch_lock, flags);
262} 262}
263EXPORT_SYMBOL_GPL(percpu_ref_switch_to_atomic);
264
265/**
266 * percpu_ref_switch_to_atomic_sync - switch a percpu_ref to atomic mode
267 * @ref: percpu_ref to switch to atomic mode
268 *
269 * Schedule switching the ref to atomic mode, and wait for the
270 * switch to complete. Caller must ensure that no other thread
271 * will switch back to percpu mode.
272 */
273void percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref)
274{
275 percpu_ref_switch_to_atomic(ref, NULL);
276 wait_event(percpu_ref_switch_waitq, !ref->confirm_switch);
277}
278EXPORT_SYMBOL_GPL(percpu_ref_switch_to_atomic_sync);
263 279
264/** 280/**
265 * percpu_ref_switch_to_percpu - switch a percpu_ref to percpu mode 281 * percpu_ref_switch_to_percpu - switch a percpu_ref to percpu mode
@@ -290,6 +306,7 @@ void percpu_ref_switch_to_percpu(struct percpu_ref *ref)
290 306
291 spin_unlock_irqrestore(&percpu_ref_switch_lock, flags); 307 spin_unlock_irqrestore(&percpu_ref_switch_lock, flags);
292} 308}
309EXPORT_SYMBOL_GPL(percpu_ref_switch_to_percpu);
293 310
294/** 311/**
295 * percpu_ref_kill_and_confirm - drop the initial ref and schedule confirmation 312 * percpu_ref_kill_and_confirm - drop the initial ref and schedule confirmation