]> Pileus Git - ~andy/linux/blobdiff - kernel/lglock.c
ixgbevf: move API neg to reset path
[~andy/linux] / kernel / lglock.c
index 6535a667a5a797a3fa1f9a9683cd9e85124692d7..86ae2aebf00432f4d681a413febebff79406889d 100644 (file)
@@ -21,7 +21,7 @@ void lg_local_lock(struct lglock *lg)
        arch_spinlock_t *lock;
 
        preempt_disable();
-       rwlock_acquire_read(&lg->lock_dep_map, 0, 0, _RET_IP_);
+       lock_acquire_shared(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
        lock = this_cpu_ptr(lg->lock);
        arch_spin_lock(lock);
 }
@@ -31,7 +31,7 @@ void lg_local_unlock(struct lglock *lg)
 {
        arch_spinlock_t *lock;
 
-       rwlock_release(&lg->lock_dep_map, 1, _RET_IP_);
+       lock_release(&lg->lock_dep_map, 1, _RET_IP_);
        lock = this_cpu_ptr(lg->lock);
        arch_spin_unlock(lock);
        preempt_enable();
@@ -43,7 +43,7 @@ void lg_local_lock_cpu(struct lglock *lg, int cpu)
        arch_spinlock_t *lock;
 
        preempt_disable();
-       rwlock_acquire_read(&lg->lock_dep_map, 0, 0, _RET_IP_);
+       lock_acquire_shared(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
        lock = per_cpu_ptr(lg->lock, cpu);
        arch_spin_lock(lock);
 }
@@ -53,7 +53,7 @@ void lg_local_unlock_cpu(struct lglock *lg, int cpu)
 {
        arch_spinlock_t *lock;
 
-       rwlock_release(&lg->lock_dep_map, 1, _RET_IP_);
+       lock_release(&lg->lock_dep_map, 1, _RET_IP_);
        lock = per_cpu_ptr(lg->lock, cpu);
        arch_spin_unlock(lock);
        preempt_enable();
@@ -65,7 +65,7 @@ void lg_global_lock(struct lglock *lg)
        int i;
 
        preempt_disable();
-       rwlock_acquire(&lg->lock_dep_map, 0, 0, _RET_IP_);
+       lock_acquire_exclusive(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
        for_each_possible_cpu(i) {
                arch_spinlock_t *lock;
                lock = per_cpu_ptr(lg->lock, i);
@@ -78,7 +78,7 @@ void lg_global_unlock(struct lglock *lg)
 {
        int i;
 
-       rwlock_release(&lg->lock_dep_map, 1, _RET_IP_);
+       lock_release(&lg->lock_dep_map, 1, _RET_IP_);
        for_each_possible_cpu(i) {
                arch_spinlock_t *lock;
                lock = per_cpu_ptr(lg->lock, i);