|
@@ -1149,8 +1149,7 @@ void bpf_prog_destroy(struct bpf_prog *fp)
|
|
|
}
|
|
|
EXPORT_SYMBOL_GPL(bpf_prog_destroy);
|
|
|
|
|
|
-static int __sk_attach_prog(struct bpf_prog *prog, struct sock *sk,
|
|
|
- bool locked)
|
|
|
+static int __sk_attach_prog(struct bpf_prog *prog, struct sock *sk)
|
|
|
{
|
|
|
struct sk_filter *fp, *old_fp;
|
|
|
|
|
@@ -1166,8 +1165,10 @@ static int __sk_attach_prog(struct bpf_prog *prog, struct sock *sk,
|
|
|
return -ENOMEM;
|
|
|
}
|
|
|
|
|
|
- old_fp = rcu_dereference_protected(sk->sk_filter, locked);
|
|
|
+ old_fp = rcu_dereference_protected(sk->sk_filter,
|
|
|
+ lockdep_sock_is_held(sk));
|
|
|
rcu_assign_pointer(sk->sk_filter, fp);
|
|
|
+
|
|
|
if (old_fp)
|
|
|
sk_filter_uncharge(sk, old_fp);
|
|
|
|
|
@@ -1246,8 +1247,7 @@ struct bpf_prog *__get_filter(struct sock_fprog *fprog, struct sock *sk)
|
|
|
* occurs or there is insufficient memory for the filter a negative
|
|
|
* errno code is returned. On success the return is zero.
|
|
|
*/
|
|
|
-int __sk_attach_filter(struct sock_fprog *fprog, struct sock *sk,
|
|
|
- bool locked)
|
|
|
+int sk_attach_filter(struct sock_fprog *fprog, struct sock *sk)
|
|
|
{
|
|
|
struct bpf_prog *prog = __get_filter(fprog, sk);
|
|
|
int err;
|
|
@@ -1255,7 +1255,7 @@ int __sk_attach_filter(struct sock_fprog *fprog, struct sock *sk,
|
|
|
if (IS_ERR(prog))
|
|
|
return PTR_ERR(prog);
|
|
|
|
|
|
- err = __sk_attach_prog(prog, sk, locked);
|
|
|
+ err = __sk_attach_prog(prog, sk);
|
|
|
if (err < 0) {
|
|
|
__bpf_prog_release(prog);
|
|
|
return err;
|
|
@@ -1263,12 +1263,7 @@ int __sk_attach_filter(struct sock_fprog *fprog, struct sock *sk,
|
|
|
|
|
|
return 0;
|
|
|
}
|
|
|
-EXPORT_SYMBOL_GPL(__sk_attach_filter);
|
|
|
-
|
|
|
-int sk_attach_filter(struct sock_fprog *fprog, struct sock *sk)
|
|
|
-{
|
|
|
- return __sk_attach_filter(fprog, sk, sock_owned_by_user(sk));
|
|
|
-}
|
|
|
+EXPORT_SYMBOL_GPL(sk_attach_filter);
|
|
|
|
|
|
int sk_reuseport_attach_filter(struct sock_fprog *fprog, struct sock *sk)
|
|
|
{
|
|
@@ -1314,7 +1309,7 @@ int sk_attach_bpf(u32 ufd, struct sock *sk)
|
|
|
if (IS_ERR(prog))
|
|
|
return PTR_ERR(prog);
|
|
|
|
|
|
- err = __sk_attach_prog(prog, sk, sock_owned_by_user(sk));
|
|
|
+ err = __sk_attach_prog(prog, sk);
|
|
|
if (err < 0) {
|
|
|
bpf_prog_put(prog);
|
|
|
return err;
|
|
@@ -2255,7 +2250,7 @@ static int __init register_sk_filter_ops(void)
|
|
|
}
|
|
|
late_initcall(register_sk_filter_ops);
|
|
|
|
|
|
-int __sk_detach_filter(struct sock *sk, bool locked)
|
|
|
+int sk_detach_filter(struct sock *sk)
|
|
|
{
|
|
|
int ret = -ENOENT;
|
|
|
struct sk_filter *filter;
|
|
@@ -2263,7 +2258,8 @@ int __sk_detach_filter(struct sock *sk, bool locked)
|
|
|
if (sock_flag(sk, SOCK_FILTER_LOCKED))
|
|
|
return -EPERM;
|
|
|
|
|
|
- filter = rcu_dereference_protected(sk->sk_filter, locked);
|
|
|
+ filter = rcu_dereference_protected(sk->sk_filter,
|
|
|
+ lockdep_sock_is_held(sk));
|
|
|
if (filter) {
|
|
|
RCU_INIT_POINTER(sk->sk_filter, NULL);
|
|
|
sk_filter_uncharge(sk, filter);
|
|
@@ -2272,12 +2268,7 @@ int __sk_detach_filter(struct sock *sk, bool locked)
|
|
|
|
|
|
return ret;
|
|
|
}
|
|
|
-EXPORT_SYMBOL_GPL(__sk_detach_filter);
|
|
|
-
|
|
|
-int sk_detach_filter(struct sock *sk)
|
|
|
-{
|
|
|
- return __sk_detach_filter(sk, sock_owned_by_user(sk));
|
|
|
-}
|
|
|
+EXPORT_SYMBOL_GPL(sk_detach_filter);
|
|
|
|
|
|
int sk_get_filter(struct sock *sk, struct sock_filter __user *ubuf,
|
|
|
unsigned int len)
|
|
@@ -2288,7 +2279,7 @@ int sk_get_filter(struct sock *sk, struct sock_filter __user *ubuf,
|
|
|
|
|
|
lock_sock(sk);
|
|
|
filter = rcu_dereference_protected(sk->sk_filter,
|
|
|
- sock_owned_by_user(sk));
|
|
|
+ lockdep_sock_is_held(sk));
|
|
|
if (!filter)
|
|
|
goto out;
|
|
|
|