Lines Matching refs:val
431 long val; in sock_set_timeout() local
450 val = MAX_SCHEDULE_TIMEOUT; in sock_set_timeout()
453 val = tv.tv_sec * HZ + DIV_ROUND_UP((unsigned long)tv.tv_usec, in sock_set_timeout()
455 WRITE_ONCE(*timeo_p, val); in sock_set_timeout()
459 static bool sk_set_prio_allowed(const struct sock *sk, int val) in sk_set_prio_allowed() argument
461 return ((val >= TC_PRIO_BESTEFFORT && val <= TC_PRIO_INTERACTIVE) || in sk_set_prio_allowed()
834 static void __sock_set_timestamps(struct sock *sk, bool val, bool new, bool ns) in __sock_set_timestamps() argument
836 sock_valbool_flag(sk, SOCK_RCVTSTAMP, val); in __sock_set_timestamps()
837 sock_valbool_flag(sk, SOCK_RCVTSTAMPNS, val && ns); in __sock_set_timestamps()
838 if (val) { in __sock_set_timestamps()
902 int val = timestamping.flags; in sock_set_timestamping() local
905 if (val & ~SOF_TIMESTAMPING_MASK) in sock_set_timestamping()
908 if (val & SOF_TIMESTAMPING_OPT_ID_TCP && in sock_set_timestamping()
909 !(val & SOF_TIMESTAMPING_OPT_ID)) in sock_set_timestamping()
912 if (val & SOF_TIMESTAMPING_OPT_ID && in sock_set_timestamping()
918 if (val & SOF_TIMESTAMPING_OPT_ID_TCP) in sock_set_timestamping()
927 if (val & SOF_TIMESTAMPING_OPT_STATS && in sock_set_timestamping()
928 !(val & SOF_TIMESTAMPING_OPT_TSONLY)) in sock_set_timestamping()
931 if (val & SOF_TIMESTAMPING_BIND_PHC) { in sock_set_timestamping()
937 WRITE_ONCE(sk->sk_tsflags, val); in sock_set_timestamping()
939 sock_valbool_flag(sk, SOCK_TIMESTAMPING_ANY, !!(val & TSFLAGS_ANY)); in sock_set_timestamping()
941 if (val & SOF_TIMESTAMPING_RX_SOFTWARE) in sock_set_timestamping()
974 static void __sock_set_rcvbuf(struct sock *sk, int val) in __sock_set_rcvbuf() argument
979 val = min_t(int, val, INT_MAX / 2); in __sock_set_rcvbuf()
992 WRITE_ONCE(sk->sk_rcvbuf, max_t(int, val * 2, SOCK_MIN_RCVBUF)); in __sock_set_rcvbuf()
995 void sock_set_rcvbuf(struct sock *sk, int val) in sock_set_rcvbuf() argument
998 __sock_set_rcvbuf(sk, val); in sock_set_rcvbuf()
1003 static void __sock_set_mark(struct sock *sk, u32 val) in __sock_set_mark() argument
1005 if (val != sk->sk_mark) { in __sock_set_mark()
1006 WRITE_ONCE(sk->sk_mark, val); in __sock_set_mark()
1011 void sock_set_mark(struct sock *sk, u32 val) in sock_set_mark() argument
1014 __sock_set_mark(sk, val); in sock_set_mark()
1193 int val; in sk_setsockopt() local
1208 if (copy_from_sockptr(&val, optval, sizeof(val))) in sk_setsockopt()
1211 valbool = val ? 1 : 0; in sk_setsockopt()
1216 if (sk_set_prio_allowed(sk, val)) { in sk_setsockopt()
1217 sock_set_priority(sk, val); in sk_setsockopt()
1228 if (val < 0) in sk_setsockopt()
1230 WRITE_ONCE(sk->sk_ll_usec, val); in sk_setsockopt()
1238 if (val > READ_ONCE(sk->sk_busy_poll_budget) && in sk_setsockopt()
1241 if (val < 0 || val > U16_MAX) in sk_setsockopt()
1243 WRITE_ONCE(sk->sk_busy_poll_budget, val); in sk_setsockopt()
1248 unsigned long ulval = (val == ~0U) ? ~0UL : (unsigned int)val; in sk_setsockopt()
1251 if (sizeof(ulval) != sizeof(val) && in sk_setsockopt()
1270 if (val < -1 || val > 1) in sk_setsockopt()
1272 if ((u8)val == SOCK_TXREHASH_DEFAULT) in sk_setsockopt()
1273 val = READ_ONCE(sock_net(sk)->core.sysctl_txrehash); in sk_setsockopt()
1277 WRITE_ONCE(sk->sk_txrehash, (u8)val); in sk_setsockopt()
1281 int (*set_peek_off)(struct sock *sk, int val); in sk_setsockopt()
1285 ret = set_peek_off(sk, val); in sk_setsockopt()
1308 if (val && !sockopt_capable(CAP_NET_ADMIN)) in sk_setsockopt()
1335 val = min_t(u32, val, READ_ONCE(sysctl_wmem_max)); in sk_setsockopt()
1340 val = min_t(int, val, INT_MAX / 2); in sk_setsockopt()
1343 max_t(int, val * 2, SOCK_MIN_SNDBUF)); in sk_setsockopt()
1357 if (val < 0) in sk_setsockopt()
1358 val = 0; in sk_setsockopt()
1367 __sock_set_rcvbuf(sk, min_t(u32, val, READ_ONCE(sysctl_rmem_max))); in sk_setsockopt()
1379 __sock_set_rcvbuf(sk, max(val, 0)); in sk_setsockopt()
1438 timestamping.flags = val; in sk_setsockopt()
1445 int (*set_rcvlowat)(struct sock *sk, int val) = NULL; in sk_setsockopt()
1447 if (val < 0) in sk_setsockopt()
1448 val = INT_MAX; in sk_setsockopt()
1452 ret = set_rcvlowat(sk, val); in sk_setsockopt()
1454 WRITE_ONCE(sk->sk_rcvlowat, val ? : 1); in sk_setsockopt()
1521 __sock_set_mark(sk, val); in sk_setsockopt()
1576 reuseport_update_incoming_cpu(sk, val); in sk_setsockopt()
1580 if (val == 1) in sk_setsockopt()
1594 if (val < 0 || val > 1) in sk_setsockopt()
1635 ret = sock_bindtoindex_locked(sk, val); in sk_setsockopt()
1639 if (val & ~SOCK_BUF_LOCK_MASK) { in sk_setsockopt()
1643 sk->sk_userlocks = val | (sk->sk_userlocks & in sk_setsockopt()
1651 if (val < 0) { in sk_setsockopt()
1656 delta = val - sk->sk_reserved_mem; in sk_setsockopt()
1725 int val; in sk_getsockopt() member
1748 v.val = sock_flag(sk, SOCK_DBG); in sk_getsockopt()
1752 v.val = sock_flag(sk, SOCK_LOCALROUTE); in sk_getsockopt()
1756 v.val = sock_flag(sk, SOCK_BROADCAST); in sk_getsockopt()
1760 v.val = READ_ONCE(sk->sk_sndbuf); in sk_getsockopt()
1764 v.val = READ_ONCE(sk->sk_rcvbuf); in sk_getsockopt()
1768 v.val = sk->sk_reuse; in sk_getsockopt()
1772 v.val = sk->sk_reuseport; in sk_getsockopt()
1776 v.val = sock_flag(sk, SOCK_KEEPOPEN); in sk_getsockopt()
1780 v.val = sk->sk_type; in sk_getsockopt()
1784 v.val = sk->sk_protocol; in sk_getsockopt()
1788 v.val = sk->sk_family; in sk_getsockopt()
1792 v.val = -sock_error(sk); in sk_getsockopt()
1793 if (v.val == 0) in sk_getsockopt()
1794 v.val = xchg(&sk->sk_err_soft, 0); in sk_getsockopt()
1798 v.val = sock_flag(sk, SOCK_URGINLINE); in sk_getsockopt()
1802 v.val = sk->sk_no_check_tx; in sk_getsockopt()
1806 v.val = READ_ONCE(sk->sk_priority); in sk_getsockopt()
1819 v.val = sock_flag(sk, SOCK_RCVTSTAMP) && in sk_getsockopt()
1825 v.val = sock_flag(sk, SOCK_RCVTSTAMPNS) && !sock_flag(sk, SOCK_TSTAMP_NEW); in sk_getsockopt()
1829 v.val = sock_flag(sk, SOCK_RCVTSTAMP) && sock_flag(sk, SOCK_TSTAMP_NEW); in sk_getsockopt()
1833 v.val = sock_flag(sk, SOCK_RCVTSTAMPNS) && sock_flag(sk, SOCK_TSTAMP_NEW); in sk_getsockopt()
1862 v.val = READ_ONCE(sk->sk_rcvlowat); in sk_getsockopt()
1866 v.val = 1; in sk_getsockopt()
1873 v.val = sk->sk_scm_credentials; in sk_getsockopt()
1880 v.val = sk->sk_scm_pidfd; in sk_getsockopt()
1887 v.val = sk->sk_scm_rights; in sk_getsockopt()
1988 v.val = sk->sk_state == TCP_LISTEN; in sk_getsockopt()
1995 v.val = sk->sk_scm_security; in sk_getsockopt()
2003 v.val = READ_ONCE(sk->sk_mark); in sk_getsockopt()
2007 v.val = sock_flag(sk, SOCK_RCVMARK); in sk_getsockopt()
2011 v.val = sock_flag(sk, SOCK_RCVPRIORITY); in sk_getsockopt()
2015 v.val = sock_flag(sk, SOCK_RXQ_OVFL); in sk_getsockopt()
2019 v.val = sock_flag(sk, SOCK_WIFI_STATUS); in sk_getsockopt()
2026 v.val = READ_ONCE(sk->sk_peek_off); in sk_getsockopt()
2029 v.val = sock_flag(sk, SOCK_NOFCS); in sk_getsockopt()
2043 v.val = sock_flag(sk, SOCK_FILTER_LOCKED); in sk_getsockopt()
2047 v.val = bpf_tell_extensions(); in sk_getsockopt()
2051 v.val = sock_flag(sk, SOCK_SELECT_ERR_QUEUE); in sk_getsockopt()
2056 v.val = READ_ONCE(sk->sk_ll_usec); in sk_getsockopt()
2059 v.val = READ_ONCE(sk->sk_prefer_busy_poll); in sk_getsockopt()
2065 if (sizeof(v.ulval) != sizeof(v.val) && len >= sizeof(v.ulval)) { in sk_getsockopt()
2070 v.val = min_t(unsigned long, ~0U, in sk_getsockopt()
2076 v.val = READ_ONCE(sk->sk_incoming_cpu); in sk_getsockopt()
2094 v.val = READ_ONCE(sk->sk_napi_id); in sk_getsockopt()
2097 if (!napi_id_valid(v.val)) in sk_getsockopt()
2098 v.val = 0; in sk_getsockopt()
2111 v.val = sock_flag(sk, SOCK_ZEROCOPY); in sk_getsockopt()
2124 v.val = READ_ONCE(sk->sk_bound_dev_if); in sk_getsockopt()
2135 v.val = sk->sk_userlocks & SOCK_BUF_LOCK_MASK; in sk_getsockopt()
2139 v.val = READ_ONCE(sk->sk_reserved_mem); in sk_getsockopt()
2147 v.val = READ_ONCE(sk->sk_txrehash); in sk_getsockopt()
3422 int sk_set_peek_off(struct sock *sk, int val) in sk_set_peek_off() argument
3424 WRITE_ONCE(sk->sk_peek_off, val); in sk_set_peek_off()
3988 res += per_cpu_ptr(net->core.prot_inuse, cpu)->val[idx]; in sock_prot_inuse_get()