@@ -6304,7 +6304,6 @@ void napi_busy_loop(unsigned int napi_id,
if (!IS_ENABLED(CONFIG_PREEMPT_RT))
preempt_enable();
rcu_read_unlock();
- cond_resched();
if (loop_end(loop_end_arg, start_time))
return;
goto restart;
@@ -6709,8 +6708,6 @@ static int napi_threaded_poll(void *data)
if (!repoll)
break;
-
- cond_resched();
}
}
return 0;
@@ -11478,7 +11475,6 @@ static void __net_exit default_device_exit_batch(struct list_head *net_list)
rtnl_lock();
list_for_each_entry(net, net_list, exit_list) {
default_device_exit_net(net);
- cond_resched();
}
list_for_each_entry(net, net_list, exit_list) {
@@ -1008,7 +1008,6 @@ static void neigh_periodic_work(struct work_struct *work)
* grows while we are preempted.
*/
write_unlock_bh(&tbl->lock);
- cond_resched();
write_lock_bh(&tbl->lock);
nht = rcu_dereference_protected(tbl->nht,
lockdep_is_held(&tbl->lock));
@@ -168,7 +168,6 @@ static void ops_exit_list(const struct pernet_operations *ops,
if (ops->exit) {
list_for_each_entry(net, net_exit_list, exit_list) {
ops->exit(net);
- cond_resched();
}
}
if (ops->exit_batch)
@@ -92,7 +92,6 @@ static void update_classid_task(struct task_struct *p, u32 classid)
task_lock(p);
fd = iterate_fd(p->files, fd, update_classid_sock, &ctx);
task_unlock(p);
- cond_resched();
} while (fd);
}
@@ -140,7 +140,6 @@ void __rtnl_unlock(void)
struct sk_buff *next = head->next;
kfree_skb(head);
- cond_resched();
head = next;
}
}
@@ -2982,8 +2982,6 @@ void __release_sock(struct sock *sk)
skb_mark_not_on_list(skb);
sk_backlog_rcv(sk, skb);
- cond_resched();
-
skb = next;
} while (skb != NULL);
@@ -389,7 +389,6 @@ inet_csk_find_open_port(const struct sock *sk, struct inet_bind_bucket **tb_ret,
goto success;
next_port:
spin_unlock_bh(&head->lock);
- cond_resched();
}
offset--;
@@ -1420,8 +1419,6 @@ void inet_csk_listen_stop(struct sock *sk)
bh_unlock_sock(child);
local_bh_enable();
sock_put(child);
-
- cond_resched();
}
if (queue->fastopenq.rskq_rst_head) {
/* Free all the reqs queued in rskq_rst_head. */
@@ -1147,7 +1147,6 @@ void inet_diag_dump_icsk(struct inet_hashinfo *hashinfo, struct sk_buff *skb,
}
if (res < 0)
break;
- cond_resched();
if (accum == SKARR_SZ) {
s_num = num + 1;
goto next_chunk;
@@ -1080,7 +1080,6 @@ int __inet_hash_connect(struct inet_timewait_death_row *death_row,
goto ok;
next_port:
spin_unlock_bh(&head->lock);
- cond_resched();
}
offset++;
@@ -288,7 +288,6 @@ void inet_twsk_purge(struct inet_hashinfo *hashinfo, int family)
for (slot = 0; slot <= hashinfo->ehash_mask; slot++) {
struct inet_ehash_bucket *head = &hashinfo->ehash[slot];
restart_rcu:
- cond_resched();
rcu_read_lock();
restart:
sk_nulls_for_each_rcu(sk, node, &head->chain) {
@@ -300,7 +300,6 @@ void inetpeer_invalidate_tree(struct inet_peer_base *base)
p = rb_next(p);
rb_erase(&peer->rb_node, &base->rb_root);
inet_putpeer(peer);
- cond_resched();
}
base->total = 0;
@@ -622,7 +622,6 @@ static void get_counters(const struct xt_table_info *t,
ADD_COUNTER(counters[i], bcnt, pcnt);
++i;
- cond_resched();
}
}
}
@@ -642,7 +641,6 @@ static void get_old_counters(const struct xt_table_info *t,
ADD_COUNTER(counters[i], tmp->bcnt, tmp->pcnt);
++i;
}
- cond_resched();
}
}
@@ -761,7 +761,6 @@ get_counters(const struct xt_table_info *t,
ADD_COUNTER(counters[i], bcnt, pcnt);
++i; /* macro does multi eval of i */
- cond_resched();
}
}
}
@@ -781,8 +780,6 @@ static void get_old_counters(const struct xt_table_info *t,
ADD_COUNTER(counters[i], tmp->bcnt, tmp->pcnt);
++i; /* macro does multi eval of i */
}
-
- cond_resched();
}
}
@@ -2424,7 +2424,6 @@ static void flush_all_nexthops(struct net *net)
while ((node = rb_first(root))) {
nh = rb_entry(node, struct nexthop, rb_node);
remove_nexthop(net, nh, NULL);
- cond_resched();
}
}
@@ -2449,8 +2449,6 @@ static void *established_get_first(struct seq_file *seq)
struct hlist_nulls_node *node;
spinlock_t *lock = inet_ehash_lockp(hinfo, st->bucket);
- cond_resched();
-
/* Lockless fast path for the common case of empty buckets */
if (empty_bucket(hinfo, st))
continue;
@@ -281,7 +281,6 @@ int udp_lib_get_port(struct sock *sk, unsigned short snum,
snum += rand;
} while (snum != first);
spin_unlock_bh(&hslot->lock);
- cond_resched();
} while (++first != last);
goto fail;
} else {
@@ -1890,7 +1889,6 @@ int udp_recvmsg(struct sock *sk, struct msghdr *msg, size_t len, int flags,
kfree_skb(skb);
/* starting over for a new packet, but check if we need to yield */
- cond_resched();
msg->msg_flags &= ~MSG_TRUNC;
goto try_again;
}
@@ -843,7 +843,6 @@ static int netlink_autobind(struct socket *sock)
bool ok;
retry:
- cond_resched();
rcu_read_lock();
ok = !__netlink_lookup(table, portid, net);
rcu_read_unlock();
@@ -772,7 +772,6 @@ static u32 qdisc_alloc_handle(struct net_device *dev)
autohandle = TC_H_MAKE(0x80000000U, 0);
if (!qdisc_lookup(dev, autohandle))
return autohandle;
- cond_resched();
} while (--i > 0);
return 0;
@@ -923,7 +922,6 @@ static int tc_fill_qdisc(struct sk_buff *skb, struct Qdisc *q, u32 clid,
u32 block_index;
__u32 qlen;
- cond_resched();
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
if (!nlh)
goto out_nlmsg_trim;
@@ -1888,7 +1886,6 @@ static int tc_fill_tclass(struct sk_buff *skb, struct Qdisc *q,
struct gnet_dump d;
const struct Qdisc_class_ops *cl_ops = q->ops->cl_ops;
- cond_resched();
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
if (!nlh)
goto out_nlmsg_trim;
@@ -2709,7 +2709,6 @@ int __sys_sendmmsg(int fd, struct mmsghdr __user *mmsg, unsigned int vlen,
++datagrams;
if (msg_data_left(&msg_sys))
break;
- cond_resched();
}
fput_light(sock->file, fput_needed);
@@ -2944,7 +2943,6 @@ static int do_recvmmsg(int fd, struct mmsghdr __user *mmsg,
/* Out of band data, return right away */
if (msg_sys.msg_flags & MSG_OOB)
break;
- cond_resched();
}
if (err == 0)