[V4,4/5] cachefiles: narrow the scope of triggering EPOLLIN events in ondemand mode
Commit Message
Don't trigger EPOLLIN when there are only reopening read requests in
xarray.
Suggested-by: Xin Yin <yinxin.x@bytedance.com>
Signed-off-by: Jia Zhu <zhujia.zj@bytedance.com>
Reviewed-by: Jingbo Xu <jefflexu@linux.alibaba.com>
---
fs/cachefiles/daemon.c | 15 +++++++++++++--
fs/cachefiles/internal.h | 12 ++++++++++++
2 files changed, 25 insertions(+), 2 deletions(-)
Comments
Jia Zhu <zhujia.zj@bytedance.com> wrote:
> + if (!xa_empty(xa)) {
> + xa_lock(xa);
> + xa_for_each_marked(xa, index, req, CACHEFILES_REQ_NEW) {
> + if (!cachefiles_ondemand_is_reopening_read(req)) {
> + mask |= EPOLLIN;
> + break;
> + }
> + }
> + xa_unlock(xa);
> + }
I wonder if there's a more efficient way to do this. I guess it depends on
how many reqs you expect to get in a queue. It might be worth taking the
rcu_read_lock before calling xa_lock() and holding it over the whole loop.
David
在 2023/3/28 22:19, David Howells 写道:
> Jia Zhu <zhujia.zj@bytedance.com> wrote:
>
>> + if (!xa_empty(xa)) {
>> + xa_lock(xa);
>> + xa_for_each_marked(xa, index, req, CACHEFILES_REQ_NEW) {
>> + if (!cachefiles_ondemand_is_reopening_read(req)) {
>> + mask |= EPOLLIN;
>> + break;
>> + }
>> + }
>> + xa_unlock(xa);
>> + }
>
> I wonder if there's a more efficient way to do this. I guess it depends on
> how many reqs you expect to get in a queue. It might be worth taking the
> rcu_read_lock before calling xa_lock() and holding it over the whole loop.
>
Thanks for the advice, will use rcu_read_lock(unlock) to replace it.
> David
>
@@ -355,14 +355,25 @@ static __poll_t cachefiles_daemon_poll(struct file *file,
struct poll_table_struct *poll)
{
struct cachefiles_cache *cache = file->private_data;
+ struct xarray *xa = &cache->reqs;
+ struct cachefiles_req *req;
+ unsigned long index;
__poll_t mask;
poll_wait(file, &cache->daemon_pollwq, poll);
mask = 0;
if (cachefiles_in_ondemand_mode(cache)) {
- if (!xa_empty(&cache->reqs))
- mask |= EPOLLIN;
+ if (!xa_empty(xa)) {
+ xa_lock(xa);
+ xa_for_each_marked(xa, index, req, CACHEFILES_REQ_NEW) {
+ if (!cachefiles_ondemand_is_reopening_read(req)) {
+ mask |= EPOLLIN;
+ break;
+ }
+ }
+ xa_unlock(xa);
+ }
} else {
if (test_bit(CACHEFILES_STATE_CHANGED, &cache->flags))
mask |= EPOLLIN;
@@ -326,6 +326,13 @@ cachefiles_ondemand_set_object_##_state(struct cachefiles_object *object) \
CACHEFILES_OBJECT_STATE_FUNCS(open);
CACHEFILES_OBJECT_STATE_FUNCS(close);
CACHEFILES_OBJECT_STATE_FUNCS(reopening);
+
+static inline bool cachefiles_ondemand_is_reopening_read(struct cachefiles_req *req)
+{
+ return cachefiles_ondemand_object_is_reopening(req->object) &&
+ req->msg.opcode == CACHEFILES_OP_READ;
+}
+
#else
static inline ssize_t cachefiles_ondemand_daemon_read(struct cachefiles_cache *cache,
char __user *_buffer, size_t buflen)
@@ -353,6 +360,11 @@ static inline int cachefiles_ondemand_init_obj_info(struct cachefiles_object *ob
{
return 0;
}
+
+static inline bool cachefiles_ondemand_is_reopening_read(struct cachefiles_req *req)
+{
+ return false;
+}
#endif
/*