4 * Copyright IBM, Corp. 2011
7 * Aneesh Kumar K.V <aneesh.kumar@linux.vnet.ibm.com>
9 * This work is licensed under the terms of the GNU GPL, version 2. See
10 * the COPYING file in the top-level directory.
15 * Not so fast! You might want to read the 9p developer docs first:
16 * https://wiki.qemu.org/Documentation/9p
19 #include "qemu/osdep.h"
20 #include "fsdev/qemu-fsdev.h"
21 #include "qemu/thread.h"
22 #include "qemu/coroutine.h"
23 #include "qemu/main-loop.h"
27 * Intended to be called from bottom-half (e.g. background I/O thread)
30 static int do_readdir(V9fsPDU
*pdu
, V9fsFidState
*fidp
, struct dirent
**dent
)
33 V9fsState
*s
= pdu
->s
;
37 entry
= s
->ops
->readdir(&s
->ctx
, &fidp
->fs
);
38 if (!entry
&& errno
) {
48 * TODO: This will be removed for performance reasons.
49 * Use v9fs_co_readdir_many() instead.
51 int coroutine_fn
v9fs_co_readdir(V9fsPDU
*pdu
, V9fsFidState
*fidp
,
56 if (v9fs_request_cancelled(pdu
)) {
59 v9fs_co_run_in_worker({
60 err
= do_readdir(pdu
, fidp
, dent
);
66 * This is solely executed on a background IO thread.
68 * See v9fs_co_readdir_many() (as its only user) below for details.
70 static int do_readdir_many(V9fsPDU
*pdu
, V9fsFidState
*fidp
,
71 struct V9fsDirEnt
**entries
, off_t offset
,
72 int32_t maxsize
, bool dostat
)
74 V9fsState
*s
= pdu
->s
;
80 struct V9fsDirEnt
*e
= NULL
;
85 v9fs_path_init(&path
);
88 * TODO: Here should be a warn_report_once() if lock failed.
90 * With a good 9p client we should not get into concurrency here,
91 * because a good client would not use the same fid for concurrent
92 * requests. We do the lock here for safety reasons though. However
93 * the client would then suffer performance issues, so better log that
96 v9fs_readdir_lock(&fidp
->fs
.dir
);
98 /* seek directory to requested initial position */
100 s
->ops
->rewinddir(&s
->ctx
, &fidp
->fs
);
102 s
->ops
->seekdir(&s
->ctx
, &fidp
->fs
, offset
);
105 /* save the directory position */
106 saved_dir_pos
= s
->ops
->telldir(&s
->ctx
, &fidp
->fs
);
107 if (saved_dir_pos
< 0) {
113 /* interrupt loop if request was cancelled by a Tflush request */
114 if (v9fs_request_cancelled(pdu
)) {
119 /* get directory entry from fs driver */
120 err
= do_readdir(pdu
, fidp
, &dent
);
126 * stop this loop as soon as it would exceed the allowed maximum
127 * response message size for the directory entries collected so far,
128 * because anything beyond that size would need to be discarded by
129 * 9p controller (main thread / top half) anyway
131 v9fs_string_init(&name
);
132 v9fs_string_sprintf(&name
, "%s", dent
->d_name
);
133 len
= v9fs_readdir_response_size(&name
);
134 v9fs_string_free(&name
);
135 if (size
+ len
> maxsize
) {
136 /* this is not an error case actually */
140 /* append next node to result chain */
142 *entries
= e
= g_malloc0(sizeof(V9fsDirEnt
));
144 e
= e
->next
= g_malloc0(sizeof(V9fsDirEnt
));
146 e
->dent
= g_malloc0(sizeof(struct dirent
));
147 memcpy(e
->dent
, dent
, sizeof(struct dirent
));
149 /* perform a full stat() for directory entry if requested by caller */
151 err
= s
->ops
->name_to_path(
152 &s
->ctx
, &fidp
->path
, dent
->d_name
, &path
159 err
= s
->ops
->lstat(&s
->ctx
, &path
, &stbuf
);
165 e
->st
= g_malloc0(sizeof(struct stat
));
166 memcpy(e
->st
, &stbuf
, sizeof(struct stat
));
170 saved_dir_pos
= dent
->d_off
;
173 /* restore (last) saved position */
174 s
->ops
->seekdir(&s
->ctx
, &fidp
->fs
, saved_dir_pos
);
177 v9fs_readdir_unlock(&fidp
->fs
.dir
);
178 v9fs_path_free(&path
);
186 * @brief Reads multiple directory entries in one rush.
188 * Retrieves the requested (max. amount of) directory entries from the fs
189 * driver. This function must only be called by the main IO thread (top half).
190 * Internally this function call will be dispatched to a background IO thread
191 * (bottom half) where it is eventually executed by the fs driver.
193 * @discussion Acquiring multiple directory entries in one rush from the fs
194 * driver, instead of retrieving each directory entry individually, is very
195 * beneficial from performance point of view. Because for every fs driver
196 * request latency is added, which in practice could lead to overall
197 * latencies of several hundred ms for reading all entries (of just a single
198 * directory) if every directory entry was individually requested from fs
201 * @note You must @b ALWAYS call @c v9fs_free_dirents(entries) after calling
202 * v9fs_co_readdir_many(), both on success and on error cases of this
203 * function, to avoid memory leaks once @p entries are no longer needed.
205 * @param pdu - the causing 9p (T_readdir) client request
206 * @param fidp - already opened directory where readdir shall be performed on
207 * @param entries - output for directory entries (must not be NULL)
208 * @param offset - initial position inside the directory the function shall
209 * seek to before retrieving the directory entries
210 * @param maxsize - maximum result message body size (in bytes)
211 * @param dostat - whether a stat() should be performed and returned for
212 * each directory entry
213 * @returns resulting response message body size (in bytes) on success,
214 * negative error code otherwise
216 int coroutine_fn
v9fs_co_readdir_many(V9fsPDU
*pdu
, V9fsFidState
*fidp
,
217 struct V9fsDirEnt
**entries
,
218 off_t offset
, int32_t maxsize
,
223 if (v9fs_request_cancelled(pdu
)) {
226 v9fs_co_run_in_worker({
227 err
= do_readdir_many(pdu
, fidp
, entries
, offset
, maxsize
, dostat
);
232 off_t
v9fs_co_telldir(V9fsPDU
*pdu
, V9fsFidState
*fidp
)
235 V9fsState
*s
= pdu
->s
;
237 if (v9fs_request_cancelled(pdu
)) {
240 v9fs_co_run_in_worker(
242 err
= s
->ops
->telldir(&s
->ctx
, &fidp
->fs
);
250 void coroutine_fn
v9fs_co_seekdir(V9fsPDU
*pdu
, V9fsFidState
*fidp
,
253 V9fsState
*s
= pdu
->s
;
254 if (v9fs_request_cancelled(pdu
)) {
257 v9fs_co_run_in_worker(
259 s
->ops
->seekdir(&s
->ctx
, &fidp
->fs
, offset
);
263 void coroutine_fn
v9fs_co_rewinddir(V9fsPDU
*pdu
, V9fsFidState
*fidp
)
265 V9fsState
*s
= pdu
->s
;
266 if (v9fs_request_cancelled(pdu
)) {
269 v9fs_co_run_in_worker(
271 s
->ops
->rewinddir(&s
->ctx
, &fidp
->fs
);
275 int coroutine_fn
v9fs_co_mkdir(V9fsPDU
*pdu
, V9fsFidState
*fidp
,
276 V9fsString
*name
, mode_t mode
, uid_t uid
,
277 gid_t gid
, struct stat
*stbuf
)
282 V9fsState
*s
= pdu
->s
;
284 if (v9fs_request_cancelled(pdu
)) {
291 v9fs_path_read_lock(s
);
292 v9fs_co_run_in_worker(
294 err
= s
->ops
->mkdir(&s
->ctx
, &fidp
->path
, name
->data
, &cred
);
298 v9fs_path_init(&path
);
299 err
= v9fs_name_to_path(s
, &fidp
->path
, name
->data
, &path
);
301 err
= s
->ops
->lstat(&s
->ctx
, &path
, stbuf
);
306 v9fs_path_free(&path
);
313 int coroutine_fn
v9fs_co_opendir(V9fsPDU
*pdu
, V9fsFidState
*fidp
)
316 V9fsState
*s
= pdu
->s
;
318 if (v9fs_request_cancelled(pdu
)) {
321 v9fs_path_read_lock(s
);
322 v9fs_co_run_in_worker(
324 err
= s
->ops
->opendir(&s
->ctx
, &fidp
->path
, &fidp
->fs
);
334 if (total_open_fd
> open_fd_hw
) {
335 v9fs_reclaim_fd(pdu
);
341 int coroutine_fn
v9fs_co_closedir(V9fsPDU
*pdu
, V9fsFidOpenState
*fs
)
344 V9fsState
*s
= pdu
->s
;
346 if (v9fs_request_cancelled(pdu
)) {
349 v9fs_co_run_in_worker(
351 err
= s
->ops
->closedir(&s
->ctx
, fs
);