 f91ce58cb2
			
		
	
	
		f91ce58cb2
		
	
	
	
	
		
			
			Free allocated directory entries in v9fs_rreaddir() if argument `entries` was passed as NULL, to avoid a memory leak. It is explicitly allowed by design for `entries` to be NULL. [1] [1] https://lore.kernel.org/all/1690923.g4PEXVpXuU@silver Reported-by: Coverity (CID 1487558) Signed-off-by: Christian Schoenebeck <qemu_oss@crudebyte.com> Reviewed-by: Greg Kurz <groug@kaod.org> Message-Id: <E1psh5T-0002XN-1C@lizzy.crudebyte.com>
		
			
				
	
	
		
			1055 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1055 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * 9P network client for VirtIO 9P test cases (based on QTest)
 | |
|  *
 | |
|  * Copyright (c) 2014 SUSE LINUX Products GmbH
 | |
|  *
 | |
|  * This work is licensed under the terms of the GNU GPL, version 2 or later.
 | |
|  * See the COPYING file in the top-level directory.
 | |
|  */
 | |
| 
 | |
| /*
 | |
|  * Not so fast! You might want to read the 9p developer docs first:
 | |
|  * https://wiki.qemu.org/Documentation/9p
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "virtio-9p-client.h"
 | |
| 
 | |
| #define QVIRTIO_9P_TIMEOUT_US (10 * 1000 * 1000)
 | |
| static QGuestAllocator *alloc;
 | |
| 
 | |
| void v9fs_set_allocator(QGuestAllocator *t_alloc)
 | |
| {
 | |
|     alloc = t_alloc;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Used to auto generate new fids. Start with arbitrary high value to avoid
 | |
|  * collision with hard coded fids in basic test code.
 | |
|  */
 | |
| static uint32_t fid_generator = 1000;
 | |
| 
 | |
| static uint32_t genfid(void)
 | |
| {
 | |
|     return fid_generator++;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Splits the @a in string by @a delim into individual (non empty) strings
 | |
|  * and outputs them to @a out. The output array @a out is NULL terminated.
 | |
|  *
 | |
|  * Output array @a out must be freed by calling split_free().
 | |
|  *
 | |
|  * @returns number of individual elements in output array @a out (without the
 | |
|  *          final NULL terminating element)
 | |
|  */
 | |
| static int split(const char *in, const char *delim, char ***out)
 | |
| {
 | |
|     int n = 0, i = 0;
 | |
|     char *tmp, *p;
 | |
| 
 | |
|     tmp = g_strdup(in);
 | |
|     for (p = strtok(tmp, delim); p != NULL; p = strtok(NULL, delim)) {
 | |
|         if (strlen(p) > 0) {
 | |
|             ++n;
 | |
|         }
 | |
|     }
 | |
|     g_free(tmp);
 | |
| 
 | |
|     *out = g_new0(char *, n + 1); /* last element NULL delimiter */
 | |
| 
 | |
|     tmp = g_strdup(in);
 | |
|     for (p = strtok(tmp, delim); p != NULL; p = strtok(NULL, delim)) {
 | |
|         if (strlen(p) > 0) {
 | |
|             (*out)[i++] = g_strdup(p);
 | |
|         }
 | |
|     }
 | |
|     g_free(tmp);
 | |
| 
 | |
|     return n;
 | |
| }
 | |
| 
 | |
| static void split_free(char ***out)
 | |
| {
 | |
|     int i;
 | |
|     if (!*out) {
 | |
|         return;
 | |
|     }
 | |
|     for (i = 0; (*out)[i]; ++i) {
 | |
|         g_free((*out)[i]);
 | |
|     }
 | |
|     g_free(*out);
 | |
|     *out = NULL;
 | |
| }
 | |
| 
 | |
| void v9fs_memwrite(P9Req *req, const void *addr, size_t len)
 | |
| {
 | |
|     qtest_memwrite(req->qts, req->t_msg + req->t_off, addr, len);
 | |
|     req->t_off += len;
 | |
| }
 | |
| 
 | |
| void v9fs_memskip(P9Req *req, size_t len)
 | |
| {
 | |
|     req->r_off += len;
 | |
| }
 | |
| 
 | |
| void v9fs_memread(P9Req *req, void *addr, size_t len)
 | |
| {
 | |
|     qtest_memread(req->qts, req->r_msg + req->r_off, addr, len);
 | |
|     req->r_off += len;
 | |
| }
 | |
| 
 | |
| void v9fs_uint8_read(P9Req *req, uint8_t *val)
 | |
| {
 | |
|     v9fs_memread(req, val, 1);
 | |
| }
 | |
| 
 | |
| void v9fs_uint16_write(P9Req *req, uint16_t val)
 | |
| {
 | |
|     uint16_t le_val = cpu_to_le16(val);
 | |
| 
 | |
|     v9fs_memwrite(req, &le_val, 2);
 | |
| }
 | |
| 
 | |
| void v9fs_uint16_read(P9Req *req, uint16_t *val)
 | |
| {
 | |
|     v9fs_memread(req, val, 2);
 | |
|     le16_to_cpus(val);
 | |
| }
 | |
| 
 | |
| void v9fs_uint32_write(P9Req *req, uint32_t val)
 | |
| {
 | |
|     uint32_t le_val = cpu_to_le32(val);
 | |
| 
 | |
|     v9fs_memwrite(req, &le_val, 4);
 | |
| }
 | |
| 
 | |
| void v9fs_uint64_write(P9Req *req, uint64_t val)
 | |
| {
 | |
|     uint64_t le_val = cpu_to_le64(val);
 | |
| 
 | |
|     v9fs_memwrite(req, &le_val, 8);
 | |
| }
 | |
| 
 | |
| void v9fs_uint32_read(P9Req *req, uint32_t *val)
 | |
| {
 | |
|     v9fs_memread(req, val, 4);
 | |
|     le32_to_cpus(val);
 | |
| }
 | |
| 
 | |
| void v9fs_uint64_read(P9Req *req, uint64_t *val)
 | |
| {
 | |
|     v9fs_memread(req, val, 8);
 | |
|     le64_to_cpus(val);
 | |
| }
 | |
| 
 | |
| /* len[2] string[len] */
 | |
| uint16_t v9fs_string_size(const char *string)
 | |
| {
 | |
|     size_t len = strlen(string);
 | |
| 
 | |
|     g_assert_cmpint(len, <=, UINT16_MAX - 2);
 | |
| 
 | |
|     return 2 + len;
 | |
| }
 | |
| 
 | |
| void v9fs_string_write(P9Req *req, const char *string)
 | |
| {
 | |
|     int len = strlen(string);
 | |
| 
 | |
|     g_assert_cmpint(len, <=, UINT16_MAX);
 | |
| 
 | |
|     v9fs_uint16_write(req, (uint16_t) len);
 | |
|     v9fs_memwrite(req, string, len);
 | |
| }
 | |
| 
 | |
| void v9fs_string_read(P9Req *req, uint16_t *len, char **string)
 | |
| {
 | |
|     uint16_t local_len;
 | |
| 
 | |
|     v9fs_uint16_read(req, &local_len);
 | |
|     if (len) {
 | |
|         *len = local_len;
 | |
|     }
 | |
|     if (string) {
 | |
|         *string = g_malloc(local_len + 1);
 | |
|         v9fs_memread(req, *string, local_len);
 | |
|         (*string)[local_len] = 0;
 | |
|     } else {
 | |
|         v9fs_memskip(req, local_len);
 | |
|     }
 | |
| }
 | |
| 
 | |
| typedef struct {
 | |
|     uint32_t size;
 | |
|     uint8_t id;
 | |
|     uint16_t tag;
 | |
| } QEMU_PACKED P9Hdr;
 | |
| 
 | |
| P9Req *v9fs_req_init(QVirtio9P *v9p, uint32_t size, uint8_t id,
 | |
|                      uint16_t tag)
 | |
| {
 | |
|     P9Req *req = g_new0(P9Req, 1);
 | |
|     uint32_t total_size = 7; /* 9P header has well-known size of 7 bytes */
 | |
|     P9Hdr hdr = {
 | |
|         .id = id,
 | |
|         .tag = cpu_to_le16(tag)
 | |
|     };
 | |
| 
 | |
|     g_assert_cmpint(total_size, <=, UINT32_MAX - size);
 | |
|     total_size += size;
 | |
|     hdr.size = cpu_to_le32(total_size);
 | |
| 
 | |
|     g_assert_cmpint(total_size, <=, P9_MAX_SIZE);
 | |
| 
 | |
|     req->qts = global_qtest;
 | |
|     req->v9p = v9p;
 | |
|     req->t_size = total_size;
 | |
|     req->t_msg = guest_alloc(alloc, req->t_size);
 | |
|     v9fs_memwrite(req, &hdr, 7);
 | |
|     req->tag = tag;
 | |
|     return req;
 | |
| }
 | |
| 
 | |
| void v9fs_req_send(P9Req *req)
 | |
| {
 | |
|     QVirtio9P *v9p = req->v9p;
 | |
| 
 | |
|     req->r_msg = guest_alloc(alloc, P9_MAX_SIZE);
 | |
|     req->free_head = qvirtqueue_add(req->qts, v9p->vq, req->t_msg, req->t_size,
 | |
|                                     false, true);
 | |
|     qvirtqueue_add(req->qts, v9p->vq, req->r_msg, P9_MAX_SIZE, true, false);
 | |
|     qvirtqueue_kick(req->qts, v9p->vdev, v9p->vq, req->free_head);
 | |
|     req->t_off = 0;
 | |
| }
 | |
| 
 | |
| static const char *rmessage_name(uint8_t id)
 | |
| {
 | |
|     return
 | |
|         id == P9_RLERROR ? "RLERROR" :
 | |
|         id == P9_RVERSION ? "RVERSION" :
 | |
|         id == P9_RATTACH ? "RATTACH" :
 | |
|         id == P9_RWALK ? "RWALK" :
 | |
|         id == P9_RLOPEN ? "RLOPEN" :
 | |
|         id == P9_RWRITE ? "RWRITE" :
 | |
|         id == P9_RMKDIR ? "RMKDIR" :
 | |
|         id == P9_RLCREATE ? "RLCREATE" :
 | |
|         id == P9_RSYMLINK ? "RSYMLINK" :
 | |
|         id == P9_RLINK ? "RLINK" :
 | |
|         id == P9_RUNLINKAT ? "RUNLINKAT" :
 | |
|         id == P9_RFLUSH ? "RFLUSH" :
 | |
|         id == P9_RREADDIR ? "READDIR" :
 | |
|         "<unknown>";
 | |
| }
 | |
| 
 | |
| void v9fs_req_wait_for_reply(P9Req *req, uint32_t *len)
 | |
| {
 | |
|     QVirtio9P *v9p = req->v9p;
 | |
| 
 | |
|     qvirtio_wait_used_elem(req->qts, v9p->vdev, v9p->vq, req->free_head, len,
 | |
|                            QVIRTIO_9P_TIMEOUT_US);
 | |
| }
 | |
| 
 | |
| void v9fs_req_recv(P9Req *req, uint8_t id)
 | |
| {
 | |
|     P9Hdr hdr;
 | |
| 
 | |
|     v9fs_memread(req, &hdr, 7);
 | |
|     hdr.size = ldl_le_p(&hdr.size);
 | |
|     hdr.tag = lduw_le_p(&hdr.tag);
 | |
| 
 | |
|     g_assert_cmpint(hdr.size, >=, 7);
 | |
|     g_assert_cmpint(hdr.size, <=, P9_MAX_SIZE);
 | |
|     g_assert_cmpint(hdr.tag, ==, req->tag);
 | |
| 
 | |
|     if (hdr.id != id) {
 | |
|         g_printerr("Received response %d (%s) instead of %d (%s)\n",
 | |
|                    hdr.id, rmessage_name(hdr.id), id, rmessage_name(id));
 | |
| 
 | |
|         if (hdr.id == P9_RLERROR) {
 | |
|             uint32_t err;
 | |
|             v9fs_uint32_read(req, &err);
 | |
|             g_printerr("Rlerror has errno %d (%s)\n", err, strerror(err));
 | |
|         }
 | |
|     }
 | |
|     g_assert_cmpint(hdr.id, ==, id);
 | |
| }
 | |
| 
 | |
| void v9fs_req_free(P9Req *req)
 | |
| {
 | |
|     guest_free(alloc, req->t_msg);
 | |
|     guest_free(alloc, req->r_msg);
 | |
|     g_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Rlerror tag[2] ecode[4] */
 | |
| void v9fs_rlerror(P9Req *req, uint32_t *err)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RLERROR);
 | |
|     v9fs_uint32_read(req, err);
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tversion tag[2] msize[4] version[s] */
 | |
| TVersionRes v9fs_tversion(TVersionOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
|     uint32_t body_size = 4;
 | |
|     uint16_t string_size;
 | |
|     uint16_t server_len;
 | |
|     g_autofree char *server_version = NULL;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
| 
 | |
|     if (!opt.msize) {
 | |
|         opt.msize = P9_MAX_SIZE;
 | |
|     }
 | |
| 
 | |
|     if (!opt.tag) {
 | |
|         opt.tag = P9_NOTAG;
 | |
|     }
 | |
| 
 | |
|     if (!opt.version) {
 | |
|         opt.version = "9P2000.L";
 | |
|     }
 | |
| 
 | |
|     string_size = v9fs_string_size(opt.version);
 | |
|     g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
 | |
|     body_size += string_size;
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TVERSION, opt.tag);
 | |
| 
 | |
|     v9fs_uint32_write(req, opt.msize);
 | |
|     v9fs_string_write(req, opt.version);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rversion(req, &server_len, &server_version);
 | |
|             g_assert_cmpmem(server_version, server_len,
 | |
|                             opt.version, strlen(opt.version));
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TVersionRes) {
 | |
|         .req = req,
 | |
|     };
 | |
| }
 | |
| 
 | |
| /* size[4] Rversion tag[2] msize[4] version[s] */
 | |
| void v9fs_rversion(P9Req *req, uint16_t *len, char **version)
 | |
| {
 | |
|     uint32_t msize;
 | |
| 
 | |
|     v9fs_req_recv(req, P9_RVERSION);
 | |
|     v9fs_uint32_read(req, &msize);
 | |
| 
 | |
|     g_assert_cmpint(msize, ==, P9_MAX_SIZE);
 | |
| 
 | |
|     if (len || version) {
 | |
|         v9fs_string_read(req, len, version);
 | |
|     }
 | |
| 
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tattach tag[2] fid[4] afid[4] uname[s] aname[s] n_uname[4] */
 | |
| TAttachRes v9fs_tattach(TAttachOpt opt)
 | |
| {
 | |
|     uint32_t err;
 | |
|     const char *uname = ""; /* ignored by QEMU */
 | |
|     const char *aname = ""; /* ignored by QEMU */
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either Rattach or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !opt.rattach.qid);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_tversion((TVersionOpt) { .client = opt.client });
 | |
|     }
 | |
| 
 | |
|     if (!opt.n_uname) {
 | |
|         opt.n_uname = getuid();
 | |
|     }
 | |
| 
 | |
|     P9Req *req = v9fs_req_init(opt.client, 4 + 4 + 2 + 2 + 4, P9_TATTACH,
 | |
|                                opt.tag);
 | |
| 
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_uint32_write(req, P9_NOFID);
 | |
|     v9fs_string_write(req, uname);
 | |
|     v9fs_string_write(req, aname);
 | |
|     v9fs_uint32_write(req, opt.n_uname);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rattach(req, opt.rattach.qid);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TAttachRes) {
 | |
|         .req = req,
 | |
|     };
 | |
| }
 | |
| 
 | |
| /* size[4] Rattach tag[2] qid[13] */
 | |
| void v9fs_rattach(P9Req *req, v9fs_qid *qid)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RATTACH);
 | |
|     if (qid) {
 | |
|         v9fs_memread(req, qid, 13);
 | |
|     }
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Twalk tag[2] fid[4] newfid[4] nwname[2] nwname*(wname[s]) */
 | |
| TWalkRes v9fs_twalk(TWalkOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     int i;
 | |
|     uint32_t body_size = 4 + 4 + 2;
 | |
|     uint32_t err;
 | |
|     char **wnames = NULL;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either high- or low-level path, both not both */
 | |
|     g_assert(!opt.path || !(opt.nwname || opt.wnames));
 | |
|     /* expecting either Rwalk or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !(opt.rwalk.nwqid || opt.rwalk.wqid));
 | |
| 
 | |
|     if (!opt.newfid) {
 | |
|         opt.newfid = genfid();
 | |
|     }
 | |
| 
 | |
|     if (opt.path) {
 | |
|         opt.nwname = split(opt.path, "/", &wnames);
 | |
|         opt.wnames = wnames;
 | |
|     }
 | |
| 
 | |
|     for (i = 0; i < opt.nwname; i++) {
 | |
|         uint16_t wname_size = v9fs_string_size(opt.wnames[i]);
 | |
| 
 | |
|         g_assert_cmpint(body_size, <=, UINT32_MAX - wname_size);
 | |
|         body_size += wname_size;
 | |
|     }
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TWALK, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_uint32_write(req, opt.newfid);
 | |
|     v9fs_uint16_write(req, opt.nwname);
 | |
|     for (i = 0; i < opt.nwname; i++) {
 | |
|         v9fs_string_write(req, opt.wnames[i]);
 | |
|     }
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rwalk(req, opt.rwalk.nwqid, opt.rwalk.wqid);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     split_free(&wnames);
 | |
| 
 | |
|     return (TWalkRes) {
 | |
|         .newfid = opt.newfid,
 | |
|         .req = req,
 | |
|     };
 | |
| }
 | |
| 
 | |
| /* size[4] Rwalk tag[2] nwqid[2] nwqid*(wqid[13]) */
 | |
| void v9fs_rwalk(P9Req *req, uint16_t *nwqid, v9fs_qid **wqid)
 | |
| {
 | |
|     uint16_t local_nwqid;
 | |
| 
 | |
|     v9fs_req_recv(req, P9_RWALK);
 | |
|     v9fs_uint16_read(req, &local_nwqid);
 | |
|     if (nwqid) {
 | |
|         *nwqid = local_nwqid;
 | |
|     }
 | |
|     if (wqid) {
 | |
|         *wqid = g_malloc(local_nwqid * 13);
 | |
|         v9fs_memread(req, *wqid, local_nwqid * 13);
 | |
|     }
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tgetattr tag[2] fid[4] request_mask[8] */
 | |
| TGetAttrRes v9fs_tgetattr(TGetAttrOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either Rgetattr or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !opt.rgetattr.attr);
 | |
| 
 | |
|     if (!opt.request_mask) {
 | |
|         opt.request_mask = P9_GETATTR_ALL;
 | |
|     }
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, 4 + 8, P9_TGETATTR, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_uint64_write(req, opt.request_mask);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rgetattr(req, opt.rgetattr.attr);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TGetAttrRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * size[4] Rgetattr tag[2] valid[8] qid[13] mode[4] uid[4] gid[4] nlink[8]
 | |
|  *                  rdev[8] size[8] blksize[8] blocks[8]
 | |
|  *                  atime_sec[8] atime_nsec[8] mtime_sec[8] mtime_nsec[8]
 | |
|  *                  ctime_sec[8] ctime_nsec[8] btime_sec[8] btime_nsec[8]
 | |
|  *                  gen[8] data_version[8]
 | |
|  */
 | |
| void v9fs_rgetattr(P9Req *req, v9fs_attr *attr)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RGETATTR);
 | |
| 
 | |
|     v9fs_uint64_read(req, &attr->valid);
 | |
|     v9fs_memread(req, &attr->qid, 13);
 | |
|     v9fs_uint32_read(req, &attr->mode);
 | |
|     v9fs_uint32_read(req, &attr->uid);
 | |
|     v9fs_uint32_read(req, &attr->gid);
 | |
|     v9fs_uint64_read(req, &attr->nlink);
 | |
|     v9fs_uint64_read(req, &attr->rdev);
 | |
|     v9fs_uint64_read(req, &attr->size);
 | |
|     v9fs_uint64_read(req, &attr->blksize);
 | |
|     v9fs_uint64_read(req, &attr->blocks);
 | |
|     v9fs_uint64_read(req, &attr->atime_sec);
 | |
|     v9fs_uint64_read(req, &attr->atime_nsec);
 | |
|     v9fs_uint64_read(req, &attr->mtime_sec);
 | |
|     v9fs_uint64_read(req, &attr->mtime_nsec);
 | |
|     v9fs_uint64_read(req, &attr->ctime_sec);
 | |
|     v9fs_uint64_read(req, &attr->ctime_nsec);
 | |
|     v9fs_uint64_read(req, &attr->btime_sec);
 | |
|     v9fs_uint64_read(req, &attr->btime_nsec);
 | |
|     v9fs_uint64_read(req, &attr->gen);
 | |
|     v9fs_uint64_read(req, &attr->data_version);
 | |
| 
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Treaddir tag[2] fid[4] offset[8] count[4] */
 | |
| TReadDirRes v9fs_treaddir(TReadDirOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either Rreaddir or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !(opt.rreaddir.count ||
 | |
|              opt.rreaddir.nentries || opt.rreaddir.entries));
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, 4 + 8 + 4, P9_TREADDIR, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_uint64_write(req, opt.offset);
 | |
|     v9fs_uint32_write(req, opt.count);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rreaddir(req, opt.rreaddir.count, opt.rreaddir.nentries,
 | |
|                           opt.rreaddir.entries);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TReadDirRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Rreaddir tag[2] count[4] data[count] */
 | |
| void v9fs_rreaddir(P9Req *req, uint32_t *count, uint32_t *nentries,
 | |
|                    struct V9fsDirent **entries)
 | |
| {
 | |
|     uint32_t local_count;
 | |
|     struct V9fsDirent *e = NULL;
 | |
|     /* only used to avoid a leak if entries was NULL */
 | |
|     struct V9fsDirent *unused_entries = NULL;
 | |
|     uint16_t slen;
 | |
|     uint32_t n = 0;
 | |
| 
 | |
|     v9fs_req_recv(req, P9_RREADDIR);
 | |
|     v9fs_uint32_read(req, &local_count);
 | |
| 
 | |
|     if (count) {
 | |
|         *count = local_count;
 | |
|     }
 | |
| 
 | |
|     for (int32_t togo = (int32_t)local_count;
 | |
|          togo >= 13 + 8 + 1 + 2;
 | |
|          togo -= 13 + 8 + 1 + 2 + slen, ++n)
 | |
|     {
 | |
|         if (!e) {
 | |
|             e = g_new(struct V9fsDirent, 1);
 | |
|             if (entries) {
 | |
|                 *entries = e;
 | |
|             } else {
 | |
|                 unused_entries = e;
 | |
|             }
 | |
|         } else {
 | |
|             e = e->next = g_new(struct V9fsDirent, 1);
 | |
|         }
 | |
|         e->next = NULL;
 | |
|         /* qid[13] offset[8] type[1] name[s] */
 | |
|         v9fs_memread(req, &e->qid, 13);
 | |
|         v9fs_uint64_read(req, &e->offset);
 | |
|         v9fs_uint8_read(req, &e->type);
 | |
|         v9fs_string_read(req, &slen, &e->name);
 | |
|     }
 | |
| 
 | |
|     if (nentries) {
 | |
|         *nentries = n;
 | |
|     }
 | |
| 
 | |
|     v9fs_free_dirents(unused_entries);
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| void v9fs_free_dirents(struct V9fsDirent *e)
 | |
| {
 | |
|     struct V9fsDirent *next = NULL;
 | |
| 
 | |
|     for (; e; e = next) {
 | |
|         next = e->next;
 | |
|         g_free(e->name);
 | |
|         g_free(e);
 | |
|     }
 | |
| }
 | |
| 
 | |
| /* size[4] Tlopen tag[2] fid[4] flags[4] */
 | |
| TLOpenRes v9fs_tlopen(TLOpenOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either Rlopen or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !(opt.rlopen.qid || opt.rlopen.iounit));
 | |
| 
 | |
|     req = v9fs_req_init(opt.client,  4 + 4, P9_TLOPEN, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_uint32_write(req, opt.flags);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rlopen(req, opt.rlopen.qid, opt.rlopen.iounit);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TLOpenRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Rlopen tag[2] qid[13] iounit[4] */
 | |
| void v9fs_rlopen(P9Req *req, v9fs_qid *qid, uint32_t *iounit)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RLOPEN);
 | |
|     if (qid) {
 | |
|         v9fs_memread(req, qid, 13);
 | |
|     } else {
 | |
|         v9fs_memskip(req, 13);
 | |
|     }
 | |
|     if (iounit) {
 | |
|         v9fs_uint32_read(req, iounit);
 | |
|     }
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Twrite tag[2] fid[4] offset[8] count[4] data[count] */
 | |
| TWriteRes v9fs_twrite(TWriteOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
|     uint32_t body_size = 4 + 8 + 4;
 | |
|     uint32_t written = 0;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
| 
 | |
|     g_assert_cmpint(body_size, <=, UINT32_MAX - opt.count);
 | |
|     body_size += opt.count;
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TWRITE, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_uint64_write(req, opt.offset);
 | |
|     v9fs_uint32_write(req, opt.count);
 | |
|     v9fs_memwrite(req, opt.data, opt.count);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rwrite(req, &written);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TWriteRes) {
 | |
|         .req = req,
 | |
|         .count = written
 | |
|     };
 | |
| }
 | |
| 
 | |
| /* size[4] Rwrite tag[2] count[4] */
 | |
| void v9fs_rwrite(P9Req *req, uint32_t *count)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RWRITE);
 | |
|     if (count) {
 | |
|         v9fs_uint32_read(req, count);
 | |
|     }
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tflush tag[2] oldtag[2] */
 | |
| TFlushRes v9fs_tflush(TFlushOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, 2, P9_TFLUSH, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.oldtag);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rflush(req);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TFlushRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Rflush tag[2] */
 | |
| void v9fs_rflush(P9Req *req)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RFLUSH);
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tmkdir tag[2] dfid[4] name[s] mode[4] gid[4] */
 | |
| TMkdirRes v9fs_tmkdir(TMkdirOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either hi-level atPath or low-level dfid, but not both */
 | |
|     g_assert(!opt.atPath || !opt.dfid);
 | |
|     /* expecting either Rmkdir or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !opt.rmkdir.qid);
 | |
| 
 | |
|     if (opt.atPath) {
 | |
|         opt.dfid = v9fs_twalk((TWalkOpt) { .client = opt.client,
 | |
|                                            .path = opt.atPath }).newfid;
 | |
|     }
 | |
| 
 | |
|     if (!opt.mode) {
 | |
|         opt.mode = 0750;
 | |
|     }
 | |
| 
 | |
|     uint32_t body_size = 4 + 4 + 4;
 | |
|     uint16_t string_size = v9fs_string_size(opt.name);
 | |
| 
 | |
|     g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
 | |
|     body_size += string_size;
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TMKDIR, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.dfid);
 | |
|     v9fs_string_write(req, opt.name);
 | |
|     v9fs_uint32_write(req, opt.mode);
 | |
|     v9fs_uint32_write(req, opt.gid);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rmkdir(req, opt.rmkdir.qid);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TMkdirRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Rmkdir tag[2] qid[13] */
 | |
| void v9fs_rmkdir(P9Req *req, v9fs_qid *qid)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RMKDIR);
 | |
|     if (qid) {
 | |
|         v9fs_memread(req, qid, 13);
 | |
|     } else {
 | |
|         v9fs_memskip(req, 13);
 | |
|     }
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tlcreate tag[2] fid[4] name[s] flags[4] mode[4] gid[4] */
 | |
| TlcreateRes v9fs_tlcreate(TlcreateOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either hi-level atPath or low-level fid, but not both */
 | |
|     g_assert(!opt.atPath || !opt.fid);
 | |
|     /* expecting either Rlcreate or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !(opt.rlcreate.qid || opt.rlcreate.iounit));
 | |
| 
 | |
|     if (opt.atPath) {
 | |
|         opt.fid = v9fs_twalk((TWalkOpt) { .client = opt.client,
 | |
|                                           .path = opt.atPath }).newfid;
 | |
|     }
 | |
| 
 | |
|     if (!opt.mode) {
 | |
|         opt.mode = 0750;
 | |
|     }
 | |
| 
 | |
|     uint32_t body_size = 4 + 4 + 4 + 4;
 | |
|     uint16_t string_size = v9fs_string_size(opt.name);
 | |
| 
 | |
|     g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
 | |
|     body_size += string_size;
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TLCREATE, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_string_write(req, opt.name);
 | |
|     v9fs_uint32_write(req, opt.flags);
 | |
|     v9fs_uint32_write(req, opt.mode);
 | |
|     v9fs_uint32_write(req, opt.gid);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rlcreate(req, opt.rlcreate.qid, opt.rlcreate.iounit);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TlcreateRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Rlcreate tag[2] qid[13] iounit[4] */
 | |
| void v9fs_rlcreate(P9Req *req, v9fs_qid *qid, uint32_t *iounit)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RLCREATE);
 | |
|     if (qid) {
 | |
|         v9fs_memread(req, qid, 13);
 | |
|     } else {
 | |
|         v9fs_memskip(req, 13);
 | |
|     }
 | |
|     if (iounit) {
 | |
|         v9fs_uint32_read(req, iounit);
 | |
|     }
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tsymlink tag[2] fid[4] name[s] symtgt[s] gid[4] */
 | |
| TsymlinkRes v9fs_tsymlink(TsymlinkOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either hi-level atPath or low-level fid, but not both */
 | |
|     g_assert(!opt.atPath || !opt.fid);
 | |
|     /* expecting either Rsymlink or Rlerror, but obviously not both */
 | |
|     g_assert(!opt.expectErr || !opt.rsymlink.qid);
 | |
| 
 | |
|     if (opt.atPath) {
 | |
|         opt.fid = v9fs_twalk((TWalkOpt) { .client = opt.client,
 | |
|                                           .path = opt.atPath }).newfid;
 | |
|     }
 | |
| 
 | |
|     uint32_t body_size = 4 + 4;
 | |
|     uint16_t string_size = v9fs_string_size(opt.name) +
 | |
|                            v9fs_string_size(opt.symtgt);
 | |
| 
 | |
|     g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
 | |
|     body_size += string_size;
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TSYMLINK, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_string_write(req, opt.name);
 | |
|     v9fs_string_write(req, opt.symtgt);
 | |
|     v9fs_uint32_write(req, opt.gid);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rsymlink(req, opt.rsymlink.qid);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TsymlinkRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Rsymlink tag[2] qid[13] */
 | |
| void v9fs_rsymlink(P9Req *req, v9fs_qid *qid)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RSYMLINK);
 | |
|     if (qid) {
 | |
|         v9fs_memread(req, qid, 13);
 | |
|     } else {
 | |
|         v9fs_memskip(req, 13);
 | |
|     }
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tlink tag[2] dfid[4] fid[4] name[s] */
 | |
| TlinkRes v9fs_tlink(TlinkOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either hi-level atPath or low-level dfid, but not both */
 | |
|     g_assert(!opt.atPath || !opt.dfid);
 | |
|     /* expecting either hi-level toPath or low-level fid, but not both */
 | |
|     g_assert(!opt.toPath || !opt.fid);
 | |
| 
 | |
|     if (opt.atPath) {
 | |
|         opt.dfid = v9fs_twalk((TWalkOpt) { .client = opt.client,
 | |
|                                            .path = opt.atPath }).newfid;
 | |
|     }
 | |
|     if (opt.toPath) {
 | |
|         opt.fid = v9fs_twalk((TWalkOpt) { .client = opt.client,
 | |
|                                           .path = opt.toPath }).newfid;
 | |
|     }
 | |
| 
 | |
|     uint32_t body_size = 4 + 4;
 | |
|     uint16_t string_size = v9fs_string_size(opt.name);
 | |
| 
 | |
|     g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
 | |
|     body_size += string_size;
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TLINK, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.dfid);
 | |
|     v9fs_uint32_write(req, opt.fid);
 | |
|     v9fs_string_write(req, opt.name);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_rlink(req);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TlinkRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Rlink tag[2] */
 | |
| void v9fs_rlink(P9Req *req)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RLINK);
 | |
|     v9fs_req_free(req);
 | |
| }
 | |
| 
 | |
| /* size[4] Tunlinkat tag[2] dirfd[4] name[s] flags[4] */
 | |
| TunlinkatRes v9fs_tunlinkat(TunlinkatOpt opt)
 | |
| {
 | |
|     P9Req *req;
 | |
|     uint32_t err;
 | |
| 
 | |
|     g_assert(opt.client);
 | |
|     /* expecting either hi-level atPath or low-level dirfd, but not both */
 | |
|     g_assert(!opt.atPath || !opt.dirfd);
 | |
| 
 | |
|     if (opt.atPath) {
 | |
|         opt.dirfd = v9fs_twalk((TWalkOpt) { .client = opt.client,
 | |
|                                             .path = opt.atPath }).newfid;
 | |
|     }
 | |
| 
 | |
|     uint32_t body_size = 4 + 4;
 | |
|     uint16_t string_size = v9fs_string_size(opt.name);
 | |
| 
 | |
|     g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
 | |
|     body_size += string_size;
 | |
| 
 | |
|     req = v9fs_req_init(opt.client, body_size, P9_TUNLINKAT, opt.tag);
 | |
|     v9fs_uint32_write(req, opt.dirfd);
 | |
|     v9fs_string_write(req, opt.name);
 | |
|     v9fs_uint32_write(req, opt.flags);
 | |
|     v9fs_req_send(req);
 | |
| 
 | |
|     if (!opt.requestOnly) {
 | |
|         v9fs_req_wait_for_reply(req, NULL);
 | |
|         if (opt.expectErr) {
 | |
|             v9fs_rlerror(req, &err);
 | |
|             g_assert_cmpint(err, ==, opt.expectErr);
 | |
|         } else {
 | |
|             v9fs_runlinkat(req);
 | |
|         }
 | |
|         req = NULL; /* request was freed */
 | |
|     }
 | |
| 
 | |
|     return (TunlinkatRes) { .req = req };
 | |
| }
 | |
| 
 | |
| /* size[4] Runlinkat tag[2] */
 | |
| void v9fs_runlinkat(P9Req *req)
 | |
| {
 | |
|     v9fs_req_recv(req, P9_RUNLINKAT);
 | |
|     v9fs_req_free(req);
 | |
| }
 |