#include #include #include #include #include #include #include #include static void vfs_backend_user_accept(struct vfs_request *req); void vfsreq_create(struct vfs_request req_) { struct vfs_request *req; if (req_.caller) { process_transition(req_.caller, PS_WAITS4FS); if (!req_.caller->reqslot) req_.caller->reqslot = kmalloc(sizeof *req); req = req_.caller->reqslot; /* (re)using a single allocation for all request a process makes */ } else { req = kmalloc(sizeof *req); } memcpy(req, &req_, sizeof *req); if (req->backend) req->backend->refcount++; if (req->type == VFSOP_OPEN && !(req->flags & OPEN_WRITE) && (req->flags & OPEN_CREATE)) { vfsreq_finish_short(req, -EINVAL); return; } // TODO if i add a handle field to vfs_request, check ->readable ->writeable here if (req->backend && req->backend->potential_handlers) { struct vfs_request **iter = &req->backend->queue; while (*iter != NULL) // find free spot in queue iter = &(*iter)->queue_next; *iter = req; vfs_backend_tryaccept(req->backend); } else { vfsreq_finish_short(req, -1); } } void vfsreq_finish(struct vfs_request *req, char __user *stored, long ret, int flags, struct process *handler) { if (req->type == VFSOP_OPEN && ret >= 0) { struct handle *h; if (!(flags & FSR_DELEGATE)) { /* default behavior - create a new handle for the file, wrap the id */ h = handle_init(HANDLE_FILE); h->backend = req->backend; req->backend->refcount++; h->file_id = stored; h->readable = OPEN_READABLE(req->flags); h->writeable = OPEN_WRITEABLE(req->flags); } else { /* delegating - moving a handle to the caller */ assert(handler); h = process_handle_take(handler, ret); // TODO don't ignore OPEN_RO } if (h) { // TODO write tests for caller getting killed while opening a file if (!req->caller) panic_unimplemented(); ret = process_handle_put(req->caller, h); if (ret < 0) ret = -EMFILE; } else { ret = -1; } } if (req->input.kern) kfree(req->input.buf_kern); if (req->backend) vfs_backend_refdown(req->backend); if (req->caller) { assert(req->caller->state == PS_WAITS4FS); regs_savereturn(&req->caller->regs, ret); process_transition(req->caller, PS_RUNNING); } else { kfree(req); } } void vfs_backend_tryaccept(struct vfs_backend *backend) { struct vfs_request *req = backend->queue; if (!req) return; if (backend->is_user && !backend->user.handler) return; backend->queue = req->queue_next; if (backend->is_user) { vfs_backend_user_accept(req); } else { assert(backend->kern.accept); backend->kern.accept(req); } } static void vfs_backend_user_accept(struct vfs_request *req) { struct process *handler; struct ufs_request res = {0}; struct virt_cpy_error cpyerr; int len; assert(req && req->backend && req->backend->user.handler); handler = req->backend->user.handler; assert(handler->state == PS_WAITS4REQUEST); // the virt_cpy calls aren't present in all kernel backends // it's a way to tell apart kernel and user backends apart // TODO check validity of memory regions somewhere else if (req->input.buf) { len = min(req->input.len, handler->awaited_req.max_len); virt_cpy(handler->pages, handler->awaited_req.buf, req->input.kern ? NULL : req->caller->pages, req->input.buf, len, &cpyerr); if (cpyerr.write_fail) panic_unimplemented(); if (cpyerr.read_fail) { vfsreq_finish_short(req, -EFAULT); return; } } else { len = req->output.len; } res.len = len; res.capacity = req->output.len; res.id = req->id; res.offset = req->offset; res.flags = req->flags; res.op = req->type; if (!virt_cpy_to(handler->pages, handler->awaited_req.res, &res, sizeof res)) { panic_unimplemented(); } struct handle *h; handle_t hid = process_handle_init(handler, HANDLE_FS_REQ, &h); if (hid < 0) panic_unimplemented(); h->req = req; process_transition(handler, PS_RUNNING); regs_savereturn(&handler->regs, hid); req->backend->user.handler = NULL; return; } void vfs_backend_refdown(struct vfs_backend *b) { assert(b); assert(b->refcount > 0); if (--(b->refcount) > 0) return; assert(!b->queue); kfree(b); }