| /* |
| * Server-side device support |
| * |
| * Copyright (C) 2007 Alexandre Julliard |
| * |
| * This library is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * This library is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with this library; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA |
| */ |
| |
| #include "config.h" |
| #include "wine/port.h" |
| |
| #include <assert.h> |
| #include <fcntl.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <stdarg.h> |
| |
| #include "ntstatus.h" |
| #define WIN32_NO_STATUS |
| #include "windef.h" |
| #include "winternl.h" |
| #include "ddk/wdm.h" |
| |
| #include "object.h" |
| #include "file.h" |
| #include "handle.h" |
| #include "request.h" |
| #include "process.h" |
| |
| /* IRP object */ |
| |
| struct irp_call |
| { |
| struct object obj; /* object header */ |
| struct list dev_entry; /* entry in device queue */ |
| struct list mgr_entry; /* entry in manager queue */ |
| struct device_file *file; /* file containing this irp */ |
| struct thread *thread; /* thread that queued the irp */ |
| struct async *async; /* pending async op */ |
| irp_params_t params; /* irp parameters */ |
| struct iosb *iosb; /* I/O status block */ |
| }; |
| |
| static void irp_call_dump( struct object *obj, int verbose ); |
| static int irp_call_signaled( struct object *obj, struct wait_queue_entry *entry ); |
| static void irp_call_destroy( struct object *obj ); |
| |
| static const struct object_ops irp_call_ops = |
| { |
| sizeof(struct irp_call), /* size */ |
| irp_call_dump, /* dump */ |
| no_get_type, /* get_type */ |
| add_queue, /* add_queue */ |
| remove_queue, /* remove_queue */ |
| irp_call_signaled, /* signaled */ |
| no_satisfied, /* satisfied */ |
| no_signal, /* signal */ |
| no_get_fd, /* get_fd */ |
| no_map_access, /* map_access */ |
| default_get_sd, /* get_sd */ |
| default_set_sd, /* set_sd */ |
| no_lookup_name, /* lookup_name */ |
| no_link_name, /* link_name */ |
| NULL, /* unlink_name */ |
| no_open_file, /* open_file */ |
| no_close_handle, /* close_handle */ |
| irp_call_destroy /* destroy */ |
| }; |
| |
| |
| /* device manager (a list of devices managed by the same client process) */ |
| |
| struct device_manager |
| { |
| struct object obj; /* object header */ |
| struct list devices; /* list of devices */ |
| struct list requests; /* list of pending irps across all devices */ |
| }; |
| |
| static void device_manager_dump( struct object *obj, int verbose ); |
| static int device_manager_signaled( struct object *obj, struct wait_queue_entry *entry ); |
| static void device_manager_destroy( struct object *obj ); |
| |
| static const struct object_ops device_manager_ops = |
| { |
| sizeof(struct device_manager), /* size */ |
| device_manager_dump, /* dump */ |
| no_get_type, /* get_type */ |
| add_queue, /* add_queue */ |
| remove_queue, /* remove_queue */ |
| device_manager_signaled, /* signaled */ |
| no_satisfied, /* satisfied */ |
| no_signal, /* signal */ |
| no_get_fd, /* get_fd */ |
| no_map_access, /* map_access */ |
| default_get_sd, /* get_sd */ |
| default_set_sd, /* set_sd */ |
| no_lookup_name, /* lookup_name */ |
| no_link_name, /* link_name */ |
| NULL, /* unlink_name */ |
| no_open_file, /* open_file */ |
| no_close_handle, /* close_handle */ |
| device_manager_destroy /* destroy */ |
| }; |
| |
| |
| /* device (a single device object) */ |
| |
| struct device |
| { |
| struct object obj; /* object header */ |
| struct device_manager *manager; /* manager for this device (or NULL if deleted) */ |
| char *unix_path; /* path to unix device if any */ |
| client_ptr_t user_ptr; /* opaque ptr for client side */ |
| struct list entry; /* entry in device manager list */ |
| struct list files; /* list of open files */ |
| }; |
| |
| static void device_dump( struct object *obj, int verbose ); |
| static struct object_type *device_get_type( struct object *obj ); |
| static void device_destroy( struct object *obj ); |
| static struct object *device_open_file( struct object *obj, unsigned int access, |
| unsigned int sharing, unsigned int options ); |
| |
| static const struct object_ops device_ops = |
| { |
| sizeof(struct device), /* size */ |
| device_dump, /* dump */ |
| device_get_type, /* get_type */ |
| no_add_queue, /* add_queue */ |
| NULL, /* remove_queue */ |
| NULL, /* signaled */ |
| no_satisfied, /* satisfied */ |
| no_signal, /* signal */ |
| no_get_fd, /* get_fd */ |
| default_fd_map_access, /* map_access */ |
| default_get_sd, /* get_sd */ |
| default_set_sd, /* set_sd */ |
| no_lookup_name, /* lookup_name */ |
| directory_link_name, /* link_name */ |
| default_unlink_name, /* unlink_name */ |
| device_open_file, /* open_file */ |
| no_close_handle, /* close_handle */ |
| device_destroy /* destroy */ |
| }; |
| |
| |
| /* device file (an open file handle to a device) */ |
| |
| struct device_file |
| { |
| struct object obj; /* object header */ |
| struct device *device; /* device for this file */ |
| struct fd *fd; /* file descriptor for irp */ |
| client_ptr_t user_ptr; /* opaque ptr for client side */ |
| struct list entry; /* entry in device list */ |
| struct list requests; /* list of pending irp requests */ |
| }; |
| |
| static void device_file_dump( struct object *obj, int verbose ); |
| static struct fd *device_file_get_fd( struct object *obj ); |
| static int device_file_close_handle( struct object *obj, struct process *process, obj_handle_t handle ); |
| static void device_file_destroy( struct object *obj ); |
| static enum server_fd_type device_file_get_fd_type( struct fd *fd ); |
| static int device_file_read( struct fd *fd, struct async *async, file_pos_t pos ); |
| static int device_file_write( struct fd *fd, struct async *async, file_pos_t pos ); |
| static int device_file_flush( struct fd *fd, struct async *async ); |
| static int device_file_ioctl( struct fd *fd, ioctl_code_t code, struct async *async ); |
| |
| static const struct object_ops device_file_ops = |
| { |
| sizeof(struct device_file), /* size */ |
| device_file_dump, /* dump */ |
| no_get_type, /* get_type */ |
| add_queue, /* add_queue */ |
| remove_queue, /* remove_queue */ |
| default_fd_signaled, /* signaled */ |
| no_satisfied, /* satisfied */ |
| no_signal, /* signal */ |
| device_file_get_fd, /* get_fd */ |
| default_fd_map_access, /* map_access */ |
| default_get_sd, /* get_sd */ |
| default_set_sd, /* set_sd */ |
| no_lookup_name, /* lookup_name */ |
| no_link_name, /* link_name */ |
| NULL, /* unlink_name */ |
| no_open_file, /* open_file */ |
| device_file_close_handle, /* close_handle */ |
| device_file_destroy /* destroy */ |
| }; |
| |
| static const struct fd_ops device_file_fd_ops = |
| { |
| default_fd_get_poll_events, /* get_poll_events */ |
| default_poll_event, /* poll_event */ |
| device_file_get_fd_type, /* get_fd_type */ |
| device_file_read, /* read */ |
| device_file_write, /* write */ |
| device_file_flush, /* flush */ |
| device_file_ioctl, /* ioctl */ |
| default_fd_queue_async, /* queue_async */ |
| default_fd_reselect_async /* reselect_async */ |
| }; |
| |
| |
| static void irp_call_dump( struct object *obj, int verbose ) |
| { |
| struct irp_call *irp = (struct irp_call *)obj; |
| fprintf( stderr, "IRP call file=%p\n", irp->file ); |
| } |
| |
| static int irp_call_signaled( struct object *obj, struct wait_queue_entry *entry ) |
| { |
| struct irp_call *irp = (struct irp_call *)obj; |
| |
| return !irp->file; /* file is cleared once the irp has completed */ |
| } |
| |
| static void irp_call_destroy( struct object *obj ) |
| { |
| struct irp_call *irp = (struct irp_call *)obj; |
| |
| if (irp->async) |
| { |
| async_terminate( irp->async, STATUS_CANCELLED ); |
| release_object( irp->async ); |
| } |
| if (irp->iosb) release_object( irp->iosb ); |
| if (irp->file) release_object( irp->file ); |
| if (irp->thread) release_object( irp->thread ); |
| } |
| |
| static struct irp_call *create_irp( struct device_file *file, const irp_params_t *params, struct async *async ) |
| { |
| struct irp_call *irp; |
| |
| if (!file->device->manager) /* it has been deleted */ |
| { |
| set_error( STATUS_FILE_DELETED ); |
| return NULL; |
| } |
| |
| if ((irp = alloc_object( &irp_call_ops ))) |
| { |
| irp->file = (struct device_file *)grab_object( file ); |
| irp->thread = NULL; |
| irp->async = NULL; |
| irp->params = *params; |
| irp->iosb = NULL; |
| |
| if (async) irp->iosb = async_get_iosb( async ); |
| if (!irp->iosb && !(irp->iosb = create_iosb( NULL, 0, 0 ))) |
| { |
| release_object( irp ); |
| irp = NULL; |
| } |
| } |
| return irp; |
| } |
| |
| static void set_irp_result( struct irp_call *irp, unsigned int status, |
| const void *out_data, data_size_t out_size, data_size_t result ) |
| { |
| struct device_file *file = irp->file; |
| struct iosb *iosb = irp->iosb; |
| |
| if (!file) return; /* already finished */ |
| |
| /* FIXME: handle the STATUS_PENDING case */ |
| iosb->status = status; |
| iosb->result = result; |
| iosb->out_size = min( iosb->out_size, out_size ); |
| if (iosb->out_size && !(iosb->out_data = memdup( out_data, iosb->out_size ))) |
| iosb->out_size = 0; |
| irp->file = NULL; |
| if (irp->async) |
| { |
| if (result) status = STATUS_ALERTED; |
| async_terminate( irp->async, status ); |
| release_object( irp->async ); |
| irp->async = NULL; |
| } |
| wake_up( &irp->obj, 0 ); |
| |
| /* remove it from the device queue */ |
| list_remove( &irp->dev_entry ); |
| release_object( irp ); /* no longer on the device queue */ |
| release_object( file ); |
| } |
| |
| |
| static void device_dump( struct object *obj, int verbose ) |
| { |
| fputs( "Device\n", stderr ); |
| } |
| |
| static struct object_type *device_get_type( struct object *obj ) |
| { |
| static const WCHAR name[] = {'D','e','v','i','c','e'}; |
| static const struct unicode_str str = { name, sizeof(name) }; |
| return get_object_type( &str ); |
| } |
| |
| static void device_destroy( struct object *obj ) |
| { |
| struct device *device = (struct device *)obj; |
| |
| assert( list_empty( &device->files )); |
| |
| free( device->unix_path ); |
| if (device->manager) list_remove( &device->entry ); |
| } |
| |
| static void add_irp_to_queue( struct device_file *file, struct irp_call *irp, struct thread *thread ) |
| { |
| struct device_manager *manager = file->device->manager; |
| |
| assert( manager ); |
| |
| grab_object( irp ); /* grab reference for queued irp */ |
| irp->thread = thread ? (struct thread *)grab_object( thread ) : NULL; |
| list_add_tail( &file->requests, &irp->dev_entry ); |
| list_add_tail( &manager->requests, &irp->mgr_entry ); |
| if (list_head( &manager->requests ) == &irp->mgr_entry) wake_up( &manager->obj, 0 ); /* first one */ |
| } |
| |
| static struct object *device_open_file( struct object *obj, unsigned int access, |
| unsigned int sharing, unsigned int options ) |
| { |
| struct device *device = (struct device *)obj; |
| struct device_file *file; |
| |
| if (!(file = alloc_object( &device_file_ops ))) return NULL; |
| |
| file->device = (struct device *)grab_object( device ); |
| file->user_ptr = 0; |
| list_init( &file->requests ); |
| list_add_tail( &device->files, &file->entry ); |
| if (device->unix_path) |
| { |
| mode_t mode = 0666; |
| access = file->obj.ops->map_access( &file->obj, access ); |
| file->fd = open_fd( NULL, device->unix_path, O_NONBLOCK | O_LARGEFILE, |
| &mode, access, sharing, options ); |
| if (file->fd) set_fd_user( file->fd, &device_file_fd_ops, &file->obj ); |
| } |
| else file->fd = alloc_pseudo_fd( &device_file_fd_ops, &file->obj, 0 ); |
| |
| if (!file->fd) |
| { |
| release_object( file ); |
| return NULL; |
| } |
| |
| allow_fd_caching( file->fd ); |
| |
| if (device->manager) |
| { |
| struct irp_call *irp; |
| irp_params_t params; |
| |
| memset( ¶ms, 0, sizeof(params) ); |
| params.create.major = IRP_MJ_CREATE; |
| params.create.access = access; |
| params.create.sharing = sharing; |
| params.create.options = options; |
| params.create.device = file->device->user_ptr; |
| |
| if ((irp = create_irp( file, ¶ms, NULL ))) |
| { |
| add_irp_to_queue( file, irp, NULL ); |
| release_object( irp ); |
| } |
| } |
| return &file->obj; |
| } |
| |
| static void device_file_dump( struct object *obj, int verbose ) |
| { |
| struct device_file *file = (struct device_file *)obj; |
| |
| fprintf( stderr, "File on device %p\n", file->device ); |
| } |
| |
| static struct fd *device_file_get_fd( struct object *obj ) |
| { |
| struct device_file *file = (struct device_file *)obj; |
| |
| return (struct fd *)grab_object( file->fd ); |
| } |
| |
| static int device_file_close_handle( struct object *obj, struct process *process, obj_handle_t handle ) |
| { |
| struct device_file *file = (struct device_file *)obj; |
| |
| if (file->device->manager && obj->handle_count == 1) /* last handle */ |
| { |
| struct irp_call *irp; |
| irp_params_t params; |
| |
| memset( ¶ms, 0, sizeof(params) ); |
| params.close.major = IRP_MJ_CLOSE; |
| params.close.file = file->user_ptr; |
| |
| if ((irp = create_irp( file, ¶ms, NULL ))) |
| { |
| add_irp_to_queue( file, irp, NULL ); |
| release_object( irp ); |
| } |
| } |
| return 1; |
| } |
| |
| static void device_file_destroy( struct object *obj ) |
| { |
| struct device_file *file = (struct device_file *)obj; |
| struct irp_call *irp, *next; |
| |
| LIST_FOR_EACH_ENTRY_SAFE( irp, next, &file->requests, struct irp_call, dev_entry ) |
| { |
| list_remove( &irp->dev_entry ); |
| release_object( irp ); /* no longer on the device queue */ |
| } |
| if (file->fd) release_object( file->fd ); |
| list_remove( &file->entry ); |
| release_object( file->device ); |
| } |
| |
| static void set_file_user_ptr( struct device_file *file, client_ptr_t ptr ) |
| { |
| struct irp_call *irp; |
| |
| if (file->user_ptr == ptr) return; /* nothing to do */ |
| |
| file->user_ptr = ptr; |
| |
| /* update already queued irps */ |
| |
| LIST_FOR_EACH_ENTRY( irp, &file->requests, struct irp_call, dev_entry ) |
| { |
| switch (irp->params.major) |
| { |
| case IRP_MJ_CLOSE: irp->params.close.file = ptr; break; |
| case IRP_MJ_READ: irp->params.read.file = ptr; break; |
| case IRP_MJ_WRITE: irp->params.write.file = ptr; break; |
| case IRP_MJ_FLUSH_BUFFERS: irp->params.flush.file = ptr; break; |
| case IRP_MJ_DEVICE_CONTROL: irp->params.ioctl.file = ptr; break; |
| } |
| } |
| } |
| |
| /* queue an irp to the device */ |
| static int queue_irp( struct device_file *file, const irp_params_t *params, struct async *async ) |
| { |
| struct irp_call *irp = create_irp( file, params, async ); |
| if (!irp) return 0; |
| |
| fd_queue_async( file->fd, async, ASYNC_TYPE_WAIT ); |
| irp->async = (struct async *)grab_object( async ); |
| add_irp_to_queue( file, irp, current ); |
| release_object( irp ); |
| set_error( STATUS_PENDING ); |
| return 1; |
| } |
| |
| static enum server_fd_type device_file_get_fd_type( struct fd *fd ) |
| { |
| return FD_TYPE_DEVICE; |
| } |
| |
| static int device_file_read( struct fd *fd, struct async *async, file_pos_t pos ) |
| { |
| struct device_file *file = get_fd_user( fd ); |
| irp_params_t params; |
| |
| memset( ¶ms, 0, sizeof(params) ); |
| params.read.major = IRP_MJ_READ; |
| params.read.key = 0; |
| params.read.pos = pos; |
| params.read.file = file->user_ptr; |
| return queue_irp( file, ¶ms, async ); |
| } |
| |
| static int device_file_write( struct fd *fd, struct async *async, file_pos_t pos ) |
| { |
| struct device_file *file = get_fd_user( fd ); |
| irp_params_t params; |
| |
| memset( ¶ms, 0, sizeof(params) ); |
| params.write.major = IRP_MJ_WRITE; |
| params.write.key = 0; |
| params.write.pos = pos; |
| params.write.file = file->user_ptr; |
| return queue_irp( file, ¶ms, async ); |
| } |
| |
| static int device_file_flush( struct fd *fd, struct async *async ) |
| { |
| struct device_file *file = get_fd_user( fd ); |
| irp_params_t params; |
| |
| memset( ¶ms, 0, sizeof(params) ); |
| params.flush.major = IRP_MJ_FLUSH_BUFFERS; |
| params.flush.file = file->user_ptr; |
| return queue_irp( file, ¶ms, NULL ); |
| } |
| |
| static int device_file_ioctl( struct fd *fd, ioctl_code_t code, struct async *async ) |
| { |
| struct device_file *file = get_fd_user( fd ); |
| irp_params_t params; |
| |
| memset( ¶ms, 0, sizeof(params) ); |
| params.ioctl.major = IRP_MJ_DEVICE_CONTROL; |
| params.ioctl.code = code; |
| params.ioctl.file = file->user_ptr; |
| return queue_irp( file, ¶ms, async ); |
| } |
| |
| static struct device *create_device( struct object *root, const struct unicode_str *name, |
| struct device_manager *manager, unsigned int attr ) |
| { |
| struct device *device; |
| |
| if ((device = create_named_object( root, &device_ops, name, attr, NULL ))) |
| { |
| if (get_error() != STATUS_OBJECT_NAME_EXISTS) |
| { |
| /* initialize it if it didn't already exist */ |
| device->unix_path = NULL; |
| device->manager = manager; |
| list_add_tail( &manager->devices, &device->entry ); |
| list_init( &device->files ); |
| } |
| } |
| return device; |
| } |
| |
| struct object *create_unix_device( struct object *root, const struct unicode_str *name, |
| const char *unix_path ) |
| { |
| struct device *device; |
| |
| if ((device = create_named_object( root, &device_ops, name, 0, NULL ))) |
| { |
| device->unix_path = strdup( unix_path ); |
| device->manager = NULL; /* no manager, requests go straight to the Unix device */ |
| list_init( &device->files ); |
| } |
| return &device->obj; |
| |
| } |
| |
| /* terminate requests when the underlying device is deleted */ |
| static void delete_file( struct device_file *file ) |
| { |
| struct irp_call *irp, *next; |
| |
| /* terminate all pending requests */ |
| LIST_FOR_EACH_ENTRY_SAFE( irp, next, &file->requests, struct irp_call, dev_entry ) |
| { |
| list_remove( &irp->mgr_entry ); |
| set_irp_result( irp, STATUS_FILE_DELETED, NULL, 0, 0 ); |
| } |
| } |
| |
| static void delete_device( struct device *device ) |
| { |
| struct device_file *file, *next; |
| |
| if (!device->manager) return; /* already deleted */ |
| |
| LIST_FOR_EACH_ENTRY_SAFE( file, next, &device->files, struct device_file, entry ) |
| delete_file( file ); |
| |
| unlink_named_object( &device->obj ); |
| list_remove( &device->entry ); |
| device->manager = NULL; |
| } |
| |
| |
| static void device_manager_dump( struct object *obj, int verbose ) |
| { |
| fprintf( stderr, "Device manager\n" ); |
| } |
| |
| static int device_manager_signaled( struct object *obj, struct wait_queue_entry *entry ) |
| { |
| struct device_manager *manager = (struct device_manager *)obj; |
| |
| return !list_empty( &manager->requests ); |
| } |
| |
| static void device_manager_destroy( struct object *obj ) |
| { |
| struct device_manager *manager = (struct device_manager *)obj; |
| struct list *ptr; |
| |
| while ((ptr = list_head( &manager->devices ))) |
| { |
| struct device *device = LIST_ENTRY( ptr, struct device, entry ); |
| delete_device( device ); |
| } |
| } |
| |
| static struct device_manager *create_device_manager(void) |
| { |
| struct device_manager *manager; |
| |
| if ((manager = alloc_object( &device_manager_ops ))) |
| { |
| list_init( &manager->devices ); |
| list_init( &manager->requests ); |
| } |
| return manager; |
| } |
| |
| |
| /* create a device manager */ |
| DECL_HANDLER(create_device_manager) |
| { |
| struct device_manager *manager = create_device_manager(); |
| |
| if (manager) |
| { |
| reply->handle = alloc_handle( current->process, manager, req->access, req->attributes ); |
| release_object( manager ); |
| } |
| } |
| |
| |
| /* create a device */ |
| DECL_HANDLER(create_device) |
| { |
| struct device *device; |
| struct unicode_str name = get_req_unicode_str(); |
| struct device_manager *manager; |
| struct object *root = NULL; |
| |
| if (!(manager = (struct device_manager *)get_handle_obj( current->process, req->manager, |
| 0, &device_manager_ops ))) |
| return; |
| |
| if (req->rootdir && !(root = get_directory_obj( current->process, req->rootdir ))) |
| { |
| release_object( manager ); |
| return; |
| } |
| |
| if ((device = create_device( root, &name, manager, req->attributes ))) |
| { |
| device->user_ptr = req->user_ptr; |
| reply->handle = alloc_handle( current->process, device, req->access, req->attributes ); |
| release_object( device ); |
| } |
| |
| if (root) release_object( root ); |
| release_object( manager ); |
| } |
| |
| |
| /* delete a device */ |
| DECL_HANDLER(delete_device) |
| { |
| struct device *device; |
| |
| if ((device = (struct device *)get_handle_obj( current->process, req->handle, 0, &device_ops ))) |
| { |
| delete_device( device ); |
| release_object( device ); |
| } |
| } |
| |
| |
| /* retrieve the next pending device irp request */ |
| DECL_HANDLER(get_next_device_request) |
| { |
| struct irp_call *irp; |
| struct device_manager *manager; |
| struct list *ptr; |
| struct iosb *iosb; |
| |
| reply->params.major = IRP_MJ_MAXIMUM_FUNCTION + 1; |
| |
| if (!(manager = (struct device_manager *)get_handle_obj( current->process, req->manager, |
| 0, &device_manager_ops ))) |
| return; |
| |
| if (req->prev) |
| { |
| if ((irp = (struct irp_call *)get_handle_obj( current->process, req->prev, 0, &irp_call_ops ))) |
| { |
| set_irp_result( irp, req->status, NULL, 0, 0 ); |
| close_handle( current->process, req->prev ); /* avoid an extra round-trip for close */ |
| release_object( irp ); |
| } |
| clear_error(); |
| } |
| |
| if ((ptr = list_head( &manager->requests ))) |
| { |
| irp = LIST_ENTRY( ptr, struct irp_call, mgr_entry ); |
| if (irp->thread) |
| { |
| reply->client_pid = get_process_id( irp->thread->process ); |
| reply->client_tid = get_thread_id( irp->thread ); |
| } |
| reply->params = irp->params; |
| iosb = irp->iosb; |
| reply->in_size = iosb->in_size; |
| reply->out_size = iosb->out_size; |
| if (iosb->in_size > get_reply_max_size()) set_error( STATUS_BUFFER_OVERFLOW ); |
| else if ((reply->next = alloc_handle( current->process, irp, 0, 0 ))) |
| { |
| set_reply_data_ptr( iosb->in_data, iosb->in_size ); |
| iosb->in_data = NULL; |
| iosb->in_size = 0; |
| list_remove( &irp->mgr_entry ); |
| list_init( &irp->mgr_entry ); |
| } |
| } |
| else set_error( STATUS_PENDING ); |
| |
| release_object( manager ); |
| } |
| |
| |
| /* store results of an async irp */ |
| DECL_HANDLER(set_irp_result) |
| { |
| struct irp_call *irp; |
| |
| if ((irp = (struct irp_call *)get_handle_obj( current->process, req->handle, 0, &irp_call_ops ))) |
| { |
| if (irp->file) set_file_user_ptr( irp->file, req->file_ptr ); |
| set_irp_result( irp, req->status, get_req_data(), get_req_data_size(), req->size ); |
| close_handle( current->process, req->handle ); /* avoid an extra round-trip for close */ |
| release_object( irp ); |
| } |
| } |