1687 lines
		
	
	
		
			47 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1687 lines
		
	
	
		
			47 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * Copyright (C) 2010 - 2015 UNISYS CORPORATION
 | |
|  * All rights reserved.
 | |
|  */
 | |
| 
 | |
| #include <linux/acpi.h>
 | |
| #include <linux/crash_dump.h>
 | |
| #include <linux/visorbus.h>
 | |
| 
 | |
| #include "visorbus_private.h"
 | |
| 
 | |
| /* {72120008-4AAB-11DC-8530-444553544200} */
 | |
| #define VISOR_SIOVM_GUID GUID_INIT(0x72120008, 0x4AAB, 0x11DC, 0x85, 0x30, \
 | |
| 				   0x44, 0x45, 0x53, 0x54, 0x42, 0x00)
 | |
| 
 | |
| static const guid_t visor_vhba_channel_guid = VISOR_VHBA_CHANNEL_GUID;
 | |
| static const guid_t visor_siovm_guid = VISOR_SIOVM_GUID;
 | |
| static const guid_t visor_controlvm_channel_guid = VISOR_CONTROLVM_CHANNEL_GUID;
 | |
| 
 | |
| #define POLLJIFFIES_CONTROLVM_FAST 1
 | |
| #define POLLJIFFIES_CONTROLVM_SLOW 100
 | |
| 
 | |
| #define MAX_CONTROLVM_PAYLOAD_BYTES (1024 * 128)
 | |
| 
 | |
| #define UNISYS_VISOR_LEAF_ID 0x40000000
 | |
| 
 | |
| /* The s-Par leaf ID returns "UnisysSpar64" encoded across ebx, ecx, edx */
 | |
| #define UNISYS_VISOR_ID_EBX 0x73696e55
 | |
| #define UNISYS_VISOR_ID_ECX 0x70537379
 | |
| #define UNISYS_VISOR_ID_EDX 0x34367261
 | |
| 
 | |
| /*
 | |
|  * When the controlvm channel is idle for at least MIN_IDLE_SECONDS, we switch
 | |
|  * to slow polling mode. As soon as we get a controlvm message, we switch back
 | |
|  * to fast polling mode.
 | |
|  */
 | |
| #define MIN_IDLE_SECONDS 10
 | |
| 
 | |
| struct parser_context {
 | |
| 	unsigned long allocbytes;
 | |
| 	unsigned long param_bytes;
 | |
| 	u8 *curr;
 | |
| 	unsigned long bytes_remaining;
 | |
| 	bool byte_stream;
 | |
| 	struct visor_controlvm_parameters_header data;
 | |
| };
 | |
| 
 | |
| /* VMCALL_CONTROLVM_ADDR: Used by all guests, not just IO. */
 | |
| #define VMCALL_CONTROLVM_ADDR 0x0501
 | |
| 
 | |
| enum vmcall_result {
 | |
| 	VMCALL_RESULT_SUCCESS = 0,
 | |
| 	VMCALL_RESULT_INVALID_PARAM = 1,
 | |
| 	VMCALL_RESULT_DATA_UNAVAILABLE = 2,
 | |
| 	VMCALL_RESULT_FAILURE_UNAVAILABLE = 3,
 | |
| 	VMCALL_RESULT_DEVICE_ERROR = 4,
 | |
| 	VMCALL_RESULT_DEVICE_NOT_READY = 5
 | |
| };
 | |
| 
 | |
| /*
 | |
|  * struct vmcall_io_controlvm_addr_params - Structure for IO VMCALLS. Has
 | |
|  *					    parameters to VMCALL_CONTROLVM_ADDR
 | |
|  *					    interface.
 | |
|  * @address:	   The Guest-relative physical address of the ControlVm channel.
 | |
|  *		   This VMCall fills this in with the appropriate address.
 | |
|  *		   Contents provided by this VMCALL (OUT).
 | |
|  * @channel_bytes: The size of the ControlVm channel in bytes This VMCall fills
 | |
|  *		   this in with the appropriate address. Contents provided by
 | |
|  *		   this VMCALL (OUT).
 | |
|  * @unused:	   Unused Bytes in the 64-Bit Aligned Struct.
 | |
|  */
 | |
| struct vmcall_io_controlvm_addr_params {
 | |
| 	u64 address;
 | |
| 	u32 channel_bytes;
 | |
| 	u8 unused[4];
 | |
| } __packed;
 | |
| 
 | |
| struct visorchipset_device {
 | |
| 	struct acpi_device *acpi_device;
 | |
| 	unsigned long poll_jiffies;
 | |
| 	/* when we got our last controlvm message */
 | |
| 	unsigned long most_recent_message_jiffies;
 | |
| 	struct delayed_work periodic_controlvm_work;
 | |
| 	struct visorchannel *controlvm_channel;
 | |
| 	unsigned long controlvm_payload_bytes_buffered;
 | |
| 	/*
 | |
| 	 * The following variables are used to handle the scenario where we are
 | |
| 	 * unable to offload the payload from a controlvm message due to memory
 | |
| 	 * requirements. In this scenario, we simply stash the controlvm
 | |
| 	 * message, then attempt to process it again the next time
 | |
| 	 * controlvm_periodic_work() runs.
 | |
| 	 */
 | |
| 	struct controlvm_message controlvm_pending_msg;
 | |
| 	bool controlvm_pending_msg_valid;
 | |
| 	struct vmcall_io_controlvm_addr_params controlvm_params;
 | |
| };
 | |
| 
 | |
| static struct visorchipset_device *chipset_dev;
 | |
| 
 | |
| struct parahotplug_request {
 | |
| 	struct list_head list;
 | |
| 	int id;
 | |
| 	unsigned long expiration;
 | |
| 	struct controlvm_message msg;
 | |
| };
 | |
| 
 | |
| /* prototypes for attributes */
 | |
| static ssize_t toolaction_show(struct device *dev,
 | |
| 			       struct device_attribute *attr,
 | |
| 			       char *buf)
 | |
| {
 | |
| 	u8 tool_action = 0;
 | |
| 	int err;
 | |
| 
 | |
| 	err = visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 				offsetof(struct visor_controlvm_channel,
 | |
| 					 tool_action),
 | |
| 				&tool_action, sizeof(u8));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return sprintf(buf, "%u\n", tool_action);
 | |
| }
 | |
| 
 | |
| static ssize_t toolaction_store(struct device *dev,
 | |
| 				struct device_attribute *attr,
 | |
| 				const char *buf, size_t count)
 | |
| {
 | |
| 	u8 tool_action;
 | |
| 	int err;
 | |
| 
 | |
| 	if (kstrtou8(buf, 10, &tool_action))
 | |
| 		return -EINVAL;
 | |
| 	err = visorchannel_write(chipset_dev->controlvm_channel,
 | |
| 				 offsetof(struct visor_controlvm_channel,
 | |
| 					  tool_action),
 | |
| 				 &tool_action, sizeof(u8));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return count;
 | |
| }
 | |
| static DEVICE_ATTR_RW(toolaction);
 | |
| 
 | |
| static ssize_t boottotool_show(struct device *dev,
 | |
| 			       struct device_attribute *attr,
 | |
| 			       char *buf)
 | |
| {
 | |
| 	struct efi_visor_indication efi_visor_indication;
 | |
| 	int err;
 | |
| 
 | |
| 	err = visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 				offsetof(struct visor_controlvm_channel,
 | |
| 					 efi_visor_ind),
 | |
| 				&efi_visor_indication,
 | |
| 				sizeof(struct efi_visor_indication));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return sprintf(buf, "%u\n", efi_visor_indication.boot_to_tool);
 | |
| }
 | |
| 
 | |
| static ssize_t boottotool_store(struct device *dev,
 | |
| 				struct device_attribute *attr,
 | |
| 				const char *buf, size_t count)
 | |
| {
 | |
| 	int val, err;
 | |
| 	struct efi_visor_indication efi_visor_indication;
 | |
| 
 | |
| 	if (kstrtoint(buf, 10, &val))
 | |
| 		return -EINVAL;
 | |
| 	efi_visor_indication.boot_to_tool = val;
 | |
| 	err = visorchannel_write(chipset_dev->controlvm_channel,
 | |
| 				 offsetof(struct visor_controlvm_channel,
 | |
| 					  efi_visor_ind),
 | |
| 				 &(efi_visor_indication),
 | |
| 				 sizeof(struct efi_visor_indication));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return count;
 | |
| }
 | |
| static DEVICE_ATTR_RW(boottotool);
 | |
| 
 | |
| static ssize_t error_show(struct device *dev, struct device_attribute *attr,
 | |
| 			  char *buf)
 | |
| {
 | |
| 	u32 error = 0;
 | |
| 	int err;
 | |
| 
 | |
| 	err = visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 				offsetof(struct visor_controlvm_channel,
 | |
| 					 installation_error),
 | |
| 				&error, sizeof(u32));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return sprintf(buf, "%u\n", error);
 | |
| }
 | |
| 
 | |
| static ssize_t error_store(struct device *dev, struct device_attribute *attr,
 | |
| 			   const char *buf, size_t count)
 | |
| {
 | |
| 	u32 error;
 | |
| 	int err;
 | |
| 
 | |
| 	if (kstrtou32(buf, 10, &error))
 | |
| 		return -EINVAL;
 | |
| 	err = visorchannel_write(chipset_dev->controlvm_channel,
 | |
| 				 offsetof(struct visor_controlvm_channel,
 | |
| 					  installation_error),
 | |
| 				 &error, sizeof(u32));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return count;
 | |
| }
 | |
| static DEVICE_ATTR_RW(error);
 | |
| 
 | |
| static ssize_t textid_show(struct device *dev, struct device_attribute *attr,
 | |
| 			   char *buf)
 | |
| {
 | |
| 	u32 text_id = 0;
 | |
| 	int err;
 | |
| 
 | |
| 	err = visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 				offsetof(struct visor_controlvm_channel,
 | |
| 					 installation_text_id),
 | |
| 				&text_id, sizeof(u32));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return sprintf(buf, "%u\n", text_id);
 | |
| }
 | |
| 
 | |
| static ssize_t textid_store(struct device *dev, struct device_attribute *attr,
 | |
| 			    const char *buf, size_t count)
 | |
| {
 | |
| 	u32 text_id;
 | |
| 	int err;
 | |
| 
 | |
| 	if (kstrtou32(buf, 10, &text_id))
 | |
| 		return -EINVAL;
 | |
| 	err = visorchannel_write(chipset_dev->controlvm_channel,
 | |
| 				 offsetof(struct visor_controlvm_channel,
 | |
| 					  installation_text_id),
 | |
| 				 &text_id, sizeof(u32));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return count;
 | |
| }
 | |
| static DEVICE_ATTR_RW(textid);
 | |
| 
 | |
| static ssize_t remaining_steps_show(struct device *dev,
 | |
| 				    struct device_attribute *attr, char *buf)
 | |
| {
 | |
| 	u16 remaining_steps = 0;
 | |
| 	int err;
 | |
| 
 | |
| 	err = visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 				offsetof(struct visor_controlvm_channel,
 | |
| 					 installation_remaining_steps),
 | |
| 				&remaining_steps, sizeof(u16));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return sprintf(buf, "%hu\n", remaining_steps);
 | |
| }
 | |
| 
 | |
| static ssize_t remaining_steps_store(struct device *dev,
 | |
| 				     struct device_attribute *attr,
 | |
| 				     const char *buf, size_t count)
 | |
| {
 | |
| 	u16 remaining_steps;
 | |
| 	int err;
 | |
| 
 | |
| 	if (kstrtou16(buf, 10, &remaining_steps))
 | |
| 		return -EINVAL;
 | |
| 	err = visorchannel_write(chipset_dev->controlvm_channel,
 | |
| 				 offsetof(struct visor_controlvm_channel,
 | |
| 					  installation_remaining_steps),
 | |
| 				 &remaining_steps, sizeof(u16));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	return count;
 | |
| }
 | |
| static DEVICE_ATTR_RW(remaining_steps);
 | |
| 
 | |
| static void controlvm_init_response(struct controlvm_message *msg,
 | |
| 				    struct controlvm_message_header *msg_hdr,
 | |
| 				    int response)
 | |
| {
 | |
| 	memset(msg, 0, sizeof(struct controlvm_message));
 | |
| 	memcpy(&msg->hdr, msg_hdr, sizeof(struct controlvm_message_header));
 | |
| 	msg->hdr.payload_bytes = 0;
 | |
| 	msg->hdr.payload_vm_offset = 0;
 | |
| 	msg->hdr.payload_max_bytes = 0;
 | |
| 	if (response < 0) {
 | |
| 		msg->hdr.flags.failed = 1;
 | |
| 		msg->hdr.completion_status = (u32)(-response);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int controlvm_respond_chipset_init(
 | |
| 				struct controlvm_message_header *msg_hdr,
 | |
| 				int response,
 | |
| 				enum visor_chipset_feature features)
 | |
| {
 | |
| 	struct controlvm_message outmsg;
 | |
| 
 | |
| 	controlvm_init_response(&outmsg, msg_hdr, response);
 | |
| 	outmsg.cmd.init_chipset.features = features;
 | |
| 	return visorchannel_signalinsert(chipset_dev->controlvm_channel,
 | |
| 					 CONTROLVM_QUEUE_REQUEST, &outmsg);
 | |
| }
 | |
| 
 | |
| static int chipset_init(struct controlvm_message *inmsg)
 | |
| {
 | |
| 	static int chipset_inited;
 | |
| 	enum visor_chipset_feature features = 0;
 | |
| 	int rc = CONTROLVM_RESP_SUCCESS;
 | |
| 	int res = 0;
 | |
| 
 | |
| 	if (chipset_inited) {
 | |
| 		rc = -CONTROLVM_RESP_ALREADY_DONE;
 | |
| 		res = -EIO;
 | |
| 		goto out_respond;
 | |
| 	}
 | |
| 	chipset_inited = 1;
 | |
| 	/*
 | |
| 	 * Set features to indicate we support parahotplug (if Command also
 | |
| 	 * supports it). Set the "reply" bit so Command knows this is a
 | |
| 	 * features-aware driver.
 | |
| 	 */
 | |
| 	features = inmsg->cmd.init_chipset.features &
 | |
| 		   VISOR_CHIPSET_FEATURE_PARA_HOTPLUG;
 | |
| 	features |= VISOR_CHIPSET_FEATURE_REPLY;
 | |
| 
 | |
| out_respond:
 | |
| 	if (inmsg->hdr.flags.response_expected)
 | |
| 		res = controlvm_respond_chipset_init(&inmsg->hdr, rc, features);
 | |
| 
 | |
| 	return res;
 | |
| }
 | |
| 
 | |
| static int controlvm_respond(struct controlvm_message_header *msg_hdr,
 | |
| 			     int response, struct visor_segment_state *state)
 | |
| {
 | |
| 	struct controlvm_message outmsg;
 | |
| 
 | |
| 	controlvm_init_response(&outmsg, msg_hdr, response);
 | |
| 	if (outmsg.hdr.flags.test_message == 1)
 | |
| 		return -EINVAL;
 | |
| 	if (state) {
 | |
| 		outmsg.cmd.device_change_state.state = *state;
 | |
| 		outmsg.cmd.device_change_state.flags.phys_device = 1;
 | |
| 	}
 | |
| 	return visorchannel_signalinsert(chipset_dev->controlvm_channel,
 | |
| 					 CONTROLVM_QUEUE_REQUEST, &outmsg);
 | |
| }
 | |
| 
 | |
| enum crash_obj_type {
 | |
| 	CRASH_DEV,
 | |
| 	CRASH_BUS,
 | |
| };
 | |
| 
 | |
| static int save_crash_message(struct controlvm_message *msg,
 | |
| 			      enum crash_obj_type cr_type)
 | |
| {
 | |
| 	u32 local_crash_msg_offset;
 | |
| 	u16 local_crash_msg_count;
 | |
| 	int err;
 | |
| 
 | |
| 	err = visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 				offsetof(struct visor_controlvm_channel,
 | |
| 					 saved_crash_message_count),
 | |
| 				&local_crash_msg_count, sizeof(u16));
 | |
| 	if (err) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to read message count\n");
 | |
| 		return err;
 | |
| 	}
 | |
| 	if (local_crash_msg_count != CONTROLVM_CRASHMSG_MAX) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"invalid number of messages\n");
 | |
| 		return -EIO;
 | |
| 	}
 | |
| 	err = visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 				offsetof(struct visor_controlvm_channel,
 | |
| 					 saved_crash_message_offset),
 | |
| 				&local_crash_msg_offset, sizeof(u32));
 | |
| 	if (err) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to read offset\n");
 | |
| 		return err;
 | |
| 	}
 | |
| 	switch (cr_type) {
 | |
| 	case CRASH_DEV:
 | |
| 		local_crash_msg_offset += sizeof(struct controlvm_message);
 | |
| 		err = visorchannel_write(chipset_dev->controlvm_channel,
 | |
| 					 local_crash_msg_offset, msg,
 | |
| 					 sizeof(struct controlvm_message));
 | |
| 		if (err) {
 | |
| 			dev_err(&chipset_dev->acpi_device->dev,
 | |
| 				"failed to write dev msg\n");
 | |
| 			return err;
 | |
| 		}
 | |
| 		break;
 | |
| 	case CRASH_BUS:
 | |
| 		err = visorchannel_write(chipset_dev->controlvm_channel,
 | |
| 					 local_crash_msg_offset, msg,
 | |
| 					 sizeof(struct controlvm_message));
 | |
| 		if (err) {
 | |
| 			dev_err(&chipset_dev->acpi_device->dev,
 | |
| 				"failed to write bus msg\n");
 | |
| 			return err;
 | |
| 		}
 | |
| 		break;
 | |
| 	default:
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"Invalid crash_obj_type\n");
 | |
| 		break;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int controlvm_responder(enum controlvm_id cmd_id,
 | |
| 			       struct controlvm_message_header *pending_msg_hdr,
 | |
| 			       int response)
 | |
| {
 | |
| 	if (pending_msg_hdr->id != (u32)cmd_id)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	return controlvm_respond(pending_msg_hdr, response, NULL);
 | |
| }
 | |
| 
 | |
| static int device_changestate_responder(enum controlvm_id cmd_id,
 | |
| 					struct visor_device *p, int response,
 | |
| 					struct visor_segment_state state)
 | |
| {
 | |
| 	struct controlvm_message outmsg;
 | |
| 
 | |
| 	if (p->pending_msg_hdr->id != cmd_id)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	controlvm_init_response(&outmsg, p->pending_msg_hdr, response);
 | |
| 	outmsg.cmd.device_change_state.bus_no = p->chipset_bus_no;
 | |
| 	outmsg.cmd.device_change_state.dev_no = p->chipset_dev_no;
 | |
| 	outmsg.cmd.device_change_state.state = state;
 | |
| 	return visorchannel_signalinsert(chipset_dev->controlvm_channel,
 | |
| 					 CONTROLVM_QUEUE_REQUEST, &outmsg);
 | |
| }
 | |
| 
 | |
| static int visorbus_create(struct controlvm_message *inmsg)
 | |
| {
 | |
| 	struct controlvm_message_packet *cmd = &inmsg->cmd;
 | |
| 	struct controlvm_message_header *pmsg_hdr;
 | |
| 	u32 bus_no = cmd->create_bus.bus_no;
 | |
| 	struct visor_device *bus_info;
 | |
| 	struct visorchannel *visorchannel;
 | |
| 	int err;
 | |
| 
 | |
| 	bus_info = visorbus_get_device_by_id(bus_no, BUS_ROOT_DEVICE, NULL);
 | |
| 	if (bus_info && bus_info->state.created == 1) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed %s: already exists\n", __func__);
 | |
| 		err = -EEXIST;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	bus_info = kzalloc(sizeof(*bus_info), GFP_KERNEL);
 | |
| 	if (!bus_info) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	INIT_LIST_HEAD(&bus_info->list_all);
 | |
| 	bus_info->chipset_bus_no = bus_no;
 | |
| 	bus_info->chipset_dev_no = BUS_ROOT_DEVICE;
 | |
| 	if (guid_equal(&cmd->create_bus.bus_inst_guid, &visor_siovm_guid)) {
 | |
| 		err = save_crash_message(inmsg, CRASH_BUS);
 | |
| 		if (err)
 | |
| 			goto err_free_bus_info;
 | |
| 	}
 | |
| 	if (inmsg->hdr.flags.response_expected == 1) {
 | |
| 		pmsg_hdr = kzalloc(sizeof(*pmsg_hdr), GFP_KERNEL);
 | |
| 		if (!pmsg_hdr) {
 | |
| 			err = -ENOMEM;
 | |
| 			goto err_free_bus_info;
 | |
| 		}
 | |
| 		memcpy(pmsg_hdr, &inmsg->hdr,
 | |
| 		       sizeof(struct controlvm_message_header));
 | |
| 		bus_info->pending_msg_hdr = pmsg_hdr;
 | |
| 	}
 | |
| 	visorchannel = visorchannel_create(cmd->create_bus.channel_addr,
 | |
| 					   GFP_KERNEL,
 | |
| 					   &cmd->create_bus.bus_data_type_guid,
 | |
| 					   false);
 | |
| 	if (!visorchannel) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto err_free_pending_msg;
 | |
| 	}
 | |
| 	bus_info->visorchannel = visorchannel;
 | |
| 	/* Response will be handled by visorbus_create_instance on success */
 | |
| 	err = visorbus_create_instance(bus_info);
 | |
| 	if (err)
 | |
| 		goto err_destroy_channel;
 | |
| 	return 0;
 | |
| 
 | |
| err_destroy_channel:
 | |
| 	visorchannel_destroy(visorchannel);
 | |
| 
 | |
| err_free_pending_msg:
 | |
| 	kfree(bus_info->pending_msg_hdr);
 | |
| 
 | |
| err_free_bus_info:
 | |
| 	kfree(bus_info);
 | |
| 
 | |
| err_respond:
 | |
| 	if (inmsg->hdr.flags.response_expected == 1)
 | |
| 		controlvm_responder(inmsg->hdr.id, &inmsg->hdr, err);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int visorbus_destroy(struct controlvm_message *inmsg)
 | |
| {
 | |
| 	struct controlvm_message_header *pmsg_hdr;
 | |
| 	u32 bus_no = inmsg->cmd.destroy_bus.bus_no;
 | |
| 	struct visor_device *bus_info;
 | |
| 	int err;
 | |
| 
 | |
| 	bus_info = visorbus_get_device_by_id(bus_no, BUS_ROOT_DEVICE, NULL);
 | |
| 	if (!bus_info) {
 | |
| 		err = -ENODEV;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (bus_info->state.created == 0) {
 | |
| 		err = -ENOENT;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (bus_info->pending_msg_hdr) {
 | |
| 		/* only non-NULL if dev is still waiting on a response */
 | |
| 		err = -EEXIST;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (inmsg->hdr.flags.response_expected == 1) {
 | |
| 		pmsg_hdr = kzalloc(sizeof(*pmsg_hdr), GFP_KERNEL);
 | |
| 		if (!pmsg_hdr) {
 | |
| 			err = -ENOMEM;
 | |
| 			goto err_respond;
 | |
| 		}
 | |
| 		memcpy(pmsg_hdr, &inmsg->hdr,
 | |
| 		       sizeof(struct controlvm_message_header));
 | |
| 		bus_info->pending_msg_hdr = pmsg_hdr;
 | |
| 	}
 | |
| 	/* Response will be handled by visorbus_remove_instance */
 | |
| 	visorbus_remove_instance(bus_info);
 | |
| 	return 0;
 | |
| 
 | |
| err_respond:
 | |
| 	if (inmsg->hdr.flags.response_expected == 1)
 | |
| 		controlvm_responder(inmsg->hdr.id, &inmsg->hdr, err);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static const guid_t *parser_id_get(struct parser_context *ctx)
 | |
| {
 | |
| 	return &ctx->data.id;
 | |
| }
 | |
| 
 | |
| static void *parser_string_get(u8 *pscan, int nscan)
 | |
| {
 | |
| 	int value_length;
 | |
| 	void *value;
 | |
| 
 | |
| 	if (nscan == 0)
 | |
| 		return NULL;
 | |
| 
 | |
| 	value_length = strnlen(pscan, nscan);
 | |
| 	value = kzalloc(value_length + 1, GFP_KERNEL);
 | |
| 	if (!value)
 | |
| 		return NULL;
 | |
| 	if (value_length > 0)
 | |
| 		memcpy(value, pscan, value_length);
 | |
| 	return value;
 | |
| }
 | |
| 
 | |
| static void *parser_name_get(struct parser_context *ctx)
 | |
| {
 | |
| 	struct visor_controlvm_parameters_header *phdr;
 | |
| 
 | |
| 	phdr = &ctx->data;
 | |
| 	if ((unsigned long)phdr->name_offset +
 | |
| 	    (unsigned long)phdr->name_length > ctx->param_bytes)
 | |
| 		return NULL;
 | |
| 	ctx->curr = (char *)&phdr + phdr->name_offset;
 | |
| 	ctx->bytes_remaining = phdr->name_length;
 | |
| 	return parser_string_get(ctx->curr, phdr->name_length);
 | |
| }
 | |
| 
 | |
| static int visorbus_configure(struct controlvm_message *inmsg,
 | |
| 			      struct parser_context *parser_ctx)
 | |
| {
 | |
| 	struct controlvm_message_packet *cmd = &inmsg->cmd;
 | |
| 	u32 bus_no;
 | |
| 	struct visor_device *bus_info;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	bus_no = cmd->configure_bus.bus_no;
 | |
| 	bus_info = visorbus_get_device_by_id(bus_no, BUS_ROOT_DEVICE, NULL);
 | |
| 	if (!bus_info) {
 | |
| 		err = -EINVAL;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (bus_info->state.created == 0) {
 | |
| 		err = -EINVAL;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (bus_info->pending_msg_hdr) {
 | |
| 		err = -EIO;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	err = visorchannel_set_clientpartition(bus_info->visorchannel,
 | |
| 					       cmd->configure_bus.guest_handle);
 | |
| 	if (err)
 | |
| 		goto err_respond;
 | |
| 	if (parser_ctx) {
 | |
| 		const guid_t *partition_guid = parser_id_get(parser_ctx);
 | |
| 
 | |
| 		guid_copy(&bus_info->partition_guid, partition_guid);
 | |
| 		bus_info->name = parser_name_get(parser_ctx);
 | |
| 	}
 | |
| 	if (inmsg->hdr.flags.response_expected == 1)
 | |
| 		controlvm_responder(inmsg->hdr.id, &inmsg->hdr, err);
 | |
| 	return 0;
 | |
| 
 | |
| err_respond:
 | |
| 	dev_err(&chipset_dev->acpi_device->dev,
 | |
| 		"%s exited with err: %d\n", __func__, err);
 | |
| 	if (inmsg->hdr.flags.response_expected == 1)
 | |
| 		controlvm_responder(inmsg->hdr.id, &inmsg->hdr, err);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int visorbus_device_create(struct controlvm_message *inmsg)
 | |
| {
 | |
| 	struct controlvm_message_packet *cmd = &inmsg->cmd;
 | |
| 	struct controlvm_message_header *pmsg_hdr;
 | |
| 	u32 bus_no = cmd->create_device.bus_no;
 | |
| 	u32 dev_no = cmd->create_device.dev_no;
 | |
| 	struct visor_device *dev_info;
 | |
| 	struct visor_device *bus_info;
 | |
| 	struct visorchannel *visorchannel;
 | |
| 	int err;
 | |
| 
 | |
| 	bus_info = visorbus_get_device_by_id(bus_no, BUS_ROOT_DEVICE, NULL);
 | |
| 	if (!bus_info) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to get bus by id: %d\n", bus_no);
 | |
| 		err = -ENODEV;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (bus_info->state.created == 0) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"bus not created, id: %d\n", bus_no);
 | |
| 		err = -EINVAL;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	dev_info = visorbus_get_device_by_id(bus_no, dev_no, NULL);
 | |
| 	if (dev_info && dev_info->state.created == 1) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to get bus by id: %d/%d\n", bus_no, dev_no);
 | |
| 		err = -EEXIST;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 
 | |
| 	dev_info = kzalloc(sizeof(*dev_info), GFP_KERNEL);
 | |
| 	if (!dev_info) {
 | |
| 		err = -ENOMEM;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	dev_info->chipset_bus_no = bus_no;
 | |
| 	dev_info->chipset_dev_no = dev_no;
 | |
| 	guid_copy(&dev_info->inst, &cmd->create_device.dev_inst_guid);
 | |
| 	dev_info->device.parent = &bus_info->device;
 | |
| 	visorchannel = visorchannel_create(cmd->create_device.channel_addr,
 | |
| 					   GFP_KERNEL,
 | |
| 					   &cmd->create_device.data_type_guid,
 | |
| 					   true);
 | |
| 	if (!visorchannel) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to create visorchannel: %d/%d\n",
 | |
| 			bus_no, dev_no);
 | |
| 		err = -ENOMEM;
 | |
| 		goto err_free_dev_info;
 | |
| 	}
 | |
| 	dev_info->visorchannel = visorchannel;
 | |
| 	guid_copy(&dev_info->channel_type_guid,
 | |
| 		  &cmd->create_device.data_type_guid);
 | |
| 	if (guid_equal(&cmd->create_device.data_type_guid,
 | |
| 		       &visor_vhba_channel_guid)) {
 | |
| 		err = save_crash_message(inmsg, CRASH_DEV);
 | |
| 		if (err)
 | |
| 			goto err_destroy_visorchannel;
 | |
| 	}
 | |
| 	if (inmsg->hdr.flags.response_expected == 1) {
 | |
| 		pmsg_hdr = kzalloc(sizeof(*pmsg_hdr), GFP_KERNEL);
 | |
| 		if (!pmsg_hdr) {
 | |
| 			err = -ENOMEM;
 | |
| 			goto err_destroy_visorchannel;
 | |
| 		}
 | |
| 		memcpy(pmsg_hdr, &inmsg->hdr,
 | |
| 		       sizeof(struct controlvm_message_header));
 | |
| 		dev_info->pending_msg_hdr = pmsg_hdr;
 | |
| 	}
 | |
| 	/* create_visor_device will send response */
 | |
| 	err = create_visor_device(dev_info);
 | |
| 	if (err)
 | |
| 		goto err_destroy_visorchannel;
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| err_destroy_visorchannel:
 | |
| 	visorchannel_destroy(visorchannel);
 | |
| 
 | |
| err_free_dev_info:
 | |
| 	kfree(dev_info);
 | |
| 
 | |
| err_respond:
 | |
| 	if (inmsg->hdr.flags.response_expected == 1)
 | |
| 		controlvm_responder(inmsg->hdr.id, &inmsg->hdr, err);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int visorbus_device_changestate(struct controlvm_message *inmsg)
 | |
| {
 | |
| 	struct controlvm_message_packet *cmd = &inmsg->cmd;
 | |
| 	struct controlvm_message_header *pmsg_hdr;
 | |
| 	u32 bus_no = cmd->device_change_state.bus_no;
 | |
| 	u32 dev_no = cmd->device_change_state.dev_no;
 | |
| 	struct visor_segment_state state = cmd->device_change_state.state;
 | |
| 	struct visor_device *dev_info;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	dev_info = visorbus_get_device_by_id(bus_no, dev_no, NULL);
 | |
| 	if (!dev_info) {
 | |
| 		err = -ENODEV;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (dev_info->state.created == 0) {
 | |
| 		err = -EINVAL;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (dev_info->pending_msg_hdr) {
 | |
| 		/* only non-NULL if dev is still waiting on a response */
 | |
| 		err = -EIO;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 
 | |
| 	if (inmsg->hdr.flags.response_expected == 1) {
 | |
| 		pmsg_hdr = kzalloc(sizeof(*pmsg_hdr), GFP_KERNEL);
 | |
| 		if (!pmsg_hdr) {
 | |
| 			err = -ENOMEM;
 | |
| 			goto err_respond;
 | |
| 		}
 | |
| 		memcpy(pmsg_hdr, &inmsg->hdr,
 | |
| 		       sizeof(struct controlvm_message_header));
 | |
| 		dev_info->pending_msg_hdr = pmsg_hdr;
 | |
| 	}
 | |
| 	if (state.alive == segment_state_running.alive &&
 | |
| 	    state.operating == segment_state_running.operating)
 | |
| 		/* Response will be sent from visorchipset_device_resume */
 | |
| 		err = visorchipset_device_resume(dev_info);
 | |
| 	/* ServerNotReady / ServerLost / SegmentStateStandby */
 | |
| 	else if (state.alive == segment_state_standby.alive &&
 | |
| 		 state.operating == segment_state_standby.operating)
 | |
| 		/*
 | |
| 		 * technically this is standby case where server is lost.
 | |
| 		 * Response will be sent from visorchipset_device_pause.
 | |
| 		 */
 | |
| 		err = visorchipset_device_pause(dev_info);
 | |
| 	if (err)
 | |
| 		goto err_respond;
 | |
| 	return 0;
 | |
| 
 | |
| err_respond:
 | |
| 	dev_err(&chipset_dev->acpi_device->dev, "failed: %d\n", err);
 | |
| 	if (inmsg->hdr.flags.response_expected == 1)
 | |
| 		controlvm_responder(inmsg->hdr.id, &inmsg->hdr, err);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int visorbus_device_destroy(struct controlvm_message *inmsg)
 | |
| {
 | |
| 	struct controlvm_message_packet *cmd = &inmsg->cmd;
 | |
| 	struct controlvm_message_header *pmsg_hdr;
 | |
| 	u32 bus_no = cmd->destroy_device.bus_no;
 | |
| 	u32 dev_no = cmd->destroy_device.dev_no;
 | |
| 	struct visor_device *dev_info;
 | |
| 	int err;
 | |
| 
 | |
| 	dev_info = visorbus_get_device_by_id(bus_no, dev_no, NULL);
 | |
| 	if (!dev_info) {
 | |
| 		err = -ENODEV;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (dev_info->state.created == 0) {
 | |
| 		err = -EINVAL;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (dev_info->pending_msg_hdr) {
 | |
| 		/* only non-NULL if dev is still waiting on a response */
 | |
| 		err = -EIO;
 | |
| 		goto err_respond;
 | |
| 	}
 | |
| 	if (inmsg->hdr.flags.response_expected == 1) {
 | |
| 		pmsg_hdr = kzalloc(sizeof(*pmsg_hdr), GFP_KERNEL);
 | |
| 		if (!pmsg_hdr) {
 | |
| 			err = -ENOMEM;
 | |
| 			goto err_respond;
 | |
| 		}
 | |
| 
 | |
| 		memcpy(pmsg_hdr, &inmsg->hdr,
 | |
| 		       sizeof(struct controlvm_message_header));
 | |
| 		dev_info->pending_msg_hdr = pmsg_hdr;
 | |
| 	}
 | |
| 	kfree(dev_info->name);
 | |
| 	remove_visor_device(dev_info);
 | |
| 	return 0;
 | |
| 
 | |
| err_respond:
 | |
| 	if (inmsg->hdr.flags.response_expected == 1)
 | |
| 		controlvm_responder(inmsg->hdr.id, &inmsg->hdr, err);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * The general parahotplug flow works as follows. The visorchipset receives
 | |
|  * a DEVICE_CHANGESTATE message from Command specifying a physical device
 | |
|  * to enable or disable. The CONTROLVM message handler calls
 | |
|  * parahotplug_process_message, which then adds the message to a global list
 | |
|  * and kicks off a udev event which causes a user level script to enable or
 | |
|  * disable the specified device. The udev script then writes to
 | |
|  * /sys/devices/platform/visorchipset/parahotplug, which causes the
 | |
|  * parahotplug store functions to get called, at which point the
 | |
|  * appropriate CONTROLVM message is retrieved from the list and responded to.
 | |
|  */
 | |
| 
 | |
| #define PARAHOTPLUG_TIMEOUT_MS 2000
 | |
| 
 | |
| /*
 | |
|  * parahotplug_next_id() - generate unique int to match an outstanding
 | |
|  *                         CONTROLVM message with a udev script /sys
 | |
|  *                         response
 | |
|  *
 | |
|  * Return: a unique integer value
 | |
|  */
 | |
| static int parahotplug_next_id(void)
 | |
| {
 | |
| 	static atomic_t id = ATOMIC_INIT(0);
 | |
| 
 | |
| 	return atomic_inc_return(&id);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * parahotplug_next_expiration() - returns the time (in jiffies) when a
 | |
|  *                                 CONTROLVM message on the list should expire
 | |
|  *                                 -- PARAHOTPLUG_TIMEOUT_MS in the future
 | |
|  *
 | |
|  * Return: expected expiration time (in jiffies)
 | |
|  */
 | |
| static unsigned long parahotplug_next_expiration(void)
 | |
| {
 | |
| 	return jiffies + msecs_to_jiffies(PARAHOTPLUG_TIMEOUT_MS);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * parahotplug_request_create() - create a parahotplug_request, which is
 | |
|  *                                basically a wrapper for a CONTROLVM_MESSAGE
 | |
|  *                                that we can stick on a list
 | |
|  * @msg: the message to insert in the request
 | |
|  *
 | |
|  * Return: the request containing the provided message
 | |
|  */
 | |
| static struct parahotplug_request *parahotplug_request_create(
 | |
| 						struct controlvm_message *msg)
 | |
| {
 | |
| 	struct parahotplug_request *req;
 | |
| 
 | |
| 	req = kmalloc(sizeof(*req), GFP_KERNEL);
 | |
| 	if (!req)
 | |
| 		return NULL;
 | |
| 	req->id = parahotplug_next_id();
 | |
| 	req->expiration = parahotplug_next_expiration();
 | |
| 	req->msg = *msg;
 | |
| 	return req;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * parahotplug_request_destroy() - free a parahotplug_request
 | |
|  * @req: the request to deallocate
 | |
|  */
 | |
| static void parahotplug_request_destroy(struct parahotplug_request *req)
 | |
| {
 | |
| 	kfree(req);
 | |
| }
 | |
| 
 | |
| static LIST_HEAD(parahotplug_request_list);
 | |
| /* lock for above */
 | |
| static DEFINE_SPINLOCK(parahotplug_request_list_lock);
 | |
| 
 | |
| /*
 | |
|  * parahotplug_request_complete() - mark request as complete
 | |
|  * @id:     the id of the request
 | |
|  * @active: indicates whether the request is assigned to active partition
 | |
|  *
 | |
|  * Called from the /sys handler, which means the user script has
 | |
|  * finished the enable/disable. Find the matching identifier, and
 | |
|  * respond to the CONTROLVM message with success.
 | |
|  *
 | |
|  * Return: 0 on success or -EINVAL on failure
 | |
|  */
 | |
| static int parahotplug_request_complete(int id, u16 active)
 | |
| {
 | |
| 	struct list_head *pos;
 | |
| 	struct list_head *tmp;
 | |
| 	struct parahotplug_request *req;
 | |
| 
 | |
| 	spin_lock(¶hotplug_request_list_lock);
 | |
| 	/* Look for a request matching "id". */
 | |
| 	list_for_each_safe(pos, tmp, ¶hotplug_request_list) {
 | |
| 		req = list_entry(pos, struct parahotplug_request, list);
 | |
| 		if (req->id == id) {
 | |
| 			/*
 | |
| 			 * Found a match. Remove it from the list and
 | |
| 			 * respond.
 | |
| 			 */
 | |
| 			list_del(pos);
 | |
| 			spin_unlock(¶hotplug_request_list_lock);
 | |
| 			req->msg.cmd.device_change_state.state.active = active;
 | |
| 			if (req->msg.hdr.flags.response_expected)
 | |
| 				controlvm_respond(
 | |
| 				       &req->msg.hdr, CONTROLVM_RESP_SUCCESS,
 | |
| 				       &req->msg.cmd.device_change_state.state);
 | |
| 			parahotplug_request_destroy(req);
 | |
| 			return 0;
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock(¶hotplug_request_list_lock);
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * devicedisabled_store() - disables the hotplug device
 | |
|  * @dev:   sysfs interface variable not utilized in this function
 | |
|  * @attr:  sysfs interface variable not utilized in this function
 | |
|  * @buf:   buffer containing the device id
 | |
|  * @count: the size of the buffer
 | |
|  *
 | |
|  * The parahotplug/devicedisabled interface gets called by our support script
 | |
|  * when an SR-IOV device has been shut down. The ID is passed to the script
 | |
|  * and then passed back when the device has been removed.
 | |
|  *
 | |
|  * Return: the size of the buffer for success or negative for error
 | |
|  */
 | |
| static ssize_t devicedisabled_store(struct device *dev,
 | |
| 				    struct device_attribute *attr,
 | |
| 				    const char *buf, size_t count)
 | |
| {
 | |
| 	unsigned int id;
 | |
| 	int err;
 | |
| 
 | |
| 	if (kstrtouint(buf, 10, &id))
 | |
| 		return -EINVAL;
 | |
| 	err = parahotplug_request_complete(id, 0);
 | |
| 	if (err < 0)
 | |
| 		return err;
 | |
| 	return count;
 | |
| }
 | |
| static DEVICE_ATTR_WO(devicedisabled);
 | |
| 
 | |
| /*
 | |
|  * deviceenabled_store() - enables the hotplug device
 | |
|  * @dev:   sysfs interface variable not utilized in this function
 | |
|  * @attr:  sysfs interface variable not utilized in this function
 | |
|  * @buf:   buffer containing the device id
 | |
|  * @count: the size of the buffer
 | |
|  *
 | |
|  * The parahotplug/deviceenabled interface gets called by our support script
 | |
|  * when an SR-IOV device has been recovered. The ID is passed to the script
 | |
|  * and then passed back when the device has been brought back up.
 | |
|  *
 | |
|  * Return: the size of the buffer for success or negative for error
 | |
|  */
 | |
| static ssize_t deviceenabled_store(struct device *dev,
 | |
| 				   struct device_attribute *attr,
 | |
| 				   const char *buf, size_t count)
 | |
| {
 | |
| 	unsigned int id;
 | |
| 
 | |
| 	if (kstrtouint(buf, 10, &id))
 | |
| 		return -EINVAL;
 | |
| 	parahotplug_request_complete(id, 1);
 | |
| 	return count;
 | |
| }
 | |
| static DEVICE_ATTR_WO(deviceenabled);
 | |
| 
 | |
| static struct attribute *visorchipset_install_attrs[] = {
 | |
| 	&dev_attr_toolaction.attr,
 | |
| 	&dev_attr_boottotool.attr,
 | |
| 	&dev_attr_error.attr,
 | |
| 	&dev_attr_textid.attr,
 | |
| 	&dev_attr_remaining_steps.attr,
 | |
| 	NULL
 | |
| };
 | |
| 
 | |
| static const struct attribute_group visorchipset_install_group = {
 | |
| 	.name = "install",
 | |
| 	.attrs = visorchipset_install_attrs
 | |
| };
 | |
| 
 | |
| static struct attribute *visorchipset_parahotplug_attrs[] = {
 | |
| 	&dev_attr_devicedisabled.attr,
 | |
| 	&dev_attr_deviceenabled.attr,
 | |
| 	NULL
 | |
| };
 | |
| 
 | |
| static const struct attribute_group visorchipset_parahotplug_group = {
 | |
| 	.name = "parahotplug",
 | |
| 	.attrs = visorchipset_parahotplug_attrs
 | |
| };
 | |
| 
 | |
| static const struct attribute_group *visorchipset_dev_groups[] = {
 | |
| 	&visorchipset_install_group,
 | |
| 	&visorchipset_parahotplug_group,
 | |
| 	NULL
 | |
| };
 | |
| 
 | |
| /*
 | |
|  * parahotplug_request_kickoff() - initiate parahotplug request
 | |
|  * @req: the request to initiate
 | |
|  *
 | |
|  * Cause uevent to run the user level script to do the disable/enable specified
 | |
|  * in the parahotplug_request.
 | |
|  */
 | |
| static int parahotplug_request_kickoff(struct parahotplug_request *req)
 | |
| {
 | |
| 	struct controlvm_message_packet *cmd = &req->msg.cmd;
 | |
| 	char env_cmd[40], env_id[40], env_state[40], env_bus[40], env_dev[40],
 | |
| 	     env_func[40];
 | |
| 	char *envp[] = { env_cmd, env_id, env_state, env_bus, env_dev,
 | |
| 			 env_func, NULL
 | |
| 	};
 | |
| 
 | |
| 	sprintf(env_cmd, "VISOR_PARAHOTPLUG=1");
 | |
| 	sprintf(env_id, "VISOR_PARAHOTPLUG_ID=%d", req->id);
 | |
| 	sprintf(env_state, "VISOR_PARAHOTPLUG_STATE=%d",
 | |
| 		cmd->device_change_state.state.active);
 | |
| 	sprintf(env_bus, "VISOR_PARAHOTPLUG_BUS=%d",
 | |
| 		cmd->device_change_state.bus_no);
 | |
| 	sprintf(env_dev, "VISOR_PARAHOTPLUG_DEVICE=%d",
 | |
| 		cmd->device_change_state.dev_no >> 3);
 | |
| 	sprintf(env_func, "VISOR_PARAHOTPLUG_FUNCTION=%d",
 | |
| 		cmd->device_change_state.dev_no & 0x7);
 | |
| 	return kobject_uevent_env(&chipset_dev->acpi_device->dev.kobj,
 | |
| 				  KOBJ_CHANGE, envp);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * parahotplug_process_message() - enables or disables a PCI device by kicking
 | |
|  *                                 off a udev script
 | |
|  * @inmsg: the message indicating whether to enable or disable
 | |
|  */
 | |
| static int parahotplug_process_message(struct controlvm_message *inmsg)
 | |
| {
 | |
| 	struct parahotplug_request *req;
 | |
| 	int err;
 | |
| 
 | |
| 	req = parahotplug_request_create(inmsg);
 | |
| 	if (!req)
 | |
| 		return -ENOMEM;
 | |
| 	/*
 | |
| 	 * For enable messages, just respond with success right away, we don't
 | |
| 	 * need to wait to see if the enable was successful.
 | |
| 	 */
 | |
| 	if (inmsg->cmd.device_change_state.state.active) {
 | |
| 		err = parahotplug_request_kickoff(req);
 | |
| 		if (err)
 | |
| 			goto err_respond;
 | |
| 		controlvm_respond(&inmsg->hdr, CONTROLVM_RESP_SUCCESS,
 | |
| 				  &inmsg->cmd.device_change_state.state);
 | |
| 		parahotplug_request_destroy(req);
 | |
| 		return 0;
 | |
| 	}
 | |
| 	/*
 | |
| 	 * For disable messages, add the request to the request list before
 | |
| 	 * kicking off the udev script. It won't get responded to until the
 | |
| 	 * script has indicated it's done.
 | |
| 	 */
 | |
| 	spin_lock(¶hotplug_request_list_lock);
 | |
| 	list_add_tail(&req->list, ¶hotplug_request_list);
 | |
| 	spin_unlock(¶hotplug_request_list_lock);
 | |
| 	err = parahotplug_request_kickoff(req);
 | |
| 	if (err)
 | |
| 		goto err_respond;
 | |
| 	return 0;
 | |
| 
 | |
| err_respond:
 | |
| 	controlvm_respond(&inmsg->hdr, err,
 | |
| 			  &inmsg->cmd.device_change_state.state);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * chipset_ready_uevent() - sends chipset_ready action
 | |
|  *
 | |
|  * Send ACTION=online for DEVPATH=/sys/devices/platform/visorchipset.
 | |
|  *
 | |
|  * Return: 0 on success, negative on failure
 | |
|  */
 | |
| static int chipset_ready_uevent(struct controlvm_message_header *msg_hdr)
 | |
| {
 | |
| 	int res;
 | |
| 
 | |
| 	res = kobject_uevent(&chipset_dev->acpi_device->dev.kobj, KOBJ_ONLINE);
 | |
| 	if (msg_hdr->flags.response_expected)
 | |
| 		controlvm_respond(msg_hdr, res, NULL);
 | |
| 	return res;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * chipset_selftest_uevent() - sends chipset_selftest action
 | |
|  *
 | |
|  * Send ACTION=online for DEVPATH=/sys/devices/platform/visorchipset.
 | |
|  *
 | |
|  * Return: 0 on success, negative on failure
 | |
|  */
 | |
| static int chipset_selftest_uevent(struct controlvm_message_header *msg_hdr)
 | |
| {
 | |
| 	char env_selftest[20];
 | |
| 	char *envp[] = { env_selftest, NULL };
 | |
| 	int res;
 | |
| 
 | |
| 	sprintf(env_selftest, "SPARSP_SELFTEST=%d", 1);
 | |
| 	res = kobject_uevent_env(&chipset_dev->acpi_device->dev.kobj,
 | |
| 				 KOBJ_CHANGE, envp);
 | |
| 	if (msg_hdr->flags.response_expected)
 | |
| 		controlvm_respond(msg_hdr, res, NULL);
 | |
| 	return res;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * chipset_notready_uevent() - sends chipset_notready action
 | |
|  *
 | |
|  * Send ACTION=offline for DEVPATH=/sys/devices/platform/visorchipset.
 | |
|  *
 | |
|  * Return: 0 on success, negative on failure
 | |
|  */
 | |
| static int chipset_notready_uevent(struct controlvm_message_header *msg_hdr)
 | |
| {
 | |
| 	int res = kobject_uevent(&chipset_dev->acpi_device->dev.kobj,
 | |
| 				 KOBJ_OFFLINE);
 | |
| 
 | |
| 	if (msg_hdr->flags.response_expected)
 | |
| 		controlvm_respond(msg_hdr, res, NULL);
 | |
| 	return res;
 | |
| }
 | |
| 
 | |
| static int unisys_vmcall(unsigned long tuple, unsigned long param)
 | |
| {
 | |
| 	int result = 0;
 | |
| 	unsigned int cpuid_eax, cpuid_ebx, cpuid_ecx, cpuid_edx;
 | |
| 	unsigned long reg_ebx;
 | |
| 	unsigned long reg_ecx;
 | |
| 
 | |
| 	reg_ebx = param & 0xFFFFFFFF;
 | |
| 	reg_ecx = param >> 32;
 | |
| 	cpuid(0x00000001, &cpuid_eax, &cpuid_ebx, &cpuid_ecx, &cpuid_edx);
 | |
| 	if (!(cpuid_ecx & 0x80000000))
 | |
| 		return -EPERM;
 | |
| 	__asm__ __volatile__(".byte 0x00f, 0x001, 0x0c1" : "=a"(result) :
 | |
| 			     "a"(tuple), "b"(reg_ebx), "c"(reg_ecx));
 | |
| 	if (result)
 | |
| 		goto error;
 | |
| 	return 0;
 | |
| 
 | |
| /* Need to convert from VMCALL error codes to Linux */
 | |
| error:
 | |
| 	switch (result) {
 | |
| 	case VMCALL_RESULT_INVALID_PARAM:
 | |
| 		return -EINVAL;
 | |
| 	case VMCALL_RESULT_DATA_UNAVAILABLE:
 | |
| 		return -ENODEV;
 | |
| 	default:
 | |
| 		return -EFAULT;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int controlvm_channel_create(struct visorchipset_device *dev)
 | |
| {
 | |
| 	struct visorchannel *chan;
 | |
| 	u64 addr;
 | |
| 	int err;
 | |
| 
 | |
| 	err = unisys_vmcall(VMCALL_CONTROLVM_ADDR,
 | |
| 			    virt_to_phys(&dev->controlvm_params));
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	addr = dev->controlvm_params.address;
 | |
| 	chan = visorchannel_create(addr, GFP_KERNEL,
 | |
| 				   &visor_controlvm_channel_guid, true);
 | |
| 	if (!chan)
 | |
| 		return -ENOMEM;
 | |
| 	dev->controlvm_channel = chan;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void setup_crash_devices_work_queue(struct work_struct *work)
 | |
| {
 | |
| 	struct controlvm_message local_crash_bus_msg;
 | |
| 	struct controlvm_message local_crash_dev_msg;
 | |
| 	struct controlvm_message msg;
 | |
| 	u32 local_crash_msg_offset;
 | |
| 	u16 local_crash_msg_count;
 | |
| 
 | |
| 	/* send init chipset msg */
 | |
| 	msg.hdr.id = CONTROLVM_CHIPSET_INIT;
 | |
| 	msg.cmd.init_chipset.bus_count = 23;
 | |
| 	msg.cmd.init_chipset.switch_count = 0;
 | |
| 	chipset_init(&msg);
 | |
| 	/* get saved message count */
 | |
| 	if (visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 			      offsetof(struct visor_controlvm_channel,
 | |
| 				       saved_crash_message_count),
 | |
| 			      &local_crash_msg_count, sizeof(u16)) < 0) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to read channel\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	if (local_crash_msg_count != CONTROLVM_CRASHMSG_MAX) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev, "invalid count\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	/* get saved crash message offset */
 | |
| 	if (visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 			      offsetof(struct visor_controlvm_channel,
 | |
| 				       saved_crash_message_offset),
 | |
| 			      &local_crash_msg_offset, sizeof(u32)) < 0) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to read channel\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	/* read create device message for storage bus offset */
 | |
| 	if (visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 			      local_crash_msg_offset,
 | |
| 			      &local_crash_bus_msg,
 | |
| 			      sizeof(struct controlvm_message)) < 0) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to read channel\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	/* read create device message for storage device */
 | |
| 	if (visorchannel_read(chipset_dev->controlvm_channel,
 | |
| 			      local_crash_msg_offset +
 | |
| 			      sizeof(struct controlvm_message),
 | |
| 			      &local_crash_dev_msg,
 | |
| 			      sizeof(struct controlvm_message)) < 0) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"failed to read channel\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	/* reuse IOVM create bus message */
 | |
| 	if (!local_crash_bus_msg.cmd.create_bus.channel_addr) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"no valid create_bus message\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	visorbus_create(&local_crash_bus_msg);
 | |
| 	/* reuse create device message for storage device */
 | |
| 	if (!local_crash_dev_msg.cmd.create_device.channel_addr) {
 | |
| 		dev_err(&chipset_dev->acpi_device->dev,
 | |
| 			"no valid create_device message\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	visorbus_device_create(&local_crash_dev_msg);
 | |
| }
 | |
| 
 | |
| void visorbus_response(struct visor_device *bus_info, int response,
 | |
| 		       int controlvm_id)
 | |
| {
 | |
| 	if (!bus_info->pending_msg_hdr)
 | |
| 		return;
 | |
| 
 | |
| 	controlvm_responder(controlvm_id, bus_info->pending_msg_hdr, response);
 | |
| 	kfree(bus_info->pending_msg_hdr);
 | |
| 	bus_info->pending_msg_hdr = NULL;
 | |
| }
 | |
| 
 | |
| void visorbus_device_changestate_response(struct visor_device *dev_info,
 | |
| 					  int response,
 | |
| 					  struct visor_segment_state state)
 | |
| {
 | |
| 	if (!dev_info->pending_msg_hdr)
 | |
| 		return;
 | |
| 
 | |
| 	device_changestate_responder(CONTROLVM_DEVICE_CHANGESTATE, dev_info,
 | |
| 				     response, state);
 | |
| 	kfree(dev_info->pending_msg_hdr);
 | |
| 	dev_info->pending_msg_hdr = NULL;
 | |
| }
 | |
| 
 | |
| static void parser_done(struct parser_context *ctx)
 | |
| {
 | |
| 	chipset_dev->controlvm_payload_bytes_buffered -= ctx->param_bytes;
 | |
| 	kfree(ctx);
 | |
| }
 | |
| 
 | |
| static struct parser_context *parser_init_stream(u64 addr, u32 bytes,
 | |
| 						 bool *retry)
 | |
| {
 | |
| 	unsigned long allocbytes;
 | |
| 	struct parser_context *ctx;
 | |
| 	void *mapping;
 | |
| 
 | |
| 	*retry = false;
 | |
| 	/* alloc an extra byte to ensure payload is \0 terminated */
 | |
| 	allocbytes = (unsigned long)bytes + 1 + (sizeof(struct parser_context) -
 | |
| 		     sizeof(struct visor_controlvm_parameters_header));
 | |
| 	if ((chipset_dev->controlvm_payload_bytes_buffered + bytes) >
 | |
| 	     MAX_CONTROLVM_PAYLOAD_BYTES) {
 | |
| 		*retry = true;
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	ctx = kzalloc(allocbytes, GFP_KERNEL);
 | |
| 	if (!ctx) {
 | |
| 		*retry = true;
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	ctx->allocbytes = allocbytes;
 | |
| 	ctx->param_bytes = bytes;
 | |
| 	mapping = memremap(addr, bytes, MEMREMAP_WB);
 | |
| 	if (!mapping)
 | |
| 		goto err_finish_ctx;
 | |
| 	memcpy(&ctx->data, mapping, bytes);
 | |
| 	memunmap(mapping);
 | |
| 	ctx->byte_stream = true;
 | |
| 	chipset_dev->controlvm_payload_bytes_buffered += ctx->param_bytes;
 | |
| 	return ctx;
 | |
| 
 | |
| err_finish_ctx:
 | |
| 	kfree(ctx);
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * handle_command() - process a controlvm message
 | |
|  * @inmsg:        the message to process
 | |
|  * @channel_addr: address of the controlvm channel
 | |
|  *
 | |
|  * Return:
 | |
|  *	0	- Successfully processed the message
 | |
|  *	-EAGAIN - ControlVM message was not processed and should be retried
 | |
|  *		  reading the next controlvm message; a scenario where this can
 | |
|  *		  occur is when we need to throttle the allocation of memory in
 | |
|  *		  which to copy out controlvm payload data.
 | |
|  *	< 0	- error: ControlVM message was processed but an error occurred.
 | |
|  */
 | |
| static int handle_command(struct controlvm_message inmsg, u64 channel_addr)
 | |
| {
 | |
| 	struct controlvm_message_packet *cmd = &inmsg.cmd;
 | |
| 	u64 parm_addr;
 | |
| 	u32 parm_bytes;
 | |
| 	struct parser_context *parser_ctx = NULL;
 | |
| 	struct controlvm_message ackmsg;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	/* create parsing context if necessary */
 | |
| 	parm_addr = channel_addr + inmsg.hdr.payload_vm_offset;
 | |
| 	parm_bytes = inmsg.hdr.payload_bytes;
 | |
| 	/*
 | |
| 	 * Parameter and channel addresses within test messages actually lie
 | |
| 	 * within our OS-controlled memory. We need to know that, because it
 | |
| 	 * makes a difference in how we compute the virtual address.
 | |
| 	 */
 | |
| 	if (parm_bytes) {
 | |
| 		bool retry;
 | |
| 
 | |
| 		parser_ctx = parser_init_stream(parm_addr, parm_bytes, &retry);
 | |
| 		if (!parser_ctx && retry)
 | |
| 			return -EAGAIN;
 | |
| 	}
 | |
| 	controlvm_init_response(&ackmsg, &inmsg.hdr, CONTROLVM_RESP_SUCCESS);
 | |
| 	err = visorchannel_signalinsert(chipset_dev->controlvm_channel,
 | |
| 					CONTROLVM_QUEUE_ACK, &ackmsg);
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	switch (inmsg.hdr.id) {
 | |
| 	case CONTROLVM_CHIPSET_INIT:
 | |
| 		err = chipset_init(&inmsg);
 | |
| 		break;
 | |
| 	case CONTROLVM_BUS_CREATE:
 | |
| 		err = visorbus_create(&inmsg);
 | |
| 		break;
 | |
| 	case CONTROLVM_BUS_DESTROY:
 | |
| 		err = visorbus_destroy(&inmsg);
 | |
| 		break;
 | |
| 	case CONTROLVM_BUS_CONFIGURE:
 | |
| 		err = visorbus_configure(&inmsg, parser_ctx);
 | |
| 		break;
 | |
| 	case CONTROLVM_DEVICE_CREATE:
 | |
| 		err = visorbus_device_create(&inmsg);
 | |
| 		break;
 | |
| 	case CONTROLVM_DEVICE_CHANGESTATE:
 | |
| 		if (cmd->device_change_state.flags.phys_device) {
 | |
| 			err = parahotplug_process_message(&inmsg);
 | |
| 		} else {
 | |
| 			/*
 | |
| 			 * save the hdr and cmd structures for later use when
 | |
| 			 * sending back the response to Command
 | |
| 			 */
 | |
| 			err = visorbus_device_changestate(&inmsg);
 | |
| 			break;
 | |
| 		}
 | |
| 		break;
 | |
| 	case CONTROLVM_DEVICE_DESTROY:
 | |
| 		err = visorbus_device_destroy(&inmsg);
 | |
| 		break;
 | |
| 	case CONTROLVM_DEVICE_CONFIGURE:
 | |
| 		/* no op just send a respond that we passed */
 | |
| 		if (inmsg.hdr.flags.response_expected)
 | |
| 			controlvm_respond(&inmsg.hdr, CONTROLVM_RESP_SUCCESS,
 | |
| 					  NULL);
 | |
| 		break;
 | |
| 	case CONTROLVM_CHIPSET_READY:
 | |
| 		err = chipset_ready_uevent(&inmsg.hdr);
 | |
| 		break;
 | |
| 	case CONTROLVM_CHIPSET_SELFTEST:
 | |
| 		err = chipset_selftest_uevent(&inmsg.hdr);
 | |
| 		break;
 | |
| 	case CONTROLVM_CHIPSET_STOP:
 | |
| 		err = chipset_notready_uevent(&inmsg.hdr);
 | |
| 		break;
 | |
| 	default:
 | |
| 		err = -ENOMSG;
 | |
| 		if (inmsg.hdr.flags.response_expected)
 | |
| 			controlvm_respond(&inmsg.hdr,
 | |
| 					  -CONTROLVM_RESP_ID_UNKNOWN, NULL);
 | |
| 		break;
 | |
| 	}
 | |
| 	if (parser_ctx) {
 | |
| 		parser_done(parser_ctx);
 | |
| 		parser_ctx = NULL;
 | |
| 	}
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * read_controlvm_event() - retreives the next message from the
 | |
|  *                          CONTROLVM_QUEUE_EVENT queue in the controlvm
 | |
|  *                          channel
 | |
|  * @msg: pointer to the retrieved message
 | |
|  *
 | |
|  * Return: 0 if valid message was retrieved or -error
 | |
|  */
 | |
| static int read_controlvm_event(struct controlvm_message *msg)
 | |
| {
 | |
| 	int err = visorchannel_signalremove(chipset_dev->controlvm_channel,
 | |
| 					    CONTROLVM_QUEUE_EVENT, msg);
 | |
| 
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 	/* got a message */
 | |
| 	if (msg->hdr.flags.test_message == 1)
 | |
| 		return -EINVAL;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * parahotplug_process_list() - remove any request from the list that's been on
 | |
|  *                              there too long and respond with an error
 | |
|  */
 | |
| static void parahotplug_process_list(void)
 | |
| {
 | |
| 	struct list_head *pos;
 | |
| 	struct list_head *tmp;
 | |
| 
 | |
| 	spin_lock(¶hotplug_request_list_lock);
 | |
| 	list_for_each_safe(pos, tmp, ¶hotplug_request_list) {
 | |
| 		struct parahotplug_request *req =
 | |
| 		    list_entry(pos, struct parahotplug_request, list);
 | |
| 
 | |
| 		if (!time_after_eq(jiffies, req->expiration))
 | |
| 			continue;
 | |
| 		list_del(pos);
 | |
| 		if (req->msg.hdr.flags.response_expected)
 | |
| 			controlvm_respond(
 | |
| 				&req->msg.hdr,
 | |
| 				CONTROLVM_RESP_DEVICE_UDEV_TIMEOUT,
 | |
| 				&req->msg.cmd.device_change_state.state);
 | |
| 		parahotplug_request_destroy(req);
 | |
| 	}
 | |
| 	spin_unlock(¶hotplug_request_list_lock);
 | |
| }
 | |
| 
 | |
| static void controlvm_periodic_work(struct work_struct *work)
 | |
| {
 | |
| 	struct controlvm_message inmsg;
 | |
| 	int count = 0;
 | |
| 	int err;
 | |
| 
 | |
| 	/* Drain the RESPONSE queue make it empty */
 | |
| 	do {
 | |
| 		err = visorchannel_signalremove(chipset_dev->controlvm_channel,
 | |
| 						CONTROLVM_QUEUE_RESPONSE,
 | |
| 						&inmsg);
 | |
| 	} while ((!err) && (++count < CONTROLVM_MESSAGE_MAX));
 | |
| 	if (err != -EAGAIN)
 | |
| 		goto schedule_out;
 | |
| 	if (chipset_dev->controlvm_pending_msg_valid) {
 | |
| 		/*
 | |
| 		 * we throttled processing of a prior msg, so try to process
 | |
| 		 * it again rather than reading a new one
 | |
| 		 */
 | |
| 		inmsg = chipset_dev->controlvm_pending_msg;
 | |
| 		chipset_dev->controlvm_pending_msg_valid = false;
 | |
| 		err = 0;
 | |
| 	} else {
 | |
| 		err = read_controlvm_event(&inmsg);
 | |
| 	}
 | |
| 	while (!err) {
 | |
| 		chipset_dev->most_recent_message_jiffies = jiffies;
 | |
| 		err = handle_command(inmsg,
 | |
| 				     visorchannel_get_physaddr
 | |
| 				     (chipset_dev->controlvm_channel));
 | |
| 		if (err == -EAGAIN) {
 | |
| 			chipset_dev->controlvm_pending_msg = inmsg;
 | |
| 			chipset_dev->controlvm_pending_msg_valid = true;
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		err = read_controlvm_event(&inmsg);
 | |
| 	}
 | |
| 	/* parahotplug_worker */
 | |
| 	parahotplug_process_list();
 | |
| 
 | |
| /*
 | |
|  * The controlvm messages are sent in a bulk. If we start receiving messages, we
 | |
|  * want the polling to be fast. If we do not receive any message for
 | |
|  * MIN_IDLE_SECONDS, we can slow down the polling.
 | |
|  */
 | |
| schedule_out:
 | |
| 	if (time_after(jiffies, chipset_dev->most_recent_message_jiffies +
 | |
| 				(HZ * MIN_IDLE_SECONDS))) {
 | |
| 		/*
 | |
| 		 * it's been longer than MIN_IDLE_SECONDS since we processed
 | |
| 		 * our last controlvm message; slow down the polling
 | |
| 		 */
 | |
| 		if (chipset_dev->poll_jiffies != POLLJIFFIES_CONTROLVM_SLOW)
 | |
| 			chipset_dev->poll_jiffies = POLLJIFFIES_CONTROLVM_SLOW;
 | |
| 	} else {
 | |
| 		if (chipset_dev->poll_jiffies != POLLJIFFIES_CONTROLVM_FAST)
 | |
| 			chipset_dev->poll_jiffies = POLLJIFFIES_CONTROLVM_FAST;
 | |
| 	}
 | |
| 	schedule_delayed_work(&chipset_dev->periodic_controlvm_work,
 | |
| 			      chipset_dev->poll_jiffies);
 | |
| }
 | |
| 
 | |
| static int visorchipset_init(struct acpi_device *acpi_device)
 | |
| {
 | |
| 	int err = -ENODEV;
 | |
| 	struct visorchannel *controlvm_channel;
 | |
| 
 | |
| 	chipset_dev = kzalloc(sizeof(*chipset_dev), GFP_KERNEL);
 | |
| 	if (!chipset_dev)
 | |
| 		goto error;
 | |
| 	err = controlvm_channel_create(chipset_dev);
 | |
| 	if (err)
 | |
| 		goto error_free_chipset_dev;
 | |
| 	acpi_device->driver_data = chipset_dev;
 | |
| 	chipset_dev->acpi_device = acpi_device;
 | |
| 	chipset_dev->poll_jiffies = POLLJIFFIES_CONTROLVM_FAST;
 | |
| 	err = sysfs_create_groups(&chipset_dev->acpi_device->dev.kobj,
 | |
| 				  visorchipset_dev_groups);
 | |
| 	if (err < 0)
 | |
| 		goto error_destroy_channel;
 | |
| 	controlvm_channel = chipset_dev->controlvm_channel;
 | |
| 	if (!visor_check_channel(visorchannel_get_header(controlvm_channel),
 | |
| 				 &chipset_dev->acpi_device->dev,
 | |
| 				 &visor_controlvm_channel_guid,
 | |
| 				 "controlvm",
 | |
| 				 sizeof(struct visor_controlvm_channel),
 | |
| 				 VISOR_CONTROLVM_CHANNEL_VERSIONID,
 | |
| 				 VISOR_CHANNEL_SIGNATURE))
 | |
| 		goto error_delete_groups;
 | |
| 	/* if booting in a crash kernel */
 | |
| 	if (is_kdump_kernel())
 | |
| 		INIT_DELAYED_WORK(&chipset_dev->periodic_controlvm_work,
 | |
| 				  setup_crash_devices_work_queue);
 | |
| 	else
 | |
| 		INIT_DELAYED_WORK(&chipset_dev->periodic_controlvm_work,
 | |
| 				  controlvm_periodic_work);
 | |
| 	chipset_dev->most_recent_message_jiffies = jiffies;
 | |
| 	chipset_dev->poll_jiffies = POLLJIFFIES_CONTROLVM_FAST;
 | |
| 	schedule_delayed_work(&chipset_dev->periodic_controlvm_work,
 | |
| 			      chipset_dev->poll_jiffies);
 | |
| 	err = visorbus_init();
 | |
| 	if (err < 0)
 | |
| 		goto error_cancel_work;
 | |
| 	return 0;
 | |
| 
 | |
| error_cancel_work:
 | |
| 	cancel_delayed_work_sync(&chipset_dev->periodic_controlvm_work);
 | |
| 
 | |
| error_delete_groups:
 | |
| 	sysfs_remove_groups(&chipset_dev->acpi_device->dev.kobj,
 | |
| 			    visorchipset_dev_groups);
 | |
| 
 | |
| error_destroy_channel:
 | |
| 	visorchannel_destroy(chipset_dev->controlvm_channel);
 | |
| 
 | |
| error_free_chipset_dev:
 | |
| 	kfree(chipset_dev);
 | |
| 
 | |
| error:
 | |
| 	dev_err(&acpi_device->dev, "failed with error %d\n", err);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int visorchipset_exit(struct acpi_device *acpi_device)
 | |
| {
 | |
| 	visorbus_exit();
 | |
| 	cancel_delayed_work_sync(&chipset_dev->periodic_controlvm_work);
 | |
| 	sysfs_remove_groups(&chipset_dev->acpi_device->dev.kobj,
 | |
| 			    visorchipset_dev_groups);
 | |
| 	visorchannel_destroy(chipset_dev->controlvm_channel);
 | |
| 	kfree(chipset_dev);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static const struct acpi_device_id unisys_device_ids[] = {
 | |
| 	{"PNP0A07", 0},
 | |
| 	{"", 0},
 | |
| };
 | |
| 
 | |
| static struct acpi_driver unisys_acpi_driver = {
 | |
| 	.name = "unisys_acpi",
 | |
| 	.class = "unisys_acpi_class",
 | |
| 	.owner = THIS_MODULE,
 | |
| 	.ids = unisys_device_ids,
 | |
| 	.ops = {
 | |
| 		.add = visorchipset_init,
 | |
| 		.remove = visorchipset_exit,
 | |
| 	},
 | |
| };
 | |
| 
 | |
| MODULE_DEVICE_TABLE(acpi, unisys_device_ids);
 | |
| 
 | |
| static __init int visorutil_spar_detect(void)
 | |
| {
 | |
| 	unsigned int eax, ebx, ecx, edx;
 | |
| 
 | |
| 	if (boot_cpu_has(X86_FEATURE_HYPERVISOR)) {
 | |
| 		/* check the ID */
 | |
| 		cpuid(UNISYS_VISOR_LEAF_ID, &eax, &ebx, &ecx, &edx);
 | |
| 		return  (ebx == UNISYS_VISOR_ID_EBX) &&
 | |
| 			(ecx == UNISYS_VISOR_ID_ECX) &&
 | |
| 			(edx == UNISYS_VISOR_ID_EDX);
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int __init init_unisys(void)
 | |
| {
 | |
| 	int result;
 | |
| 
 | |
| 	if (!visorutil_spar_detect())
 | |
| 		return -ENODEV;
 | |
| 	result = acpi_bus_register_driver(&unisys_acpi_driver);
 | |
| 	if (result)
 | |
| 		return -ENODEV;
 | |
| 	pr_info("Unisys Visorchipset Driver Loaded.\n");
 | |
| 	return 0;
 | |
| };
 | |
| 
 | |
| static void __exit exit_unisys(void)
 | |
| {
 | |
| 	acpi_bus_unregister_driver(&unisys_acpi_driver);
 | |
| }
 | |
| 
 | |
| module_init(init_unisys);
 | |
| module_exit(exit_unisys);
 | |
| 
 | |
| MODULE_AUTHOR("Unisys");
 | |
| MODULE_LICENSE("GPL");
 | |
| MODULE_DESCRIPTION("s-Par visorbus driver for virtual device buses");
 | 
