| // SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0 | 
 | /****************************************************************************** | 
 |  * | 
 |  * Module Name: evgpeutil - GPE utilities | 
 |  * | 
 |  * Copyright (C) 2000 - 2018, Intel Corp. | 
 |  * | 
 |  *****************************************************************************/ | 
 |  | 
 | #include <acpi/acpi.h> | 
 | #include "accommon.h" | 
 | #include "acevents.h" | 
 |  | 
 | #define _COMPONENT          ACPI_EVENTS | 
 | ACPI_MODULE_NAME("evgpeutil") | 
 |  | 
 | #if (!ACPI_REDUCED_HARDWARE)	/* Entire module */ | 
 | /******************************************************************************* | 
 |  * | 
 |  * FUNCTION:    acpi_ev_walk_gpe_list | 
 |  * | 
 |  * PARAMETERS:  gpe_walk_callback   - Routine called for each GPE block | 
 |  *              context             - Value passed to callback | 
 |  * | 
 |  * RETURN:      Status | 
 |  * | 
 |  * DESCRIPTION: Walk the GPE lists. | 
 |  * | 
 |  ******************************************************************************/ | 
 | acpi_status | 
 | acpi_ev_walk_gpe_list(acpi_gpe_callback gpe_walk_callback, void *context) | 
 | { | 
 | 	struct acpi_gpe_block_info *gpe_block; | 
 | 	struct acpi_gpe_xrupt_info *gpe_xrupt_info; | 
 | 	acpi_status status = AE_OK; | 
 | 	acpi_cpu_flags flags; | 
 |  | 
 | 	ACPI_FUNCTION_TRACE(ev_walk_gpe_list); | 
 |  | 
 | 	flags = acpi_os_acquire_lock(acpi_gbl_gpe_lock); | 
 |  | 
 | 	/* Walk the interrupt level descriptor list */ | 
 |  | 
 | 	gpe_xrupt_info = acpi_gbl_gpe_xrupt_list_head; | 
 | 	while (gpe_xrupt_info) { | 
 |  | 
 | 		/* Walk all Gpe Blocks attached to this interrupt level */ | 
 |  | 
 | 		gpe_block = gpe_xrupt_info->gpe_block_list_head; | 
 | 		while (gpe_block) { | 
 |  | 
 | 			/* One callback per GPE block */ | 
 |  | 
 | 			status = | 
 | 			    gpe_walk_callback(gpe_xrupt_info, gpe_block, | 
 | 					      context); | 
 | 			if (ACPI_FAILURE(status)) { | 
 | 				if (status == AE_CTRL_END) {	/* Callback abort */ | 
 | 					status = AE_OK; | 
 | 				} | 
 | 				goto unlock_and_exit; | 
 | 			} | 
 |  | 
 | 			gpe_block = gpe_block->next; | 
 | 		} | 
 |  | 
 | 		gpe_xrupt_info = gpe_xrupt_info->next; | 
 | 	} | 
 |  | 
 | unlock_and_exit: | 
 | 	acpi_os_release_lock(acpi_gbl_gpe_lock, flags); | 
 | 	return_ACPI_STATUS(status); | 
 | } | 
 |  | 
 | /******************************************************************************* | 
 |  * | 
 |  * FUNCTION:    acpi_ev_get_gpe_device | 
 |  * | 
 |  * PARAMETERS:  GPE_WALK_CALLBACK | 
 |  * | 
 |  * RETURN:      Status | 
 |  * | 
 |  * DESCRIPTION: Matches the input GPE index (0-current_gpe_count) with a GPE | 
 |  *              block device. NULL if the GPE is one of the FADT-defined GPEs. | 
 |  * | 
 |  ******************************************************************************/ | 
 |  | 
 | acpi_status | 
 | acpi_ev_get_gpe_device(struct acpi_gpe_xrupt_info *gpe_xrupt_info, | 
 | 		       struct acpi_gpe_block_info *gpe_block, void *context) | 
 | { | 
 | 	struct acpi_gpe_device_info *info = context; | 
 |  | 
 | 	/* Increment Index by the number of GPEs in this block */ | 
 |  | 
 | 	info->next_block_base_index += gpe_block->gpe_count; | 
 |  | 
 | 	if (info->index < info->next_block_base_index) { | 
 | 		/* | 
 | 		 * The GPE index is within this block, get the node. Leave the node | 
 | 		 * NULL for the FADT-defined GPEs | 
 | 		 */ | 
 | 		if ((gpe_block->node)->type == ACPI_TYPE_DEVICE) { | 
 | 			info->gpe_device = gpe_block->node; | 
 | 		} | 
 |  | 
 | 		info->status = AE_OK; | 
 | 		return (AE_CTRL_END); | 
 | 	} | 
 |  | 
 | 	return (AE_OK); | 
 | } | 
 |  | 
 | /******************************************************************************* | 
 |  * | 
 |  * FUNCTION:    acpi_ev_get_gpe_xrupt_block | 
 |  * | 
 |  * PARAMETERS:  interrupt_number            - Interrupt for a GPE block | 
 |  *              gpe_xrupt_block             - Where the block is returned | 
 |  * | 
 |  * RETURN:      Status | 
 |  * | 
 |  * DESCRIPTION: Get or Create a GPE interrupt block. There is one interrupt | 
 |  *              block per unique interrupt level used for GPEs. Should be | 
 |  *              called only when the GPE lists are semaphore locked and not | 
 |  *              subject to change. | 
 |  * | 
 |  ******************************************************************************/ | 
 |  | 
 | acpi_status | 
 | acpi_ev_get_gpe_xrupt_block(u32 interrupt_number, | 
 | 			    struct acpi_gpe_xrupt_info **gpe_xrupt_block) | 
 | { | 
 | 	struct acpi_gpe_xrupt_info *next_gpe_xrupt; | 
 | 	struct acpi_gpe_xrupt_info *gpe_xrupt; | 
 | 	acpi_status status; | 
 | 	acpi_cpu_flags flags; | 
 |  | 
 | 	ACPI_FUNCTION_TRACE(ev_get_gpe_xrupt_block); | 
 |  | 
 | 	/* No need for lock since we are not changing any list elements here */ | 
 |  | 
 | 	next_gpe_xrupt = acpi_gbl_gpe_xrupt_list_head; | 
 | 	while (next_gpe_xrupt) { | 
 | 		if (next_gpe_xrupt->interrupt_number == interrupt_number) { | 
 | 			*gpe_xrupt_block = next_gpe_xrupt; | 
 | 			return_ACPI_STATUS(AE_OK); | 
 | 		} | 
 |  | 
 | 		next_gpe_xrupt = next_gpe_xrupt->next; | 
 | 	} | 
 |  | 
 | 	/* Not found, must allocate a new xrupt descriptor */ | 
 |  | 
 | 	gpe_xrupt = ACPI_ALLOCATE_ZEROED(sizeof(struct acpi_gpe_xrupt_info)); | 
 | 	if (!gpe_xrupt) { | 
 | 		return_ACPI_STATUS(AE_NO_MEMORY); | 
 | 	} | 
 |  | 
 | 	gpe_xrupt->interrupt_number = interrupt_number; | 
 |  | 
 | 	/* Install new interrupt descriptor with spin lock */ | 
 |  | 
 | 	flags = acpi_os_acquire_lock(acpi_gbl_gpe_lock); | 
 | 	if (acpi_gbl_gpe_xrupt_list_head) { | 
 | 		next_gpe_xrupt = acpi_gbl_gpe_xrupt_list_head; | 
 | 		while (next_gpe_xrupt->next) { | 
 | 			next_gpe_xrupt = next_gpe_xrupt->next; | 
 | 		} | 
 |  | 
 | 		next_gpe_xrupt->next = gpe_xrupt; | 
 | 		gpe_xrupt->previous = next_gpe_xrupt; | 
 | 	} else { | 
 | 		acpi_gbl_gpe_xrupt_list_head = gpe_xrupt; | 
 | 	} | 
 |  | 
 | 	acpi_os_release_lock(acpi_gbl_gpe_lock, flags); | 
 |  | 
 | 	/* Install new interrupt handler if not SCI_INT */ | 
 |  | 
 | 	if (interrupt_number != acpi_gbl_FADT.sci_interrupt) { | 
 | 		status = acpi_os_install_interrupt_handler(interrupt_number, | 
 | 							   acpi_ev_gpe_xrupt_handler, | 
 | 							   gpe_xrupt); | 
 | 		if (ACPI_FAILURE(status)) { | 
 | 			ACPI_EXCEPTION((AE_INFO, status, | 
 | 					"Could not install GPE interrupt handler at level 0x%X", | 
 | 					interrupt_number)); | 
 | 			return_ACPI_STATUS(status); | 
 | 		} | 
 | 	} | 
 |  | 
 | 	*gpe_xrupt_block = gpe_xrupt; | 
 | 	return_ACPI_STATUS(AE_OK); | 
 | } | 
 |  | 
 | /******************************************************************************* | 
 |  * | 
 |  * FUNCTION:    acpi_ev_delete_gpe_xrupt | 
 |  * | 
 |  * PARAMETERS:  gpe_xrupt       - A GPE interrupt info block | 
 |  * | 
 |  * RETURN:      Status | 
 |  * | 
 |  * DESCRIPTION: Remove and free a gpe_xrupt block. Remove an associated | 
 |  *              interrupt handler if not the SCI interrupt. | 
 |  * | 
 |  ******************************************************************************/ | 
 |  | 
 | acpi_status acpi_ev_delete_gpe_xrupt(struct acpi_gpe_xrupt_info *gpe_xrupt) | 
 | { | 
 | 	acpi_status status; | 
 | 	acpi_cpu_flags flags; | 
 |  | 
 | 	ACPI_FUNCTION_TRACE(ev_delete_gpe_xrupt); | 
 |  | 
 | 	/* We never want to remove the SCI interrupt handler */ | 
 |  | 
 | 	if (gpe_xrupt->interrupt_number == acpi_gbl_FADT.sci_interrupt) { | 
 | 		gpe_xrupt->gpe_block_list_head = NULL; | 
 | 		return_ACPI_STATUS(AE_OK); | 
 | 	} | 
 |  | 
 | 	/* Disable this interrupt */ | 
 |  | 
 | 	status = | 
 | 	    acpi_os_remove_interrupt_handler(gpe_xrupt->interrupt_number, | 
 | 					     acpi_ev_gpe_xrupt_handler); | 
 | 	if (ACPI_FAILURE(status)) { | 
 | 		return_ACPI_STATUS(status); | 
 | 	} | 
 |  | 
 | 	/* Unlink the interrupt block with lock */ | 
 |  | 
 | 	flags = acpi_os_acquire_lock(acpi_gbl_gpe_lock); | 
 | 	if (gpe_xrupt->previous) { | 
 | 		gpe_xrupt->previous->next = gpe_xrupt->next; | 
 | 	} else { | 
 | 		/* No previous, update list head */ | 
 |  | 
 | 		acpi_gbl_gpe_xrupt_list_head = gpe_xrupt->next; | 
 | 	} | 
 |  | 
 | 	if (gpe_xrupt->next) { | 
 | 		gpe_xrupt->next->previous = gpe_xrupt->previous; | 
 | 	} | 
 | 	acpi_os_release_lock(acpi_gbl_gpe_lock, flags); | 
 |  | 
 | 	/* Free the block */ | 
 |  | 
 | 	ACPI_FREE(gpe_xrupt); | 
 | 	return_ACPI_STATUS(AE_OK); | 
 | } | 
 |  | 
 | /******************************************************************************* | 
 |  * | 
 |  * FUNCTION:    acpi_ev_delete_gpe_handlers | 
 |  * | 
 |  * PARAMETERS:  gpe_xrupt_info      - GPE Interrupt info | 
 |  *              gpe_block           - Gpe Block info | 
 |  * | 
 |  * RETURN:      Status | 
 |  * | 
 |  * DESCRIPTION: Delete all Handler objects found in the GPE data structs. | 
 |  *              Used only prior to termination. | 
 |  * | 
 |  ******************************************************************************/ | 
 |  | 
 | acpi_status | 
 | acpi_ev_delete_gpe_handlers(struct acpi_gpe_xrupt_info *gpe_xrupt_info, | 
 | 			    struct acpi_gpe_block_info *gpe_block, | 
 | 			    void *context) | 
 | { | 
 | 	struct acpi_gpe_event_info *gpe_event_info; | 
 | 	struct acpi_gpe_notify_info *notify; | 
 | 	struct acpi_gpe_notify_info *next; | 
 | 	u32 i; | 
 | 	u32 j; | 
 |  | 
 | 	ACPI_FUNCTION_TRACE(ev_delete_gpe_handlers); | 
 |  | 
 | 	/* Examine each GPE Register within the block */ | 
 |  | 
 | 	for (i = 0; i < gpe_block->register_count; i++) { | 
 |  | 
 | 		/* Now look at the individual GPEs in this byte register */ | 
 |  | 
 | 		for (j = 0; j < ACPI_GPE_REGISTER_WIDTH; j++) { | 
 | 			gpe_event_info = &gpe_block->event_info[((acpi_size)i * | 
 | 								 ACPI_GPE_REGISTER_WIDTH) | 
 | 								+ j]; | 
 |  | 
 | 			if ((ACPI_GPE_DISPATCH_TYPE(gpe_event_info->flags) == | 
 | 			     ACPI_GPE_DISPATCH_HANDLER) || | 
 | 			    (ACPI_GPE_DISPATCH_TYPE(gpe_event_info->flags) == | 
 | 			     ACPI_GPE_DISPATCH_RAW_HANDLER)) { | 
 |  | 
 | 				/* Delete an installed handler block */ | 
 |  | 
 | 				ACPI_FREE(gpe_event_info->dispatch.handler); | 
 | 				gpe_event_info->dispatch.handler = NULL; | 
 | 				gpe_event_info->flags &= | 
 | 				    ~ACPI_GPE_DISPATCH_MASK; | 
 | 			} else if (ACPI_GPE_DISPATCH_TYPE(gpe_event_info->flags) | 
 | 				   == ACPI_GPE_DISPATCH_NOTIFY) { | 
 |  | 
 | 				/* Delete the implicit notification device list */ | 
 |  | 
 | 				notify = gpe_event_info->dispatch.notify_list; | 
 | 				while (notify) { | 
 | 					next = notify->next; | 
 | 					ACPI_FREE(notify); | 
 | 					notify = next; | 
 | 				} | 
 |  | 
 | 				gpe_event_info->dispatch.notify_list = NULL; | 
 | 				gpe_event_info->flags &= | 
 | 				    ~ACPI_GPE_DISPATCH_MASK; | 
 | 			} | 
 | 		} | 
 | 	} | 
 |  | 
 | 	return_ACPI_STATUS(AE_OK); | 
 | } | 
 |  | 
 | #endif				/* !ACPI_REDUCED_HARDWARE */ |