FreeBSD/Linux Kernel Cross Reference
sys/dev/acpica/acpi.c
1 /*-
2 * Copyright (c) 2000 Takanori Watanabe <takawata@jp.freebsd.org>
3 * Copyright (c) 2000 Mitsuru IWASAKI <iwasaki@jp.freebsd.org>
4 * Copyright (c) 2000, 2001 Michael Smith
5 * Copyright (c) 2000 BSDi
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27 * SUCH DAMAGE.
28 */
29
30 #include <sys/cdefs.h>
31 __FBSDID("$FreeBSD$");
32
33 #include "opt_acpi.h"
34 #include <sys/param.h>
35 #include <sys/kernel.h>
36 #include <sys/proc.h>
37 #include <sys/fcntl.h>
38 #include <sys/malloc.h>
39 #include <sys/module.h>
40 #include <sys/bus.h>
41 #include <sys/conf.h>
42 #include <sys/ioccom.h>
43 #include <sys/reboot.h>
44 #include <sys/sysctl.h>
45 #include <sys/ctype.h>
46 #include <sys/linker.h>
47 #include <sys/power.h>
48 #include <sys/sbuf.h>
49 #include <sys/smp.h>
50
51 #include <machine/resource.h>
52 #include <machine/bus.h>
53 #include <sys/rman.h>
54 #include <isa/isavar.h>
55 #include <isa/pnpvar.h>
56
57 #include <contrib/dev/acpica/acpi.h>
58 #include <dev/acpica/acpivar.h>
59 #include <dev/acpica/acpiio.h>
60 #include <contrib/dev/acpica/achware.h>
61 #include <contrib/dev/acpica/acnamesp.h>
62
63 #include "pci_if.h"
64 #include <dev/pci/pcivar.h>
65 #include <dev/pci/pci_private.h>
66
67 #include <vm/vm_param.h>
68
69 MALLOC_DEFINE(M_ACPIDEV, "acpidev", "ACPI devices");
70
71 /* Hooks for the ACPI CA debugging infrastructure */
72 #define _COMPONENT ACPI_BUS
73 ACPI_MODULE_NAME("ACPI")
74
75 static d_open_t acpiopen;
76 static d_close_t acpiclose;
77 static d_ioctl_t acpiioctl;
78
79 static struct cdevsw acpi_cdevsw = {
80 .d_version = D_VERSION,
81 .d_open = acpiopen,
82 .d_close = acpiclose,
83 .d_ioctl = acpiioctl,
84 .d_name = "acpi",
85 };
86
87 /* Global mutex for locking access to the ACPI subsystem. */
88 struct mtx acpi_mutex;
89
90 /* Bitmap of device quirks. */
91 int acpi_quirks;
92
93 static int acpi_modevent(struct module *mod, int event, void *junk);
94 static void acpi_identify(driver_t *driver, device_t parent);
95 static int acpi_probe(device_t dev);
96 static int acpi_attach(device_t dev);
97 static int acpi_suspend(device_t dev);
98 static int acpi_resume(device_t dev);
99 static int acpi_shutdown(device_t dev);
100 static device_t acpi_add_child(device_t bus, int order, const char *name,
101 int unit);
102 static int acpi_print_child(device_t bus, device_t child);
103 static void acpi_probe_nomatch(device_t bus, device_t child);
104 static void acpi_driver_added(device_t dev, driver_t *driver);
105 static int acpi_read_ivar(device_t dev, device_t child, int index,
106 uintptr_t *result);
107 static int acpi_write_ivar(device_t dev, device_t child, int index,
108 uintptr_t value);
109 static struct resource_list *acpi_get_rlist(device_t dev, device_t child);
110 static int acpi_sysres_alloc(device_t dev);
111 static struct resource *acpi_alloc_resource(device_t bus, device_t child,
112 int type, int *rid, u_long start, u_long end,
113 u_long count, u_int flags);
114 static int acpi_release_resource(device_t bus, device_t child, int type,
115 int rid, struct resource *r);
116 static void acpi_delete_resource(device_t bus, device_t child, int type,
117 int rid);
118 static uint32_t acpi_isa_get_logicalid(device_t dev);
119 static int acpi_isa_get_compatid(device_t dev, uint32_t *cids, int count);
120 static char *acpi_device_id_probe(device_t bus, device_t dev, char **ids);
121 static ACPI_STATUS acpi_device_eval_obj(device_t bus, device_t dev,
122 ACPI_STRING pathname, ACPI_OBJECT_LIST *parameters,
123 ACPI_BUFFER *ret);
124 static int acpi_device_pwr_for_sleep(device_t bus, device_t dev,
125 int *dstate);
126 static ACPI_STATUS acpi_device_scan_cb(ACPI_HANDLE h, UINT32 level,
127 void *context, void **retval);
128 static ACPI_STATUS acpi_device_scan_children(device_t bus, device_t dev,
129 int max_depth, acpi_scan_cb_t user_fn, void *arg);
130 static int acpi_set_powerstate_method(device_t bus, device_t child,
131 int state);
132 static int acpi_isa_pnp_probe(device_t bus, device_t child,
133 struct isa_pnp_id *ids);
134 static void acpi_probe_children(device_t bus);
135 static int acpi_probe_order(ACPI_HANDLE handle, int *order);
136 static ACPI_STATUS acpi_probe_child(ACPI_HANDLE handle, UINT32 level,
137 void *context, void **status);
138 static BOOLEAN acpi_MatchHid(ACPI_HANDLE h, const char *hid);
139 static ACPI_STATUS acpi_EnterSleepState(struct acpi_softc *sc, int state);
140 static void acpi_shutdown_final(void *arg, int howto);
141 static void acpi_enable_fixed_events(struct acpi_softc *sc);
142 static int acpi_wake_sleep_prep(ACPI_HANDLE handle, int sstate);
143 static int acpi_wake_run_prep(ACPI_HANDLE handle, int sstate);
144 static int acpi_wake_prep_walk(int sstate);
145 static int acpi_wake_sysctl_walk(device_t dev);
146 static int acpi_wake_set_sysctl(SYSCTL_HANDLER_ARGS);
147 static void acpi_system_eventhandler_sleep(void *arg, int state);
148 static void acpi_system_eventhandler_wakeup(void *arg, int state);
149 static int acpi_supported_sleep_state_sysctl(SYSCTL_HANDLER_ARGS);
150 static int acpi_sleep_state_sysctl(SYSCTL_HANDLER_ARGS);
151 static int acpi_pm_func(u_long cmd, void *arg, ...);
152 static int acpi_child_location_str_method(device_t acdev, device_t child,
153 char *buf, size_t buflen);
154 static int acpi_child_pnpinfo_str_method(device_t acdev, device_t child,
155 char *buf, size_t buflen);
156
157 static device_method_t acpi_methods[] = {
158 /* Device interface */
159 DEVMETHOD(device_identify, acpi_identify),
160 DEVMETHOD(device_probe, acpi_probe),
161 DEVMETHOD(device_attach, acpi_attach),
162 DEVMETHOD(device_shutdown, acpi_shutdown),
163 DEVMETHOD(device_detach, bus_generic_detach),
164 DEVMETHOD(device_suspend, acpi_suspend),
165 DEVMETHOD(device_resume, acpi_resume),
166
167 /* Bus interface */
168 DEVMETHOD(bus_add_child, acpi_add_child),
169 DEVMETHOD(bus_print_child, acpi_print_child),
170 DEVMETHOD(bus_probe_nomatch, acpi_probe_nomatch),
171 DEVMETHOD(bus_driver_added, acpi_driver_added),
172 DEVMETHOD(bus_read_ivar, acpi_read_ivar),
173 DEVMETHOD(bus_write_ivar, acpi_write_ivar),
174 DEVMETHOD(bus_get_resource_list, acpi_get_rlist),
175 DEVMETHOD(bus_set_resource, bus_generic_rl_set_resource),
176 DEVMETHOD(bus_get_resource, bus_generic_rl_get_resource),
177 DEVMETHOD(bus_alloc_resource, acpi_alloc_resource),
178 DEVMETHOD(bus_release_resource, acpi_release_resource),
179 DEVMETHOD(bus_delete_resource, acpi_delete_resource),
180 DEVMETHOD(bus_child_pnpinfo_str, acpi_child_pnpinfo_str_method),
181 DEVMETHOD(bus_child_location_str, acpi_child_location_str_method),
182 DEVMETHOD(bus_activate_resource, bus_generic_activate_resource),
183 DEVMETHOD(bus_deactivate_resource, bus_generic_deactivate_resource),
184 DEVMETHOD(bus_setup_intr, bus_generic_setup_intr),
185 DEVMETHOD(bus_teardown_intr, bus_generic_teardown_intr),
186
187 /* ACPI bus */
188 DEVMETHOD(acpi_id_probe, acpi_device_id_probe),
189 DEVMETHOD(acpi_evaluate_object, acpi_device_eval_obj),
190 DEVMETHOD(acpi_pwr_for_sleep, acpi_device_pwr_for_sleep),
191 DEVMETHOD(acpi_scan_children, acpi_device_scan_children),
192
193 /* PCI emulation */
194 DEVMETHOD(pci_set_powerstate, acpi_set_powerstate_method),
195
196 /* ISA emulation */
197 DEVMETHOD(isa_pnp_probe, acpi_isa_pnp_probe),
198
199 {0, 0}
200 };
201
202 static driver_t acpi_driver = {
203 "acpi",
204 acpi_methods,
205 sizeof(struct acpi_softc),
206 };
207
208 static devclass_t acpi_devclass;
209 DRIVER_MODULE(acpi, nexus, acpi_driver, acpi_devclass, acpi_modevent, 0);
210 MODULE_VERSION(acpi, 1);
211
212 ACPI_SERIAL_DECL(acpi, "ACPI root bus");
213
214 /* Local pools for managing system resources for ACPI child devices. */
215 static struct rman acpi_rman_io, acpi_rman_mem;
216
217 #define ACPI_MINIMUM_AWAKETIME 5
218
219 static const char* sleep_state_names[] = {
220 "S0", "S1", "S2", "S3", "S4", "S5", "NONE"};
221
222 SYSCTL_NODE(_debug, OID_AUTO, acpi, CTLFLAG_RD, NULL, "ACPI debugging");
223 static char acpi_ca_version[12];
224 SYSCTL_STRING(_debug_acpi, OID_AUTO, acpi_ca_version, CTLFLAG_RD,
225 acpi_ca_version, 0, "Version of Intel ACPI-CA");
226
227 /*
228 * Allow override of whether methods execute in parallel or not.
229 * Enable this for serial behavior, which fixes "AE_ALREADY_EXISTS"
230 * errors for AML that really can't handle parallel method execution.
231 * It is off by default since this breaks recursive methods and
232 * some IBMs use such code.
233 */
234 static int acpi_serialize_methods;
235 TUNABLE_INT("hw.acpi.serialize_methods", &acpi_serialize_methods);
236
237 /* Power devices off and on in suspend and resume. XXX Remove once tested. */
238 static int acpi_do_powerstate = 1;
239 TUNABLE_INT("debug.acpi.do_powerstate", &acpi_do_powerstate);
240 SYSCTL_INT(_debug_acpi, OID_AUTO, do_powerstate, CTLFLAG_RW,
241 &acpi_do_powerstate, 1, "Turn off devices when suspending.");
242
243 /* Allow users to override quirks. */
244 TUNABLE_INT("debug.acpi.quirks", &acpi_quirks);
245
246 static int acpi_susp_bounce;
247 SYSCTL_INT(_debug_acpi, OID_AUTO, suspend_bounce, CTLFLAG_RW,
248 &acpi_susp_bounce, 0, "Don't actually suspend, just test devices.");
249
250 /*
251 * ACPI can only be loaded as a module by the loader; activating it after
252 * system bootstrap time is not useful, and can be fatal to the system.
253 * It also cannot be unloaded, since the entire system bus heirarchy hangs
254 * off it.
255 */
256 static int
257 acpi_modevent(struct module *mod, int event, void *junk)
258 {
259 switch (event) {
260 case MOD_LOAD:
261 if (!cold) {
262 printf("The ACPI driver cannot be loaded after boot.\n");
263 return (EPERM);
264 }
265 break;
266 case MOD_UNLOAD:
267 if (!cold && power_pm_get_type() == POWER_PM_TYPE_ACPI)
268 return (EBUSY);
269 break;
270 default:
271 break;
272 }
273 return (0);
274 }
275
276 /*
277 * Perform early initialization.
278 */
279 ACPI_STATUS
280 acpi_Startup(void)
281 {
282 static int started = 0;
283 ACPI_STATUS status;
284 int val;
285
286 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
287
288 /* Only run the startup code once. The MADT driver also calls this. */
289 if (started)
290 return_VALUE (AE_OK);
291 started = 1;
292
293 /*
294 * Pre-allocate space for RSDT/XSDT and DSDT tables and allow resizing
295 * if more tables exist.
296 */
297 if (ACPI_FAILURE(status = AcpiInitializeTables(NULL, 2, TRUE))) {
298 printf("ACPI: Table initialisation failed: %s\n",
299 AcpiFormatException(status));
300 return_VALUE (status);
301 }
302
303 /* Set up any quirks we have for this system. */
304 if (acpi_quirks == ACPI_Q_OK)
305 acpi_table_quirks(&acpi_quirks);
306
307 /* If the user manually set the disabled hint to 0, force-enable ACPI. */
308 if (resource_int_value("acpi", 0, "disabled", &val) == 0 && val == 0)
309 acpi_quirks &= ~ACPI_Q_BROKEN;
310 if (acpi_quirks & ACPI_Q_BROKEN) {
311 printf("ACPI disabled by blacklist. Contact your BIOS vendor.\n");
312 status = AE_SUPPORT;
313 }
314
315 return_VALUE (status);
316 }
317
318 /*
319 * Detect ACPI, perform early initialisation
320 */
321 static void
322 acpi_identify(driver_t *driver, device_t parent)
323 {
324 device_t child;
325
326 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
327
328 if (!cold)
329 return_VOID;
330
331 /* Check that we haven't been disabled with a hint. */
332 if (resource_disabled("acpi", 0))
333 return_VOID;
334
335 /* Make sure we're not being doubly invoked. */
336 if (device_find_child(parent, "acpi", 0) != NULL)
337 return_VOID;
338
339 snprintf(acpi_ca_version, sizeof(acpi_ca_version), "%x", ACPI_CA_VERSION);
340
341 /* Initialize root tables. */
342 if (ACPI_FAILURE(acpi_Startup())) {
343 printf("ACPI: Try disabling either ACPI or apic support.\n");
344 return_VOID;
345 }
346
347 /* Attach the actual ACPI device. */
348 if ((child = BUS_ADD_CHILD(parent, 10, "acpi", 0)) == NULL) {
349 device_printf(parent, "device_identify failed\n");
350 return_VOID;
351 }
352 }
353
354 /*
355 * Fetch some descriptive data from ACPI to put in our attach message.
356 */
357 static int
358 acpi_probe(device_t dev)
359 {
360 ACPI_TABLE_RSDP *rsdp;
361 ACPI_TABLE_HEADER *rsdt;
362 ACPI_PHYSICAL_ADDRESS paddr;
363 char buf[ACPI_OEM_ID_SIZE + ACPI_OEM_TABLE_ID_SIZE + 2];
364 struct sbuf sb;
365
366 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
367
368 if (power_pm_get_type() != POWER_PM_TYPE_NONE &&
369 power_pm_get_type() != POWER_PM_TYPE_ACPI) {
370 device_printf(dev, "probe failed, other PM system enabled.\n");
371 return_VALUE (ENXIO);
372 }
373
374 if ((paddr = AcpiOsGetRootPointer()) == 0 ||
375 (rsdp = AcpiOsMapMemory(paddr, sizeof(ACPI_TABLE_RSDP))) == NULL)
376 return_VALUE (ENXIO);
377 if (rsdp->Revision > 1 && rsdp->XsdtPhysicalAddress != 0)
378 paddr = (ACPI_PHYSICAL_ADDRESS)rsdp->XsdtPhysicalAddress;
379 else
380 paddr = (ACPI_PHYSICAL_ADDRESS)rsdp->RsdtPhysicalAddress;
381 AcpiOsUnmapMemory(rsdp, sizeof(ACPI_TABLE_RSDP));
382
383 if ((rsdt = AcpiOsMapMemory(paddr, sizeof(ACPI_TABLE_HEADER))) == NULL)
384 return_VALUE (ENXIO);
385 sbuf_new(&sb, buf, sizeof(buf), SBUF_FIXEDLEN);
386 sbuf_bcat(&sb, rsdt->OemId, ACPI_OEM_ID_SIZE);
387 sbuf_trim(&sb);
388 sbuf_putc(&sb, ' ');
389 sbuf_bcat(&sb, rsdt->OemTableId, ACPI_OEM_TABLE_ID_SIZE);
390 sbuf_trim(&sb);
391 sbuf_finish(&sb);
392 device_set_desc_copy(dev, sbuf_data(&sb));
393 sbuf_delete(&sb);
394 AcpiOsUnmapMemory(rsdt, sizeof(ACPI_TABLE_HEADER));
395
396 return_VALUE (0);
397 }
398
399 static int
400 acpi_attach(device_t dev)
401 {
402 struct acpi_softc *sc;
403 ACPI_TABLE_FACS *facs;
404 ACPI_STATUS status;
405 int error, state;
406 UINT32 flags;
407 UINT8 TypeA, TypeB;
408 char *env;
409
410 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
411
412 sc = device_get_softc(dev);
413 sc->acpi_dev = dev;
414 callout_init(&sc->susp_force_to, TRUE);
415
416 error = ENXIO;
417
418 /* Initialize resource manager. */
419 acpi_rman_io.rm_type = RMAN_ARRAY;
420 acpi_rman_io.rm_start = 0;
421 acpi_rman_io.rm_end = 0xffff;
422 acpi_rman_io.rm_descr = "ACPI I/O ports";
423 if (rman_init(&acpi_rman_io) != 0)
424 panic("acpi rman_init IO ports failed");
425 acpi_rman_mem.rm_type = RMAN_ARRAY;
426 acpi_rman_mem.rm_start = 0;
427 acpi_rman_mem.rm_end = ~0ul;
428 acpi_rman_mem.rm_descr = "ACPI I/O memory addresses";
429 if (rman_init(&acpi_rman_mem) != 0)
430 panic("acpi rman_init memory failed");
431
432 /* Initialise the ACPI mutex */
433 mtx_init(&acpi_mutex, "ACPI global lock", NULL, MTX_DEF);
434
435 /*
436 * Set the globals from our tunables. This is needed because ACPI-CA
437 * uses UINT8 for some values and we have no tunable_byte.
438 */
439 AcpiGbl_AllMethodsSerialized = acpi_serialize_methods;
440 AcpiGbl_EnableInterpreterSlack = TRUE;
441
442 /* Start up the ACPI CA subsystem. */
443 status = AcpiInitializeSubsystem();
444 if (ACPI_FAILURE(status)) {
445 device_printf(dev, "Could not initialize Subsystem: %s\n",
446 AcpiFormatException(status));
447 goto out;
448 }
449
450 /* Load ACPI name space. */
451 status = AcpiLoadTables();
452 if (ACPI_FAILURE(status)) {
453 device_printf(dev, "Could not load Namespace: %s\n",
454 AcpiFormatException(status));
455 goto out;
456 }
457
458 /* Install the default address space handlers. */
459 status = AcpiInstallAddressSpaceHandler(ACPI_ROOT_OBJECT,
460 ACPI_ADR_SPACE_SYSTEM_MEMORY, ACPI_DEFAULT_HANDLER, NULL, NULL);
461 if (ACPI_FAILURE(status)) {
462 device_printf(dev, "Could not initialise SystemMemory handler: %s\n",
463 AcpiFormatException(status));
464 goto out;
465 }
466 status = AcpiInstallAddressSpaceHandler(ACPI_ROOT_OBJECT,
467 ACPI_ADR_SPACE_SYSTEM_IO, ACPI_DEFAULT_HANDLER, NULL, NULL);
468 if (ACPI_FAILURE(status)) {
469 device_printf(dev, "Could not initialise SystemIO handler: %s\n",
470 AcpiFormatException(status));
471 goto out;
472 }
473 status = AcpiInstallAddressSpaceHandler(ACPI_ROOT_OBJECT,
474 ACPI_ADR_SPACE_PCI_CONFIG, ACPI_DEFAULT_HANDLER, NULL, NULL);
475 if (ACPI_FAILURE(status)) {
476 device_printf(dev, "could not initialise PciConfig handler: %s\n",
477 AcpiFormatException(status));
478 goto out;
479 }
480
481 /*
482 * Note that some systems (specifically, those with namespace evaluation
483 * issues that require the avoidance of parts of the namespace) must
484 * avoid running _INI and _STA on everything, as well as dodging the final
485 * object init pass.
486 *
487 * For these devices, we set ACPI_NO_DEVICE_INIT and ACPI_NO_OBJECT_INIT).
488 *
489 * XXX We should arrange for the object init pass after we have attached
490 * all our child devices, but on many systems it works here.
491 */
492 flags = 0;
493 if (testenv("debug.acpi.avoid"))
494 flags = ACPI_NO_DEVICE_INIT | ACPI_NO_OBJECT_INIT;
495
496 /* Bring the hardware and basic handlers online. */
497 if (ACPI_FAILURE(status = AcpiEnableSubsystem(flags))) {
498 device_printf(dev, "Could not enable ACPI: %s\n",
499 AcpiFormatException(status));
500 goto out;
501 }
502
503 /*
504 * Call the ECDT probe function to provide EC functionality before
505 * the namespace has been evaluated.
506 *
507 * XXX This happens before the sysresource devices have been probed and
508 * attached so its resources come from nexus0. In practice, this isn't
509 * a problem but should be addressed eventually.
510 */
511 acpi_ec_ecdt_probe(dev);
512
513 /* Bring device objects and regions online. */
514 if (ACPI_FAILURE(status = AcpiInitializeObjects(flags))) {
515 device_printf(dev, "Could not initialize ACPI objects: %s\n",
516 AcpiFormatException(status));
517 goto out;
518 }
519
520 /*
521 * Setup our sysctl tree.
522 *
523 * XXX: This doesn't check to make sure that none of these fail.
524 */
525 sysctl_ctx_init(&sc->acpi_sysctl_ctx);
526 sc->acpi_sysctl_tree = SYSCTL_ADD_NODE(&sc->acpi_sysctl_ctx,
527 SYSCTL_STATIC_CHILDREN(_hw), OID_AUTO,
528 device_get_name(dev), CTLFLAG_RD, 0, "");
529 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
530 OID_AUTO, "supported_sleep_state", CTLTYPE_STRING | CTLFLAG_RD,
531 0, 0, acpi_supported_sleep_state_sysctl, "A", "");
532 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
533 OID_AUTO, "power_button_state", CTLTYPE_STRING | CTLFLAG_RW,
534 &sc->acpi_power_button_sx, 0, acpi_sleep_state_sysctl, "A", "");
535 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
536 OID_AUTO, "sleep_button_state", CTLTYPE_STRING | CTLFLAG_RW,
537 &sc->acpi_sleep_button_sx, 0, acpi_sleep_state_sysctl, "A", "");
538 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
539 OID_AUTO, "lid_switch_state", CTLTYPE_STRING | CTLFLAG_RW,
540 &sc->acpi_lid_switch_sx, 0, acpi_sleep_state_sysctl, "A", "");
541 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
542 OID_AUTO, "standby_state", CTLTYPE_STRING | CTLFLAG_RW,
543 &sc->acpi_standby_sx, 0, acpi_sleep_state_sysctl, "A", "");
544 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
545 OID_AUTO, "suspend_state", CTLTYPE_STRING | CTLFLAG_RW,
546 &sc->acpi_suspend_sx, 0, acpi_sleep_state_sysctl, "A", "");
547 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
548 OID_AUTO, "sleep_delay", CTLFLAG_RW, &sc->acpi_sleep_delay, 0,
549 "sleep delay");
550 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
551 OID_AUTO, "s4bios", CTLFLAG_RW, &sc->acpi_s4bios, 0, "S4BIOS mode");
552 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
553 OID_AUTO, "verbose", CTLFLAG_RW, &sc->acpi_verbose, 0, "verbose mode");
554 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
555 OID_AUTO, "disable_on_reboot", CTLFLAG_RW,
556 &sc->acpi_do_disable, 0, "Disable ACPI when rebooting/halting system");
557 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
558 OID_AUTO, "handle_reboot", CTLFLAG_RW,
559 &sc->acpi_handle_reboot, 0, "Use ACPI Reset Register to reboot");
560
561 /*
562 * Default to 1 second before sleeping to give some machines time to
563 * stabilize.
564 */
565 sc->acpi_sleep_delay = 1;
566 if (bootverbose)
567 sc->acpi_verbose = 1;
568 if ((env = getenv("hw.acpi.verbose")) != NULL) {
569 if (strcmp(env, "") != 0)
570 sc->acpi_verbose = 1;
571 freeenv(env);
572 }
573
574 /* Only enable S4BIOS by default if the FACS says it is available. */
575 status = AcpiGetTable(ACPI_SIG_FACS, 0, (ACPI_TABLE_HEADER **)&facs);
576 if (ACPI_FAILURE(status)) {
577 device_printf(dev, "couldn't get FACS: %s\n",
578 AcpiFormatException(status));
579 error = ENXIO;
580 goto out;
581 }
582 if (facs->Flags & ACPI_FACS_S4_BIOS_PRESENT)
583 sc->acpi_s4bios = 1;
584
585 /*
586 * Dispatch the default sleep state to devices. The lid switch is set
587 * to NONE by default to avoid surprising users.
588 */
589 sc->acpi_power_button_sx = ACPI_STATE_S5;
590 sc->acpi_lid_switch_sx = ACPI_S_STATES_MAX + 1;
591 sc->acpi_standby_sx = ACPI_STATE_S1;
592 sc->acpi_suspend_sx = ACPI_STATE_S3;
593
594 /* Pick the first valid sleep state for the sleep button default. */
595 sc->acpi_sleep_button_sx = ACPI_S_STATES_MAX + 1;
596 for (state = ACPI_STATE_S1; state <= ACPI_STATE_S4; state++)
597 if (ACPI_SUCCESS(AcpiGetSleepTypeData(state, &TypeA, &TypeB))) {
598 sc->acpi_sleep_button_sx = state;
599 break;
600 }
601
602 acpi_enable_fixed_events(sc);
603
604 /*
605 * Scan the namespace and attach/initialise children.
606 */
607
608 /* Register our shutdown handler. */
609 EVENTHANDLER_REGISTER(shutdown_final, acpi_shutdown_final, sc,
610 SHUTDOWN_PRI_LAST);
611
612 /*
613 * Register our acpi event handlers.
614 * XXX should be configurable eg. via userland policy manager.
615 */
616 EVENTHANDLER_REGISTER(acpi_sleep_event, acpi_system_eventhandler_sleep,
617 sc, ACPI_EVENT_PRI_LAST);
618 EVENTHANDLER_REGISTER(acpi_wakeup_event, acpi_system_eventhandler_wakeup,
619 sc, ACPI_EVENT_PRI_LAST);
620
621 /* Flag our initial states. */
622 sc->acpi_enabled = 1;
623 sc->acpi_sstate = ACPI_STATE_S0;
624 sc->acpi_sleep_disabled = 0;
625
626 /* Create the control device */
627 sc->acpi_dev_t = make_dev(&acpi_cdevsw, 0, UID_ROOT, GID_WHEEL, 0644,
628 "acpi");
629 sc->acpi_dev_t->si_drv1 = sc;
630
631 if ((error = acpi_machdep_init(dev)))
632 goto out;
633
634 /* Register ACPI again to pass the correct argument of pm_func. */
635 power_pm_register(POWER_PM_TYPE_ACPI, acpi_pm_func, sc);
636
637 if (!acpi_disabled("bus"))
638 acpi_probe_children(dev);
639
640 error = 0;
641
642 out:
643 return_VALUE (error);
644 }
645
646 static int
647 acpi_suspend(device_t dev)
648 {
649 device_t child, *devlist;
650 int error, i, numdevs, pstate;
651
652 GIANT_REQUIRED;
653
654 /* First give child devices a chance to suspend. */
655 error = bus_generic_suspend(dev);
656 if (error)
657 return (error);
658
659 /*
660 * Now, set them into the appropriate power state, usually D3. If the
661 * device has an _SxD method for the next sleep state, use that power
662 * state instead.
663 */
664 device_get_children(dev, &devlist, &numdevs);
665 for (i = 0; i < numdevs; i++) {
666 /* If the device is not attached, we've powered it down elsewhere. */
667 child = devlist[i];
668 if (!device_is_attached(child))
669 continue;
670
671 /*
672 * Default to D3 for all sleep states. The _SxD method is optional
673 * so set the powerstate even if it's absent.
674 */
675 pstate = PCI_POWERSTATE_D3;
676 error = acpi_device_pwr_for_sleep(device_get_parent(child),
677 child, &pstate);
678 if ((error == 0 || error == ESRCH) && acpi_do_powerstate)
679 pci_set_powerstate(child, pstate);
680 }
681 free(devlist, M_TEMP);
682 error = 0;
683
684 return (error);
685 }
686
687 static int
688 acpi_resume(device_t dev)
689 {
690 ACPI_HANDLE handle;
691 int i, numdevs;
692 device_t child, *devlist;
693
694 GIANT_REQUIRED;
695
696 /*
697 * Put all devices in D0 before resuming them. Call _S0D on each one
698 * since some systems expect this.
699 */
700 device_get_children(dev, &devlist, &numdevs);
701 for (i = 0; i < numdevs; i++) {
702 child = devlist[i];
703 handle = acpi_get_handle(child);
704 if (handle)
705 AcpiEvaluateObject(handle, "_S0D", NULL, NULL);
706 if (device_is_attached(child) && acpi_do_powerstate)
707 pci_set_powerstate(child, PCI_POWERSTATE_D0);
708 }
709 free(devlist, M_TEMP);
710
711 return (bus_generic_resume(dev));
712 }
713
714 static int
715 acpi_shutdown(device_t dev)
716 {
717
718 GIANT_REQUIRED;
719
720 /* Allow children to shutdown first. */
721 bus_generic_shutdown(dev);
722
723 /*
724 * Enable any GPEs that are able to power-on the system (i.e., RTC).
725 * Also, disable any that are not valid for this state (most).
726 */
727 acpi_wake_prep_walk(ACPI_STATE_S5);
728
729 return (0);
730 }
731
732 /*
733 * Handle a new device being added
734 */
735 static device_t
736 acpi_add_child(device_t bus, int order, const char *name, int unit)
737 {
738 struct acpi_device *ad;
739 device_t child;
740
741 if ((ad = malloc(sizeof(*ad), M_ACPIDEV, M_NOWAIT | M_ZERO)) == NULL)
742 return (NULL);
743
744 resource_list_init(&ad->ad_rl);
745
746 child = device_add_child_ordered(bus, order, name, unit);
747 if (child != NULL)
748 device_set_ivars(child, ad);
749 else
750 free(ad, M_ACPIDEV);
751 return (child);
752 }
753
754 static int
755 acpi_print_child(device_t bus, device_t child)
756 {
757 struct acpi_device *adev = device_get_ivars(child);
758 struct resource_list *rl = &adev->ad_rl;
759 int retval = 0;
760
761 retval += bus_print_child_header(bus, child);
762 retval += resource_list_print_type(rl, "port", SYS_RES_IOPORT, "%#lx");
763 retval += resource_list_print_type(rl, "iomem", SYS_RES_MEMORY, "%#lx");
764 retval += resource_list_print_type(rl, "irq", SYS_RES_IRQ, "%ld");
765 retval += resource_list_print_type(rl, "drq", SYS_RES_DRQ, "%ld");
766 if (device_get_flags(child))
767 retval += printf(" flags %#x", device_get_flags(child));
768 retval += bus_print_child_footer(bus, child);
769
770 return (retval);
771 }
772
773 /*
774 * If this device is an ACPI child but no one claimed it, attempt
775 * to power it off. We'll power it back up when a driver is added.
776 *
777 * XXX Disabled for now since many necessary devices (like fdc and
778 * ATA) don't claim the devices we created for them but still expect
779 * them to be powered up.
780 */
781 static void
782 acpi_probe_nomatch(device_t bus, device_t child)
783 {
784
785 /* pci_set_powerstate(child, PCI_POWERSTATE_D3); */
786 }
787
788 /*
789 * If a new driver has a chance to probe a child, first power it up.
790 *
791 * XXX Disabled for now (see acpi_probe_nomatch for details).
792 */
793 static void
794 acpi_driver_added(device_t dev, driver_t *driver)
795 {
796 device_t child, *devlist;
797 int i, numdevs;
798
799 DEVICE_IDENTIFY(driver, dev);
800 device_get_children(dev, &devlist, &numdevs);
801 for (i = 0; i < numdevs; i++) {
802 child = devlist[i];
803 if (device_get_state(child) == DS_NOTPRESENT) {
804 /* pci_set_powerstate(child, PCI_POWERSTATE_D0); */
805 if (device_probe_and_attach(child) != 0)
806 ; /* pci_set_powerstate(child, PCI_POWERSTATE_D3); */
807 }
808 }
809 free(devlist, M_TEMP);
810 }
811
812 /* Location hint for devctl(8) */
813 static int
814 acpi_child_location_str_method(device_t cbdev, device_t child, char *buf,
815 size_t buflen)
816 {
817 struct acpi_device *dinfo = device_get_ivars(child);
818
819 if (dinfo->ad_handle)
820 snprintf(buf, buflen, "handle=%s", acpi_name(dinfo->ad_handle));
821 else
822 snprintf(buf, buflen, "unknown");
823 return (0);
824 }
825
826 /* PnP information for devctl(8) */
827 static int
828 acpi_child_pnpinfo_str_method(device_t cbdev, device_t child, char *buf,
829 size_t buflen)
830 {
831 ACPI_BUFFER adbuf = {ACPI_ALLOCATE_BUFFER, NULL};
832 ACPI_DEVICE_INFO *adinfo;
833 struct acpi_device *dinfo = device_get_ivars(child);
834 char *end;
835 int error;
836
837 error = AcpiGetObjectInfo(dinfo->ad_handle, &adbuf);
838 adinfo = (ACPI_DEVICE_INFO *) adbuf.Pointer;
839 if (error)
840 snprintf(buf, buflen, "unknown");
841 else
842 snprintf(buf, buflen, "_HID=%s _UID=%lu",
843 (adinfo->Valid & ACPI_VALID_HID) ?
844 adinfo->HardwareId.Value : "none",
845 (adinfo->Valid & ACPI_VALID_UID) ?
846 strtoul(adinfo->UniqueId.Value, &end, 10) : 0);
847 if (adinfo)
848 AcpiOsFree(adinfo);
849
850 return (0);
851 }
852
853 /*
854 * Handle per-device ivars
855 */
856 static int
857 acpi_read_ivar(device_t dev, device_t child, int index, uintptr_t *result)
858 {
859 struct acpi_device *ad;
860
861 if ((ad = device_get_ivars(child)) == NULL) {
862 printf("device has no ivars\n");
863 return (ENOENT);
864 }
865
866 /* ACPI and ISA compatibility ivars */
867 switch(index) {
868 case ACPI_IVAR_HANDLE:
869 *(ACPI_HANDLE *)result = ad->ad_handle;
870 break;
871 case ACPI_IVAR_MAGIC:
872 *(uintptr_t *)result = ad->ad_magic;
873 break;
874 case ACPI_IVAR_PRIVATE:
875 *(void **)result = ad->ad_private;
876 break;
877 case ACPI_IVAR_FLAGS:
878 *(int *)result = ad->ad_flags;
879 break;
880 case ISA_IVAR_VENDORID:
881 case ISA_IVAR_SERIAL:
882 case ISA_IVAR_COMPATID:
883 *(int *)result = -1;
884 break;
885 case ISA_IVAR_LOGICALID:
886 *(int *)result = acpi_isa_get_logicalid(child);
887 break;
888 default:
889 return (ENOENT);
890 }
891
892 return (0);
893 }
894
895 static int
896 acpi_write_ivar(device_t dev, device_t child, int index, uintptr_t value)
897 {
898 struct acpi_device *ad;
899
900 if ((ad = device_get_ivars(child)) == NULL) {
901 printf("device has no ivars\n");
902 return (ENOENT);
903 }
904
905 switch(index) {
906 case ACPI_IVAR_HANDLE:
907 ad->ad_handle = (ACPI_HANDLE)value;
908 break;
909 case ACPI_IVAR_MAGIC:
910 ad->ad_magic = (uintptr_t)value;
911 break;
912 case ACPI_IVAR_PRIVATE:
913 ad->ad_private = (void *)value;
914 break;
915 case ACPI_IVAR_FLAGS:
916 ad->ad_flags = (int)value;
917 break;
918 default:
919 panic("bad ivar write request (%d)", index);
920 return (ENOENT);
921 }
922
923 return (0);
924 }
925
926 /*
927 * Handle child resource allocation/removal
928 */
929 static struct resource_list *
930 acpi_get_rlist(device_t dev, device_t child)
931 {
932 struct acpi_device *ad;
933
934 ad = device_get_ivars(child);
935 return (&ad->ad_rl);
936 }
937
938 /*
939 * Pre-allocate/manage all memory and IO resources. Since rman can't handle
940 * duplicates, we merge any in the sysresource attach routine.
941 */
942 static int
943 acpi_sysres_alloc(device_t dev)
944 {
945 struct resource *res;
946 struct resource_list *rl;
947 struct resource_list_entry *rle;
948 struct rman *rm;
949 char *sysres_ids[] = { "PNP0C01", "PNP0C02", NULL };
950 device_t *children;
951 int child_count, i;
952
953 /*
954 * Probe/attach any sysresource devices. This would be unnecessary if we
955 * had multi-pass probe/attach.
956 */
957 if (device_get_children(dev, &children, &child_count) != 0)
958 return (ENXIO);
959 for (i = 0; i < child_count; i++) {
960 if (ACPI_ID_PROBE(dev, children[i], sysres_ids) != NULL)
961 device_probe_and_attach(children[i]);
962 }
963 free(children, M_TEMP);
964
965 rl = BUS_GET_RESOURCE_LIST(device_get_parent(dev), dev);
966 STAILQ_FOREACH(rle, rl, link) {
967 if (rle->res != NULL) {
968 device_printf(dev, "duplicate resource for %lx\n", rle->start);
969 continue;
970 }
971
972 /* Only memory and IO resources are valid here. */
973 switch (rle->type) {
974 case SYS_RES_IOPORT:
975 rm = &acpi_rman_io;
976 break;
977 case SYS_RES_MEMORY:
978 rm = &acpi_rman_mem;
979 break;
980 default:
981 continue;
982 }
983
984 /* Pre-allocate resource and add to our rman pool. */
985 res = BUS_ALLOC_RESOURCE(device_get_parent(dev), dev, rle->type,
986 &rle->rid, rle->start, rle->start + rle->count - 1, rle->count, 0);
987 if (res != NULL) {
988 rman_manage_region(rm, rman_get_start(res), rman_get_end(res));
989 rle->res = res;
990 } else
991 device_printf(dev, "reservation of %lx, %lx (%d) failed\n",
992 rle->start, rle->count, rle->type);
993 }
994 return (0);
995 }
996
997 static struct resource *
998 acpi_alloc_resource(device_t bus, device_t child, int type, int *rid,
999 u_long start, u_long end, u_long count, u_int flags)
1000 {
1001 ACPI_RESOURCE ares;
1002 struct acpi_device *ad = device_get_ivars(child);
1003 struct resource_list *rl = &ad->ad_rl;
1004 struct resource_list_entry *rle;
1005 struct resource *res;
1006 struct rman *rm;
1007
1008 res = NULL;
1009
1010 /* We only handle memory and IO resources through rman. */
1011 switch (type) {
1012 case SYS_RES_IOPORT:
1013 rm = &acpi_rman_io;
1014 break;
1015 case SYS_RES_MEMORY:
1016 rm = &acpi_rman_mem;
1017 break;
1018 default:
1019 rm = NULL;
1020 }
1021
1022 ACPI_SERIAL_BEGIN(acpi);
1023
1024 /*
1025 * If this is an allocation of the "default" range for a given RID, and
1026 * we know what the resources for this device are (i.e., they're on the
1027 * child's resource list), use those start/end values.
1028 */
1029 if (bus == device_get_parent(child) && start == 0UL && end == ~0UL) {
1030 rle = resource_list_find(rl, type, *rid);
1031 if (rle == NULL)
1032 goto out;
1033 start = rle->start;
1034 end = rle->end;
1035 count = rle->count;
1036 }
1037
1038 /*
1039 * If this is an allocation of a specific range, see if we can satisfy
1040 * the request from our system resource regions. If we can't, pass the
1041 * request up to the parent.
1042 */
1043 if (start + count - 1 == end && rm != NULL)
1044 res = rman_reserve_resource(rm, start, end, count, flags & ~RF_ACTIVE,
1045 child);
1046 if (res == NULL) {
1047 res = BUS_ALLOC_RESOURCE(device_get_parent(bus), child, type, rid,
1048 start, end, count, flags);
1049 } else {
1050 rman_set_rid(res, *rid);
1051
1052 /* If requested, activate the resource using the parent's method. */
1053 if (flags & RF_ACTIVE)
1054 if (bus_activate_resource(child, type, *rid, res) != 0) {
1055 rman_release_resource(res);
1056 res = NULL;
1057 goto out;
1058 }
1059 }
1060
1061 if (res != NULL && device_get_parent(child) == bus)
1062 switch (type) {
1063 case SYS_RES_IRQ:
1064 /*
1065 * Since bus_config_intr() takes immediate effect, we cannot
1066 * configure the interrupt associated with a device when we
1067 * parse the resources but have to defer it until a driver
1068 * actually allocates the interrupt via bus_alloc_resource().
1069 *
1070 * XXX: Should we handle the lookup failing?
1071 */
1072 if (ACPI_SUCCESS(acpi_lookup_irq_resource(child, *rid, res, &ares)))
1073 acpi_config_intr(child, &ares);
1074 break;
1075 }
1076
1077 out:
1078 ACPI_SERIAL_END(acpi);
1079 return (res);
1080 }
1081
1082 static int
1083 acpi_release_resource(device_t bus, device_t child, int type, int rid,
1084 struct resource *r)
1085 {
1086 struct rman *rm;
1087 int ret;
1088
1089 /* We only handle memory and IO resources through rman. */
1090 switch (type) {
1091 case SYS_RES_IOPORT:
1092 rm = &acpi_rman_io;
1093 break;
1094 case SYS_RES_MEMORY:
1095 rm = &acpi_rman_mem;
1096 break;
1097 default:
1098 rm = NULL;
1099 }
1100
1101 ACPI_SERIAL_BEGIN(acpi);
1102
1103 /*
1104 * If this resource belongs to one of our internal managers,
1105 * deactivate it and release it to the local pool. If it doesn't,
1106 * pass this request up to the parent.
1107 */
1108 if (rm != NULL && rman_is_region_manager(r, rm)) {
1109 if (rman_get_flags(r) & RF_ACTIVE) {
1110 ret = bus_deactivate_resource(child, type, rid, r);
1111 if (ret != 0)
1112 goto out;
1113 }
1114 ret = rman_release_resource(r);
1115 } else
1116 ret = BUS_RELEASE_RESOURCE(device_get_parent(bus), child, type, rid, r);
1117
1118 out:
1119 ACPI_SERIAL_END(acpi);
1120 return (ret);
1121 }
1122
1123 static void
1124 acpi_delete_resource(device_t bus, device_t child, int type, int rid)
1125 {
1126 struct resource_list *rl;
1127
1128 rl = acpi_get_rlist(bus, child);
1129 resource_list_delete(rl, type, rid);
1130 }
1131
1132 /* Allocate an IO port or memory resource, given its GAS. */
1133 int
1134 acpi_bus_alloc_gas(device_t dev, int *type, int *rid, ACPI_GENERIC_ADDRESS *gas,
1135 struct resource **res, u_int flags)
1136 {
1137 int error, res_type;
1138
1139 error = ENOMEM;
1140 if (type == NULL || rid == NULL || gas == NULL || res == NULL)
1141 return (EINVAL);
1142
1143 /* We only support memory and IO spaces. */
1144 switch (gas->SpaceId) {
1145 case ACPI_ADR_SPACE_SYSTEM_MEMORY:
1146 res_type = SYS_RES_MEMORY;
1147 break;
1148 case ACPI_ADR_SPACE_SYSTEM_IO:
1149 res_type = SYS_RES_IOPORT;
1150 break;
1151 default:
1152 return (EOPNOTSUPP);
1153 }
1154
1155 /*
1156 * If the register width is less than 8, assume the BIOS author means
1157 * it is a bit field and just allocate a byte.
1158 */
1159 if (gas->BitWidth && gas->BitWidth < 8)
1160 gas->BitWidth = 8;
1161
1162 /* Validate the address after we're sure we support the space. */
1163 if (gas->Address == 0 || gas->BitWidth == 0)
1164 return (EINVAL);
1165
1166 bus_set_resource(dev, res_type, *rid, gas->Address,
1167 gas->BitWidth / 8);
1168 *res = bus_alloc_resource_any(dev, res_type, rid, RF_ACTIVE | flags);
1169 if (*res != NULL) {
1170 *type = res_type;
1171 error = 0;
1172 } else
1173 bus_delete_resource(dev, res_type, *rid);
1174
1175 return (error);
1176 }
1177
1178 /* Probe _HID and _CID for compatible ISA PNP ids. */
1179 static uint32_t
1180 acpi_isa_get_logicalid(device_t dev)
1181 {
1182 ACPI_DEVICE_INFO *devinfo;
1183 ACPI_BUFFER buf;
1184 ACPI_HANDLE h;
1185 ACPI_STATUS error;
1186 u_int32_t pnpid;
1187
1188 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1189
1190 pnpid = 0;
1191 buf.Pointer = NULL;
1192 buf.Length = ACPI_ALLOCATE_BUFFER;
1193
1194 /* Fetch and validate the HID. */
1195 if ((h = acpi_get_handle(dev)) == NULL)
1196 goto out;
1197 error = AcpiGetObjectInfo(h, &buf);
1198 if (ACPI_FAILURE(error))
1199 goto out;
1200 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1201
1202 if ((devinfo->Valid & ACPI_VALID_HID) != 0)
1203 pnpid = PNP_EISAID(devinfo->HardwareId.Value);
1204
1205 out:
1206 if (buf.Pointer != NULL)
1207 AcpiOsFree(buf.Pointer);
1208 return_VALUE (pnpid);
1209 }
1210
1211 static int
1212 acpi_isa_get_compatid(device_t dev, uint32_t *cids, int count)
1213 {
1214 ACPI_DEVICE_INFO *devinfo;
1215 ACPI_BUFFER buf;
1216 ACPI_HANDLE h;
1217 ACPI_STATUS error;
1218 uint32_t *pnpid;
1219 int valid, i;
1220
1221 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1222
1223 pnpid = cids;
1224 valid = 0;
1225 buf.Pointer = NULL;
1226 buf.Length = ACPI_ALLOCATE_BUFFER;
1227
1228 /* Fetch and validate the CID */
1229 if ((h = acpi_get_handle(dev)) == NULL)
1230 goto out;
1231 error = AcpiGetObjectInfo(h, &buf);
1232 if (ACPI_FAILURE(error))
1233 goto out;
1234 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1235 if ((devinfo->Valid & ACPI_VALID_CID) == 0)
1236 goto out;
1237
1238 if (devinfo->CompatibilityId.Count < count)
1239 count = devinfo->CompatibilityId.Count;
1240 for (i = 0; i < count; i++) {
1241 if (strncmp(devinfo->CompatibilityId.Id[i].Value, "PNP", 3) != 0)
1242 continue;
1243 *pnpid++ = PNP_EISAID(devinfo->CompatibilityId.Id[i].Value);
1244 valid++;
1245 }
1246
1247 out:
1248 if (buf.Pointer != NULL)
1249 AcpiOsFree(buf.Pointer);
1250 return_VALUE (valid);
1251 }
1252
1253 static char *
1254 acpi_device_id_probe(device_t bus, device_t dev, char **ids)
1255 {
1256 ACPI_HANDLE h;
1257 int i;
1258
1259 h = acpi_get_handle(dev);
1260 if (ids == NULL || h == NULL || acpi_get_type(dev) != ACPI_TYPE_DEVICE)
1261 return (NULL);
1262
1263 /* Try to match one of the array of IDs with a HID or CID. */
1264 for (i = 0; ids[i] != NULL; i++) {
1265 if (acpi_MatchHid(h, ids[i]))
1266 return (ids[i]);
1267 }
1268 return (NULL);
1269 }
1270
1271 static ACPI_STATUS
1272 acpi_device_eval_obj(device_t bus, device_t dev, ACPI_STRING pathname,
1273 ACPI_OBJECT_LIST *parameters, ACPI_BUFFER *ret)
1274 {
1275 ACPI_HANDLE h;
1276
1277 if (dev == NULL)
1278 h = ACPI_ROOT_OBJECT;
1279 else if ((h = acpi_get_handle(dev)) == NULL)
1280 return (AE_BAD_PARAMETER);
1281 return (AcpiEvaluateObject(h, pathname, parameters, ret));
1282 }
1283
1284 static int
1285 acpi_device_pwr_for_sleep(device_t bus, device_t dev, int *dstate)
1286 {
1287 struct acpi_softc *sc;
1288 ACPI_HANDLE handle;
1289 ACPI_STATUS status;
1290 char sxd[8];
1291 int error;
1292
1293 sc = device_get_softc(bus);
1294 handle = acpi_get_handle(dev);
1295
1296 /*
1297 * XXX If we find these devices, don't try to power them down.
1298 * The serial and IRDA ports on my T23 hang the system when
1299 * set to D3 and it appears that such legacy devices may
1300 * need special handling in their drivers.
1301 */
1302 if (handle == NULL ||
1303 acpi_MatchHid(handle, "PNP0500") ||
1304 acpi_MatchHid(handle, "PNP0501") ||
1305 acpi_MatchHid(handle, "PNP0502") ||
1306 acpi_MatchHid(handle, "PNP0510") ||
1307 acpi_MatchHid(handle, "PNP0511"))
1308 return (ENXIO);
1309
1310 /*
1311 * Override next state with the value from _SxD, if present. If no
1312 * dstate argument was provided, don't fetch the return value.
1313 */
1314 snprintf(sxd, sizeof(sxd), "_S%dD", sc->acpi_sstate);
1315 if (dstate)
1316 status = acpi_GetInteger(handle, sxd, dstate);
1317 else
1318 status = AcpiEvaluateObject(handle, sxd, NULL, NULL);
1319
1320 switch (status) {
1321 case AE_OK:
1322 error = 0;
1323 break;
1324 case AE_NOT_FOUND:
1325 error = ESRCH;
1326 break;
1327 default:
1328 error = ENXIO;
1329 break;
1330 }
1331
1332 return (error);
1333 }
1334
1335 /* Callback arg for our implementation of walking the namespace. */
1336 struct acpi_device_scan_ctx {
1337 acpi_scan_cb_t user_fn;
1338 void *arg;
1339 ACPI_HANDLE parent;
1340 };
1341
1342 static ACPI_STATUS
1343 acpi_device_scan_cb(ACPI_HANDLE h, UINT32 level, void *arg, void **retval)
1344 {
1345 struct acpi_device_scan_ctx *ctx;
1346 device_t dev, old_dev;
1347 ACPI_STATUS status;
1348 ACPI_OBJECT_TYPE type;
1349
1350 /*
1351 * Skip this device if we think we'll have trouble with it or it is
1352 * the parent where the scan began.
1353 */
1354 ctx = (struct acpi_device_scan_ctx *)arg;
1355 if (acpi_avoid(h) || h == ctx->parent)
1356 return (AE_OK);
1357
1358 /* If this is not a valid device type (e.g., a method), skip it. */
1359 if (ACPI_FAILURE(AcpiGetType(h, &type)))
1360 return (AE_OK);
1361 if (type != ACPI_TYPE_DEVICE && type != ACPI_TYPE_PROCESSOR &&
1362 type != ACPI_TYPE_THERMAL && type != ACPI_TYPE_POWER)
1363 return (AE_OK);
1364
1365 /*
1366 * Call the user function with the current device. If it is unchanged
1367 * afterwards, return. Otherwise, we update the handle to the new dev.
1368 */
1369 old_dev = acpi_get_device(h);
1370 dev = old_dev;
1371 status = ctx->user_fn(h, &dev, level, ctx->arg);
1372 if (ACPI_FAILURE(status) || old_dev == dev)
1373 return (status);
1374
1375 /* Remove the old child and its connection to the handle. */
1376 if (old_dev != NULL) {
1377 device_delete_child(device_get_parent(old_dev), old_dev);
1378 AcpiDetachData(h, acpi_fake_objhandler);
1379 }
1380
1381 /* Recreate the handle association if the user created a device. */
1382 if (dev != NULL)
1383 AcpiAttachData(h, acpi_fake_objhandler, dev);
1384
1385 return (AE_OK);
1386 }
1387
1388 static ACPI_STATUS
1389 acpi_device_scan_children(device_t bus, device_t dev, int max_depth,
1390 acpi_scan_cb_t user_fn, void *arg)
1391 {
1392 ACPI_HANDLE h;
1393 struct acpi_device_scan_ctx ctx;
1394
1395 if (acpi_disabled("children"))
1396 return (AE_OK);
1397
1398 if (dev == NULL)
1399 h = ACPI_ROOT_OBJECT;
1400 else if ((h = acpi_get_handle(dev)) == NULL)
1401 return (AE_BAD_PARAMETER);
1402 ctx.user_fn = user_fn;
1403 ctx.arg = arg;
1404 ctx.parent = h;
1405 return (AcpiWalkNamespace(ACPI_TYPE_ANY, h, max_depth,
1406 acpi_device_scan_cb, &ctx, NULL));
1407 }
1408
1409 /*
1410 * Even though ACPI devices are not PCI, we use the PCI approach for setting
1411 * device power states since it's close enough to ACPI.
1412 */
1413 static int
1414 acpi_set_powerstate_method(device_t bus, device_t child, int state)
1415 {
1416 ACPI_HANDLE h;
1417 ACPI_STATUS status;
1418 int error;
1419
1420 error = 0;
1421 h = acpi_get_handle(child);
1422 if (state < ACPI_STATE_D0 || state > ACPI_STATE_D3)
1423 return (EINVAL);
1424 if (h == NULL)
1425 return (0);
1426
1427 /* Ignore errors if the power methods aren't present. */
1428 status = acpi_pwr_switch_consumer(h, state);
1429 if (ACPI_FAILURE(status) && status != AE_NOT_FOUND
1430 && status != AE_BAD_PARAMETER)
1431 device_printf(bus, "failed to set ACPI power state D%d on %s: %s\n",
1432 state, acpi_name(h), AcpiFormatException(status));
1433
1434 return (error);
1435 }
1436
1437 static int
1438 acpi_isa_pnp_probe(device_t bus, device_t child, struct isa_pnp_id *ids)
1439 {
1440 int result, cid_count, i;
1441 uint32_t lid, cids[8];
1442
1443 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1444
1445 /*
1446 * ISA-style drivers attached to ACPI may persist and
1447 * probe manually if we return ENOENT. We never want
1448 * that to happen, so don't ever return it.
1449 */
1450 result = ENXIO;
1451
1452 /* Scan the supplied IDs for a match */
1453 lid = acpi_isa_get_logicalid(child);
1454 cid_count = acpi_isa_get_compatid(child, cids, 8);
1455 while (ids && ids->ip_id) {
1456 if (lid == ids->ip_id) {
1457 result = 0;
1458 goto out;
1459 }
1460 for (i = 0; i < cid_count; i++) {
1461 if (cids[i] == ids->ip_id) {
1462 result = 0;
1463 goto out;
1464 }
1465 }
1466 ids++;
1467 }
1468
1469 out:
1470 if (result == 0 && ids->ip_desc)
1471 device_set_desc(child, ids->ip_desc);
1472
1473 return_VALUE (result);
1474 }
1475
1476 /*
1477 * Scan all of the ACPI namespace and attach child devices.
1478 *
1479 * We should only expect to find devices in the \_PR, \_TZ, \_SI, and
1480 * \_SB scopes, and \_PR and \_TZ became obsolete in the ACPI 2.0 spec.
1481 * However, in violation of the spec, some systems place their PCI link
1482 * devices in \, so we have to walk the whole namespace. We check the
1483 * type of namespace nodes, so this should be ok.
1484 */
1485 static void
1486 acpi_probe_children(device_t bus)
1487 {
1488
1489 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1490
1491 /*
1492 * Scan the namespace and insert placeholders for all the devices that
1493 * we find. We also probe/attach any early devices.
1494 *
1495 * Note that we use AcpiWalkNamespace rather than AcpiGetDevices because
1496 * we want to create nodes for all devices, not just those that are
1497 * currently present. (This assumes that we don't want to create/remove
1498 * devices as they appear, which might be smarter.)
1499 */
1500 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "namespace scan\n"));
1501 AcpiWalkNamespace(ACPI_TYPE_ANY, ACPI_ROOT_OBJECT, 100, acpi_probe_child,
1502 bus, NULL);
1503
1504 /* Pre-allocate resources for our rman from any sysresource devices. */
1505 acpi_sysres_alloc(bus);
1506
1507 /* Create any static children by calling device identify methods. */
1508 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "device identify routines\n"));
1509 bus_generic_probe(bus);
1510
1511 /* Probe/attach all children, created staticly and from the namespace. */
1512 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "first bus_generic_attach\n"));
1513 bus_generic_attach(bus);
1514
1515 /*
1516 * Some of these children may have attached others as part of their attach
1517 * process (eg. the root PCI bus driver), so rescan.
1518 */
1519 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "second bus_generic_attach\n"));
1520 bus_generic_attach(bus);
1521
1522 /* Attach wake sysctls. */
1523 acpi_wake_sysctl_walk(bus);
1524
1525 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "done attaching children\n"));
1526 return_VOID;
1527 }
1528
1529 /*
1530 * Determine the probe order for a given device and return non-zero if it
1531 * should be attached immediately.
1532 */
1533 static int
1534 acpi_probe_order(ACPI_HANDLE handle, int *order)
1535 {
1536
1537 /*
1538 * 1. I/O port and memory system resource holders
1539 * 2. Embedded controllers (to handle early accesses)
1540 * 3. PCI Link Devices
1541 */
1542 if (acpi_MatchHid(handle, "PNP0C01") || acpi_MatchHid(handle, "PNP0C02"))
1543 *order = 1;
1544 else if (acpi_MatchHid(handle, "PNP0C09"))
1545 *order = 2;
1546 else if (acpi_MatchHid(handle, "PNP0C0F"))
1547 *order = 3;
1548 return (0);
1549 }
1550
1551 /*
1552 * Evaluate a child device and determine whether we might attach a device to
1553 * it.
1554 */
1555 static ACPI_STATUS
1556 acpi_probe_child(ACPI_HANDLE handle, UINT32 level, void *context, void **status)
1557 {
1558 ACPI_OBJECT_TYPE type;
1559 ACPI_HANDLE h;
1560 device_t bus, child;
1561 int order;
1562 char *handle_str, **search;
1563 static char *scopes[] = {"\\_PR_", "\\_TZ_", "\\_SI_", "\\_SB_", NULL};
1564
1565 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1566
1567 /* Skip this device if we think we'll have trouble with it. */
1568 if (acpi_avoid(handle))
1569 return_ACPI_STATUS (AE_OK);
1570
1571 bus = (device_t)context;
1572 if (ACPI_SUCCESS(AcpiGetType(handle, &type))) {
1573 switch (type) {
1574 case ACPI_TYPE_DEVICE:
1575 case ACPI_TYPE_PROCESSOR:
1576 case ACPI_TYPE_THERMAL:
1577 case ACPI_TYPE_POWER:
1578 if (acpi_disabled("children"))
1579 break;
1580
1581 /*
1582 * Since we scan from \, be sure to skip system scope objects.
1583 * At least \_SB and \_TZ are detected as devices (ACPI-CA bug?)
1584 */
1585 handle_str = acpi_name(handle);
1586 for (search = scopes; *search != NULL; search++) {
1587 if (strcmp(handle_str, *search) == 0)
1588 break;
1589 }
1590 if (*search != NULL)
1591 break;
1592
1593 /*
1594 * Create a placeholder device for this node. Sort the placeholder
1595 * so that the probe/attach passes will run breadth-first. Orders
1596 * less than ACPI_DEV_BASE_ORDER are reserved for special objects
1597 * (i.e., system resources). Larger values are used for all other
1598 * devices.
1599 */
1600 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "scanning '%s'\n", handle_str));
1601 order = (level + 1) * ACPI_DEV_BASE_ORDER;
1602 acpi_probe_order(handle, &order);
1603 child = BUS_ADD_CHILD(bus, order, NULL, -1);
1604 if (child == NULL)
1605 break;
1606
1607 /* Associate the handle with the device_t and vice versa. */
1608 acpi_set_handle(child, handle);
1609 AcpiAttachData(handle, acpi_fake_objhandler, child);
1610
1611 /*
1612 * Check that the device is present. If it's not present,
1613 * leave it disabled (so that we have a device_t attached to
1614 * the handle, but we don't probe it).
1615 *
1616 * XXX PCI link devices sometimes report "present" but not
1617 * "functional" (i.e. if disabled). Go ahead and probe them
1618 * anyway since we may enable them later.
1619 */
1620 if (type == ACPI_TYPE_DEVICE && !acpi_DeviceIsPresent(child)) {
1621 /* Never disable PCI link devices. */
1622 if (acpi_MatchHid(handle, "PNP0C0F"))
1623 break;
1624 /*
1625 * Docking stations should remain enabled since the system
1626 * may be undocked at boot.
1627 */
1628 if (ACPI_SUCCESS(AcpiGetHandle(handle, "_DCK", &h)))
1629 break;
1630
1631 device_disable(child);
1632 break;
1633 }
1634
1635 /*
1636 * Get the device's resource settings and attach them.
1637 * Note that if the device has _PRS but no _CRS, we need
1638 * to decide when it's appropriate to try to configure the
1639 * device. Ignore the return value here; it's OK for the
1640 * device not to have any resources.
1641 */
1642 acpi_parse_resources(child, handle, &acpi_res_parse_set, NULL);
1643 break;
1644 }
1645 }
1646
1647 return_ACPI_STATUS (AE_OK);
1648 }
1649
1650 /*
1651 * AcpiAttachData() requires an object handler but never uses it. This is a
1652 * placeholder object handler so we can store a device_t in an ACPI_HANDLE.
1653 */
1654 void
1655 acpi_fake_objhandler(ACPI_HANDLE h, UINT32 fn, void *data)
1656 {
1657 }
1658
1659 static void
1660 acpi_shutdown_final(void *arg, int howto)
1661 {
1662 struct acpi_softc *sc;
1663 ACPI_STATUS status;
1664
1665 /*
1666 * XXX Shutdown code should only run on the BSP (cpuid 0).
1667 * Some chipsets do not power off the system correctly if called from
1668 * an AP.
1669 */
1670 sc = arg;
1671 if ((howto & RB_POWEROFF) != 0) {
1672 status = AcpiEnterSleepStatePrep(ACPI_STATE_S5);
1673 if (ACPI_FAILURE(status)) {
1674 printf("AcpiEnterSleepStatePrep failed - %s\n",
1675 AcpiFormatException(status));
1676 return;
1677 }
1678 printf("Powering system off using ACPI\n");
1679 ACPI_DISABLE_IRQS();
1680 status = AcpiEnterSleepState(ACPI_STATE_S5);
1681 if (ACPI_FAILURE(status)) {
1682 printf("ACPI power-off failed - %s\n", AcpiFormatException(status));
1683 } else {
1684 DELAY(1000000);
1685 printf("ACPI power-off failed - timeout\n");
1686 }
1687 } else if ((howto & RB_HALT) == 0 &&
1688 (AcpiGbl_FADT.Flags & ACPI_FADT_RESET_REGISTER) &&
1689 sc->acpi_handle_reboot) {
1690 /* Reboot using the reset register. */
1691 status = AcpiHwLowLevelWrite(
1692 AcpiGbl_FADT.ResetRegister.BitWidth,
1693 AcpiGbl_FADT.ResetValue, &AcpiGbl_FADT.ResetRegister);
1694 if (ACPI_FAILURE(status)) {
1695 printf("ACPI reset failed - %s\n", AcpiFormatException(status));
1696 } else {
1697 DELAY(1000000);
1698 printf("ACPI reset failed - timeout\n");
1699 }
1700 } else if (sc->acpi_do_disable && panicstr == NULL) {
1701 /*
1702 * Only disable ACPI if the user requested. On some systems, writing
1703 * the disable value to SMI_CMD hangs the system.
1704 */
1705 printf("Shutting down ACPI\n");
1706 AcpiTerminate();
1707 }
1708 }
1709
1710 static void
1711 acpi_enable_fixed_events(struct acpi_softc *sc)
1712 {
1713 static int first_time = 1;
1714
1715 /* Enable and clear fixed events and install handlers. */
1716 if ((AcpiGbl_FADT.Flags & ACPI_FADT_POWER_BUTTON) == 0) {
1717 AcpiClearEvent(ACPI_EVENT_POWER_BUTTON);
1718 AcpiInstallFixedEventHandler(ACPI_EVENT_POWER_BUTTON,
1719 acpi_event_power_button_sleep, sc);
1720 if (first_time)
1721 device_printf(sc->acpi_dev, "Power Button (fixed)\n");
1722 }
1723 if ((AcpiGbl_FADT.Flags & ACPI_FADT_SLEEP_BUTTON) == 0) {
1724 AcpiClearEvent(ACPI_EVENT_SLEEP_BUTTON);
1725 AcpiInstallFixedEventHandler(ACPI_EVENT_SLEEP_BUTTON,
1726 acpi_event_sleep_button_sleep, sc);
1727 if (first_time)
1728 device_printf(sc->acpi_dev, "Sleep Button (fixed)\n");
1729 }
1730
1731 first_time = 0;
1732 }
1733
1734 /*
1735 * Returns true if the device is actually present and should
1736 * be attached to. This requires the present, enabled, UI-visible
1737 * and diagnostics-passed bits to be set.
1738 */
1739 BOOLEAN
1740 acpi_DeviceIsPresent(device_t dev)
1741 {
1742 ACPI_DEVICE_INFO *devinfo;
1743 ACPI_HANDLE h;
1744 ACPI_BUFFER buf;
1745 ACPI_STATUS error;
1746 int ret;
1747
1748 ret = FALSE;
1749 if ((h = acpi_get_handle(dev)) == NULL)
1750 return (FALSE);
1751 buf.Pointer = NULL;
1752 buf.Length = ACPI_ALLOCATE_BUFFER;
1753 error = AcpiGetObjectInfo(h, &buf);
1754 if (ACPI_FAILURE(error))
1755 return (FALSE);
1756 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1757
1758 /* If no _STA method, must be present */
1759 if ((devinfo->Valid & ACPI_VALID_STA) == 0)
1760 ret = TRUE;
1761
1762 /* Return true for 'present' and 'functioning' */
1763 if (ACPI_DEVICE_PRESENT(devinfo->CurrentStatus))
1764 ret = TRUE;
1765
1766 AcpiOsFree(buf.Pointer);
1767 return (ret);
1768 }
1769
1770 /*
1771 * Returns true if the battery is actually present and inserted.
1772 */
1773 BOOLEAN
1774 acpi_BatteryIsPresent(device_t dev)
1775 {
1776 ACPI_DEVICE_INFO *devinfo;
1777 ACPI_HANDLE h;
1778 ACPI_BUFFER buf;
1779 ACPI_STATUS error;
1780 int ret;
1781
1782 ret = FALSE;
1783 if ((h = acpi_get_handle(dev)) == NULL)
1784 return (FALSE);
1785 buf.Pointer = NULL;
1786 buf.Length = ACPI_ALLOCATE_BUFFER;
1787 error = AcpiGetObjectInfo(h, &buf);
1788 if (ACPI_FAILURE(error))
1789 return (FALSE);
1790 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1791
1792 /* If no _STA method, must be present */
1793 if ((devinfo->Valid & ACPI_VALID_STA) == 0)
1794 ret = TRUE;
1795
1796 /* Return true for 'present', 'battery present', and 'functioning' */
1797 if (ACPI_BATTERY_PRESENT(devinfo->CurrentStatus))
1798 ret = TRUE;
1799
1800 AcpiOsFree(buf.Pointer);
1801 return (ret);
1802 }
1803
1804 /*
1805 * Match a HID string against a handle
1806 */
1807 static BOOLEAN
1808 acpi_MatchHid(ACPI_HANDLE h, const char *hid)
1809 {
1810 ACPI_DEVICE_INFO *devinfo;
1811 ACPI_BUFFER buf;
1812 ACPI_STATUS error;
1813 int ret, i;
1814
1815 ret = FALSE;
1816 if (hid == NULL || h == NULL)
1817 return (ret);
1818 buf.Pointer = NULL;
1819 buf.Length = ACPI_ALLOCATE_BUFFER;
1820 error = AcpiGetObjectInfo(h, &buf);
1821 if (ACPI_FAILURE(error))
1822 return (ret);
1823 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1824
1825 if ((devinfo->Valid & ACPI_VALID_HID) != 0 &&
1826 strcmp(hid, devinfo->HardwareId.Value) == 0)
1827 ret = TRUE;
1828 else if ((devinfo->Valid & ACPI_VALID_CID) != 0) {
1829 for (i = 0; i < devinfo->CompatibilityId.Count; i++) {
1830 if (strcmp(hid, devinfo->CompatibilityId.Id[i].Value) == 0) {
1831 ret = TRUE;
1832 break;
1833 }
1834 }
1835 }
1836
1837 AcpiOsFree(buf.Pointer);
1838 return (ret);
1839 }
1840
1841 /*
1842 * Return the handle of a named object within our scope, ie. that of (parent)
1843 * or one if its parents.
1844 */
1845 ACPI_STATUS
1846 acpi_GetHandleInScope(ACPI_HANDLE parent, char *path, ACPI_HANDLE *result)
1847 {
1848 ACPI_HANDLE r;
1849 ACPI_STATUS status;
1850
1851 /* Walk back up the tree to the root */
1852 for (;;) {
1853 status = AcpiGetHandle(parent, path, &r);
1854 if (ACPI_SUCCESS(status)) {
1855 *result = r;
1856 return (AE_OK);
1857 }
1858 /* XXX Return error here? */
1859 if (status != AE_NOT_FOUND)
1860 return (AE_OK);
1861 if (ACPI_FAILURE(AcpiGetParent(parent, &r)))
1862 return (AE_NOT_FOUND);
1863 parent = r;
1864 }
1865 }
1866
1867 /* Find the difference between two PM tick counts. */
1868 uint32_t
1869 acpi_TimerDelta(uint32_t end, uint32_t start)
1870 {
1871 uint32_t delta;
1872
1873 if (end >= start)
1874 delta = end - start;
1875 else if (AcpiGbl_FADT.Flags & ACPI_FADT_32BIT_TIMER)
1876 delta = ((0xFFFFFFFF - start) + end + 1);
1877 else
1878 delta = ((0x00FFFFFF - start) + end + 1) & 0x00FFFFFF;
1879 return (delta);
1880 }
1881
1882 /*
1883 * Allocate a buffer with a preset data size.
1884 */
1885 ACPI_BUFFER *
1886 acpi_AllocBuffer(int size)
1887 {
1888 ACPI_BUFFER *buf;
1889
1890 if ((buf = malloc(size + sizeof(*buf), M_ACPIDEV, M_NOWAIT)) == NULL)
1891 return (NULL);
1892 buf->Length = size;
1893 buf->Pointer = (void *)(buf + 1);
1894 return (buf);
1895 }
1896
1897 ACPI_STATUS
1898 acpi_SetInteger(ACPI_HANDLE handle, char *path, UINT32 number)
1899 {
1900 ACPI_OBJECT arg1;
1901 ACPI_OBJECT_LIST args;
1902
1903 arg1.Type = ACPI_TYPE_INTEGER;
1904 arg1.Integer.Value = number;
1905 args.Count = 1;
1906 args.Pointer = &arg1;
1907
1908 return (AcpiEvaluateObject(handle, path, &args, NULL));
1909 }
1910
1911 /*
1912 * Evaluate a path that should return an integer.
1913 */
1914 ACPI_STATUS
1915 acpi_GetInteger(ACPI_HANDLE handle, char *path, UINT32 *number)
1916 {
1917 ACPI_STATUS status;
1918 ACPI_BUFFER buf;
1919 ACPI_OBJECT param;
1920
1921 if (handle == NULL)
1922 handle = ACPI_ROOT_OBJECT;
1923
1924 /*
1925 * Assume that what we've been pointed at is an Integer object, or
1926 * a method that will return an Integer.
1927 */
1928 buf.Pointer = ¶m;
1929 buf.Length = sizeof(param);
1930 status = AcpiEvaluateObject(handle, path, NULL, &buf);
1931 if (ACPI_SUCCESS(status)) {
1932 if (param.Type == ACPI_TYPE_INTEGER)
1933 *number = param.Integer.Value;
1934 else
1935 status = AE_TYPE;
1936 }
1937
1938 /*
1939 * In some applications, a method that's expected to return an Integer
1940 * may instead return a Buffer (probably to simplify some internal
1941 * arithmetic). We'll try to fetch whatever it is, and if it's a Buffer,
1942 * convert it into an Integer as best we can.
1943 *
1944 * This is a hack.
1945 */
1946 if (status == AE_BUFFER_OVERFLOW) {
1947 if ((buf.Pointer = AcpiOsAllocate(buf.Length)) == NULL) {
1948 status = AE_NO_MEMORY;
1949 } else {
1950 status = AcpiEvaluateObject(handle, path, NULL, &buf);
1951 if (ACPI_SUCCESS(status))
1952 status = acpi_ConvertBufferToInteger(&buf, number);
1953 AcpiOsFree(buf.Pointer);
1954 }
1955 }
1956 return (status);
1957 }
1958
1959 ACPI_STATUS
1960 acpi_ConvertBufferToInteger(ACPI_BUFFER *bufp, UINT32 *number)
1961 {
1962 ACPI_OBJECT *p;
1963 UINT8 *val;
1964 int i;
1965
1966 p = (ACPI_OBJECT *)bufp->Pointer;
1967 if (p->Type == ACPI_TYPE_INTEGER) {
1968 *number = p->Integer.Value;
1969 return (AE_OK);
1970 }
1971 if (p->Type != ACPI_TYPE_BUFFER)
1972 return (AE_TYPE);
1973 if (p->Buffer.Length > sizeof(int))
1974 return (AE_BAD_DATA);
1975
1976 *number = 0;
1977 val = p->Buffer.Pointer;
1978 for (i = 0; i < p->Buffer.Length; i++)
1979 *number += val[i] << (i * 8);
1980 return (AE_OK);
1981 }
1982
1983 /*
1984 * Iterate over the elements of an a package object, calling the supplied
1985 * function for each element.
1986 *
1987 * XXX possible enhancement might be to abort traversal on error.
1988 */
1989 ACPI_STATUS
1990 acpi_ForeachPackageObject(ACPI_OBJECT *pkg,
1991 void (*func)(ACPI_OBJECT *comp, void *arg), void *arg)
1992 {
1993 ACPI_OBJECT *comp;
1994 int i;
1995
1996 if (pkg == NULL || pkg->Type != ACPI_TYPE_PACKAGE)
1997 return (AE_BAD_PARAMETER);
1998
1999 /* Iterate over components */
2000 i = 0;
2001 comp = pkg->Package.Elements;
2002 for (; i < pkg->Package.Count; i++, comp++)
2003 func(comp, arg);
2004
2005 return (AE_OK);
2006 }
2007
2008 /*
2009 * Find the (index)th resource object in a set.
2010 */
2011 ACPI_STATUS
2012 acpi_FindIndexedResource(ACPI_BUFFER *buf, int index, ACPI_RESOURCE **resp)
2013 {
2014 ACPI_RESOURCE *rp;
2015 int i;
2016
2017 rp = (ACPI_RESOURCE *)buf->Pointer;
2018 i = index;
2019 while (i-- > 0) {
2020 /* Range check */
2021 if (rp > (ACPI_RESOURCE *)((u_int8_t *)buf->Pointer + buf->Length))
2022 return (AE_BAD_PARAMETER);
2023
2024 /* Check for terminator */
2025 if (rp->Type == ACPI_RESOURCE_TYPE_END_TAG || rp->Length == 0)
2026 return (AE_NOT_FOUND);
2027 rp = ACPI_NEXT_RESOURCE(rp);
2028 }
2029 if (resp != NULL)
2030 *resp = rp;
2031
2032 return (AE_OK);
2033 }
2034
2035 /*
2036 * Append an ACPI_RESOURCE to an ACPI_BUFFER.
2037 *
2038 * Given a pointer to an ACPI_RESOURCE structure, expand the ACPI_BUFFER
2039 * provided to contain it. If the ACPI_BUFFER is empty, allocate a sensible
2040 * backing block. If the ACPI_RESOURCE is NULL, return an empty set of
2041 * resources.
2042 */
2043 #define ACPI_INITIAL_RESOURCE_BUFFER_SIZE 512
2044
2045 ACPI_STATUS
2046 acpi_AppendBufferResource(ACPI_BUFFER *buf, ACPI_RESOURCE *res)
2047 {
2048 ACPI_RESOURCE *rp;
2049 void *newp;
2050
2051 /* Initialise the buffer if necessary. */
2052 if (buf->Pointer == NULL) {
2053 buf->Length = ACPI_INITIAL_RESOURCE_BUFFER_SIZE;
2054 if ((buf->Pointer = AcpiOsAllocate(buf->Length)) == NULL)
2055 return (AE_NO_MEMORY);
2056 rp = (ACPI_RESOURCE *)buf->Pointer;
2057 rp->Type = ACPI_RESOURCE_TYPE_END_TAG;
2058 rp->Length = 0;
2059 }
2060 if (res == NULL)
2061 return (AE_OK);
2062
2063 /*
2064 * Scan the current buffer looking for the terminator.
2065 * This will either find the terminator or hit the end
2066 * of the buffer and return an error.
2067 */
2068 rp = (ACPI_RESOURCE *)buf->Pointer;
2069 for (;;) {
2070 /* Range check, don't go outside the buffer */
2071 if (rp >= (ACPI_RESOURCE *)((u_int8_t *)buf->Pointer + buf->Length))
2072 return (AE_BAD_PARAMETER);
2073 if (rp->Type == ACPI_RESOURCE_TYPE_END_TAG || rp->Length == 0)
2074 break;
2075 rp = ACPI_NEXT_RESOURCE(rp);
2076 }
2077
2078 /*
2079 * Check the size of the buffer and expand if required.
2080 *
2081 * Required size is:
2082 * size of existing resources before terminator +
2083 * size of new resource and header +
2084 * size of terminator.
2085 *
2086 * Note that this loop should really only run once, unless
2087 * for some reason we are stuffing a *really* huge resource.
2088 */
2089 while ((((u_int8_t *)rp - (u_int8_t *)buf->Pointer) +
2090 res->Length + ACPI_RS_SIZE_NO_DATA +
2091 ACPI_RS_SIZE_MIN) >= buf->Length) {
2092 if ((newp = AcpiOsAllocate(buf->Length * 2)) == NULL)
2093 return (AE_NO_MEMORY);
2094 bcopy(buf->Pointer, newp, buf->Length);
2095 rp = (ACPI_RESOURCE *)((u_int8_t *)newp +
2096 ((u_int8_t *)rp - (u_int8_t *)buf->Pointer));
2097 AcpiOsFree(buf->Pointer);
2098 buf->Pointer = newp;
2099 buf->Length += buf->Length;
2100 }
2101
2102 /* Insert the new resource. */
2103 bcopy(res, rp, res->Length + ACPI_RS_SIZE_NO_DATA);
2104
2105 /* And add the terminator. */
2106 rp = ACPI_NEXT_RESOURCE(rp);
2107 rp->Type = ACPI_RESOURCE_TYPE_END_TAG;
2108 rp->Length = 0;
2109
2110 return (AE_OK);
2111 }
2112
2113 /*
2114 * Set interrupt model.
2115 */
2116 ACPI_STATUS
2117 acpi_SetIntrModel(int model)
2118 {
2119
2120 return (acpi_SetInteger(ACPI_ROOT_OBJECT, "_PIC", model));
2121 }
2122
2123 /*
2124 * DEPRECATED. This interface has serious deficiencies and will be
2125 * removed.
2126 *
2127 * Immediately enter the sleep state. In the old model, acpiconf(8) ran
2128 * rc.suspend and rc.resume so we don't have to notify devd(8) to do this.
2129 */
2130 ACPI_STATUS
2131 acpi_SetSleepState(struct acpi_softc *sc, int state)
2132 {
2133 static int once;
2134
2135 if (!once) {
2136 printf(
2137 "warning: acpi_SetSleepState() deprecated, need to update your software\n");
2138 once = 1;
2139 }
2140 return (acpi_EnterSleepState(sc, state));
2141 }
2142
2143 static void
2144 acpi_sleep_force(void *arg)
2145 {
2146 struct acpi_softc *sc;
2147
2148 printf("acpi: suspend request timed out, forcing sleep now\n");
2149 sc = arg;
2150 if (ACPI_FAILURE(acpi_EnterSleepState(sc, sc->acpi_next_sstate)))
2151 printf("acpi: force sleep state S%d failed\n", sc->acpi_next_sstate);
2152 }
2153
2154 /*
2155 * Request that the system enter the given suspend state. All /dev/apm
2156 * devices and devd(8) will be notified. Userland then has a chance to
2157 * save state and acknowledge the request. The system sleeps once all
2158 * acks are in.
2159 */
2160 int
2161 acpi_ReqSleepState(struct acpi_softc *sc, int state)
2162 {
2163 struct apm_clone_data *clone;
2164
2165 if (state < ACPI_STATE_S1 || state > ACPI_STATE_S5)
2166 return (EINVAL);
2167
2168 /* S5 (soft-off) should be entered directly with no waiting. */
2169 if (state == ACPI_STATE_S5) {
2170 if (ACPI_SUCCESS(acpi_EnterSleepState(sc, state)))
2171 return (0);
2172 else
2173 return (ENXIO);
2174 }
2175
2176 #if !defined(__i386__)
2177 /* This platform does not support acpi suspend/resume. */
2178 return (EOPNOTSUPP);
2179 #endif
2180
2181 /* If a suspend request is already in progress, just return. */
2182 ACPI_LOCK(acpi);
2183 if (sc->acpi_next_sstate != 0) {
2184 ACPI_UNLOCK(acpi);
2185 return (0);
2186 }
2187
2188 /* Record the pending state and notify all apm devices. */
2189 sc->acpi_next_sstate = state;
2190 STAILQ_FOREACH(clone, &sc->apm_cdevs, entries) {
2191 clone->notify_status = APM_EV_NONE;
2192 if ((clone->flags & ACPI_EVF_DEVD) == 0) {
2193 selwakeuppri(&clone->sel_read, PZERO);
2194 KNOTE_UNLOCKED(&clone->sel_read.si_note, 0);
2195 }
2196 }
2197
2198 /* If devd(8) is not running, immediately enter the sleep state. */
2199 if (devctl_process_running() == FALSE) {
2200 ACPI_UNLOCK(acpi);
2201 if (ACPI_SUCCESS(acpi_EnterSleepState(sc, sc->acpi_next_sstate))) {
2202 return (0);
2203 } else {
2204 return (ENXIO);
2205 }
2206 }
2207
2208 /* Now notify devd(8) also. */
2209 acpi_UserNotify("Suspend", ACPI_ROOT_OBJECT, state);
2210
2211 /*
2212 * Set a timeout to fire if userland doesn't ack the suspend request
2213 * in time. This way we still eventually go to sleep if we were
2214 * overheating or running low on battery, even if userland is hung.
2215 * We cancel this timeout once all userland acks are in or the
2216 * suspend request is aborted.
2217 */
2218 callout_reset(&sc->susp_force_to, 10 * hz, acpi_sleep_force, sc);
2219 ACPI_UNLOCK(acpi);
2220 return (0);
2221 }
2222
2223 /*
2224 * Acknowledge (or reject) a pending sleep state. The caller has
2225 * prepared for suspend and is now ready for it to proceed. If the
2226 * error argument is non-zero, it indicates suspend should be cancelled
2227 * and gives an errno value describing why. Once all votes are in,
2228 * we suspend the system.
2229 */
2230 int
2231 acpi_AckSleepState(struct apm_clone_data *clone, int error)
2232 {
2233 struct acpi_softc *sc;
2234 int ret, sleeping;
2235
2236 #if !defined(__i386__)
2237 /* This platform does not support acpi suspend/resume. */
2238 return (EOPNOTSUPP);
2239 #endif
2240
2241 /* If no pending sleep state, return an error. */
2242 ACPI_LOCK(acpi);
2243 sc = clone->acpi_sc;
2244 if (sc->acpi_next_sstate == 0) {
2245 ACPI_UNLOCK(acpi);
2246 return (ENXIO);
2247 }
2248
2249 /* Caller wants to abort suspend process. */
2250 if (error) {
2251 sc->acpi_next_sstate = 0;
2252 callout_stop(&sc->susp_force_to);
2253 printf("acpi: listener on %s cancelled the pending suspend\n",
2254 devtoname(clone->cdev));
2255 ACPI_UNLOCK(acpi);
2256 return (0);
2257 }
2258
2259 /*
2260 * Mark this device as acking the suspend request. Then, walk through
2261 * all devices, seeing if they agree yet. We only count devices that
2262 * are writable since read-only devices couldn't ack the request.
2263 */
2264 clone->notify_status = APM_EV_ACKED;
2265 sleeping = TRUE;
2266 STAILQ_FOREACH(clone, &sc->apm_cdevs, entries) {
2267 if ((clone->flags & ACPI_EVF_WRITE) != 0 &&
2268 clone->notify_status != APM_EV_ACKED) {
2269 sleeping = FALSE;
2270 break;
2271 }
2272 }
2273
2274 /* If all devices have voted "yes", we will suspend now. */
2275 if (sleeping)
2276 callout_stop(&sc->susp_force_to);
2277 ACPI_UNLOCK(acpi);
2278 ret = 0;
2279 if (sleeping) {
2280 if (ACPI_FAILURE(acpi_EnterSleepState(sc, sc->acpi_next_sstate)))
2281 ret = ENODEV;
2282 }
2283
2284 return (ret);
2285 }
2286
2287 static void
2288 acpi_sleep_enable(void *arg)
2289 {
2290
2291 ((struct acpi_softc *)arg)->acpi_sleep_disabled = 0;
2292 }
2293
2294 enum acpi_sleep_state {
2295 ACPI_SS_NONE,
2296 ACPI_SS_GPE_SET,
2297 ACPI_SS_DEV_SUSPEND,
2298 ACPI_SS_SLP_PREP,
2299 ACPI_SS_SLEPT,
2300 };
2301
2302 /*
2303 * Enter the desired system sleep state.
2304 *
2305 * Currently we support S1-S5 but S4 is only S4BIOS
2306 */
2307 static ACPI_STATUS
2308 acpi_EnterSleepState(struct acpi_softc *sc, int state)
2309 {
2310 ACPI_STATUS status;
2311 UINT8 TypeA;
2312 UINT8 TypeB;
2313 enum acpi_sleep_state slp_state;
2314
2315 ACPI_FUNCTION_TRACE_U32((char *)(uintptr_t)__func__, state);
2316
2317 /* Re-entry once we're suspending is not allowed. */
2318 status = AE_OK;
2319 ACPI_LOCK(acpi);
2320 if (sc->acpi_sleep_disabled) {
2321 ACPI_UNLOCK(acpi);
2322 printf("acpi: suspend request ignored (not ready yet)\n");
2323 return (AE_ERROR);
2324 }
2325 sc->acpi_sleep_disabled = 1;
2326 ACPI_UNLOCK(acpi);
2327
2328 /*
2329 * Be sure to hold Giant across DEVICE_SUSPEND/RESUME since non-MPSAFE
2330 * drivers need this.
2331 */
2332 mtx_lock(&Giant);
2333 slp_state = ACPI_SS_NONE;
2334 switch (state) {
2335 case ACPI_STATE_S1:
2336 case ACPI_STATE_S2:
2337 case ACPI_STATE_S3:
2338 case ACPI_STATE_S4:
2339 status = AcpiGetSleepTypeData(state, &TypeA, &TypeB);
2340 if (status == AE_NOT_FOUND) {
2341 device_printf(sc->acpi_dev,
2342 "Sleep state S%d not supported by BIOS\n", state);
2343 break;
2344 } else if (ACPI_FAILURE(status)) {
2345 device_printf(sc->acpi_dev, "AcpiGetSleepTypeData failed - %s\n",
2346 AcpiFormatException(status));
2347 break;
2348 }
2349
2350 sc->acpi_sstate = state;
2351
2352 /* Enable any GPEs as appropriate and requested by the user. */
2353 acpi_wake_prep_walk(state);
2354 slp_state = ACPI_SS_GPE_SET;
2355
2356 /*
2357 * Inform all devices that we are going to sleep. If at least one
2358 * device fails, DEVICE_SUSPEND() automatically resumes the tree.
2359 *
2360 * XXX Note that a better two-pass approach with a 'veto' pass
2361 * followed by a "real thing" pass would be better, but the current
2362 * bus interface does not provide for this.
2363 */
2364 if (DEVICE_SUSPEND(root_bus) != 0) {
2365 device_printf(sc->acpi_dev, "device_suspend failed\n");
2366 break;
2367 }
2368 slp_state = ACPI_SS_DEV_SUSPEND;
2369
2370 /* If testing device suspend only, back out of everything here. */
2371 if (acpi_susp_bounce)
2372 break;
2373
2374 status = AcpiEnterSleepStatePrep(state);
2375 if (ACPI_FAILURE(status)) {
2376 device_printf(sc->acpi_dev, "AcpiEnterSleepStatePrep failed - %s\n",
2377 AcpiFormatException(status));
2378 break;
2379 }
2380 slp_state = ACPI_SS_SLP_PREP;
2381
2382 if (sc->acpi_sleep_delay > 0)
2383 DELAY(sc->acpi_sleep_delay * 1000000);
2384
2385 if (state != ACPI_STATE_S1) {
2386 acpi_sleep_machdep(sc, state);
2387
2388 /* Re-enable ACPI hardware on wakeup from sleep state 4. */
2389 if (state == ACPI_STATE_S4)
2390 AcpiEnable();
2391 } else {
2392 ACPI_DISABLE_IRQS();
2393 status = AcpiEnterSleepState(state);
2394 if (ACPI_FAILURE(status)) {
2395 device_printf(sc->acpi_dev, "AcpiEnterSleepState failed - %s\n",
2396 AcpiFormatException(status));
2397 break;
2398 }
2399 }
2400 slp_state = ACPI_SS_SLEPT;
2401 break;
2402 case ACPI_STATE_S5:
2403 /*
2404 * Shut down cleanly and power off. This will call us back through the
2405 * shutdown handlers.
2406 */
2407 shutdown_nice(RB_POWEROFF);
2408 break;
2409 case ACPI_STATE_S0:
2410 default:
2411 status = AE_BAD_PARAMETER;
2412 break;
2413 }
2414
2415 /*
2416 * Back out state according to how far along we got in the suspend
2417 * process. This handles both the error and success cases.
2418 */
2419 sc->acpi_next_sstate = 0;
2420 if (slp_state >= ACPI_SS_GPE_SET) {
2421 acpi_wake_prep_walk(state);
2422 sc->acpi_sstate = ACPI_STATE_S0;
2423 }
2424 if (slp_state >= ACPI_SS_SLP_PREP)
2425 AcpiLeaveSleepState(state);
2426 if (slp_state >= ACPI_SS_DEV_SUSPEND)
2427 DEVICE_RESUME(root_bus);
2428 if (slp_state >= ACPI_SS_SLEPT)
2429 acpi_enable_fixed_events(sc);
2430
2431 /* Allow another sleep request after a while. */
2432 if (state != ACPI_STATE_S5)
2433 timeout(acpi_sleep_enable, sc, hz * ACPI_MINIMUM_AWAKETIME);
2434
2435 /* Run /etc/rc.resume after we are back. */
2436 acpi_UserNotify("Resume", ACPI_ROOT_OBJECT, state);
2437
2438 mtx_unlock(&Giant);
2439 return_ACPI_STATUS (status);
2440 }
2441
2442 /* Initialize a device's wake GPE. */
2443 int
2444 acpi_wake_init(device_t dev, int type)
2445 {
2446 struct acpi_prw_data prw;
2447
2448 /* Evaluate _PRW to find the GPE. */
2449 if (acpi_parse_prw(acpi_get_handle(dev), &prw) != 0)
2450 return (ENXIO);
2451
2452 /* Set the requested type for the GPE (runtime, wake, or both). */
2453 if (ACPI_FAILURE(AcpiSetGpeType(prw.gpe_handle, prw.gpe_bit, type))) {
2454 device_printf(dev, "set GPE type failed\n");
2455 return (ENXIO);
2456 }
2457
2458 return (0);
2459 }
2460
2461 /* Enable or disable the device's wake GPE. */
2462 int
2463 acpi_wake_set_enable(device_t dev, int enable)
2464 {
2465 struct acpi_prw_data prw;
2466 ACPI_STATUS status;
2467 int flags;
2468
2469 /* Make sure the device supports waking the system and get the GPE. */
2470 if (acpi_parse_prw(acpi_get_handle(dev), &prw) != 0)
2471 return (ENXIO);
2472
2473 flags = acpi_get_flags(dev);
2474 if (enable) {
2475 status = AcpiEnableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2476 if (ACPI_FAILURE(status)) {
2477 device_printf(dev, "enable wake failed\n");
2478 return (ENXIO);
2479 }
2480 acpi_set_flags(dev, flags | ACPI_FLAG_WAKE_ENABLED);
2481 } else {
2482 status = AcpiDisableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2483 if (ACPI_FAILURE(status)) {
2484 device_printf(dev, "disable wake failed\n");
2485 return (ENXIO);
2486 }
2487 acpi_set_flags(dev, flags & ~ACPI_FLAG_WAKE_ENABLED);
2488 }
2489
2490 return (0);
2491 }
2492
2493 static int
2494 acpi_wake_sleep_prep(ACPI_HANDLE handle, int sstate)
2495 {
2496 struct acpi_prw_data prw;
2497 device_t dev;
2498
2499 /* Check that this is a wake-capable device and get its GPE. */
2500 if (acpi_parse_prw(handle, &prw) != 0)
2501 return (ENXIO);
2502 dev = acpi_get_device(handle);
2503
2504 /*
2505 * The destination sleep state must be less than (i.e., higher power)
2506 * or equal to the value specified by _PRW. If this GPE cannot be
2507 * enabled for the next sleep state, then disable it. If it can and
2508 * the user requested it be enabled, turn on any required power resources
2509 * and set _PSW.
2510 */
2511 if (sstate > prw.lowest_wake) {
2512 AcpiDisableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2513 if (bootverbose)
2514 device_printf(dev, "wake_prep disabled wake for %s (S%d)\n",
2515 acpi_name(handle), sstate);
2516 } else if (dev && (acpi_get_flags(dev) & ACPI_FLAG_WAKE_ENABLED) != 0) {
2517 acpi_pwr_wake_enable(handle, 1);
2518 acpi_SetInteger(handle, "_PSW", 1);
2519 if (bootverbose)
2520 device_printf(dev, "wake_prep enabled for %s (S%d)\n",
2521 acpi_name(handle), sstate);
2522 }
2523
2524 return (0);
2525 }
2526
2527 static int
2528 acpi_wake_run_prep(ACPI_HANDLE handle, int sstate)
2529 {
2530 struct acpi_prw_data prw;
2531 device_t dev;
2532
2533 /*
2534 * Check that this is a wake-capable device and get its GPE. Return
2535 * now if the user didn't enable this device for wake.
2536 */
2537 if (acpi_parse_prw(handle, &prw) != 0)
2538 return (ENXIO);
2539 dev = acpi_get_device(handle);
2540 if (dev == NULL || (acpi_get_flags(dev) & ACPI_FLAG_WAKE_ENABLED) == 0)
2541 return (0);
2542
2543 /*
2544 * If this GPE couldn't be enabled for the previous sleep state, it was
2545 * disabled before going to sleep so re-enable it. If it was enabled,
2546 * clear _PSW and turn off any power resources it used.
2547 */
2548 if (sstate > prw.lowest_wake) {
2549 AcpiEnableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2550 if (bootverbose)
2551 device_printf(dev, "run_prep re-enabled %s\n", acpi_name(handle));
2552 } else {
2553 acpi_SetInteger(handle, "_PSW", 0);
2554 acpi_pwr_wake_enable(handle, 0);
2555 if (bootverbose)
2556 device_printf(dev, "run_prep cleaned up for %s\n",
2557 acpi_name(handle));
2558 }
2559
2560 return (0);
2561 }
2562
2563 static ACPI_STATUS
2564 acpi_wake_prep(ACPI_HANDLE handle, UINT32 level, void *context, void **status)
2565 {
2566 int sstate;
2567
2568 /* If suspending, run the sleep prep function, otherwise wake. */
2569 sstate = *(int *)context;
2570 if (AcpiGbl_SystemAwakeAndRunning)
2571 acpi_wake_sleep_prep(handle, sstate);
2572 else
2573 acpi_wake_run_prep(handle, sstate);
2574 return (AE_OK);
2575 }
2576
2577 /* Walk the tree rooted at acpi0 to prep devices for suspend/resume. */
2578 static int
2579 acpi_wake_prep_walk(int sstate)
2580 {
2581 ACPI_HANDLE sb_handle;
2582
2583 if (ACPI_SUCCESS(AcpiGetHandle(ACPI_ROOT_OBJECT, "\\_SB_", &sb_handle)))
2584 AcpiWalkNamespace(ACPI_TYPE_DEVICE, sb_handle, 100,
2585 acpi_wake_prep, &sstate, NULL);
2586 return (0);
2587 }
2588
2589 /* Walk the tree rooted at acpi0 to attach per-device wake sysctls. */
2590 static int
2591 acpi_wake_sysctl_walk(device_t dev)
2592 {
2593 int error, i, numdevs;
2594 device_t *devlist;
2595 device_t child;
2596 ACPI_STATUS status;
2597
2598 error = device_get_children(dev, &devlist, &numdevs);
2599 if (error != 0 || numdevs == 0) {
2600 if (numdevs == 0)
2601 free(devlist, M_TEMP);
2602 return (error);
2603 }
2604 for (i = 0; i < numdevs; i++) {
2605 child = devlist[i];
2606 acpi_wake_sysctl_walk(child);
2607 if (!device_is_attached(child))
2608 continue;
2609 status = AcpiEvaluateObject(acpi_get_handle(child), "_PRW", NULL, NULL);
2610 if (ACPI_SUCCESS(status)) {
2611 SYSCTL_ADD_PROC(device_get_sysctl_ctx(child),
2612 SYSCTL_CHILDREN(device_get_sysctl_tree(child)), OID_AUTO,
2613 "wake", CTLTYPE_INT | CTLFLAG_RW, child, 0,
2614 acpi_wake_set_sysctl, "I", "Device set to wake the system");
2615 }
2616 }
2617 free(devlist, M_TEMP);
2618
2619 return (0);
2620 }
2621
2622 /* Enable or disable wake from userland. */
2623 static int
2624 acpi_wake_set_sysctl(SYSCTL_HANDLER_ARGS)
2625 {
2626 int enable, error;
2627 device_t dev;
2628
2629 dev = (device_t)arg1;
2630 enable = (acpi_get_flags(dev) & ACPI_FLAG_WAKE_ENABLED) ? 1 : 0;
2631
2632 error = sysctl_handle_int(oidp, &enable, 0, req);
2633 if (error != 0 || req->newptr == NULL)
2634 return (error);
2635 if (enable != 0 && enable != 1)
2636 return (EINVAL);
2637
2638 return (acpi_wake_set_enable(dev, enable));
2639 }
2640
2641 /* Parse a device's _PRW into a structure. */
2642 int
2643 acpi_parse_prw(ACPI_HANDLE h, struct acpi_prw_data *prw)
2644 {
2645 ACPI_STATUS status;
2646 ACPI_BUFFER prw_buffer;
2647 ACPI_OBJECT *res, *res2;
2648 int error, i, power_count;
2649
2650 if (h == NULL || prw == NULL)
2651 return (EINVAL);
2652
2653 /*
2654 * The _PRW object (7.2.9) is only required for devices that have the
2655 * ability to wake the system from a sleeping state.
2656 */
2657 error = EINVAL;
2658 prw_buffer.Pointer = NULL;
2659 prw_buffer.Length = ACPI_ALLOCATE_BUFFER;
2660 status = AcpiEvaluateObject(h, "_PRW", NULL, &prw_buffer);
2661 if (ACPI_FAILURE(status))
2662 return (ENOENT);
2663 res = (ACPI_OBJECT *)prw_buffer.Pointer;
2664 if (res == NULL)
2665 return (ENOENT);
2666 if (!ACPI_PKG_VALID(res, 2))
2667 goto out;
2668
2669 /*
2670 * Element 1 of the _PRW object:
2671 * The lowest power system sleeping state that can be entered while still
2672 * providing wake functionality. The sleeping state being entered must
2673 * be less than (i.e., higher power) or equal to this value.
2674 */
2675 if (acpi_PkgInt32(res, 1, &prw->lowest_wake) != 0)
2676 goto out;
2677
2678 /*
2679 * Element 0 of the _PRW object:
2680 */
2681 switch (res->Package.Elements[0].Type) {
2682 case ACPI_TYPE_INTEGER:
2683 /*
2684 * If the data type of this package element is numeric, then this
2685 * _PRW package element is the bit index in the GPEx_EN, in the
2686 * GPE blocks described in the FADT, of the enable bit that is
2687 * enabled for the wake event.
2688 */
2689 prw->gpe_handle = NULL;
2690 prw->gpe_bit = res->Package.Elements[0].Integer.Value;
2691 error = 0;
2692 break;
2693 case ACPI_TYPE_PACKAGE:
2694 /*
2695 * If the data type of this package element is a package, then this
2696 * _PRW package element is itself a package containing two
2697 * elements. The first is an object reference to the GPE Block
2698 * device that contains the GPE that will be triggered by the wake
2699 * event. The second element is numeric and it contains the bit
2700 * index in the GPEx_EN, in the GPE Block referenced by the
2701 * first element in the package, of the enable bit that is enabled for
2702 * the wake event.
2703 *
2704 * For example, if this field is a package then it is of the form:
2705 * Package() {\_SB.PCI0.ISA.GPE, 2}
2706 */
2707 res2 = &res->Package.Elements[0];
2708 if (!ACPI_PKG_VALID(res2, 2))
2709 goto out;
2710 prw->gpe_handle = acpi_GetReference(NULL, &res2->Package.Elements[0]);
2711 if (prw->gpe_handle == NULL)
2712 goto out;
2713 if (acpi_PkgInt32(res2, 1, &prw->gpe_bit) != 0)
2714 goto out;
2715 error = 0;
2716 break;
2717 default:
2718 goto out;
2719 }
2720
2721 /* Elements 2 to N of the _PRW object are power resources. */
2722 power_count = res->Package.Count - 2;
2723 if (power_count > ACPI_PRW_MAX_POWERRES) {
2724 printf("ACPI device %s has too many power resources\n", acpi_name(h));
2725 power_count = 0;
2726 }
2727 prw->power_res_count = power_count;
2728 for (i = 0; i < power_count; i++)
2729 prw->power_res[i] = res->Package.Elements[i];
2730
2731 out:
2732 if (prw_buffer.Pointer != NULL)
2733 AcpiOsFree(prw_buffer.Pointer);
2734 return (error);
2735 }
2736
2737 /*
2738 * ACPI Event Handlers
2739 */
2740
2741 /* System Event Handlers (registered by EVENTHANDLER_REGISTER) */
2742
2743 static void
2744 acpi_system_eventhandler_sleep(void *arg, int state)
2745 {
2746 int ret;
2747
2748 ACPI_FUNCTION_TRACE_U32((char *)(uintptr_t)__func__, state);
2749
2750 /* Check if button action is disabled. */
2751 if (state == ACPI_S_STATES_MAX + 1)
2752 return;
2753
2754 /* Request that the system prepare to enter the given suspend state. */
2755 ret = acpi_ReqSleepState((struct acpi_softc *)arg, state);
2756 if (ret != 0)
2757 printf("acpi: request to enter state S%d failed (err %d)\n",
2758 state, ret);
2759
2760 return_VOID;
2761 }
2762
2763 static void
2764 acpi_system_eventhandler_wakeup(void *arg, int state)
2765 {
2766
2767 ACPI_FUNCTION_TRACE_U32((char *)(uintptr_t)__func__, state);
2768
2769 /* Currently, nothing to do for wakeup. */
2770
2771 return_VOID;
2772 }
2773
2774 /*
2775 * ACPICA Event Handlers (FixedEvent, also called from button notify handler)
2776 */
2777 UINT32
2778 acpi_event_power_button_sleep(void *context)
2779 {
2780 struct acpi_softc *sc = (struct acpi_softc *)context;
2781
2782 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2783
2784 EVENTHANDLER_INVOKE(acpi_sleep_event, sc->acpi_power_button_sx);
2785
2786 return_VALUE (ACPI_INTERRUPT_HANDLED);
2787 }
2788
2789 UINT32
2790 acpi_event_power_button_wake(void *context)
2791 {
2792 struct acpi_softc *sc = (struct acpi_softc *)context;
2793
2794 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2795
2796 EVENTHANDLER_INVOKE(acpi_wakeup_event, sc->acpi_power_button_sx);
2797
2798 return_VALUE (ACPI_INTERRUPT_HANDLED);
2799 }
2800
2801 UINT32
2802 acpi_event_sleep_button_sleep(void *context)
2803 {
2804 struct acpi_softc *sc = (struct acpi_softc *)context;
2805
2806 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2807
2808 EVENTHANDLER_INVOKE(acpi_sleep_event, sc->acpi_sleep_button_sx);
2809
2810 return_VALUE (ACPI_INTERRUPT_HANDLED);
2811 }
2812
2813 UINT32
2814 acpi_event_sleep_button_wake(void *context)
2815 {
2816 struct acpi_softc *sc = (struct acpi_softc *)context;
2817
2818 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2819
2820 EVENTHANDLER_INVOKE(acpi_wakeup_event, sc->acpi_sleep_button_sx);
2821
2822 return_VALUE (ACPI_INTERRUPT_HANDLED);
2823 }
2824
2825 /*
2826 * XXX This static buffer is suboptimal. There is no locking so only
2827 * use this for single-threaded callers.
2828 */
2829 char *
2830 acpi_name(ACPI_HANDLE handle)
2831 {
2832 ACPI_BUFFER buf;
2833 static char data[256];
2834
2835 buf.Length = sizeof(data);
2836 buf.Pointer = data;
2837
2838 if (handle && ACPI_SUCCESS(AcpiGetName(handle, ACPI_FULL_PATHNAME, &buf)))
2839 return (data);
2840 return ("(unknown)");
2841 }
2842
2843 /*
2844 * Debugging/bug-avoidance. Avoid trying to fetch info on various
2845 * parts of the namespace.
2846 */
2847 int
2848 acpi_avoid(ACPI_HANDLE handle)
2849 {
2850 char *cp, *env, *np;
2851 int len;
2852
2853 np = acpi_name(handle);
2854 if (*np == '\\')
2855 np++;
2856 if ((env = getenv("debug.acpi.avoid")) == NULL)
2857 return (0);
2858
2859 /* Scan the avoid list checking for a match */
2860 cp = env;
2861 for (;;) {
2862 while (*cp != 0 && isspace(*cp))
2863 cp++;
2864 if (*cp == 0)
2865 break;
2866 len = 0;
2867 while (cp[len] != 0 && !isspace(cp[len]))
2868 len++;
2869 if (!strncmp(cp, np, len)) {
2870 freeenv(env);
2871 return(1);
2872 }
2873 cp += len;
2874 }
2875 freeenv(env);
2876
2877 return (0);
2878 }
2879
2880 /*
2881 * Debugging/bug-avoidance. Disable ACPI subsystem components.
2882 */
2883 int
2884 acpi_disabled(char *subsys)
2885 {
2886 char *cp, *env;
2887 int len;
2888
2889 if ((env = getenv("debug.acpi.disabled")) == NULL)
2890 return (0);
2891 if (strcmp(env, "all") == 0) {
2892 freeenv(env);
2893 return (1);
2894 }
2895
2896 /* Scan the disable list, checking for a match. */
2897 cp = env;
2898 for (;;) {
2899 while (*cp != '\0' && isspace(*cp))
2900 cp++;
2901 if (*cp == '\0')
2902 break;
2903 len = 0;
2904 while (cp[len] != '\0' && !isspace(cp[len]))
2905 len++;
2906 if (strncmp(cp, subsys, len) == 0) {
2907 freeenv(env);
2908 return (1);
2909 }
2910 cp += len;
2911 }
2912 freeenv(env);
2913
2914 return (0);
2915 }
2916
2917 /*
2918 * Control interface.
2919 *
2920 * We multiplex ioctls for all participating ACPI devices here. Individual
2921 * drivers wanting to be accessible via /dev/acpi should use the
2922 * register/deregister interface to make their handlers visible.
2923 */
2924 struct acpi_ioctl_hook
2925 {
2926 TAILQ_ENTRY(acpi_ioctl_hook) link;
2927 u_long cmd;
2928 acpi_ioctl_fn fn;
2929 void *arg;
2930 };
2931
2932 static TAILQ_HEAD(,acpi_ioctl_hook) acpi_ioctl_hooks;
2933 static int acpi_ioctl_hooks_initted;
2934
2935 int
2936 acpi_register_ioctl(u_long cmd, acpi_ioctl_fn fn, void *arg)
2937 {
2938 struct acpi_ioctl_hook *hp;
2939
2940 if ((hp = malloc(sizeof(*hp), M_ACPIDEV, M_NOWAIT)) == NULL)
2941 return (ENOMEM);
2942 hp->cmd = cmd;
2943 hp->fn = fn;
2944 hp->arg = arg;
2945
2946 ACPI_LOCK(acpi);
2947 if (acpi_ioctl_hooks_initted == 0) {
2948 TAILQ_INIT(&acpi_ioctl_hooks);
2949 acpi_ioctl_hooks_initted = 1;
2950 }
2951 TAILQ_INSERT_TAIL(&acpi_ioctl_hooks, hp, link);
2952 ACPI_UNLOCK(acpi);
2953
2954 return (0);
2955 }
2956
2957 void
2958 acpi_deregister_ioctl(u_long cmd, acpi_ioctl_fn fn)
2959 {
2960 struct acpi_ioctl_hook *hp;
2961
2962 ACPI_LOCK(acpi);
2963 TAILQ_FOREACH(hp, &acpi_ioctl_hooks, link)
2964 if (hp->cmd == cmd && hp->fn == fn)
2965 break;
2966
2967 if (hp != NULL) {
2968 TAILQ_REMOVE(&acpi_ioctl_hooks, hp, link);
2969 free(hp, M_ACPIDEV);
2970 }
2971 ACPI_UNLOCK(acpi);
2972 }
2973
2974 static int
2975 acpiopen(struct cdev *dev, int flag, int fmt, d_thread_t *td)
2976 {
2977 return (0);
2978 }
2979
2980 static int
2981 acpiclose(struct cdev *dev, int flag, int fmt, d_thread_t *td)
2982 {
2983 return (0);
2984 }
2985
2986 static int
2987 acpiioctl(struct cdev *dev, u_long cmd, caddr_t addr, int flag, d_thread_t *td)
2988 {
2989 struct acpi_softc *sc;
2990 struct acpi_ioctl_hook *hp;
2991 int error, state;
2992
2993 error = 0;
2994 hp = NULL;
2995 sc = dev->si_drv1;
2996
2997 /*
2998 * Scan the list of registered ioctls, looking for handlers.
2999 */
3000 ACPI_LOCK(acpi);
3001 if (acpi_ioctl_hooks_initted)
3002 TAILQ_FOREACH(hp, &acpi_ioctl_hooks, link) {
3003 if (hp->cmd == cmd)
3004 break;
3005 }
3006 ACPI_UNLOCK(acpi);
3007 if (hp)
3008 return (hp->fn(cmd, addr, hp->arg));
3009
3010 /*
3011 * Core ioctls are not permitted for non-writable user.
3012 * Currently, other ioctls just fetch information.
3013 * Not changing system behavior.
3014 */
3015 if ((flag & FWRITE) == 0)
3016 return (EPERM);
3017
3018 /* Core system ioctls. */
3019 switch (cmd) {
3020 case ACPIIO_REQSLPSTATE:
3021 state = *(int *)addr;
3022 if (state != ACPI_STATE_S5)
3023 error = acpi_ReqSleepState(sc, state);
3024 else {
3025 printf("power off via acpi ioctl not supported\n");
3026 error = ENXIO;
3027 }
3028 break;
3029 case ACPIIO_ACKSLPSTATE:
3030 error = *(int *)addr;
3031 error = acpi_AckSleepState(sc->acpi_clone, error);
3032 break;
3033 case ACPIIO_SETSLPSTATE: /* DEPRECATED */
3034 error = EINVAL;
3035 state = *(int *)addr;
3036 if (state >= ACPI_STATE_S0 && state <= ACPI_S_STATES_MAX)
3037 if (ACPI_SUCCESS(acpi_SetSleepState(sc, state)))
3038 error = 0;
3039 break;
3040 default:
3041 error = ENXIO;
3042 break;
3043 }
3044
3045 return (error);
3046 }
3047
3048 static int
3049 acpi_supported_sleep_state_sysctl(SYSCTL_HANDLER_ARGS)
3050 {
3051 int error;
3052 struct sbuf sb;
3053 UINT8 state, TypeA, TypeB;
3054
3055 sbuf_new(&sb, NULL, 32, SBUF_AUTOEXTEND);
3056 for (state = ACPI_STATE_S1; state < ACPI_S_STATES_MAX + 1; state++)
3057 if (ACPI_SUCCESS(AcpiGetSleepTypeData(state, &TypeA, &TypeB)))
3058 sbuf_printf(&sb, "S%d ", state);
3059 sbuf_trim(&sb);
3060 sbuf_finish(&sb);
3061 error = sysctl_handle_string(oidp, sbuf_data(&sb), sbuf_len(&sb), req);
3062 sbuf_delete(&sb);
3063 return (error);
3064 }
3065
3066 static int
3067 acpi_sleep_state_sysctl(SYSCTL_HANDLER_ARGS)
3068 {
3069 char sleep_state[10];
3070 int error;
3071 u_int new_state, old_state;
3072
3073 old_state = *(u_int *)oidp->oid_arg1;
3074 if (old_state > ACPI_S_STATES_MAX + 1)
3075 strlcpy(sleep_state, "unknown", sizeof(sleep_state));
3076 else
3077 strlcpy(sleep_state, sleep_state_names[old_state], sizeof(sleep_state));
3078 error = sysctl_handle_string(oidp, sleep_state, sizeof(sleep_state), req);
3079 if (error == 0 && req->newptr != NULL) {
3080 new_state = ACPI_STATE_S0;
3081 for (; new_state <= ACPI_S_STATES_MAX + 1; new_state++)
3082 if (strcmp(sleep_state, sleep_state_names[new_state]) == 0)
3083 break;
3084 if (new_state <= ACPI_S_STATES_MAX + 1) {
3085 if (new_state != old_state)
3086 *(u_int *)oidp->oid_arg1 = new_state;
3087 } else
3088 error = EINVAL;
3089 }
3090
3091 return (error);
3092 }
3093
3094 /* Inform devctl(4) when we receive a Notify. */
3095 void
3096 acpi_UserNotify(const char *subsystem, ACPI_HANDLE h, uint8_t notify)
3097 {
3098 char notify_buf[16];
3099 ACPI_BUFFER handle_buf;
3100 ACPI_STATUS status;
3101
3102 if (subsystem == NULL)
3103 return;
3104
3105 handle_buf.Pointer = NULL;
3106 handle_buf.Length = ACPI_ALLOCATE_BUFFER;
3107 status = AcpiNsHandleToPathname(h, &handle_buf);
3108 if (ACPI_FAILURE(status))
3109 return;
3110 snprintf(notify_buf, sizeof(notify_buf), "notify=0x%02x", notify);
3111 devctl_notify("ACPI", subsystem, handle_buf.Pointer, notify_buf);
3112 AcpiOsFree(handle_buf.Pointer);
3113 }
3114
3115 #ifdef ACPI_DEBUG
3116 /*
3117 * Support for parsing debug options from the kernel environment.
3118 *
3119 * Bits may be set in the AcpiDbgLayer and AcpiDbgLevel debug registers
3120 * by specifying the names of the bits in the debug.acpi.layer and
3121 * debug.acpi.level environment variables. Bits may be unset by
3122 * prefixing the bit name with !.
3123 */
3124 struct debugtag
3125 {
3126 char *name;
3127 UINT32 value;
3128 };
3129
3130 static struct debugtag dbg_layer[] = {
3131 {"ACPI_UTILITIES", ACPI_UTILITIES},
3132 {"ACPI_HARDWARE", ACPI_HARDWARE},
3133 {"ACPI_EVENTS", ACPI_EVENTS},
3134 {"ACPI_TABLES", ACPI_TABLES},
3135 {"ACPI_NAMESPACE", ACPI_NAMESPACE},
3136 {"ACPI_PARSER", ACPI_PARSER},
3137 {"ACPI_DISPATCHER", ACPI_DISPATCHER},
3138 {"ACPI_EXECUTER", ACPI_EXECUTER},
3139 {"ACPI_RESOURCES", ACPI_RESOURCES},
3140 {"ACPI_CA_DEBUGGER", ACPI_CA_DEBUGGER},
3141 {"ACPI_OS_SERVICES", ACPI_OS_SERVICES},
3142 {"ACPI_CA_DISASSEMBLER", ACPI_CA_DISASSEMBLER},
3143 {"ACPI_ALL_COMPONENTS", ACPI_ALL_COMPONENTS},
3144
3145 {"ACPI_AC_ADAPTER", ACPI_AC_ADAPTER},
3146 {"ACPI_BATTERY", ACPI_BATTERY},
3147 {"ACPI_BUS", ACPI_BUS},
3148 {"ACPI_BUTTON", ACPI_BUTTON},
3149 {"ACPI_EC", ACPI_EC},
3150 {"ACPI_FAN", ACPI_FAN},
3151 {"ACPI_POWERRES", ACPI_POWERRES},
3152 {"ACPI_PROCESSOR", ACPI_PROCESSOR},
3153 {"ACPI_THERMAL", ACPI_THERMAL},
3154 {"ACPI_TIMER", ACPI_TIMER},
3155 {"ACPI_ALL_DRIVERS", ACPI_ALL_DRIVERS},
3156 {NULL, 0}
3157 };
3158
3159 static struct debugtag dbg_level[] = {
3160 {"ACPI_LV_ERROR", ACPI_LV_ERROR},
3161 {"ACPI_LV_WARN", ACPI_LV_WARN},
3162 {"ACPI_LV_INIT", ACPI_LV_INIT},
3163 {"ACPI_LV_DEBUG_OBJECT", ACPI_LV_DEBUG_OBJECT},
3164 {"ACPI_LV_INFO", ACPI_LV_INFO},
3165 {"ACPI_LV_ALL_EXCEPTIONS", ACPI_LV_ALL_EXCEPTIONS},
3166
3167 /* Trace verbosity level 1 [Standard Trace Level] */
3168 {"ACPI_LV_INIT_NAMES", ACPI_LV_INIT_NAMES},
3169 {"ACPI_LV_PARSE", ACPI_LV_PARSE},
3170 {"ACPI_LV_LOAD", ACPI_LV_LOAD},
3171 {"ACPI_LV_DISPATCH", ACPI_LV_DISPATCH},
3172 {"ACPI_LV_EXEC", ACPI_LV_EXEC},
3173 {"ACPI_LV_NAMES", ACPI_LV_NAMES},
3174 {"ACPI_LV_OPREGION", ACPI_LV_OPREGION},
3175 {"ACPI_LV_BFIELD", ACPI_LV_BFIELD},
3176 {"ACPI_LV_TABLES", ACPI_LV_TABLES},
3177 {"ACPI_LV_VALUES", ACPI_LV_VALUES},
3178 {"ACPI_LV_OBJECTS", ACPI_LV_OBJECTS},
3179 {"ACPI_LV_RESOURCES", ACPI_LV_RESOURCES},
3180 {"ACPI_LV_USER_REQUESTS", ACPI_LV_USER_REQUESTS},
3181 {"ACPI_LV_PACKAGE", ACPI_LV_PACKAGE},
3182 {"ACPI_LV_VERBOSITY1", ACPI_LV_VERBOSITY1},
3183
3184 /* Trace verbosity level 2 [Function tracing and memory allocation] */
3185 {"ACPI_LV_ALLOCATIONS", ACPI_LV_ALLOCATIONS},
3186 {"ACPI_LV_FUNCTIONS", ACPI_LV_FUNCTIONS},
3187 {"ACPI_LV_OPTIMIZATIONS", ACPI_LV_OPTIMIZATIONS},
3188 {"ACPI_LV_VERBOSITY2", ACPI_LV_VERBOSITY2},
3189 {"ACPI_LV_ALL", ACPI_LV_ALL},
3190
3191 /* Trace verbosity level 3 [Threading, I/O, and Interrupts] */
3192 {"ACPI_LV_MUTEX", ACPI_LV_MUTEX},
3193 {"ACPI_LV_THREADS", ACPI_LV_THREADS},
3194 {"ACPI_LV_IO", ACPI_LV_IO},
3195 {"ACPI_LV_INTERRUPTS", ACPI_LV_INTERRUPTS},
3196 {"ACPI_LV_VERBOSITY3", ACPI_LV_VERBOSITY3},
3197
3198 /* Exceptionally verbose output -- also used in the global "DebugLevel" */
3199 {"ACPI_LV_AML_DISASSEMBLE", ACPI_LV_AML_DISASSEMBLE},
3200 {"ACPI_LV_VERBOSE_INFO", ACPI_LV_VERBOSE_INFO},
3201 {"ACPI_LV_FULL_TABLES", ACPI_LV_FULL_TABLES},
3202 {"ACPI_LV_EVENTS", ACPI_LV_EVENTS},
3203 {"ACPI_LV_VERBOSE", ACPI_LV_VERBOSE},
3204 {NULL, 0}
3205 };
3206
3207 static void
3208 acpi_parse_debug(char *cp, struct debugtag *tag, UINT32 *flag)
3209 {
3210 char *ep;
3211 int i, l;
3212 int set;
3213
3214 while (*cp) {
3215 if (isspace(*cp)) {
3216 cp++;
3217 continue;
3218 }
3219 ep = cp;
3220 while (*ep && !isspace(*ep))
3221 ep++;
3222 if (*cp == '!') {
3223 set = 0;
3224 cp++;
3225 if (cp == ep)
3226 continue;
3227 } else {
3228 set = 1;
3229 }
3230 l = ep - cp;
3231 for (i = 0; tag[i].name != NULL; i++) {
3232 if (!strncmp(cp, tag[i].name, l)) {
3233 if (set)
3234 *flag |= tag[i].value;
3235 else
3236 *flag &= ~tag[i].value;
3237 }
3238 }
3239 cp = ep;
3240 }
3241 }
3242
3243 static void
3244 acpi_set_debugging(void *junk)
3245 {
3246 char *layer, *level;
3247
3248 if (cold) {
3249 AcpiDbgLayer = 0;
3250 AcpiDbgLevel = 0;
3251 }
3252
3253 layer = getenv("debug.acpi.layer");
3254 level = getenv("debug.acpi.level");
3255 if (layer == NULL && level == NULL)
3256 return;
3257
3258 printf("ACPI set debug");
3259 if (layer != NULL) {
3260 if (strcmp("NONE", layer) != 0)
3261 printf(" layer '%s'", layer);
3262 acpi_parse_debug(layer, &dbg_layer[0], &AcpiDbgLayer);
3263 freeenv(layer);
3264 }
3265 if (level != NULL) {
3266 if (strcmp("NONE", level) != 0)
3267 printf(" level '%s'", level);
3268 acpi_parse_debug(level, &dbg_level[0], &AcpiDbgLevel);
3269 freeenv(level);
3270 }
3271 printf("\n");
3272 }
3273
3274 SYSINIT(acpi_debugging, SI_SUB_TUNABLES, SI_ORDER_ANY, acpi_set_debugging,
3275 NULL);
3276
3277 static int
3278 acpi_debug_sysctl(SYSCTL_HANDLER_ARGS)
3279 {
3280 int error, *dbg;
3281 struct debugtag *tag;
3282 struct sbuf sb;
3283
3284 if (sbuf_new(&sb, NULL, 128, SBUF_AUTOEXTEND) == NULL)
3285 return (ENOMEM);
3286 if (strcmp(oidp->oid_arg1, "debug.acpi.layer") == 0) {
3287 tag = &dbg_layer[0];
3288 dbg = &AcpiDbgLayer;
3289 } else {
3290 tag = &dbg_level[0];
3291 dbg = &AcpiDbgLevel;
3292 }
3293
3294 /* Get old values if this is a get request. */
3295 ACPI_SERIAL_BEGIN(acpi);
3296 if (*dbg == 0) {
3297 sbuf_cpy(&sb, "NONE");
3298 } else if (req->newptr == NULL) {
3299 for (; tag->name != NULL; tag++) {
3300 if ((*dbg & tag->value) == tag->value)
3301 sbuf_printf(&sb, "%s ", tag->name);
3302 }
3303 }
3304 sbuf_trim(&sb);
3305 sbuf_finish(&sb);
3306
3307 /* Copy out the old values to the user. */
3308 error = SYSCTL_OUT(req, sbuf_data(&sb), sbuf_len(&sb));
3309 sbuf_delete(&sb);
3310
3311 /* If the user is setting a string, parse it. */
3312 if (error == 0 && req->newptr != NULL) {
3313 *dbg = 0;
3314 setenv((char *)oidp->oid_arg1, (char *)req->newptr);
3315 acpi_set_debugging(NULL);
3316 }
3317 ACPI_SERIAL_END(acpi);
3318
3319 return (error);
3320 }
3321
3322 SYSCTL_PROC(_debug_acpi, OID_AUTO, layer, CTLFLAG_RW | CTLTYPE_STRING,
3323 "debug.acpi.layer", 0, acpi_debug_sysctl, "A", "");
3324 SYSCTL_PROC(_debug_acpi, OID_AUTO, level, CTLFLAG_RW | CTLTYPE_STRING,
3325 "debug.acpi.level", 0, acpi_debug_sysctl, "A", "");
3326 #endif /* ACPI_DEBUG */
3327
3328 static int
3329 acpi_pm_func(u_long cmd, void *arg, ...)
3330 {
3331 int state, acpi_state;
3332 int error;
3333 struct acpi_softc *sc;
3334 va_list ap;
3335
3336 error = 0;
3337 switch (cmd) {
3338 case POWER_CMD_SUSPEND:
3339 sc = (struct acpi_softc *)arg;
3340 if (sc == NULL) {
3341 error = EINVAL;
3342 goto out;
3343 }
3344
3345 va_start(ap, arg);
3346 state = va_arg(ap, int);
3347 va_end(ap);
3348
3349 switch (state) {
3350 case POWER_SLEEP_STATE_STANDBY:
3351 acpi_state = sc->acpi_standby_sx;
3352 break;
3353 case POWER_SLEEP_STATE_SUSPEND:
3354 acpi_state = sc->acpi_suspend_sx;
3355 break;
3356 case POWER_SLEEP_STATE_HIBERNATE:
3357 acpi_state = ACPI_STATE_S4;
3358 break;
3359 default:
3360 error = EINVAL;
3361 goto out;
3362 }
3363
3364 if (ACPI_FAILURE(acpi_EnterSleepState(sc, acpi_state)))
3365 error = ENXIO;
3366 break;
3367 default:
3368 error = EINVAL;
3369 goto out;
3370 }
3371
3372 out:
3373 return (error);
3374 }
3375
3376 static void
3377 acpi_pm_register(void *arg)
3378 {
3379 if (!cold || resource_disabled("acpi", 0))
3380 return;
3381
3382 power_pm_register(POWER_PM_TYPE_ACPI, acpi_pm_func, NULL);
3383 }
3384
3385 SYSINIT(power, SI_SUB_KLD, SI_ORDER_ANY, acpi_pm_register, 0);
Cache object: c8a12b60dc0ce6b1052583201b1aebfd
|