FreeBSD/Linux Kernel Cross Reference
sys/dev/acpica/acpi.c
1 /*-
2 * Copyright (c) 2000 Takanori Watanabe <takawata@jp.freebsd.org>
3 * Copyright (c) 2000 Mitsuru IWASAKI <iwasaki@jp.freebsd.org>
4 * Copyright (c) 2000, 2001 Michael Smith
5 * Copyright (c) 2000 BSDi
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27 * SUCH DAMAGE.
28 */
29
30 #include <sys/cdefs.h>
31 __FBSDID("$FreeBSD$");
32
33 #include "opt_acpi.h"
34 #include <sys/param.h>
35 #include <sys/kernel.h>
36 #include <sys/proc.h>
37 #include <sys/fcntl.h>
38 #include <sys/malloc.h>
39 #include <sys/module.h>
40 #include <sys/bus.h>
41 #include <sys/conf.h>
42 #include <sys/ioccom.h>
43 #include <sys/reboot.h>
44 #include <sys/sysctl.h>
45 #include <sys/ctype.h>
46 #include <sys/linker.h>
47 #include <sys/power.h>
48 #include <sys/sbuf.h>
49 #include <sys/smp.h>
50
51 #include <machine/resource.h>
52 #include <machine/bus.h>
53 #include <sys/rman.h>
54 #include <isa/isavar.h>
55 #include <isa/pnpvar.h>
56
57 #include <contrib/dev/acpica/acpi.h>
58 #include <dev/acpica/acpivar.h>
59 #include <dev/acpica/acpiio.h>
60 #include <contrib/dev/acpica/achware.h>
61 #include <contrib/dev/acpica/acnamesp.h>
62
63 #include "pci_if.h"
64 #include <dev/pci/pcivar.h>
65 #include <dev/pci/pci_private.h>
66
67 #include <vm/vm_param.h>
68
69 MALLOC_DEFINE(M_ACPIDEV, "acpidev", "ACPI devices");
70
71 /* Hooks for the ACPI CA debugging infrastructure */
72 #define _COMPONENT ACPI_BUS
73 ACPI_MODULE_NAME("ACPI")
74
75 static d_open_t acpiopen;
76 static d_close_t acpiclose;
77 static d_ioctl_t acpiioctl;
78
79 static struct cdevsw acpi_cdevsw = {
80 .d_version = D_VERSION,
81 .d_open = acpiopen,
82 .d_close = acpiclose,
83 .d_ioctl = acpiioctl,
84 .d_name = "acpi",
85 };
86
87 /* Global mutex for locking access to the ACPI subsystem. */
88 struct mtx acpi_mutex;
89
90 /* Bitmap of device quirks. */
91 int acpi_quirks;
92
93 static int acpi_modevent(struct module *mod, int event, void *junk);
94 static void acpi_identify(driver_t *driver, device_t parent);
95 static int acpi_probe(device_t dev);
96 static int acpi_attach(device_t dev);
97 static int acpi_suspend(device_t dev);
98 static int acpi_resume(device_t dev);
99 static int acpi_shutdown(device_t dev);
100 static device_t acpi_add_child(device_t bus, int order, const char *name,
101 int unit);
102 static int acpi_print_child(device_t bus, device_t child);
103 static void acpi_probe_nomatch(device_t bus, device_t child);
104 static void acpi_driver_added(device_t dev, driver_t *driver);
105 static int acpi_read_ivar(device_t dev, device_t child, int index,
106 uintptr_t *result);
107 static int acpi_write_ivar(device_t dev, device_t child, int index,
108 uintptr_t value);
109 static struct resource_list *acpi_get_rlist(device_t dev, device_t child);
110 static int acpi_sysres_alloc(device_t dev);
111 static struct resource *acpi_alloc_resource(device_t bus, device_t child,
112 int type, int *rid, u_long start, u_long end,
113 u_long count, u_int flags);
114 static int acpi_release_resource(device_t bus, device_t child, int type,
115 int rid, struct resource *r);
116 static void acpi_delete_resource(device_t bus, device_t child, int type,
117 int rid);
118 static uint32_t acpi_isa_get_logicalid(device_t dev);
119 static int acpi_isa_get_compatid(device_t dev, uint32_t *cids, int count);
120 static char *acpi_device_id_probe(device_t bus, device_t dev, char **ids);
121 static ACPI_STATUS acpi_device_eval_obj(device_t bus, device_t dev,
122 ACPI_STRING pathname, ACPI_OBJECT_LIST *parameters,
123 ACPI_BUFFER *ret);
124 static int acpi_device_pwr_for_sleep(device_t bus, device_t dev,
125 int *dstate);
126 static ACPI_STATUS acpi_device_scan_cb(ACPI_HANDLE h, UINT32 level,
127 void *context, void **retval);
128 static ACPI_STATUS acpi_device_scan_children(device_t bus, device_t dev,
129 int max_depth, acpi_scan_cb_t user_fn, void *arg);
130 static int acpi_set_powerstate_method(device_t bus, device_t child,
131 int state);
132 static int acpi_isa_pnp_probe(device_t bus, device_t child,
133 struct isa_pnp_id *ids);
134 static void acpi_probe_children(device_t bus);
135 static void acpi_probe_order(ACPI_HANDLE handle, int *order);
136 static ACPI_STATUS acpi_probe_child(ACPI_HANDLE handle, UINT32 level,
137 void *context, void **status);
138 static BOOLEAN acpi_MatchHid(ACPI_HANDLE h, const char *hid);
139 static ACPI_STATUS acpi_EnterSleepState(struct acpi_softc *sc, int state);
140 static void acpi_shutdown_final(void *arg, int howto);
141 static void acpi_enable_fixed_events(struct acpi_softc *sc);
142 static int acpi_wake_sleep_prep(ACPI_HANDLE handle, int sstate);
143 static int acpi_wake_run_prep(ACPI_HANDLE handle, int sstate);
144 static int acpi_wake_prep_walk(int sstate);
145 static int acpi_wake_sysctl_walk(device_t dev);
146 static int acpi_wake_set_sysctl(SYSCTL_HANDLER_ARGS);
147 static void acpi_system_eventhandler_sleep(void *arg, int state);
148 static void acpi_system_eventhandler_wakeup(void *arg, int state);
149 static int acpi_supported_sleep_state_sysctl(SYSCTL_HANDLER_ARGS);
150 static int acpi_sleep_state_sysctl(SYSCTL_HANDLER_ARGS);
151 static int acpi_pm_func(u_long cmd, void *arg, ...);
152 static int acpi_child_location_str_method(device_t acdev, device_t child,
153 char *buf, size_t buflen);
154 static int acpi_child_pnpinfo_str_method(device_t acdev, device_t child,
155 char *buf, size_t buflen);
156
157 static device_method_t acpi_methods[] = {
158 /* Device interface */
159 DEVMETHOD(device_identify, acpi_identify),
160 DEVMETHOD(device_probe, acpi_probe),
161 DEVMETHOD(device_attach, acpi_attach),
162 DEVMETHOD(device_shutdown, acpi_shutdown),
163 DEVMETHOD(device_detach, bus_generic_detach),
164 DEVMETHOD(device_suspend, acpi_suspend),
165 DEVMETHOD(device_resume, acpi_resume),
166
167 /* Bus interface */
168 DEVMETHOD(bus_add_child, acpi_add_child),
169 DEVMETHOD(bus_print_child, acpi_print_child),
170 DEVMETHOD(bus_probe_nomatch, acpi_probe_nomatch),
171 DEVMETHOD(bus_driver_added, acpi_driver_added),
172 DEVMETHOD(bus_read_ivar, acpi_read_ivar),
173 DEVMETHOD(bus_write_ivar, acpi_write_ivar),
174 DEVMETHOD(bus_get_resource_list, acpi_get_rlist),
175 DEVMETHOD(bus_set_resource, bus_generic_rl_set_resource),
176 DEVMETHOD(bus_get_resource, bus_generic_rl_get_resource),
177 DEVMETHOD(bus_alloc_resource, acpi_alloc_resource),
178 DEVMETHOD(bus_release_resource, acpi_release_resource),
179 DEVMETHOD(bus_delete_resource, acpi_delete_resource),
180 DEVMETHOD(bus_child_pnpinfo_str, acpi_child_pnpinfo_str_method),
181 DEVMETHOD(bus_child_location_str, acpi_child_location_str_method),
182 DEVMETHOD(bus_activate_resource, bus_generic_activate_resource),
183 DEVMETHOD(bus_deactivate_resource, bus_generic_deactivate_resource),
184 DEVMETHOD(bus_setup_intr, bus_generic_setup_intr),
185 DEVMETHOD(bus_teardown_intr, bus_generic_teardown_intr),
186
187 /* ACPI bus */
188 DEVMETHOD(acpi_id_probe, acpi_device_id_probe),
189 DEVMETHOD(acpi_evaluate_object, acpi_device_eval_obj),
190 DEVMETHOD(acpi_pwr_for_sleep, acpi_device_pwr_for_sleep),
191 DEVMETHOD(acpi_scan_children, acpi_device_scan_children),
192
193 /* PCI emulation */
194 DEVMETHOD(pci_set_powerstate, acpi_set_powerstate_method),
195
196 /* ISA emulation */
197 DEVMETHOD(isa_pnp_probe, acpi_isa_pnp_probe),
198
199 {0, 0}
200 };
201
202 static driver_t acpi_driver = {
203 "acpi",
204 acpi_methods,
205 sizeof(struct acpi_softc),
206 };
207
208 static devclass_t acpi_devclass;
209 DRIVER_MODULE(acpi, nexus, acpi_driver, acpi_devclass, acpi_modevent, 0);
210 MODULE_VERSION(acpi, 1);
211
212 ACPI_SERIAL_DECL(acpi, "ACPI root bus");
213
214 /* Local pools for managing system resources for ACPI child devices. */
215 static struct rman acpi_rman_io, acpi_rman_mem;
216
217 #define ACPI_MINIMUM_AWAKETIME 5
218
219 static const char* sleep_state_names[] = {
220 "S0", "S1", "S2", "S3", "S4", "S5", "NONE"};
221
222 SYSCTL_NODE(_debug, OID_AUTO, acpi, CTLFLAG_RD, NULL, "ACPI debugging");
223 static char acpi_ca_version[12];
224 SYSCTL_STRING(_debug_acpi, OID_AUTO, acpi_ca_version, CTLFLAG_RD,
225 acpi_ca_version, 0, "Version of Intel ACPI-CA");
226
227 /*
228 * Allow override of whether methods execute in parallel or not.
229 * Enable this for serial behavior, which fixes "AE_ALREADY_EXISTS"
230 * errors for AML that really can't handle parallel method execution.
231 * It is off by default since this breaks recursive methods and
232 * some IBMs use such code.
233 */
234 static int acpi_serialize_methods;
235 TUNABLE_INT("hw.acpi.serialize_methods", &acpi_serialize_methods);
236
237 /* Power devices off and on in suspend and resume. XXX Remove once tested. */
238 static int acpi_do_powerstate = 1;
239 TUNABLE_INT("debug.acpi.do_powerstate", &acpi_do_powerstate);
240 SYSCTL_INT(_debug_acpi, OID_AUTO, do_powerstate, CTLFLAG_RW,
241 &acpi_do_powerstate, 1, "Turn off devices when suspending.");
242
243 /* Allow users to override quirks. */
244 TUNABLE_INT("debug.acpi.quirks", &acpi_quirks);
245
246 static int acpi_susp_bounce;
247 SYSCTL_INT(_debug_acpi, OID_AUTO, suspend_bounce, CTLFLAG_RW,
248 &acpi_susp_bounce, 0, "Don't actually suspend, just test devices.");
249
250 /*
251 * ACPI can only be loaded as a module by the loader; activating it after
252 * system bootstrap time is not useful, and can be fatal to the system.
253 * It also cannot be unloaded, since the entire system bus heirarchy hangs
254 * off it.
255 */
256 static int
257 acpi_modevent(struct module *mod, int event, void *junk)
258 {
259 switch (event) {
260 case MOD_LOAD:
261 if (!cold) {
262 printf("The ACPI driver cannot be loaded after boot.\n");
263 return (EPERM);
264 }
265 break;
266 case MOD_UNLOAD:
267 if (!cold && power_pm_get_type() == POWER_PM_TYPE_ACPI)
268 return (EBUSY);
269 break;
270 default:
271 break;
272 }
273 return (0);
274 }
275
276 /*
277 * Perform early initialization.
278 */
279 ACPI_STATUS
280 acpi_Startup(void)
281 {
282 static int started = 0;
283 ACPI_STATUS status;
284 int val;
285
286 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
287
288 /* Only run the startup code once. The MADT driver also calls this. */
289 if (started)
290 return_VALUE (AE_OK);
291 started = 1;
292
293 /*
294 * Pre-allocate space for RSDT/XSDT and DSDT tables and allow resizing
295 * if more tables exist.
296 */
297 if (ACPI_FAILURE(status = AcpiInitializeTables(NULL, 2, TRUE))) {
298 printf("ACPI: Table initialisation failed: %s\n",
299 AcpiFormatException(status));
300 return_VALUE (status);
301 }
302
303 /* Set up any quirks we have for this system. */
304 if (acpi_quirks == ACPI_Q_OK)
305 acpi_table_quirks(&acpi_quirks);
306
307 /* If the user manually set the disabled hint to 0, force-enable ACPI. */
308 if (resource_int_value("acpi", 0, "disabled", &val) == 0 && val == 0)
309 acpi_quirks &= ~ACPI_Q_BROKEN;
310 if (acpi_quirks & ACPI_Q_BROKEN) {
311 printf("ACPI disabled by blacklist. Contact your BIOS vendor.\n");
312 status = AE_SUPPORT;
313 }
314
315 return_VALUE (status);
316 }
317
318 /*
319 * Detect ACPI, perform early initialisation
320 */
321 static void
322 acpi_identify(driver_t *driver, device_t parent)
323 {
324 device_t child;
325
326 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
327
328 if (!cold)
329 return_VOID;
330
331 /* Check that we haven't been disabled with a hint. */
332 if (resource_disabled("acpi", 0))
333 return_VOID;
334
335 /* Make sure we're not being doubly invoked. */
336 if (device_find_child(parent, "acpi", 0) != NULL)
337 return_VOID;
338
339 snprintf(acpi_ca_version, sizeof(acpi_ca_version), "%x", ACPI_CA_VERSION);
340
341 /* Initialize root tables. */
342 if (ACPI_FAILURE(acpi_Startup())) {
343 printf("ACPI: Try disabling either ACPI or apic support.\n");
344 return_VOID;
345 }
346
347 /* Attach the actual ACPI device. */
348 if ((child = BUS_ADD_CHILD(parent, 10, "acpi", 0)) == NULL) {
349 device_printf(parent, "device_identify failed\n");
350 return_VOID;
351 }
352 }
353
354 /*
355 * Fetch some descriptive data from ACPI to put in our attach message.
356 */
357 static int
358 acpi_probe(device_t dev)
359 {
360 ACPI_TABLE_RSDP *rsdp;
361 ACPI_TABLE_HEADER *rsdt;
362 ACPI_PHYSICAL_ADDRESS paddr;
363 char buf[ACPI_OEM_ID_SIZE + ACPI_OEM_TABLE_ID_SIZE + 2];
364 struct sbuf sb;
365
366 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
367
368 if (power_pm_get_type() != POWER_PM_TYPE_NONE &&
369 power_pm_get_type() != POWER_PM_TYPE_ACPI) {
370 device_printf(dev, "probe failed, other PM system enabled.\n");
371 return_VALUE (ENXIO);
372 }
373
374 if ((paddr = AcpiOsGetRootPointer()) == 0 ||
375 (rsdp = AcpiOsMapMemory(paddr, sizeof(ACPI_TABLE_RSDP))) == NULL)
376 return_VALUE (ENXIO);
377 if (rsdp->Revision > 1 && rsdp->XsdtPhysicalAddress != 0)
378 paddr = (ACPI_PHYSICAL_ADDRESS)rsdp->XsdtPhysicalAddress;
379 else
380 paddr = (ACPI_PHYSICAL_ADDRESS)rsdp->RsdtPhysicalAddress;
381 AcpiOsUnmapMemory(rsdp, sizeof(ACPI_TABLE_RSDP));
382
383 if ((rsdt = AcpiOsMapMemory(paddr, sizeof(ACPI_TABLE_HEADER))) == NULL)
384 return_VALUE (ENXIO);
385 sbuf_new(&sb, buf, sizeof(buf), SBUF_FIXEDLEN);
386 sbuf_bcat(&sb, rsdt->OemId, ACPI_OEM_ID_SIZE);
387 sbuf_trim(&sb);
388 sbuf_putc(&sb, ' ');
389 sbuf_bcat(&sb, rsdt->OemTableId, ACPI_OEM_TABLE_ID_SIZE);
390 sbuf_trim(&sb);
391 sbuf_finish(&sb);
392 device_set_desc_copy(dev, sbuf_data(&sb));
393 sbuf_delete(&sb);
394 AcpiOsUnmapMemory(rsdt, sizeof(ACPI_TABLE_HEADER));
395
396 return_VALUE (0);
397 }
398
399 static int
400 acpi_attach(device_t dev)
401 {
402 struct acpi_softc *sc;
403 ACPI_TABLE_FACS *facs;
404 ACPI_STATUS status;
405 int error, state;
406 UINT32 flags;
407 UINT8 TypeA, TypeB;
408 char *env;
409
410 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
411
412 sc = device_get_softc(dev);
413 sc->acpi_dev = dev;
414 callout_init(&sc->susp_force_to, TRUE);
415
416 error = ENXIO;
417
418 /* Initialize resource manager. */
419 acpi_rman_io.rm_type = RMAN_ARRAY;
420 acpi_rman_io.rm_start = 0;
421 acpi_rman_io.rm_end = 0xffff;
422 acpi_rman_io.rm_descr = "ACPI I/O ports";
423 if (rman_init(&acpi_rman_io) != 0)
424 panic("acpi rman_init IO ports failed");
425 acpi_rman_mem.rm_type = RMAN_ARRAY;
426 acpi_rman_mem.rm_start = 0;
427 acpi_rman_mem.rm_end = ~0ul;
428 acpi_rman_mem.rm_descr = "ACPI I/O memory addresses";
429 if (rman_init(&acpi_rman_mem) != 0)
430 panic("acpi rman_init memory failed");
431
432 /* Initialise the ACPI mutex */
433 mtx_init(&acpi_mutex, "ACPI global lock", NULL, MTX_DEF);
434
435 /*
436 * Set the globals from our tunables. This is needed because ACPI-CA
437 * uses UINT8 for some values and we have no tunable_byte.
438 */
439 AcpiGbl_AllMethodsSerialized = acpi_serialize_methods;
440 AcpiGbl_EnableInterpreterSlack = TRUE;
441
442 /* Start up the ACPI CA subsystem. */
443 status = AcpiInitializeSubsystem();
444 if (ACPI_FAILURE(status)) {
445 device_printf(dev, "Could not initialize Subsystem: %s\n",
446 AcpiFormatException(status));
447 goto out;
448 }
449
450 /* Load ACPI name space. */
451 status = AcpiLoadTables();
452 if (ACPI_FAILURE(status)) {
453 device_printf(dev, "Could not load Namespace: %s\n",
454 AcpiFormatException(status));
455 goto out;
456 }
457
458 /* Install the default address space handlers. */
459 status = AcpiInstallAddressSpaceHandler(ACPI_ROOT_OBJECT,
460 ACPI_ADR_SPACE_SYSTEM_MEMORY, ACPI_DEFAULT_HANDLER, NULL, NULL);
461 if (ACPI_FAILURE(status)) {
462 device_printf(dev, "Could not initialise SystemMemory handler: %s\n",
463 AcpiFormatException(status));
464 goto out;
465 }
466 status = AcpiInstallAddressSpaceHandler(ACPI_ROOT_OBJECT,
467 ACPI_ADR_SPACE_SYSTEM_IO, ACPI_DEFAULT_HANDLER, NULL, NULL);
468 if (ACPI_FAILURE(status)) {
469 device_printf(dev, "Could not initialise SystemIO handler: %s\n",
470 AcpiFormatException(status));
471 goto out;
472 }
473 status = AcpiInstallAddressSpaceHandler(ACPI_ROOT_OBJECT,
474 ACPI_ADR_SPACE_PCI_CONFIG, ACPI_DEFAULT_HANDLER, NULL, NULL);
475 if (ACPI_FAILURE(status)) {
476 device_printf(dev, "could not initialise PciConfig handler: %s\n",
477 AcpiFormatException(status));
478 goto out;
479 }
480
481 /*
482 * Note that some systems (specifically, those with namespace evaluation
483 * issues that require the avoidance of parts of the namespace) must
484 * avoid running _INI and _STA on everything, as well as dodging the final
485 * object init pass.
486 *
487 * For these devices, we set ACPI_NO_DEVICE_INIT and ACPI_NO_OBJECT_INIT).
488 *
489 * XXX We should arrange for the object init pass after we have attached
490 * all our child devices, but on many systems it works here.
491 */
492 flags = 0;
493 if (testenv("debug.acpi.avoid"))
494 flags = ACPI_NO_DEVICE_INIT | ACPI_NO_OBJECT_INIT;
495
496 /* Bring the hardware and basic handlers online. */
497 if (ACPI_FAILURE(status = AcpiEnableSubsystem(flags))) {
498 device_printf(dev, "Could not enable ACPI: %s\n",
499 AcpiFormatException(status));
500 goto out;
501 }
502
503 /*
504 * Call the ECDT probe function to provide EC functionality before
505 * the namespace has been evaluated.
506 *
507 * XXX This happens before the sysresource devices have been probed and
508 * attached so its resources come from nexus0. In practice, this isn't
509 * a problem but should be addressed eventually.
510 */
511 acpi_ec_ecdt_probe(dev);
512
513 /* Bring device objects and regions online. */
514 if (ACPI_FAILURE(status = AcpiInitializeObjects(flags))) {
515 device_printf(dev, "Could not initialize ACPI objects: %s\n",
516 AcpiFormatException(status));
517 goto out;
518 }
519
520 /*
521 * Setup our sysctl tree.
522 *
523 * XXX: This doesn't check to make sure that none of these fail.
524 */
525 sysctl_ctx_init(&sc->acpi_sysctl_ctx);
526 sc->acpi_sysctl_tree = SYSCTL_ADD_NODE(&sc->acpi_sysctl_ctx,
527 SYSCTL_STATIC_CHILDREN(_hw), OID_AUTO,
528 device_get_name(dev), CTLFLAG_RD, 0, "");
529 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
530 OID_AUTO, "supported_sleep_state", CTLTYPE_STRING | CTLFLAG_RD,
531 0, 0, acpi_supported_sleep_state_sysctl, "A", "");
532 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
533 OID_AUTO, "power_button_state", CTLTYPE_STRING | CTLFLAG_RW,
534 &sc->acpi_power_button_sx, 0, acpi_sleep_state_sysctl, "A", "");
535 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
536 OID_AUTO, "sleep_button_state", CTLTYPE_STRING | CTLFLAG_RW,
537 &sc->acpi_sleep_button_sx, 0, acpi_sleep_state_sysctl, "A", "");
538 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
539 OID_AUTO, "lid_switch_state", CTLTYPE_STRING | CTLFLAG_RW,
540 &sc->acpi_lid_switch_sx, 0, acpi_sleep_state_sysctl, "A", "");
541 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
542 OID_AUTO, "standby_state", CTLTYPE_STRING | CTLFLAG_RW,
543 &sc->acpi_standby_sx, 0, acpi_sleep_state_sysctl, "A", "");
544 SYSCTL_ADD_PROC(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
545 OID_AUTO, "suspend_state", CTLTYPE_STRING | CTLFLAG_RW,
546 &sc->acpi_suspend_sx, 0, acpi_sleep_state_sysctl, "A", "");
547 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
548 OID_AUTO, "sleep_delay", CTLFLAG_RW, &sc->acpi_sleep_delay, 0,
549 "sleep delay");
550 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
551 OID_AUTO, "s4bios", CTLFLAG_RW, &sc->acpi_s4bios, 0, "S4BIOS mode");
552 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
553 OID_AUTO, "verbose", CTLFLAG_RW, &sc->acpi_verbose, 0, "verbose mode");
554 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
555 OID_AUTO, "disable_on_reboot", CTLFLAG_RW,
556 &sc->acpi_do_disable, 0, "Disable ACPI when rebooting/halting system");
557 SYSCTL_ADD_INT(&sc->acpi_sysctl_ctx, SYSCTL_CHILDREN(sc->acpi_sysctl_tree),
558 OID_AUTO, "handle_reboot", CTLFLAG_RW,
559 &sc->acpi_handle_reboot, 0, "Use ACPI Reset Register to reboot");
560
561 /*
562 * Default to 1 second before sleeping to give some machines time to
563 * stabilize.
564 */
565 sc->acpi_sleep_delay = 1;
566 if (bootverbose)
567 sc->acpi_verbose = 1;
568 if ((env = getenv("hw.acpi.verbose")) != NULL) {
569 if (strcmp(env, "") != 0)
570 sc->acpi_verbose = 1;
571 freeenv(env);
572 }
573
574 /* Only enable S4BIOS by default if the FACS says it is available. */
575 status = AcpiGetTable(ACPI_SIG_FACS, 0, (ACPI_TABLE_HEADER **)&facs);
576 if (ACPI_FAILURE(status)) {
577 device_printf(dev, "couldn't get FACS: %s\n",
578 AcpiFormatException(status));
579 error = ENXIO;
580 goto out;
581 }
582 if (facs->Flags & ACPI_FACS_S4_BIOS_PRESENT)
583 sc->acpi_s4bios = 1;
584
585 /*
586 * Dispatch the default sleep state to devices. The lid switch is set
587 * to NONE by default to avoid surprising users.
588 */
589 sc->acpi_power_button_sx = ACPI_STATE_S5;
590 sc->acpi_lid_switch_sx = ACPI_S_STATES_MAX + 1;
591 sc->acpi_standby_sx = ACPI_STATE_S1;
592 sc->acpi_suspend_sx = ACPI_STATE_S3;
593
594 /* Pick the first valid sleep state for the sleep button default. */
595 sc->acpi_sleep_button_sx = ACPI_S_STATES_MAX + 1;
596 for (state = ACPI_STATE_S1; state <= ACPI_STATE_S4; state++)
597 if (ACPI_SUCCESS(AcpiGetSleepTypeData(state, &TypeA, &TypeB))) {
598 sc->acpi_sleep_button_sx = state;
599 break;
600 }
601
602 acpi_enable_fixed_events(sc);
603
604 /*
605 * Scan the namespace and attach/initialise children.
606 */
607
608 /* Register our shutdown handler. */
609 EVENTHANDLER_REGISTER(shutdown_final, acpi_shutdown_final, sc,
610 SHUTDOWN_PRI_LAST);
611
612 /*
613 * Register our acpi event handlers.
614 * XXX should be configurable eg. via userland policy manager.
615 */
616 EVENTHANDLER_REGISTER(acpi_sleep_event, acpi_system_eventhandler_sleep,
617 sc, ACPI_EVENT_PRI_LAST);
618 EVENTHANDLER_REGISTER(acpi_wakeup_event, acpi_system_eventhandler_wakeup,
619 sc, ACPI_EVENT_PRI_LAST);
620
621 /* Flag our initial states. */
622 sc->acpi_enabled = 1;
623 sc->acpi_sstate = ACPI_STATE_S0;
624 sc->acpi_sleep_disabled = 0;
625
626 /* Create the control device */
627 sc->acpi_dev_t = make_dev(&acpi_cdevsw, 0, UID_ROOT, GID_WHEEL, 0644,
628 "acpi");
629 sc->acpi_dev_t->si_drv1 = sc;
630
631 if ((error = acpi_machdep_init(dev)))
632 goto out;
633
634 /* Register ACPI again to pass the correct argument of pm_func. */
635 power_pm_register(POWER_PM_TYPE_ACPI, acpi_pm_func, sc);
636
637 if (!acpi_disabled("bus"))
638 acpi_probe_children(dev);
639
640 error = 0;
641
642 out:
643 return_VALUE (error);
644 }
645
646 static int
647 acpi_suspend(device_t dev)
648 {
649 device_t child, *devlist;
650 int error, i, numdevs, pstate;
651
652 GIANT_REQUIRED;
653
654 /* First give child devices a chance to suspend. */
655 error = bus_generic_suspend(dev);
656 if (error)
657 return (error);
658
659 /*
660 * Now, set them into the appropriate power state, usually D3. If the
661 * device has an _SxD method for the next sleep state, use that power
662 * state instead.
663 */
664 device_get_children(dev, &devlist, &numdevs);
665 for (i = 0; i < numdevs; i++) {
666 /* If the device is not attached, we've powered it down elsewhere. */
667 child = devlist[i];
668 if (!device_is_attached(child))
669 continue;
670
671 /*
672 * Default to D3 for all sleep states. The _SxD method is optional
673 * so set the powerstate even if it's absent.
674 */
675 pstate = PCI_POWERSTATE_D3;
676 error = acpi_device_pwr_for_sleep(device_get_parent(child),
677 child, &pstate);
678 if ((error == 0 || error == ESRCH) && acpi_do_powerstate)
679 pci_set_powerstate(child, pstate);
680 }
681 free(devlist, M_TEMP);
682 error = 0;
683
684 return (error);
685 }
686
687 static int
688 acpi_resume(device_t dev)
689 {
690 ACPI_HANDLE handle;
691 int i, numdevs;
692 device_t child, *devlist;
693
694 GIANT_REQUIRED;
695
696 /*
697 * Put all devices in D0 before resuming them. Call _S0D on each one
698 * since some systems expect this.
699 */
700 device_get_children(dev, &devlist, &numdevs);
701 for (i = 0; i < numdevs; i++) {
702 child = devlist[i];
703 handle = acpi_get_handle(child);
704 if (handle)
705 AcpiEvaluateObject(handle, "_S0D", NULL, NULL);
706 if (device_is_attached(child) && acpi_do_powerstate)
707 pci_set_powerstate(child, PCI_POWERSTATE_D0);
708 }
709 free(devlist, M_TEMP);
710
711 return (bus_generic_resume(dev));
712 }
713
714 static int
715 acpi_shutdown(device_t dev)
716 {
717
718 GIANT_REQUIRED;
719
720 /* Allow children to shutdown first. */
721 bus_generic_shutdown(dev);
722
723 /*
724 * Enable any GPEs that are able to power-on the system (i.e., RTC).
725 * Also, disable any that are not valid for this state (most).
726 */
727 acpi_wake_prep_walk(ACPI_STATE_S5);
728
729 return (0);
730 }
731
732 /*
733 * Handle a new device being added
734 */
735 static device_t
736 acpi_add_child(device_t bus, int order, const char *name, int unit)
737 {
738 struct acpi_device *ad;
739 device_t child;
740
741 if ((ad = malloc(sizeof(*ad), M_ACPIDEV, M_NOWAIT | M_ZERO)) == NULL)
742 return (NULL);
743
744 resource_list_init(&ad->ad_rl);
745
746 child = device_add_child_ordered(bus, order, name, unit);
747 if (child != NULL)
748 device_set_ivars(child, ad);
749 else
750 free(ad, M_ACPIDEV);
751 return (child);
752 }
753
754 static int
755 acpi_print_child(device_t bus, device_t child)
756 {
757 struct acpi_device *adev = device_get_ivars(child);
758 struct resource_list *rl = &adev->ad_rl;
759 int retval = 0;
760
761 retval += bus_print_child_header(bus, child);
762 retval += resource_list_print_type(rl, "port", SYS_RES_IOPORT, "%#lx");
763 retval += resource_list_print_type(rl, "iomem", SYS_RES_MEMORY, "%#lx");
764 retval += resource_list_print_type(rl, "irq", SYS_RES_IRQ, "%ld");
765 retval += resource_list_print_type(rl, "drq", SYS_RES_DRQ, "%ld");
766 if (device_get_flags(child))
767 retval += printf(" flags %#x", device_get_flags(child));
768 retval += bus_print_child_footer(bus, child);
769
770 return (retval);
771 }
772
773 /*
774 * If this device is an ACPI child but no one claimed it, attempt
775 * to power it off. We'll power it back up when a driver is added.
776 *
777 * XXX Disabled for now since many necessary devices (like fdc and
778 * ATA) don't claim the devices we created for them but still expect
779 * them to be powered up.
780 */
781 static void
782 acpi_probe_nomatch(device_t bus, device_t child)
783 {
784
785 /* pci_set_powerstate(child, PCI_POWERSTATE_D3); */
786 }
787
788 /*
789 * If a new driver has a chance to probe a child, first power it up.
790 *
791 * XXX Disabled for now (see acpi_probe_nomatch for details).
792 */
793 static void
794 acpi_driver_added(device_t dev, driver_t *driver)
795 {
796 device_t child, *devlist;
797 int i, numdevs;
798
799 DEVICE_IDENTIFY(driver, dev);
800 device_get_children(dev, &devlist, &numdevs);
801 for (i = 0; i < numdevs; i++) {
802 child = devlist[i];
803 if (device_get_state(child) == DS_NOTPRESENT) {
804 /* pci_set_powerstate(child, PCI_POWERSTATE_D0); */
805 if (device_probe_and_attach(child) != 0)
806 ; /* pci_set_powerstate(child, PCI_POWERSTATE_D3); */
807 }
808 }
809 free(devlist, M_TEMP);
810 }
811
812 /* Location hint for devctl(8) */
813 static int
814 acpi_child_location_str_method(device_t cbdev, device_t child, char *buf,
815 size_t buflen)
816 {
817 struct acpi_device *dinfo = device_get_ivars(child);
818
819 if (dinfo->ad_handle)
820 snprintf(buf, buflen, "handle=%s", acpi_name(dinfo->ad_handle));
821 else
822 snprintf(buf, buflen, "unknown");
823 return (0);
824 }
825
826 /* PnP information for devctl(8) */
827 static int
828 acpi_child_pnpinfo_str_method(device_t cbdev, device_t child, char *buf,
829 size_t buflen)
830 {
831 ACPI_BUFFER adbuf = {ACPI_ALLOCATE_BUFFER, NULL};
832 ACPI_DEVICE_INFO *adinfo;
833 struct acpi_device *dinfo = device_get_ivars(child);
834 char *end;
835 int error;
836
837 error = AcpiGetObjectInfo(dinfo->ad_handle, &adbuf);
838 adinfo = (ACPI_DEVICE_INFO *) adbuf.Pointer;
839 if (error)
840 snprintf(buf, buflen, "unknown");
841 else
842 snprintf(buf, buflen, "_HID=%s _UID=%lu",
843 (adinfo->Valid & ACPI_VALID_HID) ?
844 adinfo->HardwareId.Value : "none",
845 (adinfo->Valid & ACPI_VALID_UID) ?
846 strtoul(adinfo->UniqueId.Value, &end, 10) : 0);
847 if (adinfo)
848 AcpiOsFree(adinfo);
849
850 return (0);
851 }
852
853 /*
854 * Handle per-device ivars
855 */
856 static int
857 acpi_read_ivar(device_t dev, device_t child, int index, uintptr_t *result)
858 {
859 struct acpi_device *ad;
860
861 if ((ad = device_get_ivars(child)) == NULL) {
862 printf("device has no ivars\n");
863 return (ENOENT);
864 }
865
866 /* ACPI and ISA compatibility ivars */
867 switch(index) {
868 case ACPI_IVAR_HANDLE:
869 *(ACPI_HANDLE *)result = ad->ad_handle;
870 break;
871 case ACPI_IVAR_MAGIC:
872 *(uintptr_t *)result = ad->ad_magic;
873 break;
874 case ACPI_IVAR_PRIVATE:
875 *(void **)result = ad->ad_private;
876 break;
877 case ACPI_IVAR_FLAGS:
878 *(int *)result = ad->ad_flags;
879 break;
880 case ISA_IVAR_VENDORID:
881 case ISA_IVAR_SERIAL:
882 case ISA_IVAR_COMPATID:
883 *(int *)result = -1;
884 break;
885 case ISA_IVAR_LOGICALID:
886 *(int *)result = acpi_isa_get_logicalid(child);
887 break;
888 default:
889 return (ENOENT);
890 }
891
892 return (0);
893 }
894
895 static int
896 acpi_write_ivar(device_t dev, device_t child, int index, uintptr_t value)
897 {
898 struct acpi_device *ad;
899
900 if ((ad = device_get_ivars(child)) == NULL) {
901 printf("device has no ivars\n");
902 return (ENOENT);
903 }
904
905 switch(index) {
906 case ACPI_IVAR_HANDLE:
907 ad->ad_handle = (ACPI_HANDLE)value;
908 break;
909 case ACPI_IVAR_MAGIC:
910 ad->ad_magic = (uintptr_t)value;
911 break;
912 case ACPI_IVAR_PRIVATE:
913 ad->ad_private = (void *)value;
914 break;
915 case ACPI_IVAR_FLAGS:
916 ad->ad_flags = (int)value;
917 break;
918 default:
919 panic("bad ivar write request (%d)", index);
920 return (ENOENT);
921 }
922
923 return (0);
924 }
925
926 /*
927 * Handle child resource allocation/removal
928 */
929 static struct resource_list *
930 acpi_get_rlist(device_t dev, device_t child)
931 {
932 struct acpi_device *ad;
933
934 ad = device_get_ivars(child);
935 return (&ad->ad_rl);
936 }
937
938 /*
939 * Pre-allocate/manage all memory and IO resources. Since rman can't handle
940 * duplicates, we merge any in the sysresource attach routine.
941 */
942 static int
943 acpi_sysres_alloc(device_t dev)
944 {
945 struct resource *res;
946 struct resource_list *rl;
947 struct resource_list_entry *rle;
948 struct rman *rm;
949 char *sysres_ids[] = { "PNP0C01", "PNP0C02", NULL };
950 device_t *children;
951 int child_count, i;
952
953 /*
954 * Probe/attach any sysresource devices. This would be unnecessary if we
955 * had multi-pass probe/attach.
956 */
957 if (device_get_children(dev, &children, &child_count) != 0)
958 return (ENXIO);
959 for (i = 0; i < child_count; i++) {
960 if (ACPI_ID_PROBE(dev, children[i], sysres_ids) != NULL)
961 device_probe_and_attach(children[i]);
962 }
963 free(children, M_TEMP);
964
965 rl = BUS_GET_RESOURCE_LIST(device_get_parent(dev), dev);
966 STAILQ_FOREACH(rle, rl, link) {
967 if (rle->res != NULL) {
968 device_printf(dev, "duplicate resource for %lx\n", rle->start);
969 continue;
970 }
971
972 /* Only memory and IO resources are valid here. */
973 switch (rle->type) {
974 case SYS_RES_IOPORT:
975 rm = &acpi_rman_io;
976 break;
977 case SYS_RES_MEMORY:
978 rm = &acpi_rman_mem;
979 break;
980 default:
981 continue;
982 }
983
984 /* Pre-allocate resource and add to our rman pool. */
985 res = BUS_ALLOC_RESOURCE(device_get_parent(dev), dev, rle->type,
986 &rle->rid, rle->start, rle->start + rle->count - 1, rle->count, 0);
987 if (res != NULL) {
988 rman_manage_region(rm, rman_get_start(res), rman_get_end(res));
989 rle->res = res;
990 } else
991 device_printf(dev, "reservation of %lx, %lx (%d) failed\n",
992 rle->start, rle->count, rle->type);
993 }
994 return (0);
995 }
996
997 static struct resource *
998 acpi_alloc_resource(device_t bus, device_t child, int type, int *rid,
999 u_long start, u_long end, u_long count, u_int flags)
1000 {
1001 ACPI_RESOURCE ares;
1002 struct acpi_device *ad = device_get_ivars(child);
1003 struct resource_list *rl = &ad->ad_rl;
1004 struct resource_list_entry *rle;
1005 struct resource *res;
1006 struct rman *rm;
1007
1008 res = NULL;
1009
1010 /* We only handle memory and IO resources through rman. */
1011 switch (type) {
1012 case SYS_RES_IOPORT:
1013 rm = &acpi_rman_io;
1014 break;
1015 case SYS_RES_MEMORY:
1016 rm = &acpi_rman_mem;
1017 break;
1018 default:
1019 rm = NULL;
1020 }
1021
1022 ACPI_SERIAL_BEGIN(acpi);
1023
1024 /*
1025 * If this is an allocation of the "default" range for a given RID, and
1026 * we know what the resources for this device are (i.e., they're on the
1027 * child's resource list), use those start/end values.
1028 */
1029 if (bus == device_get_parent(child) && start == 0UL && end == ~0UL) {
1030 rle = resource_list_find(rl, type, *rid);
1031 if (rle == NULL)
1032 goto out;
1033 start = rle->start;
1034 end = rle->end;
1035 count = rle->count;
1036 }
1037
1038 /*
1039 * If this is an allocation of a specific range, see if we can satisfy
1040 * the request from our system resource regions. If we can't, pass the
1041 * request up to the parent.
1042 */
1043 if (start + count - 1 == end && rm != NULL)
1044 res = rman_reserve_resource(rm, start, end, count, flags & ~RF_ACTIVE,
1045 child);
1046 if (res == NULL) {
1047 res = BUS_ALLOC_RESOURCE(device_get_parent(bus), child, type, rid,
1048 start, end, count, flags);
1049 } else {
1050 rman_set_rid(res, *rid);
1051
1052 /* If requested, activate the resource using the parent's method. */
1053 if (flags & RF_ACTIVE)
1054 if (bus_activate_resource(child, type, *rid, res) != 0) {
1055 rman_release_resource(res);
1056 res = NULL;
1057 goto out;
1058 }
1059 }
1060
1061 if (res != NULL && device_get_parent(child) == bus)
1062 switch (type) {
1063 case SYS_RES_IRQ:
1064 /*
1065 * Since bus_config_intr() takes immediate effect, we cannot
1066 * configure the interrupt associated with a device when we
1067 * parse the resources but have to defer it until a driver
1068 * actually allocates the interrupt via bus_alloc_resource().
1069 *
1070 * XXX: Should we handle the lookup failing?
1071 */
1072 if (ACPI_SUCCESS(acpi_lookup_irq_resource(child, *rid, res, &ares)))
1073 acpi_config_intr(child, &ares);
1074 break;
1075 }
1076
1077 out:
1078 ACPI_SERIAL_END(acpi);
1079 return (res);
1080 }
1081
1082 static int
1083 acpi_release_resource(device_t bus, device_t child, int type, int rid,
1084 struct resource *r)
1085 {
1086 struct rman *rm;
1087 int ret;
1088
1089 /* We only handle memory and IO resources through rman. */
1090 switch (type) {
1091 case SYS_RES_IOPORT:
1092 rm = &acpi_rman_io;
1093 break;
1094 case SYS_RES_MEMORY:
1095 rm = &acpi_rman_mem;
1096 break;
1097 default:
1098 rm = NULL;
1099 }
1100
1101 ACPI_SERIAL_BEGIN(acpi);
1102
1103 /*
1104 * If this resource belongs to one of our internal managers,
1105 * deactivate it and release it to the local pool. If it doesn't,
1106 * pass this request up to the parent.
1107 */
1108 if (rm != NULL && rman_is_region_manager(r, rm)) {
1109 if (rman_get_flags(r) & RF_ACTIVE) {
1110 ret = bus_deactivate_resource(child, type, rid, r);
1111 if (ret != 0)
1112 goto out;
1113 }
1114 ret = rman_release_resource(r);
1115 } else
1116 ret = BUS_RELEASE_RESOURCE(device_get_parent(bus), child, type, rid, r);
1117
1118 out:
1119 ACPI_SERIAL_END(acpi);
1120 return (ret);
1121 }
1122
1123 static void
1124 acpi_delete_resource(device_t bus, device_t child, int type, int rid)
1125 {
1126 struct resource_list *rl;
1127
1128 rl = acpi_get_rlist(bus, child);
1129 resource_list_delete(rl, type, rid);
1130 }
1131
1132 /* Allocate an IO port or memory resource, given its GAS. */
1133 int
1134 acpi_bus_alloc_gas(device_t dev, int *type, int *rid, ACPI_GENERIC_ADDRESS *gas,
1135 struct resource **res, u_int flags)
1136 {
1137 int error, res_type;
1138
1139 error = ENOMEM;
1140 if (type == NULL || rid == NULL || gas == NULL || res == NULL)
1141 return (EINVAL);
1142
1143 /* We only support memory and IO spaces. */
1144 switch (gas->SpaceId) {
1145 case ACPI_ADR_SPACE_SYSTEM_MEMORY:
1146 res_type = SYS_RES_MEMORY;
1147 break;
1148 case ACPI_ADR_SPACE_SYSTEM_IO:
1149 res_type = SYS_RES_IOPORT;
1150 break;
1151 default:
1152 return (EOPNOTSUPP);
1153 }
1154
1155 /*
1156 * If the register width is less than 8, assume the BIOS author means
1157 * it is a bit field and just allocate a byte.
1158 */
1159 if (gas->BitWidth && gas->BitWidth < 8)
1160 gas->BitWidth = 8;
1161
1162 /* Validate the address after we're sure we support the space. */
1163 if (gas->Address == 0 || gas->BitWidth == 0)
1164 return (EINVAL);
1165
1166 bus_set_resource(dev, res_type, *rid, gas->Address,
1167 gas->BitWidth / 8);
1168 *res = bus_alloc_resource_any(dev, res_type, rid, RF_ACTIVE | flags);
1169 if (*res != NULL) {
1170 *type = res_type;
1171 error = 0;
1172 } else
1173 bus_delete_resource(dev, res_type, *rid);
1174
1175 return (error);
1176 }
1177
1178 /* Probe _HID and _CID for compatible ISA PNP ids. */
1179 static uint32_t
1180 acpi_isa_get_logicalid(device_t dev)
1181 {
1182 ACPI_DEVICE_INFO *devinfo;
1183 ACPI_BUFFER buf;
1184 ACPI_HANDLE h;
1185 ACPI_STATUS error;
1186 u_int32_t pnpid;
1187
1188 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1189
1190 pnpid = 0;
1191 buf.Pointer = NULL;
1192 buf.Length = ACPI_ALLOCATE_BUFFER;
1193
1194 /* Fetch and validate the HID. */
1195 if ((h = acpi_get_handle(dev)) == NULL)
1196 goto out;
1197 error = AcpiGetObjectInfo(h, &buf);
1198 if (ACPI_FAILURE(error))
1199 goto out;
1200 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1201
1202 if ((devinfo->Valid & ACPI_VALID_HID) != 0)
1203 pnpid = PNP_EISAID(devinfo->HardwareId.Value);
1204
1205 out:
1206 if (buf.Pointer != NULL)
1207 AcpiOsFree(buf.Pointer);
1208 return_VALUE (pnpid);
1209 }
1210
1211 static int
1212 acpi_isa_get_compatid(device_t dev, uint32_t *cids, int count)
1213 {
1214 ACPI_DEVICE_INFO *devinfo;
1215 ACPI_BUFFER buf;
1216 ACPI_HANDLE h;
1217 ACPI_STATUS error;
1218 uint32_t *pnpid;
1219 int valid, i;
1220
1221 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1222
1223 pnpid = cids;
1224 valid = 0;
1225 buf.Pointer = NULL;
1226 buf.Length = ACPI_ALLOCATE_BUFFER;
1227
1228 /* Fetch and validate the CID */
1229 if ((h = acpi_get_handle(dev)) == NULL)
1230 goto out;
1231 error = AcpiGetObjectInfo(h, &buf);
1232 if (ACPI_FAILURE(error))
1233 goto out;
1234 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1235 if ((devinfo->Valid & ACPI_VALID_CID) == 0)
1236 goto out;
1237
1238 if (devinfo->CompatibilityId.Count < count)
1239 count = devinfo->CompatibilityId.Count;
1240 for (i = 0; i < count; i++) {
1241 if (strncmp(devinfo->CompatibilityId.Id[i].Value, "PNP", 3) != 0)
1242 continue;
1243 *pnpid++ = PNP_EISAID(devinfo->CompatibilityId.Id[i].Value);
1244 valid++;
1245 }
1246
1247 out:
1248 if (buf.Pointer != NULL)
1249 AcpiOsFree(buf.Pointer);
1250 return_VALUE (valid);
1251 }
1252
1253 static char *
1254 acpi_device_id_probe(device_t bus, device_t dev, char **ids)
1255 {
1256 ACPI_HANDLE h;
1257 int i;
1258
1259 h = acpi_get_handle(dev);
1260 if (ids == NULL || h == NULL || acpi_get_type(dev) != ACPI_TYPE_DEVICE)
1261 return (NULL);
1262
1263 /* Try to match one of the array of IDs with a HID or CID. */
1264 for (i = 0; ids[i] != NULL; i++) {
1265 if (acpi_MatchHid(h, ids[i]))
1266 return (ids[i]);
1267 }
1268 return (NULL);
1269 }
1270
1271 static ACPI_STATUS
1272 acpi_device_eval_obj(device_t bus, device_t dev, ACPI_STRING pathname,
1273 ACPI_OBJECT_LIST *parameters, ACPI_BUFFER *ret)
1274 {
1275 ACPI_HANDLE h;
1276
1277 if (dev == NULL)
1278 h = ACPI_ROOT_OBJECT;
1279 else if ((h = acpi_get_handle(dev)) == NULL)
1280 return (AE_BAD_PARAMETER);
1281 return (AcpiEvaluateObject(h, pathname, parameters, ret));
1282 }
1283
1284 static int
1285 acpi_device_pwr_for_sleep(device_t bus, device_t dev, int *dstate)
1286 {
1287 struct acpi_softc *sc;
1288 ACPI_HANDLE handle;
1289 ACPI_STATUS status;
1290 char sxd[8];
1291 int error;
1292
1293 sc = device_get_softc(bus);
1294 handle = acpi_get_handle(dev);
1295
1296 /*
1297 * XXX If we find these devices, don't try to power them down.
1298 * The serial and IRDA ports on my T23 hang the system when
1299 * set to D3 and it appears that such legacy devices may
1300 * need special handling in their drivers.
1301 */
1302 if (handle == NULL ||
1303 acpi_MatchHid(handle, "PNP0500") ||
1304 acpi_MatchHid(handle, "PNP0501") ||
1305 acpi_MatchHid(handle, "PNP0502") ||
1306 acpi_MatchHid(handle, "PNP0510") ||
1307 acpi_MatchHid(handle, "PNP0511"))
1308 return (ENXIO);
1309
1310 /*
1311 * Override next state with the value from _SxD, if present. If no
1312 * dstate argument was provided, don't fetch the return value.
1313 */
1314 snprintf(sxd, sizeof(sxd), "_S%dD", sc->acpi_sstate);
1315 if (dstate)
1316 status = acpi_GetInteger(handle, sxd, dstate);
1317 else
1318 status = AcpiEvaluateObject(handle, sxd, NULL, NULL);
1319
1320 switch (status) {
1321 case AE_OK:
1322 error = 0;
1323 break;
1324 case AE_NOT_FOUND:
1325 error = ESRCH;
1326 break;
1327 default:
1328 error = ENXIO;
1329 break;
1330 }
1331
1332 return (error);
1333 }
1334
1335 /* Callback arg for our implementation of walking the namespace. */
1336 struct acpi_device_scan_ctx {
1337 acpi_scan_cb_t user_fn;
1338 void *arg;
1339 ACPI_HANDLE parent;
1340 };
1341
1342 static ACPI_STATUS
1343 acpi_device_scan_cb(ACPI_HANDLE h, UINT32 level, void *arg, void **retval)
1344 {
1345 struct acpi_device_scan_ctx *ctx;
1346 device_t dev, old_dev;
1347 ACPI_STATUS status;
1348 ACPI_OBJECT_TYPE type;
1349
1350 /*
1351 * Skip this device if we think we'll have trouble with it or it is
1352 * the parent where the scan began.
1353 */
1354 ctx = (struct acpi_device_scan_ctx *)arg;
1355 if (acpi_avoid(h) || h == ctx->parent)
1356 return (AE_OK);
1357
1358 /* If this is not a valid device type (e.g., a method), skip it. */
1359 if (ACPI_FAILURE(AcpiGetType(h, &type)))
1360 return (AE_OK);
1361 if (type != ACPI_TYPE_DEVICE && type != ACPI_TYPE_PROCESSOR &&
1362 type != ACPI_TYPE_THERMAL && type != ACPI_TYPE_POWER)
1363 return (AE_OK);
1364
1365 /*
1366 * Call the user function with the current device. If it is unchanged
1367 * afterwards, return. Otherwise, we update the handle to the new dev.
1368 */
1369 old_dev = acpi_get_device(h);
1370 dev = old_dev;
1371 status = ctx->user_fn(h, &dev, level, ctx->arg);
1372 if (ACPI_FAILURE(status) || old_dev == dev)
1373 return (status);
1374
1375 /* Remove the old child and its connection to the handle. */
1376 if (old_dev != NULL) {
1377 device_delete_child(device_get_parent(old_dev), old_dev);
1378 AcpiDetachData(h, acpi_fake_objhandler);
1379 }
1380
1381 /* Recreate the handle association if the user created a device. */
1382 if (dev != NULL)
1383 AcpiAttachData(h, acpi_fake_objhandler, dev);
1384
1385 return (AE_OK);
1386 }
1387
1388 static ACPI_STATUS
1389 acpi_device_scan_children(device_t bus, device_t dev, int max_depth,
1390 acpi_scan_cb_t user_fn, void *arg)
1391 {
1392 ACPI_HANDLE h;
1393 struct acpi_device_scan_ctx ctx;
1394
1395 if (acpi_disabled("children"))
1396 return (AE_OK);
1397
1398 if (dev == NULL)
1399 h = ACPI_ROOT_OBJECT;
1400 else if ((h = acpi_get_handle(dev)) == NULL)
1401 return (AE_BAD_PARAMETER);
1402 ctx.user_fn = user_fn;
1403 ctx.arg = arg;
1404 ctx.parent = h;
1405 return (AcpiWalkNamespace(ACPI_TYPE_ANY, h, max_depth,
1406 acpi_device_scan_cb, &ctx, NULL));
1407 }
1408
1409 /*
1410 * Even though ACPI devices are not PCI, we use the PCI approach for setting
1411 * device power states since it's close enough to ACPI.
1412 */
1413 static int
1414 acpi_set_powerstate_method(device_t bus, device_t child, int state)
1415 {
1416 ACPI_HANDLE h;
1417 ACPI_STATUS status;
1418 int error;
1419
1420 error = 0;
1421 h = acpi_get_handle(child);
1422 if (state < ACPI_STATE_D0 || state > ACPI_STATE_D3)
1423 return (EINVAL);
1424 if (h == NULL)
1425 return (0);
1426
1427 /* Ignore errors if the power methods aren't present. */
1428 status = acpi_pwr_switch_consumer(h, state);
1429 if (ACPI_FAILURE(status) && status != AE_NOT_FOUND
1430 && status != AE_BAD_PARAMETER)
1431 device_printf(bus, "failed to set ACPI power state D%d on %s: %s\n",
1432 state, acpi_name(h), AcpiFormatException(status));
1433
1434 return (error);
1435 }
1436
1437 static int
1438 acpi_isa_pnp_probe(device_t bus, device_t child, struct isa_pnp_id *ids)
1439 {
1440 int result, cid_count, i;
1441 uint32_t lid, cids[8];
1442
1443 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1444
1445 /*
1446 * ISA-style drivers attached to ACPI may persist and
1447 * probe manually if we return ENOENT. We never want
1448 * that to happen, so don't ever return it.
1449 */
1450 result = ENXIO;
1451
1452 /* Scan the supplied IDs for a match */
1453 lid = acpi_isa_get_logicalid(child);
1454 cid_count = acpi_isa_get_compatid(child, cids, 8);
1455 while (ids && ids->ip_id) {
1456 if (lid == ids->ip_id) {
1457 result = 0;
1458 goto out;
1459 }
1460 for (i = 0; i < cid_count; i++) {
1461 if (cids[i] == ids->ip_id) {
1462 result = 0;
1463 goto out;
1464 }
1465 }
1466 ids++;
1467 }
1468
1469 out:
1470 if (result == 0 && ids->ip_desc)
1471 device_set_desc(child, ids->ip_desc);
1472
1473 return_VALUE (result);
1474 }
1475
1476 /*
1477 * Scan all of the ACPI namespace and attach child devices.
1478 *
1479 * We should only expect to find devices in the \_PR, \_TZ, \_SI, and
1480 * \_SB scopes, and \_PR and \_TZ became obsolete in the ACPI 2.0 spec.
1481 * However, in violation of the spec, some systems place their PCI link
1482 * devices in \, so we have to walk the whole namespace. We check the
1483 * type of namespace nodes, so this should be ok.
1484 */
1485 static void
1486 acpi_probe_children(device_t bus)
1487 {
1488
1489 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1490
1491 /*
1492 * Scan the namespace and insert placeholders for all the devices that
1493 * we find. We also probe/attach any early devices.
1494 *
1495 * Note that we use AcpiWalkNamespace rather than AcpiGetDevices because
1496 * we want to create nodes for all devices, not just those that are
1497 * currently present. (This assumes that we don't want to create/remove
1498 * devices as they appear, which might be smarter.)
1499 */
1500 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "namespace scan\n"));
1501 AcpiWalkNamespace(ACPI_TYPE_ANY, ACPI_ROOT_OBJECT, 100, acpi_probe_child,
1502 bus, NULL);
1503
1504 /* Pre-allocate resources for our rman from any sysresource devices. */
1505 acpi_sysres_alloc(bus);
1506
1507 /* Create any static children by calling device identify methods. */
1508 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "device identify routines\n"));
1509 bus_generic_probe(bus);
1510
1511 /* Probe/attach all children, created staticly and from the namespace. */
1512 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "first bus_generic_attach\n"));
1513 bus_generic_attach(bus);
1514
1515 /*
1516 * Some of these children may have attached others as part of their attach
1517 * process (eg. the root PCI bus driver), so rescan.
1518 */
1519 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "second bus_generic_attach\n"));
1520 bus_generic_attach(bus);
1521
1522 /* Attach wake sysctls. */
1523 acpi_wake_sysctl_walk(bus);
1524
1525 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "done attaching children\n"));
1526 return_VOID;
1527 }
1528
1529 /*
1530 * Determine the probe order for a given device.
1531 */
1532 static void
1533 acpi_probe_order(ACPI_HANDLE handle, int *order)
1534 {
1535 ACPI_OBJECT_TYPE type;
1536
1537 /*
1538 * 1. I/O port and memory system resource holders
1539 * 2. Embedded controllers (to handle early accesses)
1540 * 3. PCI Link Devices
1541 * 100000. CPUs
1542 */
1543 AcpiGetType(handle, &type);
1544 if (acpi_MatchHid(handle, "PNP0C01") || acpi_MatchHid(handle, "PNP0C02"))
1545 *order = 1;
1546 else if (acpi_MatchHid(handle, "PNP0C09"))
1547 *order = 2;
1548 else if (acpi_MatchHid(handle, "PNP0C0F"))
1549 *order = 3;
1550 else if (type == ACPI_TYPE_PROCESSOR)
1551 *order = 100000;
1552 }
1553
1554 /*
1555 * Evaluate a child device and determine whether we might attach a device to
1556 * it.
1557 */
1558 static ACPI_STATUS
1559 acpi_probe_child(ACPI_HANDLE handle, UINT32 level, void *context, void **status)
1560 {
1561 ACPI_OBJECT_TYPE type;
1562 ACPI_HANDLE h;
1563 device_t bus, child;
1564 int order;
1565 char *handle_str, **search;
1566 static char *scopes[] = {"\\_PR_", "\\_TZ_", "\\_SI_", "\\_SB_", NULL};
1567
1568 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1569
1570 /* Skip this device if we think we'll have trouble with it. */
1571 if (acpi_avoid(handle))
1572 return_ACPI_STATUS (AE_OK);
1573
1574 bus = (device_t)context;
1575 if (ACPI_SUCCESS(AcpiGetType(handle, &type))) {
1576 switch (type) {
1577 case ACPI_TYPE_DEVICE:
1578 case ACPI_TYPE_PROCESSOR:
1579 case ACPI_TYPE_THERMAL:
1580 case ACPI_TYPE_POWER:
1581 if (acpi_disabled("children"))
1582 break;
1583
1584 /*
1585 * Since we scan from \, be sure to skip system scope objects.
1586 * At least \_SB and \_TZ are detected as devices (ACPI-CA bug?)
1587 */
1588 handle_str = acpi_name(handle);
1589 for (search = scopes; *search != NULL; search++) {
1590 if (strcmp(handle_str, *search) == 0)
1591 break;
1592 }
1593 if (*search != NULL)
1594 break;
1595
1596 /*
1597 * Create a placeholder device for this node. Sort the
1598 * placeholder so that the probe/attach passes will run
1599 * breadth-first. Orders less than ACPI_DEV_BASE_ORDER
1600 * are reserved for special objects (i.e., system
1601 * resources). CPU devices have a very high order to
1602 * ensure they are probed after other devices.
1603 */
1604 ACPI_DEBUG_PRINT((ACPI_DB_OBJECTS, "scanning '%s'\n", handle_str));
1605 order = level * 10 + 100;
1606 acpi_probe_order(handle, &order);
1607 child = BUS_ADD_CHILD(bus, order, NULL, -1);
1608 if (child == NULL)
1609 break;
1610
1611 /* Associate the handle with the device_t and vice versa. */
1612 acpi_set_handle(child, handle);
1613 AcpiAttachData(handle, acpi_fake_objhandler, child);
1614
1615 /*
1616 * Check that the device is present. If it's not present,
1617 * leave it disabled (so that we have a device_t attached to
1618 * the handle, but we don't probe it).
1619 *
1620 * XXX PCI link devices sometimes report "present" but not
1621 * "functional" (i.e. if disabled). Go ahead and probe them
1622 * anyway since we may enable them later.
1623 */
1624 if (type == ACPI_TYPE_DEVICE && !acpi_DeviceIsPresent(child)) {
1625 /* Never disable PCI link devices. */
1626 if (acpi_MatchHid(handle, "PNP0C0F"))
1627 break;
1628 /*
1629 * Docking stations should remain enabled since the system
1630 * may be undocked at boot.
1631 */
1632 if (ACPI_SUCCESS(AcpiGetHandle(handle, "_DCK", &h)))
1633 break;
1634
1635 device_disable(child);
1636 break;
1637 }
1638
1639 /*
1640 * Get the device's resource settings and attach them.
1641 * Note that if the device has _PRS but no _CRS, we need
1642 * to decide when it's appropriate to try to configure the
1643 * device. Ignore the return value here; it's OK for the
1644 * device not to have any resources.
1645 */
1646 acpi_parse_resources(child, handle, &acpi_res_parse_set, NULL);
1647 break;
1648 }
1649 }
1650
1651 return_ACPI_STATUS (AE_OK);
1652 }
1653
1654 /*
1655 * AcpiAttachData() requires an object handler but never uses it. This is a
1656 * placeholder object handler so we can store a device_t in an ACPI_HANDLE.
1657 */
1658 void
1659 acpi_fake_objhandler(ACPI_HANDLE h, UINT32 fn, void *data)
1660 {
1661 }
1662
1663 static void
1664 acpi_shutdown_final(void *arg, int howto)
1665 {
1666 struct acpi_softc *sc;
1667 ACPI_STATUS status;
1668
1669 /*
1670 * XXX Shutdown code should only run on the BSP (cpuid 0).
1671 * Some chipsets do not power off the system correctly if called from
1672 * an AP.
1673 */
1674 sc = arg;
1675 if ((howto & RB_POWEROFF) != 0) {
1676 status = AcpiEnterSleepStatePrep(ACPI_STATE_S5);
1677 if (ACPI_FAILURE(status)) {
1678 printf("AcpiEnterSleepStatePrep failed - %s\n",
1679 AcpiFormatException(status));
1680 return;
1681 }
1682 printf("Powering system off using ACPI\n");
1683 ACPI_DISABLE_IRQS();
1684 status = AcpiEnterSleepState(ACPI_STATE_S5);
1685 if (ACPI_FAILURE(status)) {
1686 printf("ACPI power-off failed - %s\n", AcpiFormatException(status));
1687 } else {
1688 DELAY(1000000);
1689 printf("ACPI power-off failed - timeout\n");
1690 }
1691 } else if ((howto & RB_HALT) == 0 &&
1692 (AcpiGbl_FADT.Flags & ACPI_FADT_RESET_REGISTER) &&
1693 sc->acpi_handle_reboot) {
1694 /* Reboot using the reset register. */
1695 status = AcpiHwLowLevelWrite(
1696 AcpiGbl_FADT.ResetRegister.BitWidth,
1697 AcpiGbl_FADT.ResetValue, &AcpiGbl_FADT.ResetRegister);
1698 if (ACPI_FAILURE(status)) {
1699 printf("ACPI reset failed - %s\n", AcpiFormatException(status));
1700 } else {
1701 DELAY(1000000);
1702 printf("ACPI reset failed - timeout\n");
1703 }
1704 } else if (sc->acpi_do_disable && panicstr == NULL) {
1705 /*
1706 * Only disable ACPI if the user requested. On some systems, writing
1707 * the disable value to SMI_CMD hangs the system.
1708 */
1709 printf("Shutting down ACPI\n");
1710 AcpiTerminate();
1711 }
1712 }
1713
1714 static void
1715 acpi_enable_fixed_events(struct acpi_softc *sc)
1716 {
1717 static int first_time = 1;
1718
1719 /* Enable and clear fixed events and install handlers. */
1720 if ((AcpiGbl_FADT.Flags & ACPI_FADT_POWER_BUTTON) == 0) {
1721 AcpiClearEvent(ACPI_EVENT_POWER_BUTTON);
1722 AcpiInstallFixedEventHandler(ACPI_EVENT_POWER_BUTTON,
1723 acpi_event_power_button_sleep, sc);
1724 if (first_time)
1725 device_printf(sc->acpi_dev, "Power Button (fixed)\n");
1726 }
1727 if ((AcpiGbl_FADT.Flags & ACPI_FADT_SLEEP_BUTTON) == 0) {
1728 AcpiClearEvent(ACPI_EVENT_SLEEP_BUTTON);
1729 AcpiInstallFixedEventHandler(ACPI_EVENT_SLEEP_BUTTON,
1730 acpi_event_sleep_button_sleep, sc);
1731 if (first_time)
1732 device_printf(sc->acpi_dev, "Sleep Button (fixed)\n");
1733 }
1734
1735 first_time = 0;
1736 }
1737
1738 /*
1739 * Returns true if the device is actually present and should
1740 * be attached to. This requires the present, enabled, UI-visible
1741 * and diagnostics-passed bits to be set.
1742 */
1743 BOOLEAN
1744 acpi_DeviceIsPresent(device_t dev)
1745 {
1746 ACPI_DEVICE_INFO *devinfo;
1747 ACPI_HANDLE h;
1748 ACPI_BUFFER buf;
1749 ACPI_STATUS error;
1750 int ret;
1751
1752 ret = FALSE;
1753 if ((h = acpi_get_handle(dev)) == NULL)
1754 return (FALSE);
1755 buf.Pointer = NULL;
1756 buf.Length = ACPI_ALLOCATE_BUFFER;
1757 error = AcpiGetObjectInfo(h, &buf);
1758 if (ACPI_FAILURE(error))
1759 return (FALSE);
1760 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1761
1762 /* If no _STA method, must be present */
1763 if ((devinfo->Valid & ACPI_VALID_STA) == 0)
1764 ret = TRUE;
1765
1766 /* Return true for 'present' and 'functioning' */
1767 if (ACPI_DEVICE_PRESENT(devinfo->CurrentStatus))
1768 ret = TRUE;
1769
1770 AcpiOsFree(buf.Pointer);
1771 return (ret);
1772 }
1773
1774 /*
1775 * Returns true if the battery is actually present and inserted.
1776 */
1777 BOOLEAN
1778 acpi_BatteryIsPresent(device_t dev)
1779 {
1780 ACPI_DEVICE_INFO *devinfo;
1781 ACPI_HANDLE h;
1782 ACPI_BUFFER buf;
1783 ACPI_STATUS error;
1784 int ret;
1785
1786 ret = FALSE;
1787 if ((h = acpi_get_handle(dev)) == NULL)
1788 return (FALSE);
1789 buf.Pointer = NULL;
1790 buf.Length = ACPI_ALLOCATE_BUFFER;
1791 error = AcpiGetObjectInfo(h, &buf);
1792 if (ACPI_FAILURE(error))
1793 return (FALSE);
1794 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1795
1796 /* If no _STA method, must be present */
1797 if ((devinfo->Valid & ACPI_VALID_STA) == 0)
1798 ret = TRUE;
1799
1800 /* Return true for 'present', 'battery present', and 'functioning' */
1801 if (ACPI_BATTERY_PRESENT(devinfo->CurrentStatus))
1802 ret = TRUE;
1803
1804 AcpiOsFree(buf.Pointer);
1805 return (ret);
1806 }
1807
1808 /*
1809 * Match a HID string against a handle
1810 */
1811 static BOOLEAN
1812 acpi_MatchHid(ACPI_HANDLE h, const char *hid)
1813 {
1814 ACPI_DEVICE_INFO *devinfo;
1815 ACPI_BUFFER buf;
1816 ACPI_STATUS error;
1817 int ret, i;
1818
1819 ret = FALSE;
1820 if (hid == NULL || h == NULL)
1821 return (ret);
1822 buf.Pointer = NULL;
1823 buf.Length = ACPI_ALLOCATE_BUFFER;
1824 error = AcpiGetObjectInfo(h, &buf);
1825 if (ACPI_FAILURE(error))
1826 return (ret);
1827 devinfo = (ACPI_DEVICE_INFO *)buf.Pointer;
1828
1829 if ((devinfo->Valid & ACPI_VALID_HID) != 0 &&
1830 strcmp(hid, devinfo->HardwareId.Value) == 0)
1831 ret = TRUE;
1832 else if ((devinfo->Valid & ACPI_VALID_CID) != 0) {
1833 for (i = 0; i < devinfo->CompatibilityId.Count; i++) {
1834 if (strcmp(hid, devinfo->CompatibilityId.Id[i].Value) == 0) {
1835 ret = TRUE;
1836 break;
1837 }
1838 }
1839 }
1840
1841 AcpiOsFree(buf.Pointer);
1842 return (ret);
1843 }
1844
1845 /*
1846 * Return the handle of a named object within our scope, ie. that of (parent)
1847 * or one if its parents.
1848 */
1849 ACPI_STATUS
1850 acpi_GetHandleInScope(ACPI_HANDLE parent, char *path, ACPI_HANDLE *result)
1851 {
1852 ACPI_HANDLE r;
1853 ACPI_STATUS status;
1854
1855 /* Walk back up the tree to the root */
1856 for (;;) {
1857 status = AcpiGetHandle(parent, path, &r);
1858 if (ACPI_SUCCESS(status)) {
1859 *result = r;
1860 return (AE_OK);
1861 }
1862 /* XXX Return error here? */
1863 if (status != AE_NOT_FOUND)
1864 return (AE_OK);
1865 if (ACPI_FAILURE(AcpiGetParent(parent, &r)))
1866 return (AE_NOT_FOUND);
1867 parent = r;
1868 }
1869 }
1870
1871 /* Find the difference between two PM tick counts. */
1872 uint32_t
1873 acpi_TimerDelta(uint32_t end, uint32_t start)
1874 {
1875 uint32_t delta;
1876
1877 if (end >= start)
1878 delta = end - start;
1879 else if (AcpiGbl_FADT.Flags & ACPI_FADT_32BIT_TIMER)
1880 delta = ((0xFFFFFFFF - start) + end + 1);
1881 else
1882 delta = ((0x00FFFFFF - start) + end + 1) & 0x00FFFFFF;
1883 return (delta);
1884 }
1885
1886 /*
1887 * Allocate a buffer with a preset data size.
1888 */
1889 ACPI_BUFFER *
1890 acpi_AllocBuffer(int size)
1891 {
1892 ACPI_BUFFER *buf;
1893
1894 if ((buf = malloc(size + sizeof(*buf), M_ACPIDEV, M_NOWAIT)) == NULL)
1895 return (NULL);
1896 buf->Length = size;
1897 buf->Pointer = (void *)(buf + 1);
1898 return (buf);
1899 }
1900
1901 ACPI_STATUS
1902 acpi_SetInteger(ACPI_HANDLE handle, char *path, UINT32 number)
1903 {
1904 ACPI_OBJECT arg1;
1905 ACPI_OBJECT_LIST args;
1906
1907 arg1.Type = ACPI_TYPE_INTEGER;
1908 arg1.Integer.Value = number;
1909 args.Count = 1;
1910 args.Pointer = &arg1;
1911
1912 return (AcpiEvaluateObject(handle, path, &args, NULL));
1913 }
1914
1915 /*
1916 * Evaluate a path that should return an integer.
1917 */
1918 ACPI_STATUS
1919 acpi_GetInteger(ACPI_HANDLE handle, char *path, UINT32 *number)
1920 {
1921 ACPI_STATUS status;
1922 ACPI_BUFFER buf;
1923 ACPI_OBJECT param;
1924
1925 if (handle == NULL)
1926 handle = ACPI_ROOT_OBJECT;
1927
1928 /*
1929 * Assume that what we've been pointed at is an Integer object, or
1930 * a method that will return an Integer.
1931 */
1932 buf.Pointer = ¶m;
1933 buf.Length = sizeof(param);
1934 status = AcpiEvaluateObject(handle, path, NULL, &buf);
1935 if (ACPI_SUCCESS(status)) {
1936 if (param.Type == ACPI_TYPE_INTEGER)
1937 *number = param.Integer.Value;
1938 else
1939 status = AE_TYPE;
1940 }
1941
1942 /*
1943 * In some applications, a method that's expected to return an Integer
1944 * may instead return a Buffer (probably to simplify some internal
1945 * arithmetic). We'll try to fetch whatever it is, and if it's a Buffer,
1946 * convert it into an Integer as best we can.
1947 *
1948 * This is a hack.
1949 */
1950 if (status == AE_BUFFER_OVERFLOW) {
1951 if ((buf.Pointer = AcpiOsAllocate(buf.Length)) == NULL) {
1952 status = AE_NO_MEMORY;
1953 } else {
1954 status = AcpiEvaluateObject(handle, path, NULL, &buf);
1955 if (ACPI_SUCCESS(status))
1956 status = acpi_ConvertBufferToInteger(&buf, number);
1957 AcpiOsFree(buf.Pointer);
1958 }
1959 }
1960 return (status);
1961 }
1962
1963 ACPI_STATUS
1964 acpi_ConvertBufferToInteger(ACPI_BUFFER *bufp, UINT32 *number)
1965 {
1966 ACPI_OBJECT *p;
1967 UINT8 *val;
1968 int i;
1969
1970 p = (ACPI_OBJECT *)bufp->Pointer;
1971 if (p->Type == ACPI_TYPE_INTEGER) {
1972 *number = p->Integer.Value;
1973 return (AE_OK);
1974 }
1975 if (p->Type != ACPI_TYPE_BUFFER)
1976 return (AE_TYPE);
1977 if (p->Buffer.Length > sizeof(int))
1978 return (AE_BAD_DATA);
1979
1980 *number = 0;
1981 val = p->Buffer.Pointer;
1982 for (i = 0; i < p->Buffer.Length; i++)
1983 *number += val[i] << (i * 8);
1984 return (AE_OK);
1985 }
1986
1987 /*
1988 * Iterate over the elements of an a package object, calling the supplied
1989 * function for each element.
1990 *
1991 * XXX possible enhancement might be to abort traversal on error.
1992 */
1993 ACPI_STATUS
1994 acpi_ForeachPackageObject(ACPI_OBJECT *pkg,
1995 void (*func)(ACPI_OBJECT *comp, void *arg), void *arg)
1996 {
1997 ACPI_OBJECT *comp;
1998 int i;
1999
2000 if (pkg == NULL || pkg->Type != ACPI_TYPE_PACKAGE)
2001 return (AE_BAD_PARAMETER);
2002
2003 /* Iterate over components */
2004 i = 0;
2005 comp = pkg->Package.Elements;
2006 for (; i < pkg->Package.Count; i++, comp++)
2007 func(comp, arg);
2008
2009 return (AE_OK);
2010 }
2011
2012 /*
2013 * Find the (index)th resource object in a set.
2014 */
2015 ACPI_STATUS
2016 acpi_FindIndexedResource(ACPI_BUFFER *buf, int index, ACPI_RESOURCE **resp)
2017 {
2018 ACPI_RESOURCE *rp;
2019 int i;
2020
2021 rp = (ACPI_RESOURCE *)buf->Pointer;
2022 i = index;
2023 while (i-- > 0) {
2024 /* Range check */
2025 if (rp > (ACPI_RESOURCE *)((u_int8_t *)buf->Pointer + buf->Length))
2026 return (AE_BAD_PARAMETER);
2027
2028 /* Check for terminator */
2029 if (rp->Type == ACPI_RESOURCE_TYPE_END_TAG || rp->Length == 0)
2030 return (AE_NOT_FOUND);
2031 rp = ACPI_NEXT_RESOURCE(rp);
2032 }
2033 if (resp != NULL)
2034 *resp = rp;
2035
2036 return (AE_OK);
2037 }
2038
2039 /*
2040 * Append an ACPI_RESOURCE to an ACPI_BUFFER.
2041 *
2042 * Given a pointer to an ACPI_RESOURCE structure, expand the ACPI_BUFFER
2043 * provided to contain it. If the ACPI_BUFFER is empty, allocate a sensible
2044 * backing block. If the ACPI_RESOURCE is NULL, return an empty set of
2045 * resources.
2046 */
2047 #define ACPI_INITIAL_RESOURCE_BUFFER_SIZE 512
2048
2049 ACPI_STATUS
2050 acpi_AppendBufferResource(ACPI_BUFFER *buf, ACPI_RESOURCE *res)
2051 {
2052 ACPI_RESOURCE *rp;
2053 void *newp;
2054
2055 /* Initialise the buffer if necessary. */
2056 if (buf->Pointer == NULL) {
2057 buf->Length = ACPI_INITIAL_RESOURCE_BUFFER_SIZE;
2058 if ((buf->Pointer = AcpiOsAllocate(buf->Length)) == NULL)
2059 return (AE_NO_MEMORY);
2060 rp = (ACPI_RESOURCE *)buf->Pointer;
2061 rp->Type = ACPI_RESOURCE_TYPE_END_TAG;
2062 rp->Length = 0;
2063 }
2064 if (res == NULL)
2065 return (AE_OK);
2066
2067 /*
2068 * Scan the current buffer looking for the terminator.
2069 * This will either find the terminator or hit the end
2070 * of the buffer and return an error.
2071 */
2072 rp = (ACPI_RESOURCE *)buf->Pointer;
2073 for (;;) {
2074 /* Range check, don't go outside the buffer */
2075 if (rp >= (ACPI_RESOURCE *)((u_int8_t *)buf->Pointer + buf->Length))
2076 return (AE_BAD_PARAMETER);
2077 if (rp->Type == ACPI_RESOURCE_TYPE_END_TAG || rp->Length == 0)
2078 break;
2079 rp = ACPI_NEXT_RESOURCE(rp);
2080 }
2081
2082 /*
2083 * Check the size of the buffer and expand if required.
2084 *
2085 * Required size is:
2086 * size of existing resources before terminator +
2087 * size of new resource and header +
2088 * size of terminator.
2089 *
2090 * Note that this loop should really only run once, unless
2091 * for some reason we are stuffing a *really* huge resource.
2092 */
2093 while ((((u_int8_t *)rp - (u_int8_t *)buf->Pointer) +
2094 res->Length + ACPI_RS_SIZE_NO_DATA +
2095 ACPI_RS_SIZE_MIN) >= buf->Length) {
2096 if ((newp = AcpiOsAllocate(buf->Length * 2)) == NULL)
2097 return (AE_NO_MEMORY);
2098 bcopy(buf->Pointer, newp, buf->Length);
2099 rp = (ACPI_RESOURCE *)((u_int8_t *)newp +
2100 ((u_int8_t *)rp - (u_int8_t *)buf->Pointer));
2101 AcpiOsFree(buf->Pointer);
2102 buf->Pointer = newp;
2103 buf->Length += buf->Length;
2104 }
2105
2106 /* Insert the new resource. */
2107 bcopy(res, rp, res->Length + ACPI_RS_SIZE_NO_DATA);
2108
2109 /* And add the terminator. */
2110 rp = ACPI_NEXT_RESOURCE(rp);
2111 rp->Type = ACPI_RESOURCE_TYPE_END_TAG;
2112 rp->Length = 0;
2113
2114 return (AE_OK);
2115 }
2116
2117 /*
2118 * Set interrupt model.
2119 */
2120 ACPI_STATUS
2121 acpi_SetIntrModel(int model)
2122 {
2123
2124 return (acpi_SetInteger(ACPI_ROOT_OBJECT, "_PIC", model));
2125 }
2126
2127 /*
2128 * DEPRECATED. This interface has serious deficiencies and will be
2129 * removed.
2130 *
2131 * Immediately enter the sleep state. In the old model, acpiconf(8) ran
2132 * rc.suspend and rc.resume so we don't have to notify devd(8) to do this.
2133 */
2134 ACPI_STATUS
2135 acpi_SetSleepState(struct acpi_softc *sc, int state)
2136 {
2137 static int once;
2138
2139 if (!once) {
2140 printf(
2141 "warning: acpi_SetSleepState() deprecated, need to update your software\n");
2142 once = 1;
2143 }
2144 return (acpi_EnterSleepState(sc, state));
2145 }
2146
2147 static void
2148 acpi_sleep_force(void *arg)
2149 {
2150 struct acpi_softc *sc;
2151
2152 printf("acpi: suspend request timed out, forcing sleep now\n");
2153 sc = arg;
2154 if (ACPI_FAILURE(acpi_EnterSleepState(sc, sc->acpi_next_sstate)))
2155 printf("acpi: force sleep state S%d failed\n", sc->acpi_next_sstate);
2156 }
2157
2158 /*
2159 * Request that the system enter the given suspend state. All /dev/apm
2160 * devices and devd(8) will be notified. Userland then has a chance to
2161 * save state and acknowledge the request. The system sleeps once all
2162 * acks are in.
2163 */
2164 int
2165 acpi_ReqSleepState(struct acpi_softc *sc, int state)
2166 {
2167 struct apm_clone_data *clone;
2168
2169 if (state < ACPI_STATE_S1 || state > ACPI_STATE_S5)
2170 return (EINVAL);
2171
2172 /* S5 (soft-off) should be entered directly with no waiting. */
2173 if (state == ACPI_STATE_S5) {
2174 if (ACPI_SUCCESS(acpi_EnterSleepState(sc, state)))
2175 return (0);
2176 else
2177 return (ENXIO);
2178 }
2179
2180 #if !defined(__i386__)
2181 /* This platform does not support acpi suspend/resume. */
2182 return (EOPNOTSUPP);
2183 #endif
2184
2185 /* If a suspend request is already in progress, just return. */
2186 ACPI_LOCK(acpi);
2187 if (sc->acpi_next_sstate != 0) {
2188 ACPI_UNLOCK(acpi);
2189 return (0);
2190 }
2191
2192 /* Record the pending state and notify all apm devices. */
2193 sc->acpi_next_sstate = state;
2194 STAILQ_FOREACH(clone, &sc->apm_cdevs, entries) {
2195 clone->notify_status = APM_EV_NONE;
2196 if ((clone->flags & ACPI_EVF_DEVD) == 0) {
2197 selwakeuppri(&clone->sel_read, PZERO);
2198 KNOTE_UNLOCKED(&clone->sel_read.si_note, 0);
2199 }
2200 }
2201
2202 /* If devd(8) is not running, immediately enter the sleep state. */
2203 if (devctl_process_running() == FALSE) {
2204 ACPI_UNLOCK(acpi);
2205 if (ACPI_SUCCESS(acpi_EnterSleepState(sc, sc->acpi_next_sstate))) {
2206 return (0);
2207 } else {
2208 return (ENXIO);
2209 }
2210 }
2211
2212 /* Now notify devd(8) also. */
2213 acpi_UserNotify("Suspend", ACPI_ROOT_OBJECT, state);
2214
2215 /*
2216 * Set a timeout to fire if userland doesn't ack the suspend request
2217 * in time. This way we still eventually go to sleep if we were
2218 * overheating or running low on battery, even if userland is hung.
2219 * We cancel this timeout once all userland acks are in or the
2220 * suspend request is aborted.
2221 */
2222 callout_reset(&sc->susp_force_to, 10 * hz, acpi_sleep_force, sc);
2223 ACPI_UNLOCK(acpi);
2224 return (0);
2225 }
2226
2227 /*
2228 * Acknowledge (or reject) a pending sleep state. The caller has
2229 * prepared for suspend and is now ready for it to proceed. If the
2230 * error argument is non-zero, it indicates suspend should be cancelled
2231 * and gives an errno value describing why. Once all votes are in,
2232 * we suspend the system.
2233 */
2234 int
2235 acpi_AckSleepState(struct apm_clone_data *clone, int error)
2236 {
2237 struct acpi_softc *sc;
2238 int ret, sleeping;
2239
2240 #if !defined(__i386__)
2241 /* This platform does not support acpi suspend/resume. */
2242 return (EOPNOTSUPP);
2243 #endif
2244
2245 /* If no pending sleep state, return an error. */
2246 ACPI_LOCK(acpi);
2247 sc = clone->acpi_sc;
2248 if (sc->acpi_next_sstate == 0) {
2249 ACPI_UNLOCK(acpi);
2250 return (ENXIO);
2251 }
2252
2253 /* Caller wants to abort suspend process. */
2254 if (error) {
2255 sc->acpi_next_sstate = 0;
2256 callout_stop(&sc->susp_force_to);
2257 printf("acpi: listener on %s cancelled the pending suspend\n",
2258 devtoname(clone->cdev));
2259 ACPI_UNLOCK(acpi);
2260 return (0);
2261 }
2262
2263 /*
2264 * Mark this device as acking the suspend request. Then, walk through
2265 * all devices, seeing if they agree yet. We only count devices that
2266 * are writable since read-only devices couldn't ack the request.
2267 */
2268 clone->notify_status = APM_EV_ACKED;
2269 sleeping = TRUE;
2270 STAILQ_FOREACH(clone, &sc->apm_cdevs, entries) {
2271 if ((clone->flags & ACPI_EVF_WRITE) != 0 &&
2272 clone->notify_status != APM_EV_ACKED) {
2273 sleeping = FALSE;
2274 break;
2275 }
2276 }
2277
2278 /* If all devices have voted "yes", we will suspend now. */
2279 if (sleeping)
2280 callout_stop(&sc->susp_force_to);
2281 ACPI_UNLOCK(acpi);
2282 ret = 0;
2283 if (sleeping) {
2284 if (ACPI_FAILURE(acpi_EnterSleepState(sc, sc->acpi_next_sstate)))
2285 ret = ENODEV;
2286 }
2287
2288 return (ret);
2289 }
2290
2291 static void
2292 acpi_sleep_enable(void *arg)
2293 {
2294
2295 ((struct acpi_softc *)arg)->acpi_sleep_disabled = 0;
2296 }
2297
2298 enum acpi_sleep_state {
2299 ACPI_SS_NONE,
2300 ACPI_SS_GPE_SET,
2301 ACPI_SS_DEV_SUSPEND,
2302 ACPI_SS_SLP_PREP,
2303 ACPI_SS_SLEPT,
2304 };
2305
2306 /*
2307 * Enter the desired system sleep state.
2308 *
2309 * Currently we support S1-S5 but S4 is only S4BIOS
2310 */
2311 static ACPI_STATUS
2312 acpi_EnterSleepState(struct acpi_softc *sc, int state)
2313 {
2314 ACPI_STATUS status;
2315 UINT8 TypeA;
2316 UINT8 TypeB;
2317 enum acpi_sleep_state slp_state;
2318
2319 ACPI_FUNCTION_TRACE_U32((char *)(uintptr_t)__func__, state);
2320
2321 /* Re-entry once we're suspending is not allowed. */
2322 status = AE_OK;
2323 ACPI_LOCK(acpi);
2324 if (sc->acpi_sleep_disabled) {
2325 ACPI_UNLOCK(acpi);
2326 printf("acpi: suspend request ignored (not ready yet)\n");
2327 return (AE_ERROR);
2328 }
2329 sc->acpi_sleep_disabled = 1;
2330 ACPI_UNLOCK(acpi);
2331
2332 /*
2333 * Be sure to hold Giant across DEVICE_SUSPEND/RESUME since non-MPSAFE
2334 * drivers need this.
2335 */
2336 mtx_lock(&Giant);
2337 slp_state = ACPI_SS_NONE;
2338 switch (state) {
2339 case ACPI_STATE_S1:
2340 case ACPI_STATE_S2:
2341 case ACPI_STATE_S3:
2342 case ACPI_STATE_S4:
2343 status = AcpiGetSleepTypeData(state, &TypeA, &TypeB);
2344 if (status == AE_NOT_FOUND) {
2345 device_printf(sc->acpi_dev,
2346 "Sleep state S%d not supported by BIOS\n", state);
2347 break;
2348 } else if (ACPI_FAILURE(status)) {
2349 device_printf(sc->acpi_dev, "AcpiGetSleepTypeData failed - %s\n",
2350 AcpiFormatException(status));
2351 break;
2352 }
2353
2354 sc->acpi_sstate = state;
2355
2356 /* Enable any GPEs as appropriate and requested by the user. */
2357 acpi_wake_prep_walk(state);
2358 slp_state = ACPI_SS_GPE_SET;
2359
2360 /*
2361 * Inform all devices that we are going to sleep. If at least one
2362 * device fails, DEVICE_SUSPEND() automatically resumes the tree.
2363 *
2364 * XXX Note that a better two-pass approach with a 'veto' pass
2365 * followed by a "real thing" pass would be better, but the current
2366 * bus interface does not provide for this.
2367 */
2368 if (DEVICE_SUSPEND(root_bus) != 0) {
2369 device_printf(sc->acpi_dev, "device_suspend failed\n");
2370 break;
2371 }
2372 slp_state = ACPI_SS_DEV_SUSPEND;
2373
2374 /* If testing device suspend only, back out of everything here. */
2375 if (acpi_susp_bounce)
2376 break;
2377
2378 status = AcpiEnterSleepStatePrep(state);
2379 if (ACPI_FAILURE(status)) {
2380 device_printf(sc->acpi_dev, "AcpiEnterSleepStatePrep failed - %s\n",
2381 AcpiFormatException(status));
2382 break;
2383 }
2384 slp_state = ACPI_SS_SLP_PREP;
2385
2386 if (sc->acpi_sleep_delay > 0)
2387 DELAY(sc->acpi_sleep_delay * 1000000);
2388
2389 if (state != ACPI_STATE_S1) {
2390 acpi_sleep_machdep(sc, state);
2391
2392 /* Re-enable ACPI hardware on wakeup from sleep state 4. */
2393 if (state == ACPI_STATE_S4)
2394 AcpiEnable();
2395 } else {
2396 ACPI_DISABLE_IRQS();
2397 status = AcpiEnterSleepState(state);
2398 if (ACPI_FAILURE(status)) {
2399 device_printf(sc->acpi_dev, "AcpiEnterSleepState failed - %s\n",
2400 AcpiFormatException(status));
2401 break;
2402 }
2403 }
2404 slp_state = ACPI_SS_SLEPT;
2405 break;
2406 case ACPI_STATE_S5:
2407 /*
2408 * Shut down cleanly and power off. This will call us back through the
2409 * shutdown handlers.
2410 */
2411 shutdown_nice(RB_POWEROFF);
2412 break;
2413 case ACPI_STATE_S0:
2414 default:
2415 status = AE_BAD_PARAMETER;
2416 break;
2417 }
2418
2419 /*
2420 * Back out state according to how far along we got in the suspend
2421 * process. This handles both the error and success cases.
2422 */
2423 sc->acpi_next_sstate = 0;
2424 if (slp_state >= ACPI_SS_GPE_SET) {
2425 acpi_wake_prep_walk(state);
2426 sc->acpi_sstate = ACPI_STATE_S0;
2427 }
2428 if (slp_state >= ACPI_SS_SLP_PREP)
2429 AcpiLeaveSleepState(state);
2430 if (slp_state >= ACPI_SS_DEV_SUSPEND)
2431 DEVICE_RESUME(root_bus);
2432 if (slp_state >= ACPI_SS_SLEPT)
2433 acpi_enable_fixed_events(sc);
2434
2435 /* Allow another sleep request after a while. */
2436 if (state != ACPI_STATE_S5)
2437 timeout(acpi_sleep_enable, sc, hz * ACPI_MINIMUM_AWAKETIME);
2438
2439 /* Run /etc/rc.resume after we are back. */
2440 acpi_UserNotify("Resume", ACPI_ROOT_OBJECT, state);
2441
2442 mtx_unlock(&Giant);
2443 return_ACPI_STATUS (status);
2444 }
2445
2446 /* Initialize a device's wake GPE. */
2447 int
2448 acpi_wake_init(device_t dev, int type)
2449 {
2450 struct acpi_prw_data prw;
2451
2452 /* Evaluate _PRW to find the GPE. */
2453 if (acpi_parse_prw(acpi_get_handle(dev), &prw) != 0)
2454 return (ENXIO);
2455
2456 /* Set the requested type for the GPE (runtime, wake, or both). */
2457 if (ACPI_FAILURE(AcpiSetGpeType(prw.gpe_handle, prw.gpe_bit, type))) {
2458 device_printf(dev, "set GPE type failed\n");
2459 return (ENXIO);
2460 }
2461
2462 return (0);
2463 }
2464
2465 /* Enable or disable the device's wake GPE. */
2466 int
2467 acpi_wake_set_enable(device_t dev, int enable)
2468 {
2469 struct acpi_prw_data prw;
2470 ACPI_STATUS status;
2471 int flags;
2472
2473 /* Make sure the device supports waking the system and get the GPE. */
2474 if (acpi_parse_prw(acpi_get_handle(dev), &prw) != 0)
2475 return (ENXIO);
2476
2477 flags = acpi_get_flags(dev);
2478 if (enable) {
2479 status = AcpiEnableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2480 if (ACPI_FAILURE(status)) {
2481 device_printf(dev, "enable wake failed\n");
2482 return (ENXIO);
2483 }
2484 acpi_set_flags(dev, flags | ACPI_FLAG_WAKE_ENABLED);
2485 } else {
2486 status = AcpiDisableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2487 if (ACPI_FAILURE(status)) {
2488 device_printf(dev, "disable wake failed\n");
2489 return (ENXIO);
2490 }
2491 acpi_set_flags(dev, flags & ~ACPI_FLAG_WAKE_ENABLED);
2492 }
2493
2494 return (0);
2495 }
2496
2497 static int
2498 acpi_wake_sleep_prep(ACPI_HANDLE handle, int sstate)
2499 {
2500 struct acpi_prw_data prw;
2501 device_t dev;
2502
2503 /* Check that this is a wake-capable device and get its GPE. */
2504 if (acpi_parse_prw(handle, &prw) != 0)
2505 return (ENXIO);
2506 dev = acpi_get_device(handle);
2507
2508 /*
2509 * The destination sleep state must be less than (i.e., higher power)
2510 * or equal to the value specified by _PRW. If this GPE cannot be
2511 * enabled for the next sleep state, then disable it. If it can and
2512 * the user requested it be enabled, turn on any required power resources
2513 * and set _PSW.
2514 */
2515 if (sstate > prw.lowest_wake) {
2516 AcpiDisableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2517 if (bootverbose)
2518 device_printf(dev, "wake_prep disabled wake for %s (S%d)\n",
2519 acpi_name(handle), sstate);
2520 } else if (dev && (acpi_get_flags(dev) & ACPI_FLAG_WAKE_ENABLED) != 0) {
2521 acpi_pwr_wake_enable(handle, 1);
2522 acpi_SetInteger(handle, "_PSW", 1);
2523 if (bootverbose)
2524 device_printf(dev, "wake_prep enabled for %s (S%d)\n",
2525 acpi_name(handle), sstate);
2526 }
2527
2528 return (0);
2529 }
2530
2531 static int
2532 acpi_wake_run_prep(ACPI_HANDLE handle, int sstate)
2533 {
2534 struct acpi_prw_data prw;
2535 device_t dev;
2536
2537 /*
2538 * Check that this is a wake-capable device and get its GPE. Return
2539 * now if the user didn't enable this device for wake.
2540 */
2541 if (acpi_parse_prw(handle, &prw) != 0)
2542 return (ENXIO);
2543 dev = acpi_get_device(handle);
2544 if (dev == NULL || (acpi_get_flags(dev) & ACPI_FLAG_WAKE_ENABLED) == 0)
2545 return (0);
2546
2547 /*
2548 * If this GPE couldn't be enabled for the previous sleep state, it was
2549 * disabled before going to sleep so re-enable it. If it was enabled,
2550 * clear _PSW and turn off any power resources it used.
2551 */
2552 if (sstate > prw.lowest_wake) {
2553 AcpiEnableGpe(prw.gpe_handle, prw.gpe_bit, ACPI_NOT_ISR);
2554 if (bootverbose)
2555 device_printf(dev, "run_prep re-enabled %s\n", acpi_name(handle));
2556 } else {
2557 acpi_SetInteger(handle, "_PSW", 0);
2558 acpi_pwr_wake_enable(handle, 0);
2559 if (bootverbose)
2560 device_printf(dev, "run_prep cleaned up for %s\n",
2561 acpi_name(handle));
2562 }
2563
2564 return (0);
2565 }
2566
2567 static ACPI_STATUS
2568 acpi_wake_prep(ACPI_HANDLE handle, UINT32 level, void *context, void **status)
2569 {
2570 int sstate;
2571
2572 /* If suspending, run the sleep prep function, otherwise wake. */
2573 sstate = *(int *)context;
2574 if (AcpiGbl_SystemAwakeAndRunning)
2575 acpi_wake_sleep_prep(handle, sstate);
2576 else
2577 acpi_wake_run_prep(handle, sstate);
2578 return (AE_OK);
2579 }
2580
2581 /* Walk the tree rooted at acpi0 to prep devices for suspend/resume. */
2582 static int
2583 acpi_wake_prep_walk(int sstate)
2584 {
2585 ACPI_HANDLE sb_handle;
2586
2587 if (ACPI_SUCCESS(AcpiGetHandle(ACPI_ROOT_OBJECT, "\\_SB_", &sb_handle)))
2588 AcpiWalkNamespace(ACPI_TYPE_DEVICE, sb_handle, 100,
2589 acpi_wake_prep, &sstate, NULL);
2590 return (0);
2591 }
2592
2593 /* Walk the tree rooted at acpi0 to attach per-device wake sysctls. */
2594 static int
2595 acpi_wake_sysctl_walk(device_t dev)
2596 {
2597 int error, i, numdevs;
2598 device_t *devlist;
2599 device_t child;
2600 ACPI_STATUS status;
2601
2602 error = device_get_children(dev, &devlist, &numdevs);
2603 if (error != 0 || numdevs == 0) {
2604 if (numdevs == 0)
2605 free(devlist, M_TEMP);
2606 return (error);
2607 }
2608 for (i = 0; i < numdevs; i++) {
2609 child = devlist[i];
2610 acpi_wake_sysctl_walk(child);
2611 if (!device_is_attached(child))
2612 continue;
2613 status = AcpiEvaluateObject(acpi_get_handle(child), "_PRW", NULL, NULL);
2614 if (ACPI_SUCCESS(status)) {
2615 SYSCTL_ADD_PROC(device_get_sysctl_ctx(child),
2616 SYSCTL_CHILDREN(device_get_sysctl_tree(child)), OID_AUTO,
2617 "wake", CTLTYPE_INT | CTLFLAG_RW, child, 0,
2618 acpi_wake_set_sysctl, "I", "Device set to wake the system");
2619 }
2620 }
2621 free(devlist, M_TEMP);
2622
2623 return (0);
2624 }
2625
2626 /* Enable or disable wake from userland. */
2627 static int
2628 acpi_wake_set_sysctl(SYSCTL_HANDLER_ARGS)
2629 {
2630 int enable, error;
2631 device_t dev;
2632
2633 dev = (device_t)arg1;
2634 enable = (acpi_get_flags(dev) & ACPI_FLAG_WAKE_ENABLED) ? 1 : 0;
2635
2636 error = sysctl_handle_int(oidp, &enable, 0, req);
2637 if (error != 0 || req->newptr == NULL)
2638 return (error);
2639 if (enable != 0 && enable != 1)
2640 return (EINVAL);
2641
2642 return (acpi_wake_set_enable(dev, enable));
2643 }
2644
2645 /* Parse a device's _PRW into a structure. */
2646 int
2647 acpi_parse_prw(ACPI_HANDLE h, struct acpi_prw_data *prw)
2648 {
2649 ACPI_STATUS status;
2650 ACPI_BUFFER prw_buffer;
2651 ACPI_OBJECT *res, *res2;
2652 int error, i, power_count;
2653
2654 if (h == NULL || prw == NULL)
2655 return (EINVAL);
2656
2657 /*
2658 * The _PRW object (7.2.9) is only required for devices that have the
2659 * ability to wake the system from a sleeping state.
2660 */
2661 error = EINVAL;
2662 prw_buffer.Pointer = NULL;
2663 prw_buffer.Length = ACPI_ALLOCATE_BUFFER;
2664 status = AcpiEvaluateObject(h, "_PRW", NULL, &prw_buffer);
2665 if (ACPI_FAILURE(status))
2666 return (ENOENT);
2667 res = (ACPI_OBJECT *)prw_buffer.Pointer;
2668 if (res == NULL)
2669 return (ENOENT);
2670 if (!ACPI_PKG_VALID(res, 2))
2671 goto out;
2672
2673 /*
2674 * Element 1 of the _PRW object:
2675 * The lowest power system sleeping state that can be entered while still
2676 * providing wake functionality. The sleeping state being entered must
2677 * be less than (i.e., higher power) or equal to this value.
2678 */
2679 if (acpi_PkgInt32(res, 1, &prw->lowest_wake) != 0)
2680 goto out;
2681
2682 /*
2683 * Element 0 of the _PRW object:
2684 */
2685 switch (res->Package.Elements[0].Type) {
2686 case ACPI_TYPE_INTEGER:
2687 /*
2688 * If the data type of this package element is numeric, then this
2689 * _PRW package element is the bit index in the GPEx_EN, in the
2690 * GPE blocks described in the FADT, of the enable bit that is
2691 * enabled for the wake event.
2692 */
2693 prw->gpe_handle = NULL;
2694 prw->gpe_bit = res->Package.Elements[0].Integer.Value;
2695 error = 0;
2696 break;
2697 case ACPI_TYPE_PACKAGE:
2698 /*
2699 * If the data type of this package element is a package, then this
2700 * _PRW package element is itself a package containing two
2701 * elements. The first is an object reference to the GPE Block
2702 * device that contains the GPE that will be triggered by the wake
2703 * event. The second element is numeric and it contains the bit
2704 * index in the GPEx_EN, in the GPE Block referenced by the
2705 * first element in the package, of the enable bit that is enabled for
2706 * the wake event.
2707 *
2708 * For example, if this field is a package then it is of the form:
2709 * Package() {\_SB.PCI0.ISA.GPE, 2}
2710 */
2711 res2 = &res->Package.Elements[0];
2712 if (!ACPI_PKG_VALID(res2, 2))
2713 goto out;
2714 prw->gpe_handle = acpi_GetReference(NULL, &res2->Package.Elements[0]);
2715 if (prw->gpe_handle == NULL)
2716 goto out;
2717 if (acpi_PkgInt32(res2, 1, &prw->gpe_bit) != 0)
2718 goto out;
2719 error = 0;
2720 break;
2721 default:
2722 goto out;
2723 }
2724
2725 /* Elements 2 to N of the _PRW object are power resources. */
2726 power_count = res->Package.Count - 2;
2727 if (power_count > ACPI_PRW_MAX_POWERRES) {
2728 printf("ACPI device %s has too many power resources\n", acpi_name(h));
2729 power_count = 0;
2730 }
2731 prw->power_res_count = power_count;
2732 for (i = 0; i < power_count; i++)
2733 prw->power_res[i] = res->Package.Elements[i];
2734
2735 out:
2736 if (prw_buffer.Pointer != NULL)
2737 AcpiOsFree(prw_buffer.Pointer);
2738 return (error);
2739 }
2740
2741 /*
2742 * ACPI Event Handlers
2743 */
2744
2745 /* System Event Handlers (registered by EVENTHANDLER_REGISTER) */
2746
2747 static void
2748 acpi_system_eventhandler_sleep(void *arg, int state)
2749 {
2750 int ret;
2751
2752 ACPI_FUNCTION_TRACE_U32((char *)(uintptr_t)__func__, state);
2753
2754 /* Check if button action is disabled. */
2755 if (state == ACPI_S_STATES_MAX + 1)
2756 return;
2757
2758 /* Request that the system prepare to enter the given suspend state. */
2759 ret = acpi_ReqSleepState((struct acpi_softc *)arg, state);
2760 if (ret != 0)
2761 printf("acpi: request to enter state S%d failed (err %d)\n",
2762 state, ret);
2763
2764 return_VOID;
2765 }
2766
2767 static void
2768 acpi_system_eventhandler_wakeup(void *arg, int state)
2769 {
2770
2771 ACPI_FUNCTION_TRACE_U32((char *)(uintptr_t)__func__, state);
2772
2773 /* Currently, nothing to do for wakeup. */
2774
2775 return_VOID;
2776 }
2777
2778 /*
2779 * ACPICA Event Handlers (FixedEvent, also called from button notify handler)
2780 */
2781 UINT32
2782 acpi_event_power_button_sleep(void *context)
2783 {
2784 struct acpi_softc *sc = (struct acpi_softc *)context;
2785
2786 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2787
2788 EVENTHANDLER_INVOKE(acpi_sleep_event, sc->acpi_power_button_sx);
2789
2790 return_VALUE (ACPI_INTERRUPT_HANDLED);
2791 }
2792
2793 UINT32
2794 acpi_event_power_button_wake(void *context)
2795 {
2796 struct acpi_softc *sc = (struct acpi_softc *)context;
2797
2798 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2799
2800 EVENTHANDLER_INVOKE(acpi_wakeup_event, sc->acpi_power_button_sx);
2801
2802 return_VALUE (ACPI_INTERRUPT_HANDLED);
2803 }
2804
2805 UINT32
2806 acpi_event_sleep_button_sleep(void *context)
2807 {
2808 struct acpi_softc *sc = (struct acpi_softc *)context;
2809
2810 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2811
2812 EVENTHANDLER_INVOKE(acpi_sleep_event, sc->acpi_sleep_button_sx);
2813
2814 return_VALUE (ACPI_INTERRUPT_HANDLED);
2815 }
2816
2817 UINT32
2818 acpi_event_sleep_button_wake(void *context)
2819 {
2820 struct acpi_softc *sc = (struct acpi_softc *)context;
2821
2822 ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
2823
2824 EVENTHANDLER_INVOKE(acpi_wakeup_event, sc->acpi_sleep_button_sx);
2825
2826 return_VALUE (ACPI_INTERRUPT_HANDLED);
2827 }
2828
2829 /*
2830 * XXX This static buffer is suboptimal. There is no locking so only
2831 * use this for single-threaded callers.
2832 */
2833 char *
2834 acpi_name(ACPI_HANDLE handle)
2835 {
2836 ACPI_BUFFER buf;
2837 static char data[256];
2838
2839 buf.Length = sizeof(data);
2840 buf.Pointer = data;
2841
2842 if (handle && ACPI_SUCCESS(AcpiGetName(handle, ACPI_FULL_PATHNAME, &buf)))
2843 return (data);
2844 return ("(unknown)");
2845 }
2846
2847 /*
2848 * Debugging/bug-avoidance. Avoid trying to fetch info on various
2849 * parts of the namespace.
2850 */
2851 int
2852 acpi_avoid(ACPI_HANDLE handle)
2853 {
2854 char *cp, *env, *np;
2855 int len;
2856
2857 np = acpi_name(handle);
2858 if (*np == '\\')
2859 np++;
2860 if ((env = getenv("debug.acpi.avoid")) == NULL)
2861 return (0);
2862
2863 /* Scan the avoid list checking for a match */
2864 cp = env;
2865 for (;;) {
2866 while (*cp != 0 && isspace(*cp))
2867 cp++;
2868 if (*cp == 0)
2869 break;
2870 len = 0;
2871 while (cp[len] != 0 && !isspace(cp[len]))
2872 len++;
2873 if (!strncmp(cp, np, len)) {
2874 freeenv(env);
2875 return(1);
2876 }
2877 cp += len;
2878 }
2879 freeenv(env);
2880
2881 return (0);
2882 }
2883
2884 /*
2885 * Debugging/bug-avoidance. Disable ACPI subsystem components.
2886 */
2887 int
2888 acpi_disabled(char *subsys)
2889 {
2890 char *cp, *env;
2891 int len;
2892
2893 if ((env = getenv("debug.acpi.disabled")) == NULL)
2894 return (0);
2895 if (strcmp(env, "all") == 0) {
2896 freeenv(env);
2897 return (1);
2898 }
2899
2900 /* Scan the disable list, checking for a match. */
2901 cp = env;
2902 for (;;) {
2903 while (*cp != '\0' && isspace(*cp))
2904 cp++;
2905 if (*cp == '\0')
2906 break;
2907 len = 0;
2908 while (cp[len] != '\0' && !isspace(cp[len]))
2909 len++;
2910 if (strncmp(cp, subsys, len) == 0) {
2911 freeenv(env);
2912 return (1);
2913 }
2914 cp += len;
2915 }
2916 freeenv(env);
2917
2918 return (0);
2919 }
2920
2921 /*
2922 * Control interface.
2923 *
2924 * We multiplex ioctls for all participating ACPI devices here. Individual
2925 * drivers wanting to be accessible via /dev/acpi should use the
2926 * register/deregister interface to make their handlers visible.
2927 */
2928 struct acpi_ioctl_hook
2929 {
2930 TAILQ_ENTRY(acpi_ioctl_hook) link;
2931 u_long cmd;
2932 acpi_ioctl_fn fn;
2933 void *arg;
2934 };
2935
2936 static TAILQ_HEAD(,acpi_ioctl_hook) acpi_ioctl_hooks;
2937 static int acpi_ioctl_hooks_initted;
2938
2939 int
2940 acpi_register_ioctl(u_long cmd, acpi_ioctl_fn fn, void *arg)
2941 {
2942 struct acpi_ioctl_hook *hp;
2943
2944 if ((hp = malloc(sizeof(*hp), M_ACPIDEV, M_NOWAIT)) == NULL)
2945 return (ENOMEM);
2946 hp->cmd = cmd;
2947 hp->fn = fn;
2948 hp->arg = arg;
2949
2950 ACPI_LOCK(acpi);
2951 if (acpi_ioctl_hooks_initted == 0) {
2952 TAILQ_INIT(&acpi_ioctl_hooks);
2953 acpi_ioctl_hooks_initted = 1;
2954 }
2955 TAILQ_INSERT_TAIL(&acpi_ioctl_hooks, hp, link);
2956 ACPI_UNLOCK(acpi);
2957
2958 return (0);
2959 }
2960
2961 void
2962 acpi_deregister_ioctl(u_long cmd, acpi_ioctl_fn fn)
2963 {
2964 struct acpi_ioctl_hook *hp;
2965
2966 ACPI_LOCK(acpi);
2967 TAILQ_FOREACH(hp, &acpi_ioctl_hooks, link)
2968 if (hp->cmd == cmd && hp->fn == fn)
2969 break;
2970
2971 if (hp != NULL) {
2972 TAILQ_REMOVE(&acpi_ioctl_hooks, hp, link);
2973 free(hp, M_ACPIDEV);
2974 }
2975 ACPI_UNLOCK(acpi);
2976 }
2977
2978 static int
2979 acpiopen(struct cdev *dev, int flag, int fmt, d_thread_t *td)
2980 {
2981 return (0);
2982 }
2983
2984 static int
2985 acpiclose(struct cdev *dev, int flag, int fmt, d_thread_t *td)
2986 {
2987 return (0);
2988 }
2989
2990 static int
2991 acpiioctl(struct cdev *dev, u_long cmd, caddr_t addr, int flag, d_thread_t *td)
2992 {
2993 struct acpi_softc *sc;
2994 struct acpi_ioctl_hook *hp;
2995 int error, state;
2996
2997 error = 0;
2998 hp = NULL;
2999 sc = dev->si_drv1;
3000
3001 /*
3002 * Scan the list of registered ioctls, looking for handlers.
3003 */
3004 ACPI_LOCK(acpi);
3005 if (acpi_ioctl_hooks_initted)
3006 TAILQ_FOREACH(hp, &acpi_ioctl_hooks, link) {
3007 if (hp->cmd == cmd)
3008 break;
3009 }
3010 ACPI_UNLOCK(acpi);
3011 if (hp)
3012 return (hp->fn(cmd, addr, hp->arg));
3013
3014 /*
3015 * Core ioctls are not permitted for non-writable user.
3016 * Currently, other ioctls just fetch information.
3017 * Not changing system behavior.
3018 */
3019 if ((flag & FWRITE) == 0)
3020 return (EPERM);
3021
3022 /* Core system ioctls. */
3023 switch (cmd) {
3024 case ACPIIO_REQSLPSTATE:
3025 state = *(int *)addr;
3026 if (state != ACPI_STATE_S5)
3027 error = acpi_ReqSleepState(sc, state);
3028 else {
3029 printf("power off via acpi ioctl not supported\n");
3030 error = ENXIO;
3031 }
3032 break;
3033 case ACPIIO_ACKSLPSTATE:
3034 error = *(int *)addr;
3035 error = acpi_AckSleepState(sc->acpi_clone, error);
3036 break;
3037 case ACPIIO_SETSLPSTATE: /* DEPRECATED */
3038 error = EINVAL;
3039 state = *(int *)addr;
3040 if (state >= ACPI_STATE_S0 && state <= ACPI_S_STATES_MAX)
3041 if (ACPI_SUCCESS(acpi_SetSleepState(sc, state)))
3042 error = 0;
3043 break;
3044 default:
3045 error = ENXIO;
3046 break;
3047 }
3048
3049 return (error);
3050 }
3051
3052 static int
3053 acpi_supported_sleep_state_sysctl(SYSCTL_HANDLER_ARGS)
3054 {
3055 int error;
3056 struct sbuf sb;
3057 UINT8 state, TypeA, TypeB;
3058
3059 sbuf_new(&sb, NULL, 32, SBUF_AUTOEXTEND);
3060 for (state = ACPI_STATE_S1; state < ACPI_S_STATES_MAX + 1; state++)
3061 if (ACPI_SUCCESS(AcpiGetSleepTypeData(state, &TypeA, &TypeB)))
3062 sbuf_printf(&sb, "S%d ", state);
3063 sbuf_trim(&sb);
3064 sbuf_finish(&sb);
3065 error = sysctl_handle_string(oidp, sbuf_data(&sb), sbuf_len(&sb), req);
3066 sbuf_delete(&sb);
3067 return (error);
3068 }
3069
3070 static int
3071 acpi_sleep_state_sysctl(SYSCTL_HANDLER_ARGS)
3072 {
3073 char sleep_state[10];
3074 int error;
3075 u_int new_state, old_state;
3076
3077 old_state = *(u_int *)oidp->oid_arg1;
3078 if (old_state > ACPI_S_STATES_MAX + 1)
3079 strlcpy(sleep_state, "unknown", sizeof(sleep_state));
3080 else
3081 strlcpy(sleep_state, sleep_state_names[old_state], sizeof(sleep_state));
3082 error = sysctl_handle_string(oidp, sleep_state, sizeof(sleep_state), req);
3083 if (error == 0 && req->newptr != NULL) {
3084 new_state = ACPI_STATE_S0;
3085 for (; new_state <= ACPI_S_STATES_MAX + 1; new_state++)
3086 if (strcmp(sleep_state, sleep_state_names[new_state]) == 0)
3087 break;
3088 if (new_state <= ACPI_S_STATES_MAX + 1) {
3089 if (new_state != old_state)
3090 *(u_int *)oidp->oid_arg1 = new_state;
3091 } else
3092 error = EINVAL;
3093 }
3094
3095 return (error);
3096 }
3097
3098 /* Inform devctl(4) when we receive a Notify. */
3099 void
3100 acpi_UserNotify(const char *subsystem, ACPI_HANDLE h, uint8_t notify)
3101 {
3102 char notify_buf[16];
3103 ACPI_BUFFER handle_buf;
3104 ACPI_STATUS status;
3105
3106 if (subsystem == NULL)
3107 return;
3108
3109 handle_buf.Pointer = NULL;
3110 handle_buf.Length = ACPI_ALLOCATE_BUFFER;
3111 status = AcpiNsHandleToPathname(h, &handle_buf);
3112 if (ACPI_FAILURE(status))
3113 return;
3114 snprintf(notify_buf, sizeof(notify_buf), "notify=0x%02x", notify);
3115 devctl_notify("ACPI", subsystem, handle_buf.Pointer, notify_buf);
3116 AcpiOsFree(handle_buf.Pointer);
3117 }
3118
3119 #ifdef ACPI_DEBUG
3120 /*
3121 * Support for parsing debug options from the kernel environment.
3122 *
3123 * Bits may be set in the AcpiDbgLayer and AcpiDbgLevel debug registers
3124 * by specifying the names of the bits in the debug.acpi.layer and
3125 * debug.acpi.level environment variables. Bits may be unset by
3126 * prefixing the bit name with !.
3127 */
3128 struct debugtag
3129 {
3130 char *name;
3131 UINT32 value;
3132 };
3133
3134 static struct debugtag dbg_layer[] = {
3135 {"ACPI_UTILITIES", ACPI_UTILITIES},
3136 {"ACPI_HARDWARE", ACPI_HARDWARE},
3137 {"ACPI_EVENTS", ACPI_EVENTS},
3138 {"ACPI_TABLES", ACPI_TABLES},
3139 {"ACPI_NAMESPACE", ACPI_NAMESPACE},
3140 {"ACPI_PARSER", ACPI_PARSER},
3141 {"ACPI_DISPATCHER", ACPI_DISPATCHER},
3142 {"ACPI_EXECUTER", ACPI_EXECUTER},
3143 {"ACPI_RESOURCES", ACPI_RESOURCES},
3144 {"ACPI_CA_DEBUGGER", ACPI_CA_DEBUGGER},
3145 {"ACPI_OS_SERVICES", ACPI_OS_SERVICES},
3146 {"ACPI_CA_DISASSEMBLER", ACPI_CA_DISASSEMBLER},
3147 {"ACPI_ALL_COMPONENTS", ACPI_ALL_COMPONENTS},
3148
3149 {"ACPI_AC_ADAPTER", ACPI_AC_ADAPTER},
3150 {"ACPI_BATTERY", ACPI_BATTERY},
3151 {"ACPI_BUS", ACPI_BUS},
3152 {"ACPI_BUTTON", ACPI_BUTTON},
3153 {"ACPI_EC", ACPI_EC},
3154 {"ACPI_FAN", ACPI_FAN},
3155 {"ACPI_POWERRES", ACPI_POWERRES},
3156 {"ACPI_PROCESSOR", ACPI_PROCESSOR},
3157 {"ACPI_THERMAL", ACPI_THERMAL},
3158 {"ACPI_TIMER", ACPI_TIMER},
3159 {"ACPI_ALL_DRIVERS", ACPI_ALL_DRIVERS},
3160 {NULL, 0}
3161 };
3162
3163 static struct debugtag dbg_level[] = {
3164 {"ACPI_LV_ERROR", ACPI_LV_ERROR},
3165 {"ACPI_LV_WARN", ACPI_LV_WARN},
3166 {"ACPI_LV_INIT", ACPI_LV_INIT},
3167 {"ACPI_LV_DEBUG_OBJECT", ACPI_LV_DEBUG_OBJECT},
3168 {"ACPI_LV_INFO", ACPI_LV_INFO},
3169 {"ACPI_LV_ALL_EXCEPTIONS", ACPI_LV_ALL_EXCEPTIONS},
3170
3171 /* Trace verbosity level 1 [Standard Trace Level] */
3172 {"ACPI_LV_INIT_NAMES", ACPI_LV_INIT_NAMES},
3173 {"ACPI_LV_PARSE", ACPI_LV_PARSE},
3174 {"ACPI_LV_LOAD", ACPI_LV_LOAD},
3175 {"ACPI_LV_DISPATCH", ACPI_LV_DISPATCH},
3176 {"ACPI_LV_EXEC", ACPI_LV_EXEC},
3177 {"ACPI_LV_NAMES", ACPI_LV_NAMES},
3178 {"ACPI_LV_OPREGION", ACPI_LV_OPREGION},
3179 {"ACPI_LV_BFIELD", ACPI_LV_BFIELD},
3180 {"ACPI_LV_TABLES", ACPI_LV_TABLES},
3181 {"ACPI_LV_VALUES", ACPI_LV_VALUES},
3182 {"ACPI_LV_OBJECTS", ACPI_LV_OBJECTS},
3183 {"ACPI_LV_RESOURCES", ACPI_LV_RESOURCES},
3184 {"ACPI_LV_USER_REQUESTS", ACPI_LV_USER_REQUESTS},
3185 {"ACPI_LV_PACKAGE", ACPI_LV_PACKAGE},
3186 {"ACPI_LV_VERBOSITY1", ACPI_LV_VERBOSITY1},
3187
3188 /* Trace verbosity level 2 [Function tracing and memory allocation] */
3189 {"ACPI_LV_ALLOCATIONS", ACPI_LV_ALLOCATIONS},
3190 {"ACPI_LV_FUNCTIONS", ACPI_LV_FUNCTIONS},
3191 {"ACPI_LV_OPTIMIZATIONS", ACPI_LV_OPTIMIZATIONS},
3192 {"ACPI_LV_VERBOSITY2", ACPI_LV_VERBOSITY2},
3193 {"ACPI_LV_ALL", ACPI_LV_ALL},
3194
3195 /* Trace verbosity level 3 [Threading, I/O, and Interrupts] */
3196 {"ACPI_LV_MUTEX", ACPI_LV_MUTEX},
3197 {"ACPI_LV_THREADS", ACPI_LV_THREADS},
3198 {"ACPI_LV_IO", ACPI_LV_IO},
3199 {"ACPI_LV_INTERRUPTS", ACPI_LV_INTERRUPTS},
3200 {"ACPI_LV_VERBOSITY3", ACPI_LV_VERBOSITY3},
3201
3202 /* Exceptionally verbose output -- also used in the global "DebugLevel" */
3203 {"ACPI_LV_AML_DISASSEMBLE", ACPI_LV_AML_DISASSEMBLE},
3204 {"ACPI_LV_VERBOSE_INFO", ACPI_LV_VERBOSE_INFO},
3205 {"ACPI_LV_FULL_TABLES", ACPI_LV_FULL_TABLES},
3206 {"ACPI_LV_EVENTS", ACPI_LV_EVENTS},
3207 {"ACPI_LV_VERBOSE", ACPI_LV_VERBOSE},
3208 {NULL, 0}
3209 };
3210
3211 static void
3212 acpi_parse_debug(char *cp, struct debugtag *tag, UINT32 *flag)
3213 {
3214 char *ep;
3215 int i, l;
3216 int set;
3217
3218 while (*cp) {
3219 if (isspace(*cp)) {
3220 cp++;
3221 continue;
3222 }
3223 ep = cp;
3224 while (*ep && !isspace(*ep))
3225 ep++;
3226 if (*cp == '!') {
3227 set = 0;
3228 cp++;
3229 if (cp == ep)
3230 continue;
3231 } else {
3232 set = 1;
3233 }
3234 l = ep - cp;
3235 for (i = 0; tag[i].name != NULL; i++) {
3236 if (!strncmp(cp, tag[i].name, l)) {
3237 if (set)
3238 *flag |= tag[i].value;
3239 else
3240 *flag &= ~tag[i].value;
3241 }
3242 }
3243 cp = ep;
3244 }
3245 }
3246
3247 static void
3248 acpi_set_debugging(void *junk)
3249 {
3250 char *layer, *level;
3251
3252 if (cold) {
3253 AcpiDbgLayer = 0;
3254 AcpiDbgLevel = 0;
3255 }
3256
3257 layer = getenv("debug.acpi.layer");
3258 level = getenv("debug.acpi.level");
3259 if (layer == NULL && level == NULL)
3260 return;
3261
3262 printf("ACPI set debug");
3263 if (layer != NULL) {
3264 if (strcmp("NONE", layer) != 0)
3265 printf(" layer '%s'", layer);
3266 acpi_parse_debug(layer, &dbg_layer[0], &AcpiDbgLayer);
3267 freeenv(layer);
3268 }
3269 if (level != NULL) {
3270 if (strcmp("NONE", level) != 0)
3271 printf(" level '%s'", level);
3272 acpi_parse_debug(level, &dbg_level[0], &AcpiDbgLevel);
3273 freeenv(level);
3274 }
3275 printf("\n");
3276 }
3277
3278 SYSINIT(acpi_debugging, SI_SUB_TUNABLES, SI_ORDER_ANY, acpi_set_debugging,
3279 NULL);
3280
3281 static int
3282 acpi_debug_sysctl(SYSCTL_HANDLER_ARGS)
3283 {
3284 int error, *dbg;
3285 struct debugtag *tag;
3286 struct sbuf sb;
3287
3288 if (sbuf_new(&sb, NULL, 128, SBUF_AUTOEXTEND) == NULL)
3289 return (ENOMEM);
3290 if (strcmp(oidp->oid_arg1, "debug.acpi.layer") == 0) {
3291 tag = &dbg_layer[0];
3292 dbg = &AcpiDbgLayer;
3293 } else {
3294 tag = &dbg_level[0];
3295 dbg = &AcpiDbgLevel;
3296 }
3297
3298 /* Get old values if this is a get request. */
3299 ACPI_SERIAL_BEGIN(acpi);
3300 if (*dbg == 0) {
3301 sbuf_cpy(&sb, "NONE");
3302 } else if (req->newptr == NULL) {
3303 for (; tag->name != NULL; tag++) {
3304 if ((*dbg & tag->value) == tag->value)
3305 sbuf_printf(&sb, "%s ", tag->name);
3306 }
3307 }
3308 sbuf_trim(&sb);
3309 sbuf_finish(&sb);
3310
3311 /* Copy out the old values to the user. */
3312 error = SYSCTL_OUT(req, sbuf_data(&sb), sbuf_len(&sb));
3313 sbuf_delete(&sb);
3314
3315 /* If the user is setting a string, parse it. */
3316 if (error == 0 && req->newptr != NULL) {
3317 *dbg = 0;
3318 setenv((char *)oidp->oid_arg1, (char *)req->newptr);
3319 acpi_set_debugging(NULL);
3320 }
3321 ACPI_SERIAL_END(acpi);
3322
3323 return (error);
3324 }
3325
3326 SYSCTL_PROC(_debug_acpi, OID_AUTO, layer, CTLFLAG_RW | CTLTYPE_STRING,
3327 "debug.acpi.layer", 0, acpi_debug_sysctl, "A", "");
3328 SYSCTL_PROC(_debug_acpi, OID_AUTO, level, CTLFLAG_RW | CTLTYPE_STRING,
3329 "debug.acpi.level", 0, acpi_debug_sysctl, "A", "");
3330 #endif /* ACPI_DEBUG */
3331
3332 static int
3333 acpi_pm_func(u_long cmd, void *arg, ...)
3334 {
3335 int state, acpi_state;
3336 int error;
3337 struct acpi_softc *sc;
3338 va_list ap;
3339
3340 error = 0;
3341 switch (cmd) {
3342 case POWER_CMD_SUSPEND:
3343 sc = (struct acpi_softc *)arg;
3344 if (sc == NULL) {
3345 error = EINVAL;
3346 goto out;
3347 }
3348
3349 va_start(ap, arg);
3350 state = va_arg(ap, int);
3351 va_end(ap);
3352
3353 switch (state) {
3354 case POWER_SLEEP_STATE_STANDBY:
3355 acpi_state = sc->acpi_standby_sx;
3356 break;
3357 case POWER_SLEEP_STATE_SUSPEND:
3358 acpi_state = sc->acpi_suspend_sx;
3359 break;
3360 case POWER_SLEEP_STATE_HIBERNATE:
3361 acpi_state = ACPI_STATE_S4;
3362 break;
3363 default:
3364 error = EINVAL;
3365 goto out;
3366 }
3367
3368 if (ACPI_FAILURE(acpi_EnterSleepState(sc, acpi_state)))
3369 error = ENXIO;
3370 break;
3371 default:
3372 error = EINVAL;
3373 goto out;
3374 }
3375
3376 out:
3377 return (error);
3378 }
3379
3380 static void
3381 acpi_pm_register(void *arg)
3382 {
3383 if (!cold || resource_disabled("acpi", 0))
3384 return;
3385
3386 power_pm_register(POWER_PM_TYPE_ACPI, acpi_pm_func, NULL);
3387 }
3388
3389 SYSINIT(power, SI_SUB_KLD, SI_ORDER_ANY, acpi_pm_register, 0);
Cache object: c6efab3bf65bdb07893d577198f2ee51
|