2 * Copyright (C) 2005, 2006 IBM Corporation
3 * Copyright (C) 2014, 2015 Intel Corporation
6 * Leendert van Doorn <leendert@watson.ibm.com>
7 * Kylene Hall <kjhall@us.ibm.com>
9 * Maintained by: <tpmdd-devel@lists.sourceforge.net>
11 * Device driver for TCG/TCPA TPM (trusted platform module).
12 * Specifications at www.trustedcomputinggroup.org
14 * This device driver implements the TPM interface as defined in
15 * the TCG TPM Interface Spec version 1.2, revision 1.0.
17 * This program is free software; you can redistribute it and/or
18 * modify it under the terms of the GNU General Public License as
19 * published by the Free Software Foundation, version 2 of the
22 #include <linux/init.h>
23 #include <linux/module.h>
24 #include <linux/moduleparam.h>
25 #include <linux/pnp.h>
26 #include <linux/slab.h>
27 #include <linux/interrupt.h>
28 #include <linux/wait.h>
29 #include <linux/acpi.h>
30 #include <linux/freezer.h>
31 #include <acpi/actbl2.h>
35 TPM_ACCESS_VALID = 0x80,
36 TPM_ACCESS_ACTIVE_LOCALITY = 0x20,
37 TPM_ACCESS_REQUEST_PENDING = 0x04,
38 TPM_ACCESS_REQUEST_USE = 0x02,
43 TPM_STS_COMMAND_READY = 0x40,
45 TPM_STS_DATA_AVAIL = 0x10,
46 TPM_STS_DATA_EXPECT = 0x08,
50 TPM_GLOBAL_INT_ENABLE = 0x80000000,
51 TPM_INTF_BURST_COUNT_STATIC = 0x100,
52 TPM_INTF_CMD_READY_INT = 0x080,
53 TPM_INTF_INT_EDGE_FALLING = 0x040,
54 TPM_INTF_INT_EDGE_RISING = 0x020,
55 TPM_INTF_INT_LEVEL_LOW = 0x010,
56 TPM_INTF_INT_LEVEL_HIGH = 0x008,
57 TPM_INTF_LOCALITY_CHANGE_INT = 0x004,
58 TPM_INTF_STS_VALID_INT = 0x002,
59 TPM_INTF_DATA_AVAIL_INT = 0x001,
63 TIS_MEM_BASE = 0xFED40000,
65 TIS_SHORT_TIMEOUT = 750, /* ms */
66 TIS_LONG_TIMEOUT = 2000, /* 2 sec */
75 static struct tpm_info tis_default_info = {
76 .start = TIS_MEM_BASE,
81 /* Some timeout values are needed before it is known whether the chip is
84 #define TIS_TIMEOUT_A_MAX max(TIS_SHORT_TIMEOUT, TPM2_TIMEOUT_A)
85 #define TIS_TIMEOUT_B_MAX max(TIS_LONG_TIMEOUT, TPM2_TIMEOUT_B)
86 #define TIS_TIMEOUT_C_MAX max(TIS_SHORT_TIMEOUT, TPM2_TIMEOUT_C)
87 #define TIS_TIMEOUT_D_MAX max(TIS_SHORT_TIMEOUT, TPM2_TIMEOUT_D)
89 #define TPM_ACCESS(l) (0x0000 | ((l) << 12))
90 #define TPM_INT_ENABLE(l) (0x0008 | ((l) << 12))
91 #define TPM_INT_VECTOR(l) (0x000C | ((l) << 12))
92 #define TPM_INT_STATUS(l) (0x0010 | ((l) << 12))
93 #define TPM_INTF_CAPS(l) (0x0014 | ((l) << 12))
94 #define TPM_STS(l) (0x0018 | ((l) << 12))
95 #define TPM_STS3(l) (0x001b | ((l) << 12))
96 #define TPM_DATA_FIFO(l) (0x0024 | ((l) << 12))
98 #define TPM_DID_VID(l) (0x0F00 | ((l) << 12))
99 #define TPM_RID(l) (0x0F04 | ((l) << 12))
105 #if defined(CONFIG_PNP) && defined(CONFIG_ACPI)
106 static int has_hid(struct acpi_device *dev, const char *hid)
108 struct acpi_hardware_id *id;
110 list_for_each_entry(id, &dev->pnp.ids, list)
111 if (!strcmp(hid, id->id))
117 static inline int is_itpm(struct acpi_device *dev)
119 return has_hid(dev, "INTC0102");
122 static inline int is_fifo(struct acpi_device *dev)
124 struct acpi_table_tpm2 *tbl;
128 if (!has_hid(dev, "MSFT0101"))
131 st = acpi_get_table(ACPI_SIG_TPM2, 1,
132 (struct acpi_table_header **) &tbl);
133 if (ACPI_FAILURE(st)) {
134 dev_err(&dev->dev, "failed to get TPM2 ACPI table\n");
138 if (le32_to_cpu(tbl->start_method) != TPM2_START_FIFO)
145 static inline int is_itpm(struct acpi_device *dev)
150 static inline int is_fifo(struct acpi_device *dev)
156 /* Before we attempt to access the TPM we must see that the valid bit is set.
157 * The specification says that this bit is 0 at reset and remains 0 until the
158 * 'TPM has gone through its self test and initialization and has established
159 * correct values in the other bits.' */
160 static int wait_startup(struct tpm_chip *chip, int l)
162 unsigned long stop = jiffies + chip->vendor.timeout_a;
164 if (ioread8(chip->vendor.iobase + TPM_ACCESS(l)) &
168 } while (time_before(jiffies, stop));
172 static int check_locality(struct tpm_chip *chip, int l)
174 if ((ioread8(chip->vendor.iobase + TPM_ACCESS(l)) &
175 (TPM_ACCESS_ACTIVE_LOCALITY | TPM_ACCESS_VALID)) ==
176 (TPM_ACCESS_ACTIVE_LOCALITY | TPM_ACCESS_VALID))
177 return chip->vendor.locality = l;
182 static void release_locality(struct tpm_chip *chip, int l, int force)
184 if (force || (ioread8(chip->vendor.iobase + TPM_ACCESS(l)) &
185 (TPM_ACCESS_REQUEST_PENDING | TPM_ACCESS_VALID)) ==
186 (TPM_ACCESS_REQUEST_PENDING | TPM_ACCESS_VALID))
187 iowrite8(TPM_ACCESS_ACTIVE_LOCALITY,
188 chip->vendor.iobase + TPM_ACCESS(l));
191 static int request_locality(struct tpm_chip *chip, int l)
193 unsigned long stop, timeout;
196 if (check_locality(chip, l) >= 0)
199 iowrite8(TPM_ACCESS_REQUEST_USE,
200 chip->vendor.iobase + TPM_ACCESS(l));
202 stop = jiffies + chip->vendor.timeout_a;
204 if (chip->vendor.irq) {
206 timeout = stop - jiffies;
207 if ((long)timeout <= 0)
209 rc = wait_event_interruptible_timeout(chip->vendor.int_queue,
215 if (rc == -ERESTARTSYS && freezing(current)) {
216 clear_thread_flag(TIF_SIGPENDING);
220 /* wait for burstcount */
222 if (check_locality(chip, l) >= 0)
226 while (time_before(jiffies, stop));
231 static u8 tpm_tis_status(struct tpm_chip *chip)
233 return ioread8(chip->vendor.iobase +
234 TPM_STS(chip->vendor.locality));
237 static void tpm_tis_ready(struct tpm_chip *chip)
239 /* this causes the current command to be aborted */
240 iowrite8(TPM_STS_COMMAND_READY,
241 chip->vendor.iobase + TPM_STS(chip->vendor.locality));
244 static int get_burstcount(struct tpm_chip *chip)
249 /* wait for burstcount */
250 /* which timeout value, spec has 2 answers (c & d) */
251 stop = jiffies + chip->vendor.timeout_d;
253 burstcnt = ioread8(chip->vendor.iobase +
254 TPM_STS(chip->vendor.locality) + 1);
255 burstcnt += ioread8(chip->vendor.iobase +
256 TPM_STS(chip->vendor.locality) +
261 } while (time_before(jiffies, stop));
265 static int recv_data(struct tpm_chip *chip, u8 *buf, size_t count)
267 int size = 0, burstcnt;
268 while (size < count &&
269 wait_for_tpm_stat(chip,
270 TPM_STS_DATA_AVAIL | TPM_STS_VALID,
271 chip->vendor.timeout_c,
272 &chip->vendor.read_queue, true)
274 burstcnt = get_burstcount(chip);
275 for (; burstcnt > 0 && size < count; burstcnt--)
276 buf[size++] = ioread8(chip->vendor.iobase +
277 TPM_DATA_FIFO(chip->vendor.
283 static int tpm_tis_recv(struct tpm_chip *chip, u8 *buf, size_t count)
286 int expected, status;
288 if (count < TPM_HEADER_SIZE) {
293 /* read first 10 bytes, including tag, paramsize, and result */
295 recv_data(chip, buf, TPM_HEADER_SIZE)) < TPM_HEADER_SIZE) {
296 dev_err(chip->pdev, "Unable to read header\n");
300 expected = be32_to_cpu(*(__be32 *) (buf + 2));
301 if (expected > count) {
307 recv_data(chip, &buf[TPM_HEADER_SIZE],
308 expected - TPM_HEADER_SIZE)) < expected) {
309 dev_err(chip->pdev, "Unable to read remainder of result\n");
314 wait_for_tpm_stat(chip, TPM_STS_VALID, chip->vendor.timeout_c,
315 &chip->vendor.int_queue, false);
316 status = tpm_tis_status(chip);
317 if (status & TPM_STS_DATA_AVAIL) { /* retry? */
318 dev_err(chip->pdev, "Error left over data\n");
325 release_locality(chip, chip->vendor.locality, 0);
330 module_param(itpm, bool, 0444);
331 MODULE_PARM_DESC(itpm, "Force iTPM workarounds (found on some Lenovo laptops)");
334 * If interrupts are used (signaled by an irq set in the vendor structure)
335 * tpm.c can skip polling for the data to be available as the interrupt is
338 static int tpm_tis_send_data(struct tpm_chip *chip, u8 *buf, size_t len)
340 int rc, status, burstcnt;
343 if (request_locality(chip, 0) < 0)
346 status = tpm_tis_status(chip);
347 if ((status & TPM_STS_COMMAND_READY) == 0) {
349 if (wait_for_tpm_stat
350 (chip, TPM_STS_COMMAND_READY, chip->vendor.timeout_b,
351 &chip->vendor.int_queue, false) < 0) {
357 while (count < len - 1) {
358 burstcnt = get_burstcount(chip);
359 for (; burstcnt > 0 && count < len - 1; burstcnt--) {
360 iowrite8(buf[count], chip->vendor.iobase +
361 TPM_DATA_FIFO(chip->vendor.locality));
365 wait_for_tpm_stat(chip, TPM_STS_VALID, chip->vendor.timeout_c,
366 &chip->vendor.int_queue, false);
367 status = tpm_tis_status(chip);
368 if (!itpm && (status & TPM_STS_DATA_EXPECT) == 0) {
374 /* write last byte */
376 chip->vendor.iobase + TPM_DATA_FIFO(chip->vendor.locality));
377 wait_for_tpm_stat(chip, TPM_STS_VALID, chip->vendor.timeout_c,
378 &chip->vendor.int_queue, false);
379 status = tpm_tis_status(chip);
380 if ((status & TPM_STS_DATA_EXPECT) != 0) {
389 release_locality(chip, chip->vendor.locality, 0);
393 static void disable_interrupts(struct tpm_chip *chip)
398 ioread32(chip->vendor.iobase +
399 TPM_INT_ENABLE(chip->vendor.locality));
400 intmask &= ~TPM_GLOBAL_INT_ENABLE;
402 chip->vendor.iobase +
403 TPM_INT_ENABLE(chip->vendor.locality));
404 free_irq(chip->vendor.irq, chip);
405 chip->vendor.irq = 0;
409 * If interrupts are used (signaled by an irq set in the vendor structure)
410 * tpm.c can skip polling for the data to be available as the interrupt is
413 static int tpm_tis_send_main(struct tpm_chip *chip, u8 *buf, size_t len)
419 rc = tpm_tis_send_data(chip, buf, len);
425 chip->vendor.iobase + TPM_STS(chip->vendor.locality));
427 if (chip->vendor.irq) {
428 ordinal = be32_to_cpu(*((__be32 *) (buf + 6)));
430 if (chip->flags & TPM_CHIP_FLAG_TPM2)
431 dur = tpm2_calc_ordinal_duration(chip, ordinal);
433 dur = tpm_calc_ordinal_duration(chip, ordinal);
435 if (wait_for_tpm_stat
436 (chip, TPM_STS_DATA_AVAIL | TPM_STS_VALID, dur,
437 &chip->vendor.read_queue, false) < 0) {
445 release_locality(chip, chip->vendor.locality, 0);
449 static int tpm_tis_send(struct tpm_chip *chip, u8 *buf, size_t len)
452 struct priv_data *priv = chip->vendor.priv;
454 if (!chip->vendor.irq || priv->irq_tested)
455 return tpm_tis_send_main(chip, buf, len);
457 /* Verify receipt of the expected IRQ */
458 irq = chip->vendor.irq;
459 chip->vendor.irq = 0;
460 rc = tpm_tis_send_main(chip, buf, len);
461 chip->vendor.irq = irq;
462 if (!priv->irq_tested)
464 if (!priv->irq_tested) {
465 disable_interrupts(chip);
467 FW_BUG "TPM interrupt not working, polling instead\n");
469 priv->irq_tested = true;
473 struct tis_vendor_timeout_override {
475 unsigned long timeout_us[4];
478 static const struct tis_vendor_timeout_override vendor_timeout_overrides[] = {
480 { 0x32041114, { (TIS_SHORT_TIMEOUT*1000), (TIS_LONG_TIMEOUT*1000),
481 (TIS_SHORT_TIMEOUT*1000), (TIS_SHORT_TIMEOUT*1000) } },
484 static bool tpm_tis_update_timeouts(struct tpm_chip *chip,
485 unsigned long *timeout_cap)
490 did_vid = ioread32(chip->vendor.iobase + TPM_DID_VID(0));
492 for (i = 0; i != ARRAY_SIZE(vendor_timeout_overrides); i++) {
493 if (vendor_timeout_overrides[i].did_vid != did_vid)
495 memcpy(timeout_cap, vendor_timeout_overrides[i].timeout_us,
496 sizeof(vendor_timeout_overrides[i].timeout_us));
504 * Early probing for iTPM with STS_DATA_EXPECT flaw.
505 * Try sending command without itpm flag set and if that
506 * fails, repeat with itpm flag set.
508 static int probe_itpm(struct tpm_chip *chip)
511 u8 cmd_getticks[] = {
512 0x00, 0xc1, 0x00, 0x00, 0x00, 0x0a,
513 0x00, 0x00, 0x00, 0xf1
515 size_t len = sizeof(cmd_getticks);
516 bool rem_itpm = itpm;
517 u16 vendor = ioread16(chip->vendor.iobase + TPM_DID_VID(0));
519 /* probe only iTPMS */
520 if (vendor != TPM_VID_INTEL)
525 rc = tpm_tis_send_data(chip, cmd_getticks, len);
530 release_locality(chip, chip->vendor.locality, 0);
534 rc = tpm_tis_send_data(chip, cmd_getticks, len);
536 dev_info(chip->pdev, "Detected an iTPM.\n");
544 release_locality(chip, chip->vendor.locality, 0);
549 static bool tpm_tis_req_canceled(struct tpm_chip *chip, u8 status)
551 switch (chip->vendor.manufacturer_id) {
552 case TPM_VID_WINBOND:
553 return ((status == TPM_STS_VALID) ||
554 (status == (TPM_STS_VALID | TPM_STS_COMMAND_READY)));
556 return (status == (TPM_STS_VALID | TPM_STS_COMMAND_READY));
558 return (status == TPM_STS_COMMAND_READY);
562 static const struct tpm_class_ops tpm_tis = {
563 .status = tpm_tis_status,
564 .recv = tpm_tis_recv,
565 .send = tpm_tis_send,
566 .cancel = tpm_tis_ready,
567 .update_timeouts = tpm_tis_update_timeouts,
568 .req_complete_mask = TPM_STS_DATA_AVAIL | TPM_STS_VALID,
569 .req_complete_val = TPM_STS_DATA_AVAIL | TPM_STS_VALID,
570 .req_canceled = tpm_tis_req_canceled,
573 static irqreturn_t tis_int_probe(int irq, void *dev_id)
575 struct tpm_chip *chip = dev_id;
578 interrupt = ioread32(chip->vendor.iobase +
579 TPM_INT_STATUS(chip->vendor.locality));
584 chip->vendor.probed_irq = irq;
586 /* Clear interrupts handled with TPM_EOI */
588 chip->vendor.iobase +
589 TPM_INT_STATUS(chip->vendor.locality));
593 static irqreturn_t tis_int_handler(int dummy, void *dev_id)
595 struct tpm_chip *chip = dev_id;
599 interrupt = ioread32(chip->vendor.iobase +
600 TPM_INT_STATUS(chip->vendor.locality));
605 ((struct priv_data *)chip->vendor.priv)->irq_tested = true;
606 if (interrupt & TPM_INTF_DATA_AVAIL_INT)
607 wake_up_interruptible(&chip->vendor.read_queue);
608 if (interrupt & TPM_INTF_LOCALITY_CHANGE_INT)
609 for (i = 0; i < 5; i++)
610 if (check_locality(chip, i) >= 0)
613 (TPM_INTF_LOCALITY_CHANGE_INT | TPM_INTF_STS_VALID_INT |
614 TPM_INTF_CMD_READY_INT))
615 wake_up_interruptible(&chip->vendor.int_queue);
617 /* Clear interrupts handled with TPM_EOI */
619 chip->vendor.iobase +
620 TPM_INT_STATUS(chip->vendor.locality));
621 ioread32(chip->vendor.iobase + TPM_INT_STATUS(chip->vendor.locality));
625 static bool interrupts = true;
626 module_param(interrupts, bool, 0444);
627 MODULE_PARM_DESC(interrupts, "Enable interrupts");
629 static void tpm_tis_remove(struct tpm_chip *chip)
631 if (chip->flags & TPM_CHIP_FLAG_TPM2)
632 tpm2_shutdown(chip, TPM2_SU_CLEAR);
634 iowrite32(~TPM_GLOBAL_INT_ENABLE &
635 ioread32(chip->vendor.iobase +
636 TPM_INT_ENABLE(chip->vendor.
638 chip->vendor.iobase +
639 TPM_INT_ENABLE(chip->vendor.locality));
640 release_locality(chip, chip->vendor.locality, 1);
643 static int tpm_tis_init(struct device *dev, struct tpm_info *tpm_info,
644 acpi_handle acpi_dev_handle)
646 u32 vendor, intfcaps, intmask;
647 int rc, i, irq_s, irq_e, probe;
648 struct tpm_chip *chip;
649 struct priv_data *priv;
651 priv = devm_kzalloc(dev, sizeof(struct priv_data), GFP_KERNEL);
655 chip = tpmm_chip_alloc(dev, &tpm_tis);
657 return PTR_ERR(chip);
659 chip->vendor.priv = priv;
661 chip->acpi_dev_handle = acpi_dev_handle;
664 chip->vendor.iobase = devm_ioremap(dev, tpm_info->start, tpm_info->len);
665 if (!chip->vendor.iobase)
668 /* Maximum timeouts */
669 chip->vendor.timeout_a = TIS_TIMEOUT_A_MAX;
670 chip->vendor.timeout_b = TIS_TIMEOUT_B_MAX;
671 chip->vendor.timeout_c = TIS_TIMEOUT_C_MAX;
672 chip->vendor.timeout_d = TIS_TIMEOUT_D_MAX;
674 if (wait_startup(chip, 0) != 0) {
679 if (request_locality(chip, 0) != 0) {
684 rc = tpm2_probe(chip);
688 vendor = ioread32(chip->vendor.iobase + TPM_DID_VID(0));
689 chip->vendor.manufacturer_id = vendor;
691 dev_info(dev, "%s TPM (device-id 0x%X, rev-id %d)\n",
692 (chip->flags & TPM_CHIP_FLAG_TPM2) ? "2.0" : "1.2",
693 vendor >> 16, ioread8(chip->vendor.iobase + TPM_RID(0)));
696 probe = probe_itpm(chip);
705 dev_info(dev, "Intel iTPM workaround enabled\n");
708 /* Figure out the capabilities */
710 ioread32(chip->vendor.iobase +
711 TPM_INTF_CAPS(chip->vendor.locality));
712 dev_dbg(dev, "TPM interface capabilities (0x%x):\n",
714 if (intfcaps & TPM_INTF_BURST_COUNT_STATIC)
715 dev_dbg(dev, "\tBurst Count Static\n");
716 if (intfcaps & TPM_INTF_CMD_READY_INT)
717 dev_dbg(dev, "\tCommand Ready Int Support\n");
718 if (intfcaps & TPM_INTF_INT_EDGE_FALLING)
719 dev_dbg(dev, "\tInterrupt Edge Falling\n");
720 if (intfcaps & TPM_INTF_INT_EDGE_RISING)
721 dev_dbg(dev, "\tInterrupt Edge Rising\n");
722 if (intfcaps & TPM_INTF_INT_LEVEL_LOW)
723 dev_dbg(dev, "\tInterrupt Level Low\n");
724 if (intfcaps & TPM_INTF_INT_LEVEL_HIGH)
725 dev_dbg(dev, "\tInterrupt Level High\n");
726 if (intfcaps & TPM_INTF_LOCALITY_CHANGE_INT)
727 dev_dbg(dev, "\tLocality Change Int Support\n");
728 if (intfcaps & TPM_INTF_STS_VALID_INT)
729 dev_dbg(dev, "\tSts Valid Int Support\n");
730 if (intfcaps & TPM_INTF_DATA_AVAIL_INT)
731 dev_dbg(dev, "\tData Avail Int Support\n");
733 /* INTERRUPT Setup */
734 init_waitqueue_head(&chip->vendor.read_queue);
735 init_waitqueue_head(&chip->vendor.int_queue);
738 ioread32(chip->vendor.iobase +
739 TPM_INT_ENABLE(chip->vendor.locality));
741 intmask |= TPM_INTF_CMD_READY_INT
742 | TPM_INTF_LOCALITY_CHANGE_INT | TPM_INTF_DATA_AVAIL_INT
743 | TPM_INTF_STS_VALID_INT;
746 chip->vendor.iobase +
747 TPM_INT_ENABLE(chip->vendor.locality));
749 chip->vendor.irq = tpm_info->irq;
750 if (interrupts && !chip->vendor.irq) {
752 ioread8(chip->vendor.iobase +
753 TPM_INT_VECTOR(chip->vendor.locality));
761 for (i = irq_s; i <= irq_e && chip->vendor.irq == 0; i++) {
762 iowrite8(i, chip->vendor.iobase +
763 TPM_INT_VECTOR(chip->vendor.locality));
765 (dev, i, tis_int_probe, IRQF_SHARED,
766 chip->devname, chip) != 0) {
768 "Unable to request irq: %d for probe\n",
773 /* Clear all existing */
775 (chip->vendor.iobase +
776 TPM_INT_STATUS(chip->vendor.locality)),
777 chip->vendor.iobase +
778 TPM_INT_STATUS(chip->vendor.locality));
781 iowrite32(intmask | TPM_GLOBAL_INT_ENABLE,
782 chip->vendor.iobase +
783 TPM_INT_ENABLE(chip->vendor.locality));
785 chip->vendor.probed_irq = 0;
787 /* Generate Interrupts */
788 if (chip->flags & TPM_CHIP_FLAG_TPM2)
789 tpm2_gen_interrupt(chip);
791 tpm_gen_interrupt(chip);
793 chip->vendor.irq = chip->vendor.probed_irq;
795 /* free_irq will call into tis_int_probe;
796 clear all irqs we haven't seen while doing
799 (chip->vendor.iobase +
800 TPM_INT_STATUS(chip->vendor.locality)),
801 chip->vendor.iobase +
802 TPM_INT_STATUS(chip->vendor.locality));
806 chip->vendor.iobase +
807 TPM_INT_ENABLE(chip->vendor.locality));
810 if (chip->vendor.irq) {
811 iowrite8(chip->vendor.irq,
812 chip->vendor.iobase +
813 TPM_INT_VECTOR(chip->vendor.locality));
815 (dev, chip->vendor.irq, tis_int_handler, IRQF_SHARED,
816 chip->devname, chip) != 0) {
818 "Unable to request irq: %d for use\n",
820 chip->vendor.irq = 0;
822 /* Clear all existing */
824 (chip->vendor.iobase +
825 TPM_INT_STATUS(chip->vendor.locality)),
826 chip->vendor.iobase +
827 TPM_INT_STATUS(chip->vendor.locality));
830 iowrite32(intmask | TPM_GLOBAL_INT_ENABLE,
831 chip->vendor.iobase +
832 TPM_INT_ENABLE(chip->vendor.locality));
836 if (chip->flags & TPM_CHIP_FLAG_TPM2) {
837 chip->vendor.timeout_a = msecs_to_jiffies(TPM2_TIMEOUT_A);
838 chip->vendor.timeout_b = msecs_to_jiffies(TPM2_TIMEOUT_B);
839 chip->vendor.timeout_c = msecs_to_jiffies(TPM2_TIMEOUT_C);
840 chip->vendor.timeout_d = msecs_to_jiffies(TPM2_TIMEOUT_D);
841 chip->vendor.duration[TPM_SHORT] =
842 msecs_to_jiffies(TPM2_DURATION_SHORT);
843 chip->vendor.duration[TPM_MEDIUM] =
844 msecs_to_jiffies(TPM2_DURATION_MEDIUM);
845 chip->vendor.duration[TPM_LONG] =
846 msecs_to_jiffies(TPM2_DURATION_LONG);
848 rc = tpm2_do_selftest(chip);
849 if (rc == TPM2_RC_INITIALIZE) {
850 dev_warn(dev, "Firmware has not started TPM\n");
851 rc = tpm2_startup(chip, TPM2_SU_CLEAR);
853 rc = tpm2_do_selftest(chip);
857 dev_err(dev, "TPM self test failed\n");
863 if (tpm_get_timeouts(chip)) {
864 dev_err(dev, "Could not get TPM timeouts and durations\n");
869 if (tpm_do_selftest(chip)) {
870 dev_err(dev, "TPM self test failed\n");
876 return tpm_chip_register(chip);
878 tpm_tis_remove(chip);
882 #ifdef CONFIG_PM_SLEEP
883 static void tpm_tis_reenable_interrupts(struct tpm_chip *chip)
887 /* reenable interrupts that device may have lost or
888 BIOS/firmware may have disabled */
889 iowrite8(chip->vendor.irq, chip->vendor.iobase +
890 TPM_INT_VECTOR(chip->vendor.locality));
893 ioread32(chip->vendor.iobase +
894 TPM_INT_ENABLE(chip->vendor.locality));
896 intmask |= TPM_INTF_CMD_READY_INT
897 | TPM_INTF_LOCALITY_CHANGE_INT | TPM_INTF_DATA_AVAIL_INT
898 | TPM_INTF_STS_VALID_INT | TPM_GLOBAL_INT_ENABLE;
901 chip->vendor.iobase + TPM_INT_ENABLE(chip->vendor.locality));
904 static int tpm_tis_resume(struct device *dev)
906 struct tpm_chip *chip = dev_get_drvdata(dev);
909 if (chip->vendor.irq)
910 tpm_tis_reenable_interrupts(chip);
912 ret = tpm_pm_resume(dev);
916 /* TPM 1.2 requires self-test on resume. This function actually returns
917 * an error code but for unknown reason it isn't handled.
919 if (!(chip->flags & TPM_CHIP_FLAG_TPM2))
920 tpm_do_selftest(chip);
926 static SIMPLE_DEV_PM_OPS(tpm_tis_pm, tpm_pm_suspend, tpm_tis_resume);
929 static int tpm_tis_pnp_init(struct pnp_dev *pnp_dev,
930 const struct pnp_device_id *pnp_id)
932 struct tpm_info tpm_info = tis_default_info;
933 acpi_handle acpi_dev_handle = NULL;
935 tpm_info.start = pnp_mem_start(pnp_dev, 0);
936 tpm_info.len = pnp_mem_len(pnp_dev, 0);
938 if (pnp_irq_valid(pnp_dev, 0))
939 tpm_info.irq = pnp_irq(pnp_dev, 0);
944 if (pnp_acpi_device(pnp_dev)) {
945 if (is_itpm(pnp_acpi_device(pnp_dev)))
948 acpi_dev_handle = pnp_acpi_device(pnp_dev)->handle;
952 return tpm_tis_init(&pnp_dev->dev, &tpm_info, acpi_dev_handle);
955 static struct pnp_device_id tpm_pnp_tbl[] = {
956 {"PNP0C31", 0}, /* TPM */
957 {"ATM1200", 0}, /* Atmel */
958 {"IFX0102", 0}, /* Infineon */
959 {"BCM0101", 0}, /* Broadcom */
960 {"BCM0102", 0}, /* Broadcom */
961 {"NSC1200", 0}, /* National */
962 {"ICO0102", 0}, /* Intel */
964 {"", 0}, /* User Specified */
965 {"", 0} /* Terminator */
967 MODULE_DEVICE_TABLE(pnp, tpm_pnp_tbl);
969 static void tpm_tis_pnp_remove(struct pnp_dev *dev)
971 struct tpm_chip *chip = pnp_get_drvdata(dev);
973 tpm_chip_unregister(chip);
974 tpm_tis_remove(chip);
977 static struct pnp_driver tis_pnp_driver = {
979 .id_table = tpm_pnp_tbl,
980 .probe = tpm_tis_pnp_init,
981 .remove = tpm_tis_pnp_remove,
987 #define TIS_HID_USR_IDX sizeof(tpm_pnp_tbl)/sizeof(struct pnp_device_id) -2
988 module_param_string(hid, tpm_pnp_tbl[TIS_HID_USR_IDX].id,
989 sizeof(tpm_pnp_tbl[TIS_HID_USR_IDX].id), 0444);
990 MODULE_PARM_DESC(hid, "Set additional specific HID for this driver to probe");
994 static int tpm_check_resource(struct acpi_resource *ares, void *data)
996 struct tpm_info *tpm_info = (struct tpm_info *) data;
999 if (acpi_dev_resource_interrupt(ares, 0, &res)) {
1000 tpm_info->irq = res.start;
1001 } else if (acpi_dev_resource_memory(ares, &res)) {
1002 tpm_info->start = res.start;
1003 tpm_info->len = resource_size(&res);
1009 static int tpm_tis_acpi_init(struct acpi_device *acpi_dev)
1011 struct list_head resources;
1012 struct tpm_info tpm_info = tis_default_info;
1015 if (!is_fifo(acpi_dev))
1018 INIT_LIST_HEAD(&resources);
1019 ret = acpi_dev_get_resources(acpi_dev, &resources, tpm_check_resource,
1024 acpi_dev_free_resource_list(&resources);
1029 if (is_itpm(acpi_dev))
1032 return tpm_tis_init(&acpi_dev->dev, &tpm_info, acpi_dev->handle);
1035 static int tpm_tis_acpi_remove(struct acpi_device *dev)
1037 struct tpm_chip *chip = dev_get_drvdata(&dev->dev);
1039 tpm_chip_unregister(chip);
1040 tpm_tis_remove(chip);
1045 static struct acpi_device_id tpm_acpi_tbl[] = {
1046 {"MSFT0101", 0}, /* TPM 2.0 */
1048 {"", 0}, /* User Specified */
1049 {"", 0} /* Terminator */
1051 MODULE_DEVICE_TABLE(acpi, tpm_acpi_tbl);
1053 static struct acpi_driver tis_acpi_driver = {
1055 .ids = tpm_acpi_tbl,
1057 .add = tpm_tis_acpi_init,
1058 .remove = tpm_tis_acpi_remove,
1066 static struct platform_driver tis_drv = {
1073 static struct platform_device *pdev;
1076 module_param(force, bool, 0444);
1077 MODULE_PARM_DESC(force, "Force device probe rather than using ACPI entry");
1078 static int __init init_tis(void)
1083 rc = pnp_register_driver(&tis_pnp_driver);
1090 rc = acpi_bus_register_driver(&tis_acpi_driver);
1093 pnp_unregister_driver(&tis_pnp_driver);
1102 rc = platform_driver_register(&tis_drv);
1105 pdev = platform_device_register_simple("tpm_tis", -1, NULL, 0);
1110 rc = tpm_tis_init(&pdev->dev, &tis_default_info, NULL);
1115 platform_device_unregister(pdev);
1117 platform_driver_unregister(&tis_drv);
1121 static void __exit cleanup_tis(void)
1123 struct tpm_chip *chip;
1124 #if defined(CONFIG_PNP) || defined(CONFIG_ACPI)
1127 acpi_bus_unregister_driver(&tis_acpi_driver);
1130 pnp_unregister_driver(&tis_pnp_driver);
1135 chip = dev_get_drvdata(&pdev->dev);
1136 tpm_chip_unregister(chip);
1137 tpm_tis_remove(chip);
1138 platform_device_unregister(pdev);
1139 platform_driver_unregister(&tis_drv);
1142 module_init(init_tis);
1143 module_exit(cleanup_tis);
1144 MODULE_AUTHOR("Leendert van Doorn (leendert@watson.ibm.com)");
1145 MODULE_DESCRIPTION("TPM Driver");
1146 MODULE_VERSION("2.0");
1147 MODULE_LICENSE("GPL");