diff options
Diffstat (limited to 'drivers/crypto/ccp/ccp-pci.c')
| -rw-r--r-- | drivers/crypto/ccp/ccp-pci.c | 360 | 
1 files changed, 360 insertions, 0 deletions
diff --git a/drivers/crypto/ccp/ccp-pci.c b/drivers/crypto/ccp/ccp-pci.c new file mode 100644 index 00000000000..0d746236df5 --- /dev/null +++ b/drivers/crypto/ccp/ccp-pci.c @@ -0,0 +1,360 @@ +/* + * AMD Cryptographic Coprocessor (CCP) driver + * + * Copyright (C) 2013 Advanced Micro Devices, Inc. + * + * Author: Tom Lendacky <thomas.lendacky@amd.com> + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License version 2 as + * published by the Free Software Foundation. + */ + +#include <linux/module.h> +#include <linux/kernel.h> +#include <linux/pci.h> +#include <linux/pci_ids.h> +#include <linux/kthread.h> +#include <linux/sched.h> +#include <linux/interrupt.h> +#include <linux/spinlock.h> +#include <linux/delay.h> +#include <linux/ccp.h> + +#include "ccp-dev.h" + +#define IO_BAR				2 +#define MSIX_VECTORS			2 + +struct ccp_msix { +	u32 vector; +	char name[16]; +}; + +struct ccp_pci { +	int msix_count; +	struct ccp_msix msix[MSIX_VECTORS]; +}; + +static int ccp_get_msix_irqs(struct ccp_device *ccp) +{ +	struct ccp_pci *ccp_pci = ccp->dev_specific; +	struct device *dev = ccp->dev; +	struct pci_dev *pdev = container_of(dev, struct pci_dev, dev); +	struct msix_entry msix_entry[MSIX_VECTORS]; +	unsigned int name_len = sizeof(ccp_pci->msix[0].name) - 1; +	int v, ret; + +	for (v = 0; v < ARRAY_SIZE(msix_entry); v++) +		msix_entry[v].entry = v; + +	ret = pci_enable_msix_range(pdev, msix_entry, 1, v); +	if (ret < 0) +		return ret; + +	ccp_pci->msix_count = ret; +	for (v = 0; v < ccp_pci->msix_count; v++) { +		/* Set the interrupt names and request the irqs */ +		snprintf(ccp_pci->msix[v].name, name_len, "ccp-%u", v); +		ccp_pci->msix[v].vector = msix_entry[v].vector; +		ret = request_irq(ccp_pci->msix[v].vector, ccp_irq_handler, +				  0, ccp_pci->msix[v].name, dev); +		if (ret) { +			dev_notice(dev, "unable to allocate MSI-X IRQ (%d)\n", +				   ret); +			goto e_irq; +		} +	} + +	return 0; + +e_irq: +	while (v--) +		free_irq(ccp_pci->msix[v].vector, dev); + +	pci_disable_msix(pdev); + +	ccp_pci->msix_count = 0; + +	return ret; +} + +static int ccp_get_msi_irq(struct ccp_device *ccp) +{ +	struct device *dev = ccp->dev; +	struct pci_dev *pdev = container_of(dev, struct pci_dev, dev); +	int ret; + +	ret = pci_enable_msi(pdev); +	if (ret) +		return ret; + +	ret = request_irq(pdev->irq, ccp_irq_handler, 0, "ccp", dev); +	if (ret) { +		dev_notice(dev, "unable to allocate MSI IRQ (%d)\n", ret); +		goto e_msi; +	} + +	return 0; + +e_msi: +	pci_disable_msi(pdev); + +	return ret; +} + +static int ccp_get_irqs(struct ccp_device *ccp) +{ +	struct device *dev = ccp->dev; +	int ret; + +	ret = ccp_get_msix_irqs(ccp); +	if (!ret) +		return 0; + +	/* Couldn't get MSI-X vectors, try MSI */ +	dev_notice(dev, "could not enable MSI-X (%d), trying MSI\n", ret); +	ret = ccp_get_msi_irq(ccp); +	if (!ret) +		return 0; + +	/* Couldn't get MSI interrupt */ +	dev_notice(dev, "could not enable MSI (%d)\n", ret); + +	return ret; +} + +static void ccp_free_irqs(struct ccp_device *ccp) +{ +	struct ccp_pci *ccp_pci = ccp->dev_specific; +	struct device *dev = ccp->dev; +	struct pci_dev *pdev = container_of(dev, struct pci_dev, dev); + +	if (ccp_pci->msix_count) { +		while (ccp_pci->msix_count--) +			free_irq(ccp_pci->msix[ccp_pci->msix_count].vector, +				 dev); +		pci_disable_msix(pdev); +	} else { +		free_irq(pdev->irq, dev); +		pci_disable_msi(pdev); +	} +} + +static int ccp_find_mmio_area(struct ccp_device *ccp) +{ +	struct device *dev = ccp->dev; +	struct pci_dev *pdev = container_of(dev, struct pci_dev, dev); +	resource_size_t io_len; +	unsigned long io_flags; +	int bar; + +	io_flags = pci_resource_flags(pdev, IO_BAR); +	io_len = pci_resource_len(pdev, IO_BAR); +	if ((io_flags & IORESOURCE_MEM) && (io_len >= (IO_OFFSET + 0x800))) +		return IO_BAR; + +	for (bar = 0; bar < PCI_STD_RESOURCE_END; bar++) { +		io_flags = pci_resource_flags(pdev, bar); +		io_len = pci_resource_len(pdev, bar); +		if ((io_flags & IORESOURCE_MEM) && +		    (io_len >= (IO_OFFSET + 0x800))) +			return bar; +	} + +	return -EIO; +} + +static int ccp_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) +{ +	struct ccp_device *ccp; +	struct ccp_pci *ccp_pci; +	struct device *dev = &pdev->dev; +	unsigned int bar; +	int ret; + +	ret = -ENOMEM; +	ccp = ccp_alloc_struct(dev); +	if (!ccp) +		goto e_err; + +	ccp_pci = kzalloc(sizeof(*ccp_pci), GFP_KERNEL); +	if (!ccp_pci) { +		ret = -ENOMEM; +		goto e_free1; +	} +	ccp->dev_specific = ccp_pci; +	ccp->get_irq = ccp_get_irqs; +	ccp->free_irq = ccp_free_irqs; + +	ret = pci_request_regions(pdev, "ccp"); +	if (ret) { +		dev_err(dev, "pci_request_regions failed (%d)\n", ret); +		goto e_free2; +	} + +	ret = pci_enable_device(pdev); +	if (ret) { +		dev_err(dev, "pci_enable_device failed (%d)\n", ret); +		goto e_regions; +	} + +	pci_set_master(pdev); + +	ret = ccp_find_mmio_area(ccp); +	if (ret < 0) +		goto e_device; +	bar = ret; + +	ret = -EIO; +	ccp->io_map = pci_iomap(pdev, bar, 0); +	if (ccp->io_map == NULL) { +		dev_err(dev, "pci_iomap failed\n"); +		goto e_device; +	} +	ccp->io_regs = ccp->io_map + IO_OFFSET; + +	ret = dma_set_mask(dev, DMA_BIT_MASK(48)); +	if (ret == 0) { +		ret = dma_set_coherent_mask(dev, DMA_BIT_MASK(48)); +		if (ret) { +			dev_err(dev, +				"pci_set_consistent_dma_mask failed (%d)\n", +				ret); +			goto e_bar0; +		} +	} else { +		ret = dma_set_mask(dev, DMA_BIT_MASK(32)); +		if (ret) { +			dev_err(dev, "pci_set_dma_mask failed (%d)\n", ret); +			goto e_bar0; +		} +	} + +	dev_set_drvdata(dev, ccp); + +	ret = ccp_init(ccp); +	if (ret) +		goto e_bar0; + +	dev_notice(dev, "enabled\n"); + +	return 0; + +e_bar0: +	pci_iounmap(pdev, ccp->io_map); + +e_device: +	pci_disable_device(pdev); + +e_regions: +	pci_release_regions(pdev); + +e_free2: +	kfree(ccp_pci); + +e_free1: +	kfree(ccp); + +e_err: +	dev_notice(dev, "initialization failed\n"); +	return ret; +} + +static void ccp_pci_remove(struct pci_dev *pdev) +{ +	struct device *dev = &pdev->dev; +	struct ccp_device *ccp = dev_get_drvdata(dev); + +	if (!ccp) +		return; + +	ccp_destroy(ccp); + +	pci_iounmap(pdev, ccp->io_map); + +	pci_disable_device(pdev); + +	pci_release_regions(pdev); + +	kfree(ccp); + +	dev_notice(dev, "disabled\n"); +} + +#ifdef CONFIG_PM +static int ccp_pci_suspend(struct pci_dev *pdev, pm_message_t state) +{ +	struct device *dev = &pdev->dev; +	struct ccp_device *ccp = dev_get_drvdata(dev); +	unsigned long flags; +	unsigned int i; + +	spin_lock_irqsave(&ccp->cmd_lock, flags); + +	ccp->suspending = 1; + +	/* Wake all the queue kthreads to prepare for suspend */ +	for (i = 0; i < ccp->cmd_q_count; i++) +		wake_up_process(ccp->cmd_q[i].kthread); + +	spin_unlock_irqrestore(&ccp->cmd_lock, flags); + +	/* Wait for all queue kthreads to say they're done */ +	while (!ccp_queues_suspended(ccp)) +		wait_event_interruptible(ccp->suspend_queue, +					 ccp_queues_suspended(ccp)); + +	return 0; +} + +static int ccp_pci_resume(struct pci_dev *pdev) +{ +	struct device *dev = &pdev->dev; +	struct ccp_device *ccp = dev_get_drvdata(dev); +	unsigned long flags; +	unsigned int i; + +	spin_lock_irqsave(&ccp->cmd_lock, flags); + +	ccp->suspending = 0; + +	/* Wake up all the kthreads */ +	for (i = 0; i < ccp->cmd_q_count; i++) { +		ccp->cmd_q[i].suspended = 0; +		wake_up_process(ccp->cmd_q[i].kthread); +	} + +	spin_unlock_irqrestore(&ccp->cmd_lock, flags); + +	return 0; +} +#endif + +static DEFINE_PCI_DEVICE_TABLE(ccp_pci_table) = { +	{ PCI_VDEVICE(AMD, 0x1537), }, +	/* Last entry must be zero */ +	{ 0, } +}; +MODULE_DEVICE_TABLE(pci, ccp_pci_table); + +static struct pci_driver ccp_pci_driver = { +	.name = "AMD Cryptographic Coprocessor", +	.id_table = ccp_pci_table, +	.probe = ccp_pci_probe, +	.remove = ccp_pci_remove, +#ifdef CONFIG_PM +	.suspend = ccp_pci_suspend, +	.resume = ccp_pci_resume, +#endif +}; + +int ccp_pci_init(void) +{ +	return pci_register_driver(&ccp_pci_driver); +} + +void ccp_pci_exit(void) +{ +	pci_unregister_driver(&ccp_pci_driver); +}  | 
