/* * Moorestown platform Langwell chip GPIO driver * * Copyright (c) 2008 - 2009, Intel Corporation. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. */ /* Supports: * Moorestown platform Langwell chip. * Medfield platform Penwell chip. * Whitney point. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* * Langwell chip has 64 pins and thus there are 2 32bit registers to control * each feature, while Penwell chip has 96 pins for each block, and need 3 32bit * registers to control them, so we only define the order here instead of a * structure, to get a bit offset for a pin (use GPDR as an example): * * nreg = ngpio / 32; * reg = offset / 32; * bit = offset % 32; * reg_addr = reg_base + GPDR * nreg * 4 + reg * 4; * * so the bit of reg_addr is to control pin offset's GPDR feature */ enum GPIO_REG { GPLR = 0, /* pin level read-only */ GPDR, /* pin direction */ GPSR, /* pin set */ GPCR, /* pin clear */ GRER, /* rising edge detect */ GFER, /* falling edge detect */ GEDR, /* edge detect result */ GAFR, /* alt function */ }; struct lnw_gpio { struct gpio_chip chip; void *reg_base; spinlock_t lock; struct pci_dev *pdev; struct irq_domain *domain; }; static void __iomem *gpio_reg(struct gpio_chip *chip, unsigned offset, enum GPIO_REG reg_type) { struct lnw_gpio *lnw = container_of(chip, struct lnw_gpio, chip); unsigned nreg = chip->ngpio / 32; u8 reg = offset / 32; void __iomem *ptr; ptr = (void __iomem *)(lnw->reg_base + reg_type * nreg * 4 + reg * 4); return ptr; } static void __iomem *gpio_reg_2bit(struct gpio_chip *chip, unsigned offset, enum GPIO_REG reg_type) { struct lnw_gpio *lnw = container_of(chip, struct lnw_gpio, chip); unsigned nreg = chip->ngpio / 32; u8 reg = offset / 16; void __iomem *ptr; ptr = (void __iomem *)(lnw->reg_base + reg_type * nreg * 4 + reg * 4); return ptr; } static int lnw_gpio_request(struct gpio_chip *chip, unsigned offset) { void __iomem *gafr = gpio_reg_2bit(chip, offset, GAFR); u32 value = readl(gafr); int shift = (offset % 16) << 1, af = (value >> shift) & 3; if (af) { value &= ~(3 << shift); writel(value, gafr); } return 0; } static int lnw_gpio_get(struct gpio_chip *chip, unsigned offset) { void __iomem *gplr = gpio_reg(chip, offset, GPLR); return readl(gplr) & BIT(offset % 32); } static void lnw_gpio_set(struct gpio_chip *chip, unsigned offset, int value) { void __iomem *gpsr, *gpcr; if (value) { gpsr = gpio_reg(chip, offset, GPSR); writel(BIT(offset % 32), gpsr); } else { gpcr = gpio_reg(chip, offset, GPCR); writel(BIT(offset % 32), gpcr); } } static int lnw_gpio_direction_input(struct gpio_chip *chip, unsigned offset) { struct lnw_gpio *lnw = container_of(chip, struct lnw_gpio, chip); void __iomem *gpdr = gpio_reg(chip, offset, GPDR); u32 value; unsigned long flags; if (lnw->pdev) pm_runtime_get(&lnw->pdev->dev); spin_lock_irqsave(&lnw->lock, flags); value = readl(gpdr); value &= ~BIT(offset % 32); writel(value, gpdr); spin_unlock_irqrestore(&lnw->lock, flags); if (lnw->pdev) pm_runtime_put(&lnw->pdev->dev); return 0; } static int lnw_gpio_direction_output(struct gpio_chip *chip, unsigned offset, int value) { struct lnw_gpio *lnw = container_of(chip, struct lnw_gpio, chip); void __iomem *gpdr = gpio_reg(chip, offset, GPDR); unsigned long flags; lnw_gpio_set(chip, offset, value); if (lnw->pdev) pm_runtime_get(&lnw->pdev->dev); spin_lock_irqsave(&lnw->lock, flags); value = readl(gpdr); value |= BIT(offset % 32); writel(value, gpdr); spin_unlock_irqrestore(&lnw->lock, flags); if (lnw->pdev) pm_runtime_put(&lnw->pdev->dev); return 0; } static int lnw_gpio_to_irq(struct gpio_chip *chip, unsigned offset) { struct lnw_gpio *lnw = container_of(chip, struct lnw_gpio, chip); return irq_create_mapping(lnw->domain, offset); } static int lnw_irq_type(struct irq_data *d, unsigned type) { struct lnw_gpio *lnw = irq_data_get_irq_chip_data(d); u32 gpio = irqd_to_hwirq(d); unsigned long flags; u32 value; void __iomem *grer = gpio_reg(&lnw->chip, gpio, GRER); void __iomem *gfer = gpio_reg(&lnw->chip, gpio, GFER); if (gpio >= lnw->chip.ngpio) return -EINVAL; if (lnw->pdev) pm_runtime_get(&lnw->pdev->dev); spin_lock_irqsave(&lnw->lock, flags); if (type & IRQ_TYPE_EDGE_RISING) value = readl(grer) | BIT(gpio % 32); else value = readl(grer) & (~BIT(gpio % 32)); writel(value, grer); if (type & IRQ_TYPE_EDGE_FALLING) value = readl(gfer) | BIT(gpio % 32); else value = readl(gfer) & (~BIT(gpio % 32)); writel(value, gfer); spin_unlock_irqrestore(&lnw->lock, flags); if (lnw->pdev) pm_runtime_put(&lnw->pdev->dev); return 0; } static void lnw_irq_unmask(struct irq_data *d) { } static void lnw_irq_mask(struct irq_data *d) { } static struct irq_chip lnw_irqchip = { .name = "LNW-GPIO", .irq_mask = lnw_irq_mask, .irq_unmask = lnw_irq_unmask, .irq_set_type = lnw_irq_type, }; static DEFINE_PCI_DEVICE_TABLE(lnw_gpio_ids) = { /* pin number */ { PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x080f), .driver_data = 64 }, { PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x081f), .driver_data = 96 }, { PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x081a), .driver_data = 96 }, { 0, } }; MODULE_DEVICE_TABLE(pci, lnw_gpio_ids); static void lnw_irq_handler(unsigned irq, struct irq_desc *desc) { struct irq_data *data = irq_desc_get_irq_data(desc); struct lnw_gpio *lnw = irq_data_get_irq_handler_data(data); struct irq_chip *chip = irq_data_get_irq_chip(data); u32 base, gpio, mask; unsigned long pending; void __iomem *gedr; /* check GPIO controller to check which pin triggered the interrupt */ for (base = 0; base < lnw->chip.ngpio; base += 32) { gedr = gpio_reg(&lnw->chip, base, GEDR); while ((pending = readl(gedr))) { gpio = __ffs(pending); mask = BIT(gpio); /* Clear before handling so we can't lose an edge */ writel(mask, gedr); generic_handle_irq(irq_find_mapping(lnw->domain, base + gpio)); } } chip->irq_eoi(data); } static void lnw_irq_init_hw(struct lnw_gpio *lnw) { void __iomem *reg; unsigned base; for (base = 0; base < lnw->chip.ngpio; base += 32) { /* Clear the rising-edge detect register */ reg = gpio_reg(&lnw->chip, base, GRER); writel(0, reg); /* Clear the falling-edge detect register */ reg = gpio_reg(&lnw->chip, base, GFER); writel(0, reg); /* Clear the edge detect status register */ reg = gpio_reg(&lnw->chip, base, GEDR); writel(~0, reg); } } static int lnw_gpio_irq_map(struct irq_domain *d, unsigned int virq, irq_hw_number_t hw) { struct lnw_gpio *lnw = d->host_data; irq_set_chip_and_handler_name(virq, &lnw_irqchip, handle_simple_irq, "demux"); irq_set_chip_data(virq, lnw); irq_set_irq_type(virq, IRQ_TYPE_NONE); return 0; } static const struct irq_domain_ops lnw_gpio_irq_ops = { .map = lnw_gpio_irq_map, .xlate = irq_domain_xlate_twocell, }; #ifdef CONFIG_PM static int lnw_gpio_runtime_resume(struct device *dev) { return 0; } static int lnw_gpio_runtime_suspend(struct device *dev) { return 0; } static int lnw_gpio_runtime_idle(struct device *dev) { int err = pm_schedule_suspend(dev, 500); if (!err) return 0; return -EBUSY; } #else #define lnw_gpio_runtime_suspend NULL #define lnw_gpio_runtime_resume NULL #define lnw_gpio_runtime_idle NULL #endif static const struct dev_pm_ops lnw_gpio_pm_ops = { .runtime_suspend = lnw_gpio_runtime_suspend, .runtime_resume = lnw_gpio_runtime_resume, .runtime_idle = lnw_gpio_runtime_idle, }; static int __devinit lnw_gpio_probe(struct pci_dev *pdev, const struct pci_device_id *id) { void *base; resource_size_t start, len; struct lnw_gpio *lnw; u32 gpio_base; int retval; int ngpio = id->driver_data; retval = pci_enable_device(pdev); if (retval) return retval; retval = pci_request_regions(pdev, "langwell_gpio"); if (retval) { dev_err(&pdev->dev, "error requesting resources\n"); goto err2; } /* get the gpio_base from bar1 */ start = pci_resource_start(pdev, 1); len = pci_resource_len(pdev, 1); base = ioremap_nocache(start, len); if (!base) { dev_err(&pdev->dev, "error mapping bar1\n"); retval = -EFAULT; goto err3; } gpio_base = *((u32 *)base + 1); /* release the IO mapping, since we already get the info from bar1 */ iounmap(base); /* get the register base from bar0 */ start = pci_resource_start(pdev, 0); len = pci_resource_len(pdev, 0); base = devm_ioremap_nocache(&pdev->dev, start, len); if (!base) { dev_err(&pdev->dev, "error mapping bar0\n"); retval = -EFAULT; goto err3; } lnw = devm_kzalloc(&pdev->dev, sizeof(struct lnw_gpio), GFP_KERNEL); if (!lnw) { dev_err(&pdev->dev, "can't allocate langwell_gpio chip data\n"); retval = -ENOMEM; goto err3; } lnw->domain = irq_domain_add_linear(pdev->dev.of_node, ngpio, &lnw_gpio_irq_ops, lnw); if (!lnw->domain) { retval = -ENOMEM; goto err3; } lnw->reg_base = base; lnw->chip.label = dev_name(&pdev->dev); lnw->chip.request = lnw_gpio_request; lnw->chip.direction_input = lnw_gpio_direction_input; lnw->chip.direction_output = lnw_gpio_direction_output; lnw->chip.get = lnw_gpio_get; lnw->chip.set = lnw_gpio_set; lnw->chip.to_irq = lnw_gpio_to_irq; lnw->chip.base = gpio_base; lnw->chip.ngpio = ngpio; lnw->chip.can_sleep = 0; lnw->pdev = pdev; pci_set_drvdata(pdev, lnw); retval = gpiochip_add(&lnw->chip); if (retval) { dev_err(&pdev->dev, "langwell gpiochip_add error %d\n", retval); goto err3; } lnw_irq_init_hw(lnw); irq_set_handler_data(pdev->irq, lnw); irq_set_chained_handler(pdev->irq, lnw_irq_handler); spin_lock_init(&lnw->lock); pm_runtime_put_noidle(&pdev->dev); pm_runtime_allow(&pdev->dev); return 0; err3: pci_release_regions(pdev); err2: pci_disable_device(pdev); return retval; } static struct pci_driver lnw_gpio_driver = { .name = "langwell_gpio", .id_table = lnw_gpio_ids, .probe = lnw_gpio_probe, .driver = { .pm = &lnw_gpio_pm_ops, }, }; static int __devinit wp_gpio_probe(struct platform_device *pdev) { struct lnw_gpio *lnw; struct gpio_chip *gc; struct resource *rc; int retval = 0; rc = platform_get_resource(pdev, IORESOURCE_MEM, 0); if (!rc) return -EINVAL; lnw = kzalloc(sizeof(struct lnw_gpio), GFP_KERNEL); if (!lnw) { dev_err(&pdev->dev, "can't allocate whitneypoint_gpio chip data\n"); return -ENOMEM; } lnw->reg_base = ioremap_nocache(rc->start, resource_size(rc)); if (lnw->reg_base == NULL) { retval = -EINVAL; goto err_kmalloc; } spin_lock_init(&lnw->lock); gc = &lnw->chip; gc->label = dev_name(&pdev->dev); gc->owner = THIS_MODULE; gc->direction_input = lnw_gpio_direction_input; gc->direction_output = lnw_gpio_direction_output; gc->get = lnw_gpio_get; gc->set = lnw_gpio_set; gc->to_irq = NULL; gc->base = 0; gc->ngpio = 64; gc->can_sleep = 0; retval = gpiochip_add(gc); if (retval) { dev_err(&pdev->dev, "whitneypoint gpiochip_add error %d\n", retval); goto err_ioremap; } platform_set_drvdata(pdev, lnw); return 0; err_ioremap: iounmap(lnw->reg_base); err_kmalloc: kfree(lnw); return retval; } static int __devexit wp_gpio_remove(struct platform_device *pdev) { struct lnw_gpio *lnw = platform_get_drvdata(pdev); int err; err = gpiochip_remove(&lnw->chip); if (err) dev_err(&pdev->dev, "failed to remove gpio_chip.\n"); iounmap(lnw->reg_base); kfree(lnw); platform_set_drvdata(pdev, NULL); return 0; } static struct platform_driver wp_gpio_driver = { .probe = wp_gpio_probe, .remove = __devexit_p(wp_gpio_remove), .driver = { .name = "wp_gpio", .owner = THIS_MODULE, }, }; static int __init lnw_gpio_init(void) { int ret; ret = pci_register_driver(&lnw_gpio_driver); if (ret < 0) return ret; ret = platform_driver_register(&wp_gpio_driver); if (ret < 0) pci_unregister_driver(&lnw_gpio_driver); return ret; } device_initcall(lnw_gpio_init);