/* * PCMCIA 16-bit resource management functions * * The initial developer of the original code is David A. Hinds * <dahinds@users.sourceforge.net>. Portions created by David A. Hinds * are Copyright (C) 1999 David A. Hinds. All Rights Reserved. * * Copyright (C) 1999 David A. Hinds * Copyright (C) 2004-2005 Dominik Brodowski * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. * */ #include <linux/module.h> #include <linux/kernel.h> #include <linux/interrupt.h> #include <linux/delay.h> #include <linux/pci.h> #include <linux/device.h> #include <linux/netdevice.h> #include <linux/slab.h> #include <asm/irq.h> #include <pcmcia/cs_types.h> #include <pcmcia/ss.h> #include <pcmcia/cs.h> #include <pcmcia/cistpl.h> #include <pcmcia/cisreg.h> #include <pcmcia/ds.h> #include "cs_internal.h" /* Access speed for IO windows */ static int io_speed; module_param(io_speed, int, 0444); int pcmcia_validate_mem(struct pcmcia_socket *s) { if (s->resource_ops->validate_mem) return s->resource_ops->validate_mem(s); /* if there is no callback, we can assume that everything is OK */ return 0; } struct resource *pcmcia_find_mem_region(u_long base, u_long num, u_long align, int low, struct pcmcia_socket *s) { if (s->resource_ops->find_mem) return s->resource_ops->find_mem(base, num, align, low, s); return NULL; } /** alloc_io_space * * Special stuff for managing IO windows, because they are scarce */ static int alloc_io_space(struct pcmcia_socket *s, u_int attr, unsigned int *base, unsigned int num, u_int lines) { unsigned int align; align = (*base) ? (lines ? 1<<lines : 0) : 1; if (align && (align < num)) { if (*base) { dev_dbg(&s->dev, "odd IO request: num %#x align %#x\n", num, align); align = 0; } else while (align && (align < num)) align <<= 1; } if (*base & ~(align-1)) { dev_dbg(&s->dev, "odd IO request: base %#x align %#x\n", *base, align); align = 0; } return s->resource_ops->find_io(s, attr, base, num, align); } /* alloc_io_space */ static void release_io_space(struct pcmcia_socket *s, unsigned int base, unsigned int num) { int i; for (i = 0; i < MAX_IO_WIN; i++) { if (!s->io[i].res) continue; if ((s->io[i].res->start <= base) && (s->io[i].res->end >= base+num-1)) { s->io[i].InUse -= num; /* Free the window if no one else is using it */ if (s->io[i].InUse == 0) { release_resource(s->io[i].res); kfree(s->io[i].res); s->io[i].res = NULL; } } } } /* release_io_space */ /** pccard_access_configuration_register * * Access_configuration_register() reads and writes configuration * registers in attribute memory. Memory window 0 is reserved for * this and the tuple reading services. */ int pcmcia_access_configuration_register(struct pcmcia_device *p_dev, conf_reg_t *reg) { struct pcmcia_socket *s; config_t *c; int addr; u_char val; int ret = 0; if (!p_dev || !p_dev->function_config) return -EINVAL; s = p_dev->socket; mutex_lock(&s->ops_mutex); c = p_dev->function_config; if (!(c->state & CONFIG_LOCKED)) { dev_dbg(&s->dev, "Configuration isnt't locked\n"); mutex_unlock(&s->ops_mutex); return -EACCES; } addr = (c->ConfigBase + reg->Offset) >> 1; switch (reg->Action) { case CS_READ: ret = pcmcia_read_cis_mem(s, 1, addr, 1, &val); reg->Value = val; break; case CS_WRITE: val = reg->Value; pcmcia_write_cis_mem(s, 1, addr, 1, &val); break; default: dev_dbg(&s->dev, "Invalid conf register request\n"); ret = -EINVAL; break; } mutex_unlock(&s->ops_mutex); return ret; } /* pcmcia_access_configuration_register */ EXPORT_SYMBOL(pcmcia_access_configuration_register); int pcmcia_map_mem_page(struct pcmcia_device *p_dev, window_handle_t wh, memreq_t *req) { struct pcmcia_socket *s = p_dev->socket; int ret; wh--; if (wh >= MAX_WIN) return -EINVAL; if (req->Page != 0) { dev_dbg(&s->dev, "failure: requested page is zero\n"); return -EINVAL; } mutex_lock(&s->ops_mutex); s->win[wh].card_start = req->CardOffset; ret = s->ops->set_mem_map(s, &s->win[wh]); if (ret) dev_warn(&s->dev, "failed to set_mem_map\n"); mutex_unlock(&s->ops_mutex); return ret; } /* pcmcia_map_mem_page */ EXPORT_SYMBOL(pcmcia_map_mem_page); /** pcmcia_modify_configuration * * Modify a locked socket configuration */ int pcmcia_modify_configuration(struct pcmcia_device *p_dev, modconf_t *mod) { struct pcmcia_socket *s; config_t *c; int ret; s = p_dev->socket; mutex_lock(&s->ops_mutex); c = p_dev->function_config; if (!(s->state & SOCKET_PRESENT)) { dev_dbg(&s->dev, "No card present\n"); ret = -ENODEV; goto unlock; } if (!(c->state & CONFIG_LOCKED)) { dev_dbg(&s->dev, "Configuration isnt't locked\n"); ret = -EACCES; goto unlock; } if (mod->Attributes & (CONF_IRQ_CHANGE_VALID | CONF_VCC_CHANGE_VALID)) { dev_dbg(&s->dev, "changing Vcc or IRQ is not allowed at this time\n"); ret = -EINVAL; goto unlock; } /* We only allow changing Vpp1 and Vpp2 to the same value */ if ((mod->Attributes & CONF_VPP1_CHANGE_VALID) && (mod->Attributes & CONF_VPP2_CHANGE_VALID)) { if (mod->Vpp1 != mod->Vpp2) { dev_dbg(&s->dev, "Vpp1 and Vpp2 must be the same\n"); ret = -EINVAL; goto unlock; } s->socket.Vpp = mod->Vpp1; if (s->ops->set_socket(s, &s->socket)) { dev_printk(KERN_WARNING, &s->dev, "Unable to set VPP\n"); ret = -EIO; goto unlock; } } else if ((mod->Attributes & CONF_VPP1_CHANGE_VALID) || (mod->Attributes & CONF_VPP2_CHANGE_VALID)) { dev_dbg(&s->dev, "changing Vcc is not allowed at this time\n"); ret = -EINVAL; goto unlock; } if (mod->Attributes & CONF_IO_CHANGE_WIDTH) { pccard_io_map io_off = { 0, 0, 0, 0, 1 }; pccard_io_map io_on; int i; io_on.speed = io_speed; for (i = 0; i < MAX_IO_WIN; i++) { if (!s->io[i].res) continue; io_off.map = i; io_on.map = i; io_on.flags = MAP_ACTIVE | IO_DATA_PATH_WIDTH_8; io_on.start = s->io[i].res->start; io_on.stop = s->io[i].res->end; s->ops->set_io_map(s, &io_off); mdelay(40); s->ops->set_io_map(s, &io_on); } } ret = 0; unlock: mutex_unlock(&s->ops_mutex); return ret; } /* modify_configuration */ EXPORT_SYMBOL(pcmcia_modify_configuration); int pcmcia_release_configuration(struct pcmcia_device *p_dev) { pccard_io_map io = { 0, 0, 0, 0, 1 }; struct pcmcia_socket *s = p_dev->socket; config_t *c; int i; mutex_lock(&s->ops_mutex); c = p_dev->function_config; if (p_dev->_locked) { p_dev->_locked = 0; if (--(s->lock_count) == 0) { s->socket.flags = SS_OUTPUT_ENA; /* Is this correct? */ s->socket.Vpp = 0; s->socket.io_irq = 0; s->ops->set_socket(s, &s->socket); } } if (c->state & CONFIG_LOCKED) { c->state &= ~CONFIG_LOCKED; if (c->state & CONFIG_IO_REQ) for (i = 0; i < MAX_IO_WIN; i++) { if (!s->io[i].res) continue; s->io[i].Config--; if (s->io[i].Config != 0) continue; io.map = i; s->ops->set_io_map(s, &io); } } mutex_unlock(&s->ops_mutex); return 0; } /* pcmcia_release_configuration */ /** pcmcia_release_io * * Release_io() releases the I/O ranges allocated by a client. This * may be invoked some time after a card ejection has already dumped * the actual socket configuration, so if the client is "stale", we * don't bother checking the port ranges against the current socket * values. */ static int pcmcia_release_io(struct pcmcia_device *p_dev, io_req_t *req) { struct pcmcia_socket *s = p_dev->socket; int ret = -EINVAL; config_t *c; mutex_lock(&s->ops_mutex); c = p_dev->function_config; if (!p_dev->_io) goto out; p_dev->_io = 0; if ((c->io.BasePort1 != req->BasePort1) || (c->io.NumPorts1 != req->NumPorts1) || (c->io.BasePort2 != req->BasePort2) || (c->io.NumPorts2 != req->NumPorts2)) goto out; c->state &= ~CONFIG_IO_REQ; release_io_space(s, req->BasePort1, req->NumPorts1); if (req->NumPorts2) release_io_space(s, req->BasePort2, req->NumPorts2); out: mutex_unlock(&s->ops_mutex); return ret; } /* pcmcia_release_io */ int pcmcia_release_window(struct pcmcia_device *p_dev, window_handle_t wh) { struct pcmcia_socket *s = p_dev->socket; pccard_mem_map *win; wh--; if (wh >= MAX_WIN) return -EINVAL; mutex_lock(&s->ops_mutex); win = &s->win[wh]; if (!(p_dev->_win & CLIENT_WIN_REQ(wh))) { dev_dbg(&s->dev, "not releasing unknown window\n"); mutex_unlock(&s->ops_mutex); return -EINVAL; } /* Shut down memory window */ win->flags &= ~MAP_ACTIVE; s->ops->set_mem_map(s, win); s->state &= ~SOCKET_WIN_REQ(wh); /* Release system memory */ if (win->res) { release_resource(win->res); kfree(win->res); win->res = NULL; } p_dev->_win &= ~CLIENT_WIN_REQ(wh); mutex_unlock(&s->ops_mutex); return 0; } /* pcmcia_release_window */ EXPORT_SYMBOL(pcmcia_release_window); int pcmcia_request_configuration(struct pcmcia_device *p_dev, config_req_t *req) { int i; u_int base; struct pcmcia_socket *s = p_dev->socket; config_t *c; pccard_io_map iomap; if (!(s->state & SOCKET_PRESENT)) return -ENODEV; if (req->IntType & INT_CARDBUS) { dev_dbg(&s->dev, "IntType may not be INT_CARDBUS\n"); return -EINVAL; } mutex_lock(&s->ops_mutex); c = p_dev->function_config; if (c->state & CONFIG_LOCKED) { mutex_unlock(&s->ops_mutex); dev_dbg(&s->dev, "Configuration is locked\n"); return -EACCES; } /* Do power control. We don't allow changes in Vcc. */ s->socket.Vpp = req->Vpp; if (s->ops->set_socket(s, &s->socket)) { mutex_unlock(&s->ops_mutex); dev_printk(KERN_WARNING, &s->dev, "Unable to set socket state\n"); return -EINVAL; } /* Pick memory or I/O card, DMA mode, interrupt */ c->IntType = req->IntType; c->Attributes = req->Attributes; if (req->IntType & INT_MEMORY_AND_IO) s->socket.flags |= SS_IOCARD; if (req->IntType & INT_ZOOMED_VIDEO) s->socket.flags |= SS_ZVCARD | SS_IOCARD; if (req->Attributes & CONF_ENABLE_DMA) s->socket.flags |= SS_DMA_MODE; if (req->Attributes & CONF_ENABLE_SPKR) s->socket.flags |= SS_SPKR_ENA; if (req->Attributes & CONF_ENABLE_IRQ) s->socket.io_irq = s->pcmcia_irq; else s->socket.io_irq = 0; s->ops->set_socket(s, &s->socket); s->lock_count++; /* Set up CIS configuration registers */ base = c->ConfigBase = req->ConfigBase; c->CardValues = req->Present; if (req->Present & PRESENT_COPY) { c->Copy = req->Copy; pcmcia_write_cis_mem(s, 1, (base + CISREG_SCR)>>1, 1, &c->Copy); } if (req->Present & PRESENT_OPTION) { if (s->functions == 1) { c->Option = req->ConfigIndex & COR_CONFIG_MASK; } else { c->Option = req->ConfigIndex & COR_MFC_CONFIG_MASK; c->Option |= COR_FUNC_ENA|COR_IREQ_ENA; if (req->Present & PRESENT_IOBASE_0) c->Option |= COR_ADDR_DECODE; } if ((req->Attributes & CONF_ENABLE_IRQ) && !(req->Attributes & CONF_ENABLE_PULSE_IRQ)) c->Option |= COR_LEVEL_REQ; pcmcia_write_cis_mem(s, 1, (base + CISREG_COR)>>1, 1, &c->Option); mdelay(40); } if (req->Present & PRESENT_STATUS) { c->Status = req->Status; pcmcia_write_cis_mem(s, 1, (base + CISREG_CCSR)>>1, 1, &c->Status); } if (req->Present & PRESENT_PIN_REPLACE) { c->Pin = req->Pin; pcmcia_write_cis_mem(s, 1, (base + CISREG_PRR)>>1, 1, &c->Pin); } if (req->Present & PRESENT_EXT_STATUS) { c->ExtStatus = req->ExtStatus; pcmcia_write_cis_mem(s, 1, (base + CISREG_ESR)>>1, 1, &c->ExtStatus); } if (req->Present & PRESENT_IOBASE_0) { u_char b = c->io.BasePort1 & 0xff; pcmcia_write_cis_mem(s, 1, (base + CISREG_IOBASE_0)>>1, 1, &b); b = (c->io.BasePort1 >> 8) & 0xff; pcmcia_write_cis_mem(s, 1, (base + CISREG_IOBASE_1)>>1, 1, &b); } if (req->Present & PRESENT_IOSIZE) { u_char b = c->io.NumPorts1 + c->io.NumPorts2 - 1; pcmcia_write_cis_mem(s, 1, (base + CISREG_IOSIZE)>>1, 1, &b); } /* Configure I/O windows */ if (c->state & CONFIG_IO_REQ) { iomap.speed = io_speed; for (i = 0; i < MAX_IO_WIN; i++) if (s->io[i].res) { iomap.map = i; iomap.flags = MAP_ACTIVE; switch (s->io[i].res->flags & IO_DATA_PATH_WIDTH) { case IO_DATA_PATH_WIDTH_16: iomap.flags |= MAP_16BIT; break; case IO_DATA_PATH_WIDTH_AUTO: iomap.flags |= MAP_AUTOSZ; break; default: break; } iomap.start = s->io[i].res->start; iomap.stop = s->io[i].res->end; s->ops->set_io_map(s, &iomap); s->io[i].Config++; } } c->state |= CONFIG_LOCKED; p_dev->_locked = 1; mutex_unlock(&s->ops_mutex); return 0; } /* pcmcia_request_configuration */ EXPORT_SYMBOL(pcmcia_request_configuration); /** pcmcia_request_io * * Request_io() reserves ranges of port addresses for a socket. * I have not implemented range sharing or alias addressing. */ int pcmcia_request_io(struct pcmcia_device *p_dev, io_req_t *req) { struct pcmcia_socket *s = p_dev->socket; config_t *c; int ret = -EINVAL; mutex_lock(&s->ops_mutex); if (!(s->state & SOCKET_PRESENT)) { dev_dbg(&s->dev, "No card present\n"); goto out; } if (!req) goto out; c = p_dev->function_config; if (c->state & CONFIG_LOCKED) { dev_dbg(&s->dev, "Configuration is locked\n"); goto out; } if (c->state & CONFIG_IO_REQ) { dev_dbg(&s->dev, "IO already configured\n"); goto out; } if (req->Attributes1 & (IO_SHARED | IO_FORCE_ALIAS_ACCESS)) { dev_dbg(&s->dev, "bad attribute setting for IO region 1\n"); goto out; } if ((req->NumPorts2 > 0) && (req->Attributes2 & (IO_SHARED | IO_FORCE_ALIAS_ACCESS))) { dev_dbg(&s->dev, "bad attribute setting for IO region 2\n"); goto out; } dev_dbg(&s->dev, "trying to allocate resource 1\n"); ret = alloc_io_space(s, req->Attributes1, &req->BasePort1, req->NumPorts1, req->IOAddrLines); if (ret) { dev_dbg(&s->dev, "allocation of resource 1 failed\n"); goto out; } if (req->NumPorts2) { dev_dbg(&s->dev, "trying to allocate resource 2\n"); ret = alloc_io_space(s, req->Attributes2, &req->BasePort2, req->NumPorts2, req->IOAddrLines); if (ret) { dev_dbg(&s->dev, "allocation of resource 2 failed\n"); release_io_space(s, req->BasePort1, req->NumPorts1); goto out; } } c->io = *req; c->state |= CONFIG_IO_REQ; p_dev->_io = 1; dev_dbg(&s->dev, "allocating resources succeeded: %d\n", ret); out: mutex_unlock(&s->ops_mutex); return ret; } /* pcmcia_request_io */ EXPORT_SYMBOL(pcmcia_request_io); /** * pcmcia_request_irq() - attempt to request a IRQ for a PCMCIA device * * pcmcia_request_irq() is a wrapper around request_irq which will allow * the PCMCIA core to clean up the registration in pcmcia_disable_device(). * Drivers are free to use request_irq() directly, but then they need to * call free_irq themselfves, too. Also, only IRQF_SHARED capable IRQ * handlers are allowed. */ int __must_check pcmcia_request_irq(struct pcmcia_device *p_dev, irq_handler_t handler) { int ret; if (!p_dev->irq) return -EINVAL; ret = request_irq(p_dev->irq, handler, IRQF_SHARED, p_dev->devname, p_dev->priv); if (!ret) p_dev->_irq = 1; return ret; } EXPORT_SYMBOL(pcmcia_request_irq); /** * pcmcia_request_exclusive_irq() - attempt to request an exclusive IRQ first * * pcmcia_request_exclusive_irq() is a wrapper around request_irq which * attempts first to request an exclusive IRQ. If it fails, it also accepts * a shared IRQ, but prints out a warning. PCMCIA drivers should allow for * IRQ sharing and either use request_irq directly (then they need to call * free_irq themselves, too), or the pcmcia_request_irq() function. */ int __must_check __pcmcia_request_exclusive_irq(struct pcmcia_device *p_dev, irq_handler_t handler) { int ret; if (!p_dev->irq) return -EINVAL; ret = request_irq(p_dev->irq, handler, 0, p_dev->devname, p_dev->priv); if (ret) { ret = pcmcia_request_irq(p_dev, handler); dev_printk(KERN_WARNING, &p_dev->dev, "pcmcia: " "request for exclusive IRQ could not be fulfilled.\n"); dev_printk(KERN_WARNING, &p_dev->dev, "pcmcia: the driver " "needs updating to supported shared IRQ lines.\n"); } if (ret) dev_printk(KERN_INFO, &p_dev->dev, "request_irq() failed\n"); else p_dev->_irq = 1; return ret; } /* pcmcia_request_exclusive_irq */ EXPORT_SYMBOL(__pcmcia_request_exclusive_irq); #ifdef CONFIG_PCMCIA_PROBE /* mask of IRQs already reserved by other cards, we should avoid using them */ static u8 pcmcia_used_irq[NR_IRQS]; static irqreturn_t test_action(int cpl, void *dev_id) { return IRQ_NONE; } /** * pcmcia_setup_isa_irq() - determine whether an ISA IRQ can be used * @p_dev - the associated PCMCIA device * * locking note: must be called with ops_mutex locked. */ static int pcmcia_setup_isa_irq(struct pcmcia_device *p_dev, int type) { struct pcmcia_socket *s = p_dev->socket; unsigned int try, irq; u32 mask = s->irq_mask; int ret = -ENODEV; for (try = 0; try < 64; try++) { irq = try % 32; /* marked as available by driver, not blocked by userspace? */ if (!((mask >> irq) & 1)) continue; /* avoid an IRQ which is already used by another PCMCIA card */ if ((try < 32) && pcmcia_used_irq[irq]) continue; /* register the correct driver, if possible, to check whether * registering a dummy handle works, i.e. if the IRQ isn't * marked as used by the kernel resource management core */ ret = request_irq(irq, test_action, type, p_dev->devname, p_dev); if (!ret) { free_irq(irq, p_dev); p_dev->irq = s->pcmcia_irq = irq; pcmcia_used_irq[irq]++; break; } } return ret; } void pcmcia_cleanup_irq(struct pcmcia_socket *s) { pcmcia_used_irq[s->pcmcia_irq]--; s->pcmcia_irq = 0; } #else /* CONFIG_PCMCIA_PROBE */ static int pcmcia_setup_isa_irq(struct pcmcia_device *p_dev, int type) { return -EINVAL; } void pcmcia_cleanup_irq(struct pcmcia_socket *s) { s->pcmcia_irq = 0; return; } #endif /* CONFIG_PCMCIA_PROBE */ /** * pcmcia_setup_irq() - determine IRQ to be used for device * @p_dev - the associated PCMCIA device * * locking note: must be called with ops_mutex locked. */ int pcmcia_setup_irq(struct pcmcia_device *p_dev) { struct pcmcia_socket *s = p_dev->socket; if (p_dev->irq) return 0; /* already assigned? */ if (s->pcmcia_irq) { p_dev->irq = s->pcmcia_irq; return 0; } /* prefer an exclusive ISA irq */ if (!pcmcia_setup_isa_irq(p_dev, 0)) return 0; /* but accept a shared ISA irq */ if (!pcmcia_setup_isa_irq(p_dev, IRQF_SHARED)) return 0; /* but use the PCI irq otherwise */ if (s->pci_irq) { p_dev->irq = s->pcmcia_irq = s->pci_irq; return 0; } return -EINVAL; } /** pcmcia_request_window * * Request_window() establishes a mapping between card memory space * and system memory space. */ int pcmcia_request_window(struct pcmcia_device *p_dev, win_req_t *req, window_handle_t *wh) { struct pcmcia_socket *s = p_dev->socket; pccard_mem_map *win; u_long align; int w; if (!(s->state & SOCKET_PRESENT)) { dev_dbg(&s->dev, "No card present\n"); return -ENODEV; } if (req->Attributes & (WIN_PAGED | WIN_SHARED)) { dev_dbg(&s->dev, "bad attribute setting for iomem region\n"); return -EINVAL; } /* Window size defaults to smallest available */ if (req->Size == 0) req->Size = s->map_size; align = (((s->features & SS_CAP_MEM_ALIGN) || (req->Attributes & WIN_STRICT_ALIGN)) ? req->Size : s->map_size); if (req->Size & (s->map_size-1)) { dev_dbg(&s->dev, "invalid map size\n"); return -EINVAL; } if ((req->Base && (s->features & SS_CAP_STATIC_MAP)) || (req->Base & (align-1))) { dev_dbg(&s->dev, "invalid base address\n"); return -EINVAL; } if (req->Base) align = 0; /* Allocate system memory window */ for (w = 0; w < MAX_WIN; w++) if (!(s->state & SOCKET_WIN_REQ(w))) break; if (w == MAX_WIN) { dev_dbg(&s->dev, "all windows are used already\n"); return -EINVAL; } mutex_lock(&s->ops_mutex); win = &s->win[w]; if (!(s->features & SS_CAP_STATIC_MAP)) { win->res = pcmcia_find_mem_region(req->Base, req->Size, align, (req->Attributes & WIN_MAP_BELOW_1MB), s); if (!win->res) { dev_dbg(&s->dev, "allocating mem region failed\n"); mutex_unlock(&s->ops_mutex); return -EINVAL; } } p_dev->_win |= CLIENT_WIN_REQ(w); /* Configure the socket controller */ win->map = w+1; win->flags = 0; win->speed = req->AccessSpeed; if (req->Attributes & WIN_MEMORY_TYPE) win->flags |= MAP_ATTRIB; if (req->Attributes & WIN_ENABLE) win->flags |= MAP_ACTIVE; if (req->Attributes & WIN_DATA_WIDTH_16) win->flags |= MAP_16BIT; if (req->Attributes & WIN_USE_WAIT) win->flags |= MAP_USE_WAIT; win->card_start = 0; if (s->ops->set_mem_map(s, win) != 0) { dev_dbg(&s->dev, "failed to set memory mapping\n"); mutex_unlock(&s->ops_mutex); return -EIO; } s->state |= SOCKET_WIN_REQ(w); /* Return window handle */ if (s->features & SS_CAP_STATIC_MAP) req->Base = win->static_start; else req->Base = win->res->start; mutex_unlock(&s->ops_mutex); *wh = w + 1; return 0; } /* pcmcia_request_window */ EXPORT_SYMBOL(pcmcia_request_window); void pcmcia_disable_device(struct pcmcia_device *p_dev) { pcmcia_release_configuration(p_dev); pcmcia_release_io(p_dev, &p_dev->io); if (p_dev->_irq) free_irq(p_dev->irq, p_dev->priv); if (p_dev->win) pcmcia_release_window(p_dev, p_dev->win); } EXPORT_SYMBOL(pcmcia_disable_device);