/* * linux/drivers/mmc/host/glamo-mmc.c - Glamo MMC driver * * Copyright (C) 2007 Openmoko, Inc, Andy Green <andy@openmoko.com> * Based on S3C MMC driver that was: * Copyright (C) 2004-2006 maintech GmbH, Thomas Kleffel <tk@maintech.de> * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. */ #include <linux/module.h> #include <linux/mmc/mmc.h> #include <linux/mmc/sd.h> #include <linux/mmc/host.h> #include <linux/platform_device.h> #include <linux/irq.h> #include <linux/delay.h> #include <linux/interrupt.h> #include <linux/workqueue.h> #include <linux/crc7.h> #include <linux/scatterlist.h> #include <linux/io.h> #include <linux/regulator/consumer.h> #include <linux/mfd/glamo.h> #include "glamo-core.h" #include "glamo-regs.h" #define DRIVER_NAME "glamo-mci" struct glamo_mci_host { struct glamo_mmc_platform_data *pdata; struct platform_device *pdev; struct glamo_core *core; struct mmc_host *mmc; struct resource *mmio_mem; struct resource *data_mem; void __iomem *mmio_base; u16 __iomem *data_base; struct regulator *regulator; struct mmc_request *mrq; unsigned int clk_rate; unsigned short vdd; char power_mode; unsigned char request_counter; struct timer_list disable_timer; struct work_struct irq_work; struct work_struct read_work; unsigned clk_enabled : 1; }; static void glamo_mci_send_request(struct mmc_host *mmc, struct mmc_request* mrq); static void glamo_mci_send_command(struct glamo_mci_host *host, struct mmc_command *cmd); /* * Max SD clock rate * * held at /(3 + 1) due to concerns of 100R recommended series resistor * allows 16MHz @ 4-bit --> 8MBytes/sec raw * * you can override this on kernel commandline using * * glamo_mci.sd_max_clk=10000000 * * for example */ static int sd_max_clk = 21000000; module_param(sd_max_clk, int, 0644); /* * Slow SD clock rate * * you can override this on kernel commandline using * * glamo_mci.sd_slow_ratio=8 * * for example * * platform callback is used to decide effective clock rate, if not * defined then max is used, if defined and returns nonzero, rate is * divided by this factor */ static int sd_slow_ratio = 8; module_param(sd_slow_ratio, int, 0644); /* * Post-power SD clock rate * * you can override this on kernel commandline using * * glamo_mci.sd_post_power_clock=1000000 * * for example * * After changing power to card, clock is held at this rate until first bulk * transfer completes */ static int sd_post_power_clock = 1000000; module_param(sd_post_power_clock, int, 0644); static inline void glamo_reg_write(struct glamo_mci_host *glamo, u_int16_t reg, u_int16_t val) { writew(val, glamo->mmio_base + reg); } static inline u_int16_t glamo_reg_read(struct glamo_mci_host *glamo, u_int16_t reg) { return readw(glamo->mmio_base + reg); } static void glamo_reg_set_bit_mask(struct glamo_mci_host *glamo, u_int16_t reg, u_int16_t mask, u_int16_t val) { u_int16_t tmp; val &= mask; tmp = glamo_reg_read(glamo, reg); tmp &= ~mask; tmp |= val; glamo_reg_write(glamo, reg, tmp); } static void glamo_mci_clock_disable(struct glamo_mci_host *host) { if (host->clk_enabled) { glamo_engine_div_disable(host->core, GLAMO_ENGINE_MMC); host->clk_enabled = 0; } } static void glamo_mci_clock_enable(struct glamo_mci_host *host) { del_timer_sync(&host->disable_timer); if (!host->clk_enabled) { glamo_engine_div_enable(host->core, GLAMO_ENGINE_MMC); host->clk_enabled = 1; } } static void glamo_mci_disable_timer(unsigned long data) { struct glamo_mci_host *host = (struct glamo_mci_host *)data; glamo_mci_clock_disable(host); } static void do_pio_read(struct glamo_mci_host *host, struct mmc_data *data) { struct scatterlist *sg; u16 __iomem *from_ptr = host->data_base; void *sg_pointer; dev_dbg(&host->pdev->dev, "pio_read():\n"); for (sg = data->sg; sg; sg = sg_next(sg)) { sg_pointer = page_address(sg_page(sg)) + sg->offset; memcpy(sg_pointer, from_ptr, sg->length); from_ptr += sg->length >> 1; data->bytes_xfered += sg->length; } dev_dbg(&host->pdev->dev, "pio_read(): " "complete (no more data).\n"); } static void do_pio_write(struct glamo_mci_host *host, struct mmc_data *data) { struct scatterlist *sg; u16 __iomem *to_ptr = host->data_base; void *sg_pointer; dev_dbg(&host->pdev->dev, "pio_write():\n"); for (sg = data->sg; sg; sg = sg_next(sg)) { sg_pointer = page_address(sg_page(sg)) + sg->offset; data->bytes_xfered += sg->length; memcpy(to_ptr, sg_pointer, sg->length); to_ptr += sg->length >> 1; } dev_dbg(&host->pdev->dev, "pio_write(): complete\n"); } static int glamo_mci_set_card_clock(struct glamo_mci_host *host, int freq) { int real_rate = 0; if (freq) { glamo_mci_clock_enable(host); real_rate = glamo_engine_reclock(host->core, GLAMO_ENGINE_MMC, freq); } else { glamo_mci_clock_disable(host); } return real_rate; } static void glamo_mci_request_done(struct glamo_mci_host *host, struct mmc_request *mrq) { mod_timer(&host->disable_timer, jiffies + HZ / 16); mmc_request_done(host->mmc, mrq); } static void glamo_mci_irq_worker(struct work_struct *work) { struct glamo_mci_host *host = container_of(work, struct glamo_mci_host, irq_work); struct mmc_command *cmd; uint16_t status; if (!host->mrq || !host->mrq->cmd) return; cmd = host->mrq->cmd; #if 0 if (cmd->data->flags & MMC_DATA_READ) { return; } #endif status = glamo_reg_read(host, GLAMO_REG_MMC_RB_STAT1); dev_dbg(&host->pdev->dev, "status = 0x%04x\n", status); /* we ignore a data timeout report if we are also told the data came */ if (status & GLAMO_STAT1_MMC_RB_DRDY) status &= ~GLAMO_STAT1_MMC_DTOUT; if (status & (GLAMO_STAT1_MMC_RTOUT | GLAMO_STAT1_MMC_DTOUT)) cmd->error = -ETIMEDOUT; if (status & (GLAMO_STAT1_MMC_BWERR | GLAMO_STAT1_MMC_BRERR)) { cmd->error = -EILSEQ; } if (cmd->error) { dev_info(&host->pdev->dev, "Error after cmd: 0x%x\n", status); goto done; } /* issue STOP if we have been given one to use */ if (host->mrq->stop) { glamo_mci_send_command(host, host->mrq->stop); } if (cmd->data->flags & MMC_DATA_READ) do_pio_read(host, cmd->data); done: host->mrq = NULL; glamo_mci_request_done(host, cmd->mrq); } static void glamo_mci_read_worker(struct work_struct *work) { struct glamo_mci_host *host = container_of(work, struct glamo_mci_host, read_work); struct mmc_command *cmd; uint16_t status; uint16_t blocks_ready; size_t data_read = 0; size_t data_ready; struct scatterlist *sg; u16 __iomem *from_ptr = host->data_base; void *sg_pointer; cmd = host->mrq->cmd; sg = cmd->data->sg; do { status = glamo_reg_read(host, GLAMO_REG_MMC_RB_STAT1); if (status & (GLAMO_STAT1_MMC_RTOUT | GLAMO_STAT1_MMC_DTOUT)) cmd->error = -ETIMEDOUT; if (status & (GLAMO_STAT1_MMC_BWERR | GLAMO_STAT1_MMC_BRERR)) cmd->error = -EILSEQ; if (cmd->error) { dev_info(&host->pdev->dev, "Error after cmd: 0x%x\n", status); goto done; } blocks_ready = glamo_reg_read(host, GLAMO_REG_MMC_RB_BLKCNT); data_ready = blocks_ready * cmd->data->blksz; if (data_ready == data_read) yield(); while(sg && data_read + sg->length <= data_ready) { sg_pointer = page_address(sg_page(sg)) + sg->offset; memcpy(sg_pointer, from_ptr, sg->length); from_ptr += sg->length >> 1; data_read += sg->length; sg = sg_next(sg); } } while(sg); cmd->data->bytes_xfered = data_read; do { status = glamo_reg_read(host, GLAMO_REG_MMC_RB_STAT1); } while (!(status & GLAMO_STAT1_MMC_IDLE)); if (host->mrq->stop) glamo_mci_send_command(host, host->mrq->stop); do { status = glamo_reg_read(host, GLAMO_REG_MMC_RB_STAT1); } while (!(status & GLAMO_STAT1_MMC_IDLE)); done: host->mrq = NULL; glamo_mci_request_done(host, cmd->mrq); } static irqreturn_t glamo_mci_irq(int irq, void *devid) { struct glamo_mci_host *host = (struct glamo_mci_host*)devid; schedule_work(&host->irq_work); return IRQ_HANDLED; } static void glamo_mci_send_command(struct glamo_mci_host *host, struct mmc_command *cmd) { u8 u8a[6]; u16 fire = 0; unsigned int timeout = 1000000; u16 * reg_resp = (u16 *)(host->mmio_base + GLAMO_REG_MMC_CMD_RSP1); u16 status; int triggers_int = 1; /* if we can't do it, reject as busy */ if (!glamo_reg_read(host, GLAMO_REG_MMC_RB_STAT1) & GLAMO_STAT1_MMC_IDLE) { cmd->error = -EBUSY; return; } /* create an array in wire order for CRC computation */ u8a[0] = 0x40 | (cmd->opcode & 0x3f); u8a[1] = (u8)(cmd->arg >> 24); u8a[2] = (u8)(cmd->arg >> 16); u8a[3] = (u8)(cmd->arg >> 8); u8a[4] = (u8)cmd->arg; u8a[5] = (crc7(0, u8a, 5) << 1) | 0x01; /* crc7 on first 5 bytes of packet */ /* issue the wire-order array including CRC in register order */ glamo_reg_write(host, GLAMO_REG_MMC_CMD_REG1, ((u8a[4] << 8) | u8a[5])); glamo_reg_write(host, GLAMO_REG_MMC_CMD_REG2, ((u8a[2] << 8) | u8a[3])); glamo_reg_write(host, GLAMO_REG_MMC_CMD_REG3, ((u8a[0] << 8) | u8a[1])); /* command index toggle */ fire |= (host->request_counter & 1) << 12; /* set type of command */ switch (mmc_cmd_type(cmd)) { case MMC_CMD_BC: fire |= GLAMO_FIRE_MMC_CMDT_BNR; break; case MMC_CMD_BCR: fire |= GLAMO_FIRE_MMC_CMDT_BR; break; case MMC_CMD_AC: fire |= GLAMO_FIRE_MMC_CMDT_AND; break; case MMC_CMD_ADTC: fire |= GLAMO_FIRE_MMC_CMDT_AD; break; } /* * if it expects a response, set the type expected * * R1, Length : 48bit, Normal response * R1b, Length : 48bit, same R1, but added card busy status * R2, Length : 136bit (really 128 bits with CRC snipped) * R3, Length : 48bit (OCR register value) * R4, Length : 48bit, SDIO_OP_CONDITION, Reverse SDIO Card * R5, Length : 48bit, IO_RW_DIRECTION, Reverse SDIO Card * R6, Length : 48bit (RCA register) * R7, Length : 48bit (interface condition, VHS(voltage supplied), * check pattern, CRC7) */ switch (mmc_resp_type(cmd)) { case MMC_RSP_R1: /* same index as R6 and R7 */ fire |= GLAMO_FIRE_MMC_RSPT_R1; break; case MMC_RSP_R1B: fire |= GLAMO_FIRE_MMC_RSPT_R1b; break; case MMC_RSP_R2: fire |= GLAMO_FIRE_MMC_RSPT_R2; break; case MMC_RSP_R3: fire |= GLAMO_FIRE_MMC_RSPT_R3; break; /* R4 and R5 supported by chip not defined in linux/mmc/core.h (sdio) */ } /* * From the command index, set up the command class in the host ctrllr * * missing guys present on chip but couldn't figure out how to use yet: * 0x0 "stream read" * 0x9 "cancel running command" */ switch (cmd->opcode) { case MMC_READ_SINGLE_BLOCK: fire |= GLAMO_FIRE_MMC_CC_SBR; /* single block read */ break; case MMC_SWITCH: /* 64 byte payload */ case SD_APP_SEND_SCR: case MMC_READ_MULTIPLE_BLOCK: /* we will get an interrupt off this */ if (!cmd->mrq->stop) /* multiblock no stop */ fire |= GLAMO_FIRE_MMC_CC_MBRNS; else /* multiblock with stop */ fire |= GLAMO_FIRE_MMC_CC_MBRS; break; case MMC_WRITE_BLOCK: fire |= GLAMO_FIRE_MMC_CC_SBW; /* single block write */ break; case MMC_WRITE_MULTIPLE_BLOCK: if (cmd->mrq->stop) /* multiblock with stop */ fire |= GLAMO_FIRE_MMC_CC_MBWS; else /* multiblock NO stop-- 'RESERVED'? */ fire |= GLAMO_FIRE_MMC_CC_MBWNS; break; case MMC_STOP_TRANSMISSION: fire |= GLAMO_FIRE_MMC_CC_STOP; /* STOP */ triggers_int = 0; break; default: fire |= GLAMO_FIRE_MMC_CC_BASIC; /* "basic command" */ triggers_int = 0; break; } if (cmd->data) host->mrq = cmd->mrq; /* always largest timeout */ glamo_reg_write(host, GLAMO_REG_MMC_TIMEOUT, 0xfff); /* Generate interrupt on txfer */ glamo_reg_set_bit_mask(host, GLAMO_REG_MMC_BASIC, 0xff36, 0x0800 | GLAMO_BASIC_MMC_NO_CLK_RD_WAIT | GLAMO_BASIC_MMC_EN_COMPL_INT | GLAMO_BASIC_MMC_EN_DATA_PUPS | GLAMO_BASIC_MMC_EN_CMD_PUP); /* send the command out on the wire */ /* dev_info(&host->pdev->dev, "Using FIRE %04X\n", fire); */ glamo_reg_write(host, GLAMO_REG_MMC_CMD_FIRE, fire); /* we are deselecting card? because it isn't going to ack then... */ if ((cmd->opcode == 7) && (cmd->arg == 0)) return; /* * we must spin until response is ready or timed out * -- we don't get interrupts unless there is a bulk rx */ do status = glamo_reg_read(host, GLAMO_REG_MMC_RB_STAT1); while (((((status >> 15) & 1) != (host->request_counter & 1)) || (!(status & (GLAMO_STAT1_MMC_RB_RRDY | GLAMO_STAT1_MMC_RTOUT | GLAMO_STAT1_MMC_DTOUT | GLAMO_STAT1_MMC_BWERR | GLAMO_STAT1_MMC_BRERR)))) && (timeout--)); if ((status & (GLAMO_STAT1_MMC_RTOUT | GLAMO_STAT1_MMC_DTOUT)) || (timeout == 0)) { cmd->error = -ETIMEDOUT; } else if (status & (GLAMO_STAT1_MMC_BWERR | GLAMO_STAT1_MMC_BRERR)) { cmd->error = -EILSEQ; } if (cmd->flags & MMC_RSP_PRESENT) { if (cmd->flags & MMC_RSP_136) { cmd->resp[3] = readw(®_resp[0]) | (readw(®_resp[1]) << 16); cmd->resp[2] = readw(®_resp[2]) | (readw(®_resp[3]) << 16); cmd->resp[1] = readw(®_resp[4]) | (readw(®_resp[5]) << 16); cmd->resp[0] = readw(®_resp[6]) | (readw(®_resp[7]) << 16); } else { cmd->resp[0] = (readw(®_resp[0]) >> 8) | (readw(®_resp[1]) << 8) | ((readw(®_resp[2])) << 24); } } #if 0 /* We'll only get an interrupt when all data has been transfered. By starting to copy data when it's avaiable we can increase throughput by up to 30%. */ if (cmd->data && (cmd->data->flags & MMC_DATA_READ)) schedule_work(&host->read_work); #endif } static int glamo_mci_prepare_pio(struct glamo_mci_host *host, struct mmc_data *data) { /* set up the block info */ glamo_reg_write(host, GLAMO_REG_MMC_DATBLKLEN, data->blksz); glamo_reg_write(host, GLAMO_REG_MMC_DATBLKCNT, data->blocks); data->bytes_xfered = 0; /* if write, prep the write into the shared RAM before the command */ if (data->flags & MMC_DATA_WRITE) { do_pio_write(host, data); } dev_dbg(&host->pdev->dev, "(blksz=%d, count=%d)\n", data->blksz, data->blocks); return 0; } static int glamo_mci_irq_poll(struct glamo_mci_host *host, struct mmc_command *cmd) { int timeout = 1000000; /* * if the glamo INT# line isn't wired (*cough* it can happen) * I'm afraid we have to spin on the IRQ status bit and "be * our own INT# line" */ /* * we have faith we will get an "interrupt"... * but something insane like suspend problems can mean * we spin here forever, so we timeout after a LONG time */ while ((!(readw(host->core->base + GLAMO_REG_IRQ_STATUS) & GLAMO_IRQ_MMC)) && (timeout--)); if (timeout < 0) { if (cmd->data->error) cmd->data->error = -ETIMEDOUT; dev_err(&host->pdev->dev, "Payload timeout\n"); return -ETIMEDOUT; } /* ack this interrupt source */ writew(GLAMO_IRQ_MMC, host->core->base + GLAMO_REG_IRQ_CLEAR); /* yay we are an interrupt controller! -- call the ISR * it will stop clock to card */ glamo_mci_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), host); return 0; } static void glamo_mci_send_request(struct mmc_host *mmc, struct mmc_request *mrq) { struct glamo_mci_host *host = mmc_priv(mmc); struct mmc_command *cmd = mrq->cmd; glamo_mci_clock_enable(host); host->request_counter++; if (cmd->data) { if(glamo_mci_prepare_pio(host, cmd->data)) { cmd->error = -EIO; cmd->data->error = -EIO; goto done; } } dev_dbg(&host->pdev->dev,"cmd 0x%x, " "arg 0x%x data=%p mrq->stop=%p flags 0x%x\n", cmd->opcode, cmd->arg, cmd->data, cmd->mrq->stop, cmd->flags); glamo_mci_send_command(host, cmd); /* * if we don't have bulk data to take care of, we're done */ if (!cmd->data || cmd->error) goto done; if (!host->core->irq_works) { if (glamo_mci_irq_poll(host, mrq->cmd)) goto done; } /* * Otherwise can can use the interrupt as async completion -- * if there is read data coming, or we wait for write data to complete, * exit without mmc_request_done() as the payload interrupt * will service it */ dev_dbg(&host->pdev->dev, "Waiting for payload data\n"); return; done: glamo_mci_request_done(host, mrq); } static void glamo_mci_set_power_mode(struct glamo_mci_host *host, unsigned char power_mode) { int ret; if (power_mode == host->power_mode) return; switch(power_mode) { case MMC_POWER_UP: if (host->power_mode == MMC_POWER_OFF) { ret = regulator_enable(host->regulator); if (ret) dev_err(&host->pdev->dev, "Failed to enable regulator: %d\n", ret); } break; case MMC_POWER_ON: break; case MMC_POWER_OFF: default: glamo_engine_disable(host->core, GLAMO_ENGINE_MMC); ret = regulator_disable(host->regulator); if (ret) dev_warn(&host->pdev->dev, "Failed to disable regulator: %d\n", ret); break; } host->power_mode = power_mode; } static void glamo_mci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios) { struct glamo_mci_host *host = mmc_priv(mmc); int bus_width = 0; int rate; int sd_drive; int ret; /* Set power */ glamo_mci_set_power_mode(host, ios->power_mode); if (host->vdd != ios->vdd) { ret = mmc_regulator_set_ocr(host->regulator, ios->vdd); if (ret) dev_err(&host->pdev->dev, "Failed to set regulator voltage: %d\n", ret); else host->vdd = ios->vdd; } rate = glamo_mci_set_card_clock(host, ios->clock); if ((ios->power_mode == MMC_POWER_ON) || (ios->power_mode == MMC_POWER_UP)) { dev_info(&host->pdev->dev, "powered (vdd = %hu) clk: %dkHz div=%hu (req: %ukHz). " "Bus width=%d\n", ios->vdd, rate / 1000, 0, ios->clock / 1000, (int)ios->bus_width); } else { dev_info(&host->pdev->dev, "glamo_mci_set_ios: power down.\n"); } /* set bus width */ if (ios->bus_width == MMC_BUS_WIDTH_4) bus_width = GLAMO_BASIC_MMC_EN_4BIT_DATA; sd_drive = (rate * 4) / host->clk_rate; if (sd_drive > 3) sd_drive = 3; glamo_reg_set_bit_mask(host, GLAMO_REG_MMC_BASIC, GLAMO_BASIC_MMC_EN_4BIT_DATA | 0xb0, bus_width | sd_drive << 6); } /* * no physical write protect supported by us */ static int glamo_mci_get_ro(struct mmc_host *mmc) { return 0; } static struct mmc_host_ops glamo_mci_ops = { .request = glamo_mci_send_request, .set_ios = glamo_mci_set_ios, .get_ro = glamo_mci_get_ro, }; static int glamo_mci_probe(struct platform_device *pdev) { struct mmc_host *mmc; struct glamo_mci_host *host; struct glamo_core *core = dev_get_drvdata(pdev->dev.parent); int ret; dev_info(&pdev->dev, "glamo_mci driver (C)2007 Openmoko, Inc\n"); mmc = mmc_alloc_host(sizeof(struct glamo_mci_host), &pdev->dev); if (!mmc) { ret = -ENOMEM; goto probe_out; } host = mmc_priv(mmc); host->mmc = mmc; host->pdev = pdev; if (core->pdata) host->pdata = core->pdata->mmc_data; host->power_mode = MMC_POWER_OFF; host->clk_enabled = 0; host->core = core; INIT_WORK(&host->irq_work, glamo_mci_irq_worker); INIT_WORK(&host->read_work, glamo_mci_read_worker); host->regulator = regulator_get(pdev->dev.parent, "SD_3V3"); if (!host->regulator) { dev_err(&pdev->dev, "Cannot proceed without regulator.\n"); ret = -ENODEV; goto probe_free_host; } host->mmio_mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); if (!host->mmio_mem) { dev_err(&pdev->dev, "failed to get io memory region resouce.\n"); ret = -ENOENT; goto probe_regulator_put; } host->mmio_mem = request_mem_region(host->mmio_mem->start, resource_size(host->mmio_mem), pdev->name); if (!host->mmio_mem) { dev_err(&pdev->dev, "failed to request io memory region.\n"); ret = -ENOENT; goto probe_regulator_put; } host->mmio_base = ioremap(host->mmio_mem->start, resource_size(host->mmio_mem)); if (!host->mmio_base) { dev_err(&pdev->dev, "failed to ioremap() io memory region.\n"); ret = -EINVAL; goto probe_free_mem_region_mmio; } /* Get ahold of our data buffer we use for data in and out on MMC */ host->data_mem = platform_get_resource(pdev, IORESOURCE_MEM, 1); if (!host->data_mem) { dev_err(&pdev->dev, "failed to get io memory region resource.\n"); ret = -ENOENT; goto probe_iounmap_mmio; } host->data_mem = request_mem_region(host->data_mem->start, resource_size(host->data_mem), pdev->name); if (!host->data_mem) { dev_err(&pdev->dev, "failed to request io memory region.\n"); ret = -ENOENT; goto probe_iounmap_mmio; } host->data_base = ioremap(host->data_mem->start, resource_size(host->data_mem)); if (host->data_base == 0) { dev_err(&pdev->dev, "failed to ioremap() io memory region.\n"); ret = -EINVAL; goto probe_free_mem_region_data; } ret = request_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), glamo_mci_irq, IRQF_SHARED, pdev->name, host); if (ret) { dev_err(&pdev->dev, "failed to register irq.\n"); goto probe_iounmap_data; } host->vdd = 0; host->clk_rate = glamo_pll_rate(host->core, GLAMO_PLL1); /* explain our host controller capabilities */ mmc->ops = &glamo_mci_ops; mmc->ocr_avail = mmc_regulator_get_ocrmask(host->regulator); mmc->caps = MMC_CAP_4_BIT_DATA | MMC_CAP_MMC_HIGHSPEED | MMC_CAP_SD_HIGHSPEED; mmc->f_min = host->clk_rate / 256; mmc->f_max = sd_max_clk; mmc->max_blk_count = (1 << 16) - 1; /* GLAMO_REG_MMC_RB_BLKCNT */ mmc->max_blk_size = (1 << 12) - 1; /* GLAMO_REG_MMC_RB_BLKLEN */ mmc->max_req_size = resource_size(host->data_mem); mmc->max_seg_size = mmc->max_req_size; mmc->max_phys_segs = 128; mmc->max_hw_segs = 128; if (mmc->ocr_avail < 0) { dev_warn(&pdev->dev, "Failed to get ocr list for regulator: %d.\n", mmc->ocr_avail); mmc->ocr_avail = MMC_VDD_32_33 | MMC_VDD_33_34; } platform_set_drvdata(pdev, mmc); glamo_engine_enable(host->core, GLAMO_ENGINE_MMC); glamo_engine_reset(host->core, GLAMO_ENGINE_MMC); glamo_reg_write(host, GLAMO_REG_MMC_WDATADS1, (u16)(host->data_mem->start)); glamo_reg_write(host, GLAMO_REG_MMC_WDATADS2, (u16)(host->data_mem->start >> 16)); glamo_reg_write(host, GLAMO_REG_MMC_RDATADS1, (u16)(host->data_mem->start)); glamo_reg_write(host, GLAMO_REG_MMC_RDATADS2, (u16)(host->data_mem->start >> 16)); setup_timer(&host->disable_timer, glamo_mci_disable_timer, (unsigned long)host); if ((ret = mmc_add_host(mmc))) { dev_err(&pdev->dev, "failed to add mmc host.\n"); goto probe_freeirq; } dev_info(&pdev->dev,"initialisation done.\n"); return 0; probe_freeirq: free_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), host); probe_iounmap_data: iounmap(host->data_base); probe_free_mem_region_data: release_mem_region(host->data_mem->start, resource_size(host->data_mem)); probe_iounmap_mmio: iounmap(host->mmio_base); probe_free_mem_region_mmio: release_mem_region(host->mmio_mem->start, resource_size(host->mmio_mem)); probe_regulator_put: regulator_put(host->regulator); probe_free_host: mmc_free_host(mmc); probe_out: return ret; } static int glamo_mci_remove(struct platform_device *pdev) { struct mmc_host *mmc = platform_get_drvdata(pdev); struct glamo_mci_host *host = mmc_priv(mmc); free_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), host); mmc_remove_host(mmc); iounmap(host->mmio_base); iounmap(host->data_base); release_mem_region(host->mmio_mem->start, resource_size(host->mmio_mem)); release_mem_region(host->data_mem->start, resource_size(host->data_mem)); regulator_put(host->regulator); mmc_free_host(mmc); glamo_engine_disable(host->core, GLAMO_ENGINE_MMC); return 0; } #ifdef CONFIG_PM static int glamo_mci_suspend(struct device *dev) { struct mmc_host *mmc = dev_get_drvdata(dev); struct glamo_mci_host *host = mmc_priv(mmc); int ret; cancel_work_sync(&host->irq_work); ret = mmc_suspend_host(mmc, PMSG_SUSPEND); glamo_mci_clock_enable(host); return ret; } static int glamo_mci_resume(struct device *dev) { struct mmc_host *mmc = dev_get_drvdata(dev); struct glamo_mci_host *host = mmc_priv(mmc); int ret; glamo_engine_enable(host->core, GLAMO_ENGINE_MMC); glamo_engine_reset(host->core, GLAMO_ENGINE_MMC); glamo_reg_write(host, GLAMO_REG_MMC_WDATADS1, (u16)(host->data_mem->start)); glamo_reg_write(host, GLAMO_REG_MMC_WDATADS2, (u16)(host->data_mem->start >> 16)); glamo_reg_write(host, GLAMO_REG_MMC_RDATADS1, (u16)(host->data_mem->start)); glamo_reg_write(host, GLAMO_REG_MMC_RDATADS2, (u16)(host->data_mem->start >> 16)); mdelay(5); ret = mmc_resume_host(host->mmc); /* glamo_mci_clock_disable(host);*/ return 0; } static struct dev_pm_ops glamo_mci_pm_ops = { .suspend = glamo_mci_suspend, .resume = glamo_mci_resume, }; #define GLAMO_MCI_PM_OPS (&glamo_mci_pm_ops) #else /* CONFIG_PM */ #define GLAMO_MCI_PM_OPS NULL #endif /* CONFIG_PM */ static struct platform_driver glamo_mci_driver = { .probe = glamo_mci_probe, .remove = glamo_mci_remove, .driver = { .name = "glamo-mci", .owner = THIS_MODULE, .pm = GLAMO_MCI_PM_OPS, }, }; static int __init glamo_mci_init(void) { platform_driver_register(&glamo_mci_driver); return 0; } static void __exit glamo_mci_exit(void) { platform_driver_unregister(&glamo_mci_driver); } module_init(glamo_mci_init); module_exit(glamo_mci_exit); MODULE_DESCRIPTION("Glamo MMC/SD Card Interface driver"); MODULE_LICENSE("GPL"); MODULE_AUTHOR("Andy Green <andy@openmoko.com>");