aboutsummaryrefslogtreecommitdiffstats
path: root/target/linux/ramips/files-4.14/drivers/mmc/host
Commit message (Collapse)AuthorAgeFilesLines
* Revert "ramips: mmc: Sync with staging drivers"John Crispin2018-05-155-2693/+3418
| | | | | | This reverts commit 2d401925b9bdf34de800fff57f593a3665de2a6a. Signed-off-by: John Crispin <john@phrozen.org>
* ramips: mmc: Sync with staging driversRosen Penev2018-05-155-3418/+2693
| | | | | | | | | | | | | Mostly whitespace cleanups. Some unneeded code was removed. MMC init was also moved to the probe function as in 6069bdd0871a20b5adce8d2f677946e05a2da609 The cleanup commits are over 100, making it hard to do them individually. Tested on GnuBee PC1 with an SD card being used as swap. Signed-off-by: Rosen Penev <rosenp@gmail.com>
* ramips: move mtk-mmc init to probe function to avoid breaking NAND flashFelix Fietkau2018-05-081-36/+28
| | | | | | | The driver messes with the pin control settings - MMC and NAND are mutually exclusive Signed-off-by: Felix Fietkau <nbd@nbd.name>
* ramips: mtk-mmc: don't advertise highmem support.NeilBrown2018-04-051-0/+1
| | | | | | | | | | | | | | | | of_dma_configure() sets a default ->dma_mask of DMA_BIT_MASK(32), claiming devices can DMA from the full 32bit address space. The mtk-mmc driver does not support access to highmem pages, so it is really limited to the bottom 512M (actually 448M due to 64M of IO space). Setting ->dma_mask to NULL causes mmc_setup_queue() to fall-back to using BLK_BOUNCE_HIGH to tell the block layer to use a bounce-buffer for any highmem pages requiring IO. Signed-off-by: NeilBrown <neil@brown.name> Signed-off-by: Felix Fietkau <nbd@nbd.name>
* ramips: fix BUG_ON on mtk-mmc driver probeFelix Fietkau2018-04-051-3/+1
| | | | | | | Our .dts files only have one device defined and this is unlikely to change, so statically initialize host->id to 0. Signed-off-by: Felix Fietkau <nbd@nbd.name>
* ramips: move mtk-mmc driver code out of patches-*/ and into files-*/Felix Fietkau2018-04-057-0/+4758
Signed-off-by: Felix Fietkau <nbd@nbd.name>