Commit ef5eeea6 authored by Boris Brezillon's avatar Boris Brezillon

mtd: nand: brcm: switch to mtd_ooblayout_ops

Implementing the mtd_ooblayout_ops interface is the new way of exposing
ECC/OOB layout to MTD users.
Signed-off-by: default avatarBoris Brezillon <boris.brezillon@free-electrons.com>
parent c8766e81
...@@ -780,127 +780,183 @@ static inline bool is_hamming_ecc(struct brcmnand_cfg *cfg) ...@@ -780,127 +780,183 @@ static inline bool is_hamming_ecc(struct brcmnand_cfg *cfg)
} }
/* /*
* Returns a nand_ecclayout strucutre for the given layout/configuration. * Set mtd->ooblayout to the appropriate mtd_ooblayout_ops given
* Returns NULL on failure. * the layout/configuration.
* Returns -ERRCODE on failure.
*/ */
static struct nand_ecclayout *brcmnand_create_layout(int ecc_level, static int brcmnand_hamming_ooblayout_ecc(struct mtd_info *mtd, int section,
struct brcmnand_host *host) struct mtd_oob_region *oobregion)
{ {
struct nand_chip *chip = mtd_to_nand(mtd);
struct brcmnand_host *host = nand_get_controller_data(chip);
struct brcmnand_cfg *cfg = &host->hwcfg; struct brcmnand_cfg *cfg = &host->hwcfg;
int i, j; int sas = cfg->spare_area_size << cfg->sector_size_1k;
struct nand_ecclayout *layout; int sectors = cfg->page_size / (512 << cfg->sector_size_1k);
int req;
int sectors;
int sas;
int idx1, idx2;
layout = devm_kzalloc(&host->pdev->dev, sizeof(*layout), GFP_KERNEL);
if (!layout)
return NULL;
sectors = cfg->page_size / (512 << cfg->sector_size_1k);
sas = cfg->spare_area_size << cfg->sector_size_1k;
/* Hamming */
if (is_hamming_ecc(cfg)) {
for (i = 0, idx1 = 0, idx2 = 0; i < sectors; i++) {
/* First sector of each page may have BBI */
if (i == 0) {
layout->oobfree[idx2].offset = i * sas + 1;
/* Small-page NAND use byte 6 for BBI */
if (cfg->page_size == 512)
layout->oobfree[idx2].offset--;
layout->oobfree[idx2].length = 5;
} else {
layout->oobfree[idx2].offset = i * sas;
layout->oobfree[idx2].length = 6;
}
idx2++;
layout->eccpos[idx1++] = i * sas + 6;
layout->eccpos[idx1++] = i * sas + 7;
layout->eccpos[idx1++] = i * sas + 8;
layout->oobfree[idx2].offset = i * sas + 9;
layout->oobfree[idx2].length = 7;
idx2++;
/* Leave zero-terminated entry for OOBFREE */
if (idx1 >= MTD_MAX_ECCPOS_ENTRIES_LARGE ||
idx2 >= MTD_MAX_OOBFREE_ENTRIES_LARGE - 1)
break;
}
return layout; if (section >= sectors)
} return -ERANGE;
oobregion->offset = (section * sas) + 6;
oobregion->length = 3;
return 0;
}
static int brcmnand_hamming_ooblayout_free(struct mtd_info *mtd, int section,
struct mtd_oob_region *oobregion)
{
struct nand_chip *chip = mtd_to_nand(mtd);
struct brcmnand_host *host = nand_get_controller_data(chip);
struct brcmnand_cfg *cfg = &host->hwcfg;
int sas = cfg->spare_area_size << cfg->sector_size_1k;
int sectors = cfg->page_size / (512 << cfg->sector_size_1k);
if (section >= sectors * 2)
return -ERANGE;
oobregion->offset = (section / 2) * sas;
if (section & 1) {
oobregion->offset += 9;
oobregion->length = 7;
} else {
oobregion->length = 6;
/* First sector of each page may have BBI */
if (!section) {
/* /*
* CONTROLLER_VERSION: * Small-page NAND use byte 6 for BBI while large-page
* < v5.0: ECC_REQ = ceil(BCH_T * 13/8) * NAND use byte 0.
* >= v5.0: ECC_REQ = ceil(BCH_T * 14/8)
* But we will just be conservative.
*/ */
req = DIV_ROUND_UP(ecc_level * 14, 8); if (cfg->page_size > 512)
if (req >= sas) { oobregion->offset++;
dev_err(&host->pdev->dev, oobregion->length--;
"error: ECC too large for OOB (ECC bytes %d, spare sector %d)\n", }
req, sas);
return NULL;
} }
layout->eccbytes = req * sectors; return 0;
for (i = 0, idx1 = 0, idx2 = 0; i < sectors; i++) { }
for (j = sas - req; j < sas && idx1 <
MTD_MAX_ECCPOS_ENTRIES_LARGE; j++, idx1++)
layout->eccpos[idx1] = i * sas + j;
/* First sector of each page may have BBI */ static const struct mtd_ooblayout_ops brcmnand_hamming_ooblayout_ops = {
if (i == 0) { .ecc = brcmnand_hamming_ooblayout_ecc,
if (cfg->page_size == 512 && (sas - req >= 6)) { .free = brcmnand_hamming_ooblayout_free,
/* Small-page NAND use byte 6 for BBI */ };
layout->oobfree[idx2].offset = 0;
layout->oobfree[idx2].length = 5; static int brcmnand_bch_ooblayout_ecc(struct mtd_info *mtd, int section,
idx2++; struct mtd_oob_region *oobregion)
if (sas - req > 6) { {
layout->oobfree[idx2].offset = 6; struct nand_chip *chip = mtd_to_nand(mtd);
layout->oobfree[idx2].length = struct brcmnand_host *host = nand_get_controller_data(chip);
sas - req - 6; struct brcmnand_cfg *cfg = &host->hwcfg;
idx2++; int sas = cfg->spare_area_size << cfg->sector_size_1k;
} int sectors = cfg->page_size / (512 << cfg->sector_size_1k);
} else if (sas > req + 1) {
layout->oobfree[idx2].offset = i * sas + 1; if (section >= sectors)
layout->oobfree[idx2].length = sas - req - 1; return -ERANGE;
idx2++;
} oobregion->offset = (section * (sas + 1)) - chip->ecc.bytes;
} else if (sas > req) { oobregion->length = chip->ecc.bytes;
layout->oobfree[idx2].offset = i * sas;
layout->oobfree[idx2].length = sas - req; return 0;
idx2++; }
}
/* Leave zero-terminated entry for OOBFREE */ static int brcmnand_bch_ooblayout_free_lp(struct mtd_info *mtd, int section,
if (idx1 >= MTD_MAX_ECCPOS_ENTRIES_LARGE || struct mtd_oob_region *oobregion)
idx2 >= MTD_MAX_OOBFREE_ENTRIES_LARGE - 1) {
break; struct nand_chip *chip = mtd_to_nand(mtd);
struct brcmnand_host *host = nand_get_controller_data(chip);
struct brcmnand_cfg *cfg = &host->hwcfg;
int sas = cfg->spare_area_size << cfg->sector_size_1k;
int sectors = cfg->page_size / (512 << cfg->sector_size_1k);
if (section >= sectors)
return -ERANGE;
if (sas <= chip->ecc.bytes)
return 0;
oobregion->offset = section * sas;
oobregion->length = sas - chip->ecc.bytes;
if (!section) {
oobregion->offset++;
oobregion->length--;
}
return 0;
}
static int brcmnand_bch_ooblayout_free_sp(struct mtd_info *mtd, int section,
struct mtd_oob_region *oobregion)
{
struct nand_chip *chip = mtd_to_nand(mtd);
struct brcmnand_host *host = nand_get_controller_data(chip);
struct brcmnand_cfg *cfg = &host->hwcfg;
int sas = cfg->spare_area_size << cfg->sector_size_1k;
if (section > 1 || sas - chip->ecc.bytes < 6 ||
(section && sas - chip->ecc.bytes == 6))
return -ERANGE;
if (!section) {
oobregion->offset = 0;
oobregion->length = 5;
} else {
oobregion->offset = 6;
oobregion->length = sas - chip->ecc.bytes - 6;
} }
return layout; return 0;
} }
static struct nand_ecclayout *brcmstb_choose_ecc_layout( static const struct mtd_ooblayout_ops brcmnand_bch_lp_ooblayout_ops = {
struct brcmnand_host *host) .ecc = brcmnand_bch_ooblayout_ecc,
.free = brcmnand_bch_ooblayout_free_lp,
};
static const struct mtd_ooblayout_ops brcmnand_bch_sp_ooblayout_ops = {
.ecc = brcmnand_bch_ooblayout_ecc,
.free = brcmnand_bch_ooblayout_free_sp,
};
static int brcmstb_choose_ecc_layout(struct brcmnand_host *host)
{ {
struct nand_ecclayout *layout;
struct brcmnand_cfg *p = &host->hwcfg; struct brcmnand_cfg *p = &host->hwcfg;
struct mtd_info *mtd = nand_to_mtd(&host->chip);
struct nand_ecc_ctrl *ecc = &host->chip.ecc;
unsigned int ecc_level = p->ecc_level; unsigned int ecc_level = p->ecc_level;
int sas = p->spare_area_size << p->sector_size_1k;
int sectors = p->page_size / (512 << p->sector_size_1k);
if (p->sector_size_1k) if (p->sector_size_1k)
ecc_level <<= 1; ecc_level <<= 1;
layout = brcmnand_create_layout(ecc_level, host); if (is_hamming_ecc(p)) {
if (!layout) { ecc->bytes = 3 * sectors;
mtd_set_ooblayout(mtd, &brcmnand_hamming_ooblayout_ops);
return 0;
}
/*
* CONTROLLER_VERSION:
* < v5.0: ECC_REQ = ceil(BCH_T * 13/8)
* >= v5.0: ECC_REQ = ceil(BCH_T * 14/8)
* But we will just be conservative.
*/
ecc->bytes = DIV_ROUND_UP(ecc_level * 14, 8);
if (p->page_size == 512)
mtd_set_ooblayout(mtd, &brcmnand_bch_sp_ooblayout_ops);
else
mtd_set_ooblayout(mtd, &brcmnand_bch_lp_ooblayout_ops);
if (ecc->bytes >= sas) {
dev_err(&host->pdev->dev, dev_err(&host->pdev->dev,
"no proper ecc_layout for this NAND cfg\n"); "error: ECC too large for OOB (ECC bytes %d, spare sector %d)\n",
return NULL; ecc->bytes, sas);
return -EINVAL;
} }
return layout; return 0;
} }
static void brcmnand_wp(struct mtd_info *mtd, int wp) static void brcmnand_wp(struct mtd_info *mtd, int wp)
...@@ -2010,9 +2066,9 @@ static int brcmnand_init_cs(struct brcmnand_host *host, struct device_node *dn) ...@@ -2010,9 +2066,9 @@ static int brcmnand_init_cs(struct brcmnand_host *host, struct device_node *dn)
/* only use our internal HW threshold */ /* only use our internal HW threshold */
mtd->bitflip_threshold = 1; mtd->bitflip_threshold = 1;
chip->ecc.layout = brcmstb_choose_ecc_layout(host); ret = brcmstb_choose_ecc_layout(host);
if (!chip->ecc.layout) if (ret)
return -ENXIO; return ret;
if (nand_scan_tail(mtd)) if (nand_scan_tail(mtd))
return -ENXIO; return -ENXIO;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment