mtdcore.c 30.4 KB
Newer Older
Linus Torvalds's avatar
Linus Torvalds committed
1 2 3 4
/*
 * Core registration and callback routines for MTD
 * drivers and users.
 *
5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21
 * Copyright © 1999-2010 David Woodhouse <dwmw2@infradead.org>
 * Copyright © 2006      Red Hat UK Limited 
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License as published by
 * the Free Software Foundation; either version 2 of the License, or
 * (at your option) any later version.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA
 *
Linus Torvalds's avatar
Linus Torvalds committed
22 23 24 25 26
 */

#include <linux/module.h>
#include <linux/kernel.h>
#include <linux/ptrace.h>
27
#include <linux/seq_file.h>
Linus Torvalds's avatar
Linus Torvalds committed
28 29 30 31
#include <linux/string.h>
#include <linux/timer.h>
#include <linux/major.h>
#include <linux/fs.h>
32
#include <linux/err.h>
Linus Torvalds's avatar
Linus Torvalds committed
33 34 35
#include <linux/ioctl.h>
#include <linux/init.h>
#include <linux/proc_fs.h>
36
#include <linux/idr.h>
37
#include <linux/backing-dev.h>
Tejun Heo's avatar
Tejun Heo committed
38
#include <linux/gfp.h>
Linus Torvalds's avatar
Linus Torvalds committed
39 40

#include <linux/mtd/mtd.h>
41
#include <linux/mtd/partitions.h>
Linus Torvalds's avatar
Linus Torvalds committed
42

43
#include "mtdcore.h"
44 45 46 47
/*
 * backing device capabilities for non-mappable devices (such as NAND flash)
 * - permits private mappings, copies are taken of the data
 */
48
static struct backing_dev_info mtd_bdi_unmappable = {
49 50 51 52 53 54 55 56
	.capabilities	= BDI_CAP_MAP_COPY,
};

/*
 * backing device capabilities for R/O mappable devices (such as ROM)
 * - permits private mappings, copies are taken of the data
 * - permits non-writable shared mappings
 */
57
static struct backing_dev_info mtd_bdi_ro_mappable = {
58 59 60 61 62 63 64 65 66
	.capabilities	= (BDI_CAP_MAP_COPY | BDI_CAP_MAP_DIRECT |
			   BDI_CAP_EXEC_MAP | BDI_CAP_READ_MAP),
};

/*
 * backing device capabilities for writable mappable devices (such as RAM)
 * - permits private mappings, copies are taken of the data
 * - permits non-writable shared mappings
 */
67
static struct backing_dev_info mtd_bdi_rw_mappable = {
68 69 70 71
	.capabilities	= (BDI_CAP_MAP_COPY | BDI_CAP_MAP_DIRECT |
			   BDI_CAP_EXEC_MAP | BDI_CAP_READ_MAP |
			   BDI_CAP_WRITE_MAP),
};
72

73 74 75 76 77 78 79 80 81
static int mtd_cls_suspend(struct device *dev, pm_message_t state);
static int mtd_cls_resume(struct device *dev);

static struct class mtd_class = {
	.name = "mtd",
	.owner = THIS_MODULE,
	.suspend = mtd_cls_suspend,
	.resume = mtd_cls_resume,
};
David Brownell's avatar
David Brownell committed
82

83 84
static DEFINE_IDR(mtd_idr);

85
/* These are exported solely for the purpose of mtd_blkdevs.c. You
Linus Torvalds's avatar
Linus Torvalds committed
86
   should not use them for _anything_ else */
Ingo Molnar's avatar
Ingo Molnar committed
87
DEFINE_MUTEX(mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
88
EXPORT_SYMBOL_GPL(mtd_table_mutex);
89 90 91 92 93 94

struct mtd_info *__mtd_next_device(int i)
{
	return idr_get_next(&mtd_idr, &i);
}
EXPORT_SYMBOL_GPL(__mtd_next_device);
Linus Torvalds's avatar
Linus Torvalds committed
95 96 97

static LIST_HEAD(mtd_notifiers);

David Brownell's avatar
David Brownell committed
98 99 100 101 102 103 104 105 106 107 108 109

#if defined(CONFIG_MTD_CHAR) || defined(CONFIG_MTD_CHAR_MODULE)
#define MTD_DEVT(index) MKDEV(MTD_CHAR_MAJOR, (index)*2)
#else
#define MTD_DEVT(index) 0
#endif

/* REVISIT once MTD uses the driver model better, whoever allocates
 * the mtd_info will probably want to use the release() hook...
 */
static void mtd_release(struct device *dev)
{
110
	struct mtd_info __maybe_unused *mtd = dev_get_drvdata(dev);
Artem Bityutskiy's avatar
Artem Bityutskiy committed
111
	dev_t index = MTD_DEVT(mtd->index);
David Brownell's avatar
David Brownell committed
112 113

	/* remove /dev/mtdXro node if needed */
114
	if (index)
115 116 117 118 119
		device_destroy(&mtd_class, index + 1);
}

static int mtd_cls_suspend(struct device *dev, pm_message_t state)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
120
	struct mtd_info *mtd = dev_get_drvdata(dev);
121

Artem Bityutskiy's avatar
Artem Bityutskiy committed
122
	return mtd ? mtd_suspend(mtd) : 0;
123 124 125 126
}

static int mtd_cls_resume(struct device *dev)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
127
	struct mtd_info *mtd = dev_get_drvdata(dev);
128

129
	if (mtd)
130
		mtd_resume(mtd);
131
	return 0;
David Brownell's avatar
David Brownell committed
132 133 134 135 136
}

static ssize_t mtd_type_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
137
	struct mtd_info *mtd = dev_get_drvdata(dev);
David Brownell's avatar
David Brownell committed
138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167
	char *type;

	switch (mtd->type) {
	case MTD_ABSENT:
		type = "absent";
		break;
	case MTD_RAM:
		type = "ram";
		break;
	case MTD_ROM:
		type = "rom";
		break;
	case MTD_NORFLASH:
		type = "nor";
		break;
	case MTD_NANDFLASH:
		type = "nand";
		break;
	case MTD_DATAFLASH:
		type = "dataflash";
		break;
	case MTD_UBIVOLUME:
		type = "ubi";
		break;
	default:
		type = "unknown";
	}

	return snprintf(buf, PAGE_SIZE, "%s\n", type);
}
168 169 170 171 172
static DEVICE_ATTR(type, S_IRUGO, mtd_type_show, NULL);

static ssize_t mtd_flags_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
173
	struct mtd_info *mtd = dev_get_drvdata(dev);
174 175 176 177 178 179 180 181 182

	return snprintf(buf, PAGE_SIZE, "0x%lx\n", (unsigned long)mtd->flags);

}
static DEVICE_ATTR(flags, S_IRUGO, mtd_flags_show, NULL);

static ssize_t mtd_size_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
183
	struct mtd_info *mtd = dev_get_drvdata(dev);
184 185 186 187 188 189 190 191 192 193

	return snprintf(buf, PAGE_SIZE, "%llu\n",
		(unsigned long long)mtd->size);

}
static DEVICE_ATTR(size, S_IRUGO, mtd_size_show, NULL);

static ssize_t mtd_erasesize_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
194
	struct mtd_info *mtd = dev_get_drvdata(dev);
195 196 197 198 199 200 201 202 203

	return snprintf(buf, PAGE_SIZE, "%lu\n", (unsigned long)mtd->erasesize);

}
static DEVICE_ATTR(erasesize, S_IRUGO, mtd_erasesize_show, NULL);

static ssize_t mtd_writesize_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
204
	struct mtd_info *mtd = dev_get_drvdata(dev);
205 206 207 208 209 210

	return snprintf(buf, PAGE_SIZE, "%lu\n", (unsigned long)mtd->writesize);

}
static DEVICE_ATTR(writesize, S_IRUGO, mtd_writesize_show, NULL);

211 212 213
static ssize_t mtd_subpagesize_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
214
	struct mtd_info *mtd = dev_get_drvdata(dev);
215 216 217 218 219 220 221
	unsigned int subpagesize = mtd->writesize >> mtd->subpage_sft;

	return snprintf(buf, PAGE_SIZE, "%u\n", subpagesize);

}
static DEVICE_ATTR(subpagesize, S_IRUGO, mtd_subpagesize_show, NULL);

222 223 224
static ssize_t mtd_oobsize_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
225
	struct mtd_info *mtd = dev_get_drvdata(dev);
226 227 228 229 230 231 232 233 234

	return snprintf(buf, PAGE_SIZE, "%lu\n", (unsigned long)mtd->oobsize);

}
static DEVICE_ATTR(oobsize, S_IRUGO, mtd_oobsize_show, NULL);

static ssize_t mtd_numeraseregions_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
235
	struct mtd_info *mtd = dev_get_drvdata(dev);
236 237 238 239 240 241 242 243 244 245

	return snprintf(buf, PAGE_SIZE, "%u\n", mtd->numeraseregions);

}
static DEVICE_ATTR(numeraseregions, S_IRUGO, mtd_numeraseregions_show,
	NULL);

static ssize_t mtd_name_show(struct device *dev,
		struct device_attribute *attr, char *buf)
{
Artem Bityutskiy's avatar
Artem Bityutskiy committed
246
	struct mtd_info *mtd = dev_get_drvdata(dev);
247 248 249 250 251

	return snprintf(buf, PAGE_SIZE, "%s\n", mtd->name);

}
static DEVICE_ATTR(name, S_IRUGO, mtd_name_show, NULL);
David Brownell's avatar
David Brownell committed
252

253 254 255 256 257 258 259 260 261
static ssize_t mtd_ecc_strength_show(struct device *dev,
				     struct device_attribute *attr, char *buf)
{
	struct mtd_info *mtd = dev_get_drvdata(dev);

	return snprintf(buf, PAGE_SIZE, "%u\n", mtd->ecc_strength);
}
static DEVICE_ATTR(ecc_strength, S_IRUGO, mtd_ecc_strength_show, NULL);

262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289
static ssize_t mtd_bitflip_threshold_show(struct device *dev,
					  struct device_attribute *attr,
					  char *buf)
{
	struct mtd_info *mtd = dev_get_drvdata(dev);

	return snprintf(buf, PAGE_SIZE, "%u\n", mtd->bitflip_threshold);
}

static ssize_t mtd_bitflip_threshold_store(struct device *dev,
					   struct device_attribute *attr,
					   const char *buf, size_t count)
{
	struct mtd_info *mtd = dev_get_drvdata(dev);
	unsigned int bitflip_threshold;
	int retval;

	retval = kstrtouint(buf, 0, &bitflip_threshold);
	if (retval)
		return retval;

	mtd->bitflip_threshold = bitflip_threshold;
	return count;
}
static DEVICE_ATTR(bitflip_threshold, S_IRUGO | S_IWUSR,
		   mtd_bitflip_threshold_show,
		   mtd_bitflip_threshold_store);

David Brownell's avatar
David Brownell committed
290
static struct attribute *mtd_attrs[] = {
291 292 293 294 295
	&dev_attr_type.attr,
	&dev_attr_flags.attr,
	&dev_attr_size.attr,
	&dev_attr_erasesize.attr,
	&dev_attr_writesize.attr,
296
	&dev_attr_subpagesize.attr,
297 298 299
	&dev_attr_oobsize.attr,
	&dev_attr_numeraseregions.attr,
	&dev_attr_name.attr,
300
	&dev_attr_ecc_strength.attr,
301
	&dev_attr_bitflip_threshold.attr,
David Brownell's avatar
David Brownell committed
302 303 304
	NULL,
};

305
static struct attribute_group mtd_group = {
David Brownell's avatar
David Brownell committed
306 307 308
	.attrs		= mtd_attrs,
};

309
static const struct attribute_group *mtd_groups[] = {
David Brownell's avatar
David Brownell committed
310 311 312 313 314 315 316 317 318 319
	&mtd_group,
	NULL,
};

static struct device_type mtd_devtype = {
	.name		= "mtd",
	.groups		= mtd_groups,
	.release	= mtd_release,
};

Linus Torvalds's avatar
Linus Torvalds committed
320 321 322 323 324 325 326
/**
 *	add_mtd_device - register an MTD device
 *	@mtd: pointer to new MTD device info structure
 *
 *	Add a device to the list of MTD devices present in the system, and
 *	notify each currently active MTD 'user' of its arrival. Returns
 *	zero on success or 1 on failure, which currently will only happen
327
 *	if there is insufficient memory or a sysfs error.
Linus Torvalds's avatar
Linus Torvalds committed
328 329 330 331
 */

int add_mtd_device(struct mtd_info *mtd)
{
332 333
	struct mtd_notifier *not;
	int i, error;
Linus Torvalds's avatar
Linus Torvalds committed
334

335 336 337 338 339 340 341 342 343 344 345 346 347 348
	if (!mtd->backing_dev_info) {
		switch (mtd->type) {
		case MTD_RAM:
			mtd->backing_dev_info = &mtd_bdi_rw_mappable;
			break;
		case MTD_ROM:
			mtd->backing_dev_info = &mtd_bdi_ro_mappable;
			break;
		default:
			mtd->backing_dev_info = &mtd_bdi_unmappable;
			break;
		}
	}

349
	BUG_ON(mtd->writesize == 0);
Ingo Molnar's avatar
Ingo Molnar committed
350
	mutex_lock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
351

352 353 354 355 356
	do {
		if (!idr_pre_get(&mtd_idr, GFP_KERNEL))
			goto fail_locked;
		error = idr_get_new(&mtd_idr, mtd, &i);
	} while (error == -EAGAIN);
357

358 359
	if (error)
		goto fail_locked;
David Brownell's avatar
David Brownell committed
360

361 362 363
	mtd->index = i;
	mtd->usecount = 0;

364 365 366 367
	/* default value if not set by driver */
	if (mtd->bitflip_threshold == 0)
		mtd->bitflip_threshold = mtd->ecc_strength;

368 369 370 371 372 373 374 375 376 377 378 379 380 381
	if (is_power_of_2(mtd->erasesize))
		mtd->erasesize_shift = ffs(mtd->erasesize) - 1;
	else
		mtd->erasesize_shift = 0;

	if (is_power_of_2(mtd->writesize))
		mtd->writesize_shift = ffs(mtd->writesize) - 1;
	else
		mtd->writesize_shift = 0;

	mtd->erasesize_mask = (1 << mtd->erasesize_shift) - 1;
	mtd->writesize_mask = (1 << mtd->writesize_shift) - 1;

	/* Some chips always power up locked. Unlock them now */
382 383 384
	if ((mtd->flags & MTD_WRITEABLE) && (mtd->flags & MTD_POWERUP_LOCK)) {
		error = mtd_unlock(mtd, 0, mtd->size);
		if (error && error != -EOPNOTSUPP)
385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405
			printk(KERN_WARNING
			       "%s: unlock failed, writes may not work\n",
			       mtd->name);
	}

	/* Caller should have set dev.parent to match the
	 * physical device.
	 */
	mtd->dev.type = &mtd_devtype;
	mtd->dev.class = &mtd_class;
	mtd->dev.devt = MTD_DEVT(i);
	dev_set_name(&mtd->dev, "mtd%d", i);
	dev_set_drvdata(&mtd->dev, mtd);
	if (device_register(&mtd->dev) != 0)
		goto fail_added;

	if (MTD_DEVT(i))
		device_create(&mtd_class, mtd->dev.parent,
			      MTD_DEVT(i) + 1,
			      NULL, "mtd%dro", i);

406
	pr_debug("mtd: Giving out device %d to %s\n", i, mtd->name);
407 408 409 410 411 412 413 414 415 416 417 418
	/* No need to get a refcount on the module containing
	   the notifier, since we hold the mtd_table_mutex */
	list_for_each_entry(not, &mtd_notifiers, list)
		not->add(mtd);

	mutex_unlock(&mtd_table_mutex);
	/* We _know_ we aren't being removed, because
	   our caller is still holding us here. So none
	   of this try_ nonsense, and no bitching about it
	   either. :) */
	__module_get(THIS_MODULE);
	return 0;
419

420 421 422
fail_added:
	idr_remove(&mtd_idr, i);
fail_locked:
Ingo Molnar's avatar
Ingo Molnar committed
423
	mutex_unlock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
424 425 426 427 428 429 430 431 432 433 434 435 436
	return 1;
}

/**
 *	del_mtd_device - unregister an MTD device
 *	@mtd: pointer to MTD device info structure
 *
 *	Remove a device from the list of MTD devices present in the system,
 *	and notify each currently active MTD 'user' of its departure.
 *	Returns zero on success or 1 on failure, which currently will happen
 *	if the requested device does not appear to be present in the list.
 */

437
int del_mtd_device(struct mtd_info *mtd)
Linus Torvalds's avatar
Linus Torvalds committed
438 439
{
	int ret;
440
	struct mtd_notifier *not;
441

Ingo Molnar's avatar
Ingo Molnar committed
442
	mutex_lock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
443

444
	if (idr_find(&mtd_idr, mtd->index) != mtd) {
Linus Torvalds's avatar
Linus Torvalds committed
445
		ret = -ENODEV;
446 447 448 449 450 451 452 453 454
		goto out_error;
	}

	/* No need to get a refcount on the module containing
		the notifier, since we hold the mtd_table_mutex */
	list_for_each_entry(not, &mtd_notifiers, list)
		not->remove(mtd);

	if (mtd->usecount) {
455
		printk(KERN_NOTICE "Removing MTD device #%d (%s) with use count %d\n",
Linus Torvalds's avatar
Linus Torvalds committed
456 457 458
		       mtd->index, mtd->name, mtd->usecount);
		ret = -EBUSY;
	} else {
459 460
		device_unregister(&mtd->dev);

461
		idr_remove(&mtd_idr, mtd->index);
Linus Torvalds's avatar
Linus Torvalds committed
462 463 464 465 466

		module_put(THIS_MODULE);
		ret = 0;
	}

467
out_error:
Ingo Molnar's avatar
Ingo Molnar committed
468
	mutex_unlock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
469 470 471
	return ret;
}

472 473 474 475 476 477
/**
 * mtd_device_parse_register - parse partitions and register an MTD device.
 *
 * @mtd: the MTD device to register
 * @types: the list of MTD partition probes to try, see
 *         'parse_mtd_partitions()' for more information
478
 * @parser_data: MTD partition parser-specific data
479 480 481 482 483 484 485 486 487 488 489 490 491 492
 * @parts: fallback partition information to register, if parsing fails;
 *         only valid if %nr_parts > %0
 * @nr_parts: the number of partitions in parts, if zero then the full
 *            MTD device is registered if no partition info is found
 *
 * This function aggregates MTD partitions parsing (done by
 * 'parse_mtd_partitions()') and MTD device and partitions registering. It
 * basically follows the most common pattern found in many MTD drivers:
 *
 * * It first tries to probe partitions on MTD device @mtd using parsers
 *   specified in @types (if @types is %NULL, then the default list of parsers
 *   is used, see 'parse_mtd_partitions()' for more information). If none are
 *   found this functions tries to fallback to information specified in
 *   @parts/@nr_parts.
Brian Norris's avatar
Brian Norris committed
493
 * * If any partitioning info was found, this function registers the found
494 495 496 497 498 499 500
 *   partitions.
 * * If no partitions were found this function just registers the MTD device
 *   @mtd and exits.
 *
 * Returns zero in case of success and a negative error code in case of failure.
 */
int mtd_device_parse_register(struct mtd_info *mtd, const char **types,
501
			      struct mtd_part_parser_data *parser_data,
502 503 504 505 506 507
			      const struct mtd_partition *parts,
			      int nr_parts)
{
	int err;
	struct mtd_partition *real_parts;

508
	err = parse_mtd_partitions(mtd, types, &real_parts, parser_data);
509
	if (err <= 0 && nr_parts && parts) {
510 511
		real_parts = kmemdup(parts, sizeof(*parts) * nr_parts,
				     GFP_KERNEL);
512
		if (!real_parts)
513
			err = -ENOMEM;
514 515
		else
			err = nr_parts;
516 517 518 519 520 521 522 523 524 525 526 527 528 529 530
	}

	if (err > 0) {
		err = add_mtd_partitions(mtd, real_parts, err);
		kfree(real_parts);
	} else if (err == 0) {
		err = add_mtd_device(mtd);
		if (err == 1)
			err = -ENODEV;
	}

	return err;
}
EXPORT_SYMBOL_GPL(mtd_device_parse_register);

531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551
/**
 * mtd_device_unregister - unregister an existing MTD device.
 *
 * @master: the MTD device to unregister.  This will unregister both the master
 *          and any partitions if registered.
 */
int mtd_device_unregister(struct mtd_info *master)
{
	int err;

	err = del_mtd_partitions(master);
	if (err)
		return err;

	if (!device_is_registered(&master->dev))
		return 0;

	return del_mtd_device(master);
}
EXPORT_SYMBOL_GPL(mtd_device_unregister);

Linus Torvalds's avatar
Linus Torvalds committed
552 553 554 555 556 557 558 559 560 561
/**
 *	register_mtd_user - register a 'user' of MTD devices.
 *	@new: pointer to notifier info structure
 *
 *	Registers a pair of callbacks function to be called upon addition
 *	or removal of MTD devices. Causes the 'add' callback to be immediately
 *	invoked for each MTD device currently present in the system.
 */
void register_mtd_user (struct mtd_notifier *new)
{
562
	struct mtd_info *mtd;
Linus Torvalds's avatar
Linus Torvalds committed
563

Ingo Molnar's avatar
Ingo Molnar committed
564
	mutex_lock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
565 566 567

	list_add(&new->list, &mtd_notifiers);

568
	__module_get(THIS_MODULE);
569

570 571
	mtd_for_each_device(mtd)
		new->add(mtd);
Linus Torvalds's avatar
Linus Torvalds committed
572

Ingo Molnar's avatar
Ingo Molnar committed
573
	mutex_unlock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
574
}
575
EXPORT_SYMBOL_GPL(register_mtd_user);
Linus Torvalds's avatar
Linus Torvalds committed
576 577

/**
578 579
 *	unregister_mtd_user - unregister a 'user' of MTD devices.
 *	@old: pointer to notifier info structure
Linus Torvalds's avatar
Linus Torvalds committed
580 581 582 583 584 585 586 587
 *
 *	Removes a callback function pair from the list of 'users' to be
 *	notified upon addition or removal of MTD devices. Causes the
 *	'remove' callback to be immediately invoked for each MTD device
 *	currently present in the system.
 */
int unregister_mtd_user (struct mtd_notifier *old)
{
588
	struct mtd_info *mtd;
Linus Torvalds's avatar
Linus Torvalds committed
589

Ingo Molnar's avatar
Ingo Molnar committed
590
	mutex_lock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
591 592 593

	module_put(THIS_MODULE);

594 595
	mtd_for_each_device(mtd)
		old->remove(mtd);
596

Linus Torvalds's avatar
Linus Torvalds committed
597
	list_del(&old->list);
Ingo Molnar's avatar
Ingo Molnar committed
598
	mutex_unlock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
599 600
	return 0;
}
601
EXPORT_SYMBOL_GPL(unregister_mtd_user);
Linus Torvalds's avatar
Linus Torvalds committed
602 603 604 605 606 607 608 609 610

/**
 *	get_mtd_device - obtain a validated handle for an MTD device
 *	@mtd: last known address of the required MTD device
 *	@num: internal device number of the required MTD device
 *
 *	Given a number and NULL address, return the num'th entry in the device
 *	table, if any.	Given an address and num == -1, search the device table
 *	for a device with that address and return if it's still present. Given
611 612
 *	both, return the num'th driver only if its address matches. Return
 *	error code if not.
Linus Torvalds's avatar
Linus Torvalds committed
613 614 615
 */
struct mtd_info *get_mtd_device(struct mtd_info *mtd, int num)
{
616 617
	struct mtd_info *ret = NULL, *other;
	int err = -ENODEV;
Linus Torvalds's avatar
Linus Torvalds committed
618

Ingo Molnar's avatar
Ingo Molnar committed
619
	mutex_lock(&mtd_table_mutex);
Linus Torvalds's avatar
Linus Torvalds committed
620 621

	if (num == -1) {
622 623 624 625 626 627
		mtd_for_each_device(other) {
			if (other == mtd) {
				ret = mtd;
				break;
			}
		}
628 629
	} else if (num >= 0) {
		ret = idr_find(&mtd_idr, num);
Linus Torvalds's avatar
Linus Torvalds committed
630 631 632 633
		if (mtd && mtd != ret)
			ret = NULL;
	}

634 635 636
	if (!ret) {
		ret = ERR_PTR(err);
		goto out;
637
	}
Linus Torvalds's avatar
Linus Torvalds committed
638

639 640 641 642
	err = __get_mtd_device(ret);
	if (err)
		ret = ERR_PTR(err);
out:
643 644
	mutex_unlock(&mtd_table_mutex);
	return ret;
645
}
646
EXPORT_SYMBOL_GPL(get_mtd_device);
Linus Torvalds's avatar
Linus Torvalds committed
647

648 649 650 651 652 653 654 655

int __get_mtd_device(struct mtd_info *mtd)
{
	int err;

	if (!try_module_get(mtd->owner))
		return -ENODEV;

656 657
	if (mtd->_get_device) {
		err = mtd->_get_device(mtd);
658 659 660 661 662 663 664 665

		if (err) {
			module_put(mtd->owner);
			return err;
		}
	}
	mtd->usecount++;
	return 0;
Linus Torvalds's avatar
Linus Torvalds committed
666
}
667
EXPORT_SYMBOL_GPL(__get_mtd_device);
Linus Torvalds's avatar
Linus Torvalds committed
668

669 670 671 672 673 674 675 676 677 678
/**
 *	get_mtd_device_nm - obtain a validated handle for an MTD device by
 *	device name
 *	@name: MTD device name to open
 *
 * 	This function returns MTD device description structure in case of
 * 	success and an error code in case of failure.
 */
struct mtd_info *get_mtd_device_nm(const char *name)
{
679 680
	int err = -ENODEV;
	struct mtd_info *mtd = NULL, *other;
681 682 683

	mutex_lock(&mtd_table_mutex);

684 685 686
	mtd_for_each_device(other) {
		if (!strcmp(name, other->name)) {
			mtd = other;
687 688 689 690
			break;
		}
	}

691
	if (!mtd)
692 693
		goto out_unlock;

694 695
	err = __get_mtd_device(mtd);
	if (err)
696 697
		goto out_unlock;

698 699
	mutex_unlock(&mtd_table_mutex);
	return mtd;
700 701 702

out_unlock:
	mutex_unlock(&mtd_table_mutex);
703
	return ERR_PTR(err);
704
}
705
EXPORT_SYMBOL_GPL(get_mtd_device_nm);
706

Linus Torvalds's avatar
Linus Torvalds committed
707 708
void put_mtd_device(struct mtd_info *mtd)
{
Ingo Molnar's avatar
Ingo Molnar committed
709
	mutex_lock(&mtd_table_mutex);
710 711 712 713
	__put_mtd_device(mtd);
	mutex_unlock(&mtd_table_mutex);

}
714
EXPORT_SYMBOL_GPL(put_mtd_device);
715 716 717 718 719 720

void __put_mtd_device(struct mtd_info *mtd)
{
	--mtd->usecount;
	BUG_ON(mtd->usecount < 0);

721 722
	if (mtd->_put_device)
		mtd->_put_device(mtd);
Linus Torvalds's avatar
Linus Torvalds committed
723 724 725

	module_put(mtd->owner);
}
726
EXPORT_SYMBOL_GPL(__put_mtd_device);
Linus Torvalds's avatar
Linus Torvalds committed
727

728 729 730 731 732 733 734 735 736 737 738
/*
 * Erase is an asynchronous operation.  Device drivers are supposed
 * to call instr->callback() whenever the operation completes, even
 * if it completes with a failure.
 * Callers are supposed to pass a callback function and wait for it
 * to be called before writing to the block.
 */
int mtd_erase(struct mtd_info *mtd, struct erase_info *instr)
{
	if (instr->addr > mtd->size || instr->len > mtd->size - instr->addr)
		return -EINVAL;
739 740
	if (!(mtd->flags & MTD_WRITEABLE))
		return -EROFS;
741
	instr->fail_addr = MTD_FAIL_ADDR_UNKNOWN;
742 743 744 745 746
	if (!instr->len) {
		instr->state = MTD_ERASE_DONE;
		mtd_erase_callback(instr);
		return 0;
	}
747 748 749 750 751 752 753 754 755 756 757
	return mtd->_erase(mtd, instr);
}
EXPORT_SYMBOL_GPL(mtd_erase);

/*
 * This stuff for eXecute-In-Place. phys is optional and may be set to NULL.
 */
int mtd_point(struct mtd_info *mtd, loff_t from, size_t len, size_t *retlen,
	      void **virt, resource_size_t *phys)
{
	*retlen = 0;
758 759 760
	*virt = NULL;
	if (phys)
		*phys = 0;
761 762 763 764
	if (!mtd->_point)
		return -EOPNOTSUPP;
	if (from < 0 || from > mtd->size || len > mtd->size - from)
		return -EINVAL;
765 766
	if (!len)
		return 0;
767 768 769 770 771 772 773 774 775 776 777
	return mtd->_point(mtd, from, len, retlen, virt, phys);
}
EXPORT_SYMBOL_GPL(mtd_point);

/* We probably shouldn't allow XIP if the unpoint isn't a NULL */
int mtd_unpoint(struct mtd_info *mtd, loff_t from, size_t len)
{
	if (!mtd->_point)
		return -EOPNOTSUPP;
	if (from < 0 || from > mtd->size || len > mtd->size - from)
		return -EINVAL;
778 779
	if (!len)
		return 0;
780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802
	return mtd->_unpoint(mtd, from, len);
}
EXPORT_SYMBOL_GPL(mtd_unpoint);

/*
 * Allow NOMMU mmap() to directly map the device (if not NULL)
 * - return the address to which the offset maps
 * - return -ENOSYS to indicate refusal to do the mapping
 */
unsigned long mtd_get_unmapped_area(struct mtd_info *mtd, unsigned long len,
				    unsigned long offset, unsigned long flags)
{
	if (!mtd->_get_unmapped_area)
		return -EOPNOTSUPP;
	if (offset > mtd->size || len > mtd->size - offset)
		return -EINVAL;
	return mtd->_get_unmapped_area(mtd, len, offset, flags);
}
EXPORT_SYMBOL_GPL(mtd_get_unmapped_area);

int mtd_read(struct mtd_info *mtd, loff_t from, size_t len, size_t *retlen,
	     u_char *buf)
{
803
	*retlen = 0;
804 805
	if (from < 0 || from > mtd->size || len > mtd->size - from)
		return -EINVAL;
806 807
	if (!len)
		return 0;
808 809 810 811 812 813 814 815 816 817
	return mtd->_read(mtd, from, len, retlen, buf);
}
EXPORT_SYMBOL_GPL(mtd_read);

int mtd_write(struct mtd_info *mtd, loff_t to, size_t len, size_t *retlen,
	      const u_char *buf)
{
	*retlen = 0;
	if (to < 0 || to > mtd->size || len > mtd->size - to)
		return -EINVAL;
818 819
	if (!mtd->_write || !(mtd->flags & MTD_WRITEABLE))
		return -EROFS;
820 821
	if (!len)
		return 0;
822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840
	return mtd->_write(mtd, to, len, retlen, buf);
}
EXPORT_SYMBOL_GPL(mtd_write);

/*
 * In blackbox flight recorder like scenarios we want to make successful writes
 * in interrupt context. panic_write() is only intended to be called when its
 * known the kernel is about to panic and we need the write to succeed. Since
 * the kernel is not going to be running for much longer, this function can
 * break locks and delay to ensure the write succeeds (but not sleep).
 */
int mtd_panic_write(struct mtd_info *mtd, loff_t to, size_t len, size_t *retlen,
		    const u_char *buf)
{
	*retlen = 0;
	if (!mtd->_panic_write)
		return -EOPNOTSUPP;
	if (to < 0 || to > mtd->size || len > mtd->size - to)
		return -EINVAL;
841 842
	if (!(mtd->flags & MTD_WRITEABLE))
		return -EROFS;
843 844
	if (!len)
		return 0;
845 846 847 848
	return mtd->_panic_write(mtd, to, len, retlen, buf);
}
EXPORT_SYMBOL_GPL(mtd_panic_write);

849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921
/*
 * Method to access the protection register area, present in some flash
 * devices. The user data is one time programmable but the factory data is read
 * only.
 */
int mtd_get_fact_prot_info(struct mtd_info *mtd, struct otp_info *buf,
			   size_t len)
{
	if (!mtd->_get_fact_prot_info)
		return -EOPNOTSUPP;
	if (!len)
		return 0;
	return mtd->_get_fact_prot_info(mtd, buf, len);
}
EXPORT_SYMBOL_GPL(mtd_get_fact_prot_info);

int mtd_read_fact_prot_reg(struct mtd_info *mtd, loff_t from, size_t len,
			   size_t *retlen, u_char *buf)
{
	*retlen = 0;
	if (!mtd->_read_fact_prot_reg)
		return -EOPNOTSUPP;
	if (!len)
		return 0;
	return mtd->_read_fact_prot_reg(mtd, from, len, retlen, buf);
}
EXPORT_SYMBOL_GPL(mtd_read_fact_prot_reg);

int mtd_get_user_prot_info(struct mtd_info *mtd, struct otp_info *buf,
			   size_t len)
{
	if (!mtd->_get_user_prot_info)
		return -EOPNOTSUPP;
	if (!len)
		return 0;
	return mtd->_get_user_prot_info(mtd, buf, len);
}
EXPORT_SYMBOL_GPL(mtd_get_user_prot_info);

int mtd_read_user_prot_reg(struct mtd_info *mtd, loff_t from, size_t len,
			   size_t *retlen, u_char *buf)
{
	*retlen = 0;
	if (!mtd->_read_user_prot_reg)
		return -EOPNOTSUPP;
	if (!len)
		return 0;
	return mtd->_read_user_prot_reg(mtd, from, len, retlen, buf);
}
EXPORT_SYMBOL_GPL(mtd_read_user_prot_reg);

int mtd_write_user_prot_reg(struct mtd_info *mtd, loff_t to, size_t len,
			    size_t *retlen, u_char *buf)
{
	*retlen = 0;
	if (!mtd->_write_user_prot_reg)
		return -EOPNOTSUPP;
	if (!len)
		return 0;
	return mtd->_write_user_prot_reg(mtd, to, len, retlen, buf);
}
EXPORT_SYMBOL_GPL(mtd_write_user_prot_reg);

int mtd_lock_user_prot_reg(struct mtd_info *mtd, loff_t from, size_t len)
{
	if (!mtd->_lock_user_prot_reg)
		return -EOPNOTSUPP;
	if (!len)
		return 0;
	return mtd->_lock_user_prot_reg(mtd, from, len);
}
EXPORT_SYMBOL_GPL(mtd_lock_user_prot_reg);

922 923 924 925 926 927 928
/* Chip-supported device locking */
int mtd_lock(struct mtd_info *mtd, loff_t ofs, uint64_t len)
{
	if (!mtd->_lock)
		return -EOPNOTSUPP;
	if (ofs < 0 || ofs > mtd->size || len > mtd->size - ofs)
		return -EINVAL;
929 930
	if (!len)
		return 0;
931 932 933 934 935 936 937 938 939 940
	return mtd->_lock(mtd, ofs, len);
}
EXPORT_SYMBOL_GPL(mtd_lock);

int mtd_unlock(struct mtd_info *mtd, loff_t ofs, uint64_t len)
{
	if (!mtd->_unlock)
		return -EOPNOTSUPP;
	if (ofs < 0 || ofs > mtd->size || len > mtd->size - ofs)
		return -EINVAL;
941 942
	if (!len)
		return 0;
943 944 945 946 947 948 949 950 951 952
	return mtd->_unlock(mtd, ofs, len);
}
EXPORT_SYMBOL_GPL(mtd_unlock);

int mtd_is_locked(struct mtd_info *mtd, loff_t ofs, uint64_t len)
{
	if (!mtd->_is_locked)
		return -EOPNOTSUPP;
	if (ofs < 0 || ofs > mtd->size || len > mtd->size - ofs)
		return -EINVAL;
953 954
	if (!len)
		return 0;
955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974
	return mtd->_is_locked(mtd, ofs, len);
}
EXPORT_SYMBOL_GPL(mtd_is_locked);

int mtd_block_isbad(struct mtd_info *mtd, loff_t ofs)
{
	if (!mtd->_block_isbad)
		return 0;
	if (ofs < 0 || ofs > mtd->size)
		return -EINVAL;
	return mtd->_block_isbad(mtd, ofs);
}
EXPORT_SYMBOL_GPL(mtd_block_isbad);

int mtd_block_markbad(struct mtd_info *mtd, loff_t ofs)
{
	if (!mtd->_block_markbad)
		return -EOPNOTSUPP;
	if (ofs < 0 || ofs > mtd->size)
		return -EINVAL;
975 976
	if (!(mtd->flags & MTD_WRITEABLE))
		return -EROFS;
977 978 979 980
	return mtd->_block_markbad(mtd, ofs);
}
EXPORT_SYMBOL_GPL(mtd_block_markbad);

981 982 983 984 985 986 987 988 989 990
/*
 * default_mtd_writev - the default writev method
 * @mtd: mtd device description object pointer
 * @vecs: the vectors to write
 * @count: count of vectors in @vecs
 * @to: the MTD device offset to write to
 * @retlen: on exit contains the count of bytes written to the MTD device.
 *
 * This function returns zero in case of success and a negative error code in
 * case of failure.
Linus Torvalds's avatar
Linus Torvalds committed
991
 */
992 993
static int default_mtd_writev(struct mtd_info *mtd, const struct kvec *vecs,
			      unsigned long count, loff_t to, size_t *retlen)
Linus Torvalds's avatar
Linus Torvalds committed
994 995 996 997 998
{
	unsigned long i;
	size_t totlen = 0, thislen;
	int ret = 0;

999 1000 1001 1002 1003 1004 1005 1006 1007
	for (i = 0; i < count; i++) {
		if (!vecs[i].iov_len)
			continue;
		ret = mtd_write(mtd, to, vecs[i].iov_len, &thislen,
				vecs[i].iov_base);
		totlen += thislen;
		if (ret || thislen != vecs[i].iov_len)
			break;
		to += vecs[i].iov_len;
Linus Torvalds's avatar
Linus Torvalds committed
1008
	}
1009
	*retlen = totlen;
Linus Torvalds's avatar
Linus Torvalds committed
1010 1011
	return ret;
}
1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027

/*
 * mtd_writev - the vector-based MTD write method
 * @mtd: mtd device description object pointer
 * @vecs: the vectors to write
 * @count: count of vectors in @vecs
 * @to: the MTD device offset to write to
 * @retlen: on exit contains the count of bytes written to the MTD device.
 *
 * This function returns zero in case of success and a negative error code in
 * case of failure.
 */
int mtd_writev(struct mtd_info *mtd, const struct kvec *vecs,
	       unsigned long count, loff_t to, size_t *retlen)
{
	*retlen = 0;
1028 1029
	if (!(mtd->flags & MTD_WRITEABLE))
		return -EROFS;
1030
	if (!mtd->_writev)
1031
		return default_mtd_writev(mtd, vecs, count, to, retlen);
1032
	return mtd->_writev(mtd, vecs, count, to, retlen);
1033 1034
}
EXPORT_SYMBOL_GPL(mtd_writev);
Linus Torvalds's avatar
Linus Torvalds committed
1035

1036 1037
/**
 * mtd_kmalloc_up_to - allocate a contiguous buffer up to the specified size
1038 1039
 * @mtd: mtd device description object pointer
 * @size: a pointer to the ideal or maximum size of the allocation, points
1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084
 *        to the actual allocation size on success.
 *
 * This routine attempts to allocate a contiguous kernel buffer up to
 * the specified size, backing off the size of the request exponentially
 * until the request succeeds or until the allocation size falls below
 * the system page size. This attempts to make sure it does not adversely
 * impact system performance, so when allocating more than one page, we
 * ask the memory allocator to avoid re-trying, swapping, writing back
 * or performing I/O.
 *
 * Note, this function also makes sure that the allocated buffer is aligned to
 * the MTD device's min. I/O unit, i.e. the "mtd->writesize" value.
 *
 * This is called, for example by mtd_{read,write} and jffs2_scan_medium,
 * to handle smaller (i.e. degraded) buffer allocations under low- or
 * fragmented-memory situations where such reduced allocations, from a
 * requested ideal, are allowed.
 *
 * Returns a pointer to the allocated buffer on success; otherwise, NULL.
 */
void *mtd_kmalloc_up_to(const struct mtd_info *mtd, size_t *size)
{
	gfp_t flags = __GFP_NOWARN | __GFP_WAIT |
		       __GFP_NORETRY | __GFP_NO_KSWAPD;
	size_t min_alloc = max_t(size_t, mtd->writesize, PAGE_SIZE);
	void *kbuf;

	*size = min_t(size_t, *size, KMALLOC_MAX_SIZE);

	while (*size > min_alloc) {
		kbuf = kmalloc(*size, flags);
		if (kbuf)
			return kbuf;

		*size >>= 1;
		*size = ALIGN(*size, mtd->writesize);
	}

	/*
	 * For the last resort allocation allow 'kmalloc()' to do all sorts of
	 * things (write-back, dropping caches, etc) by using GFP_KERNEL.
	 */
	return kmalloc(*size, GFP_KERNEL);
}
EXPORT_SYMBOL_GPL(mtd_kmalloc_up_to);
Linus Torvalds's avatar
Linus Torvalds committed
1085

1086 1087
#ifdef CONFIG_PROC_FS

Linus Torvalds's avatar
Linus Torvalds committed
1088 1089 1090 1091 1092
/*====================================================================*/
/* Support for /proc/mtd */

static struct proc_dir_entry *proc_mtd;

1093
static int mtd_proc_show(struct seq_file *m, void *v)
Linus Torvalds's avatar
Linus Torvalds committed
1094
{
1095
	struct mtd_info *mtd;
Linus Torvalds's avatar
Linus Torvalds committed
1096

1097
	seq_puts(m, "dev:    size   erasesize  name\n");
Ingo Molnar's avatar
Ingo Molnar committed
1098
	mutex_lock(&mtd_table_mutex);
1099
	mtd_for_each_device(mtd) {
1100 1101 1102
		seq_printf(m, "mtd%d: %8.8llx %8.8x \"%s\"\n",
			   mtd->index, (unsigned long long)mtd->size,
			   mtd->erasesize, mtd->name);
1103
	}
Ingo Molnar's avatar
Ingo Molnar committed
1104
	mutex_unlock(&mtd_table_mutex);
1105
	return 0;
Linus Torvalds's avatar
Linus Torvalds committed
1106 1107
}

1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118
static int mtd_proc_open(struct inode *inode, struct file *file)
{
	return single_open(file, mtd_proc_show, NULL);
}

static const struct file_operations mtd_proc_ops = {
	.open		= mtd_proc_open,
	.read		= seq_read,
	.llseek		= seq_lseek,
	.release	= single_release,
};
1119 1120
#endif /* CONFIG_PROC_FS */

Linus Torvalds's avatar
Linus Torvalds committed
1121 1122 1123
/*====================================================================*/
/* Init code */

1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137
static int __init mtd_bdi_init(struct backing_dev_info *bdi, const char *name)
{
	int ret;

	ret = bdi_init(bdi);
	if (!ret)
		ret = bdi_register(bdi, NULL, name);

	if (ret)
		bdi_destroy(bdi);

	return ret;
}

Linus Torvalds's avatar
Linus Torvalds committed
1138 1139
static int __init init_mtd(void)
{
1140
	int ret;
1141

1142
	ret = class_register(&mtd_class);
1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156
	if (ret)
		goto err_reg;

	ret = mtd_bdi_init(&mtd_bdi_unmappable, "mtd-unmap");
	if (ret)
		goto err_bdi1;

	ret = mtd_bdi_init(&mtd_bdi_ro_mappable, "mtd-romap");
	if (ret)
		goto err_bdi2;

	ret = mtd_bdi_init(&mtd_bdi_rw_mappable, "mtd-rwmap");
	if (ret)
		goto err_bdi3;
1157

1158
#ifdef CONFIG_PROC_FS
1159
	proc_mtd = proc_create("mtd", 0, NULL, &mtd_proc_ops);
1160
#endif /* CONFIG_PROC_FS */
Linus Torvalds's avatar
Linus Torvalds committed
1161
	return 0;
1162 1163 1164 1165 1166 1167 1168 1169 1170 1171

err_bdi3:
	bdi_destroy(&mtd_bdi_ro_mappable);
err_bdi2:
	bdi_destroy(&mtd_bdi_unmappable);
err_bdi1:
	class_unregister(&mtd_class);
err_reg:
	pr_err("Error registering mtd class or bdi: %d\n", ret);
	return ret;
Linus Torvalds's avatar
Linus Torvalds committed
1172 1173 1174 1175
}

static void __exit cleanup_mtd(void)
{
1176
#ifdef CONFIG_PROC_FS
1177
	if (proc_mtd)
Linus Torvalds's avatar
Linus Torvalds committed
1178
		remove_proc_entry( "mtd", NULL);
1179
#endif /* CONFIG_PROC_FS */
1180
	class_unregister(&mtd_class);
1181 1182 1183
	bdi_destroy(&mtd_bdi_unmappable);
	bdi_destroy(&mtd_bdi_ro_mappable);
	bdi_destroy(&mtd_bdi_rw_mappable);
Linus Torvalds's avatar
Linus Torvalds committed
1184 1185 1186 1187 1188 1189 1190 1191
}

module_init(init_mtd);
module_exit(cleanup_mtd);

MODULE_LICENSE("GPL");
MODULE_AUTHOR("David Woodhouse <dwmw2@infradead.org>");
MODULE_DESCRIPTION("Core MTD registration and access routines");