dm-table.c 18.5 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14
/*
 * Copyright (C) 2001 Sistina Software (UK) Limited.
 *
 * This file is released under the GPL.
 */

#include "dm.h"

#include <linux/module.h>
#include <linux/vmalloc.h>
#include <linux/blkdev.h>
#include <linux/namei.h>
#include <linux/ctype.h>
#include <linux/slab.h>
15
#include <linux/interrupt.h>
16 17 18 19 20 21 22 23 24 25 26
#include <asm/atomic.h>

#define MAX_DEPTH 16
#define NODE_SIZE L1_CACHE_BYTES
#define KEYS_PER_NODE (NODE_SIZE / sizeof(sector_t))
#define CHILDREN_PER_NODE (KEYS_PER_NODE + 1)

struct dm_table {
	atomic_t holders;

	/* btree table */
27 28
	unsigned int depth;
	unsigned int counts[MAX_DEPTH];	/* in nodes */
29 30
	sector_t *index[MAX_DEPTH];

31 32
	unsigned int num_targets;
	unsigned int num_allocated;
33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51
	sector_t *highs;
	struct dm_target *targets;

	/*
	 * Indicates the rw permissions for the new logical
	 * device.  This should be a combination of FMODE_READ
	 * and FMODE_WRITE.
	 */
	int mode;

	/* a list of devices used by this table */
	struct list_head devices;

	/*
	 * These are optimistic limits taken from all the
	 * targets, some targets will need smaller limits.
	 */
	struct io_restrictions limits;

52 53 54
	/* events get handed up using this callback */
	void (*event_fn)(void *);
	void *event_context;
55 56 57 58 59 60 61 62 63 64
};

/*
 * Similar to ceiling(log_size(n))
 */
static unsigned int int_log(unsigned long n, unsigned long base)
{
	int result = 0;

	while (n > 1) {
Joe Thornber's avatar
Joe Thornber committed
65
		n = dm_div_up(n, base);
66 67 68 69 70 71
		result++;
	}

	return result;
}

72 73 74 75
/*
 * Returns the minimum that is _not_ zero, unless both are zero.
 */
#define min_not_zero(l, r) (l == 0) ? r : ((r == 0) ? l : min(l, r))
76 77 78 79 80 81 82

/*
 * Combine two io_restrictions, always taking the lower value.
 */
static void combine_restrictions_low(struct io_restrictions *lhs,
				     struct io_restrictions *rhs)
{
83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98
	lhs->max_sectors =
		min_not_zero(lhs->max_sectors, rhs->max_sectors);

	lhs->max_phys_segments =
		min_not_zero(lhs->max_phys_segments, rhs->max_phys_segments);

	lhs->max_hw_segments =
		min_not_zero(lhs->max_hw_segments, rhs->max_hw_segments);

	lhs->hardsect_size = max(lhs->hardsect_size, rhs->hardsect_size);

	lhs->max_segment_size =
		min_not_zero(lhs->max_segment_size, rhs->max_segment_size);

	lhs->seg_boundary_mask =
		min_not_zero(lhs->seg_boundary_mask, rhs->seg_boundary_mask);
99 100 101 102 103
}

/*
 * Calculate the index of the child node of the n'th node k'th key.
 */
104
static inline unsigned int get_child(unsigned int n, unsigned int k)
105 106 107 108 109 110 111
{
	return (n * CHILDREN_PER_NODE) + k;
}

/*
 * Return the n'th node of level l from table t.
 */
112 113
static inline sector_t *get_node(struct dm_table *t,
				 unsigned int l, unsigned int n)
114 115 116 117 118 119 120 121
{
	return t->index[l] + (n * KEYS_PER_NODE);
}

/*
 * Return the highest key that you could lookup from the n'th
 * node on level l of the btree.
 */
122
static sector_t high(struct dm_table *t, unsigned int l, unsigned int n)
123 124 125 126 127 128 129 130 131 132 133 134 135 136
{
	for (; l < t->depth - 1; l++)
		n = get_child(n, CHILDREN_PER_NODE - 1);

	if (n >= t->counts[l])
		return (sector_t) - 1;

	return get_node(t, l, n)[KEYS_PER_NODE - 1];
}

/*
 * Fills in a level of the btree based on the highs of the level
 * below it.
 */
137
static int setup_btree_index(unsigned int l, struct dm_table *t)
138
{
139
	unsigned int n, k;
140 141
	sector_t *node;

142
	for (n = 0U; n < t->counts[l]; n++) {
143 144
		node = get_node(t, l, n);

145
		for (k = 0U; k < KEYS_PER_NODE; k++)
146 147 148 149 150 151
			node[k] = high(t, l + 1, get_child(n, k));
	}

	return 0;
}

152
void *dm_vcalloc(unsigned long nmemb, unsigned long elem_size)
153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170
{
	unsigned long size;
	void *addr;

	/*
	 * Check that we're not going to overflow.
	 */
	if (nmemb > (ULONG_MAX / elem_size))
		return NULL;

	size = nmemb * elem_size;
	addr = vmalloc(size);
	if (addr)
		memset(addr, 0, size);

	return addr;
}

171 172 173 174
/*
 * highs, and targets are managed as dynamic arrays during a
 * table load.
 */
175
static int alloc_targets(struct dm_table *t, unsigned int num)
176 177 178 179 180 181 182 183
{
	sector_t *n_highs;
	struct dm_target *n_targets;
	int n = t->num_targets;

	/*
	 * Allocate both the target array and offset array at once.
	 */
184 185
	n_highs = (sector_t *) dm_vcalloc(num, sizeof(struct dm_target) +
					  sizeof(sector_t));
186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205
	if (!n_highs)
		return -ENOMEM;

	n_targets = (struct dm_target *) (n_highs + num);

	if (n) {
		memcpy(n_highs, t->highs, sizeof(*n_highs) * n);
		memcpy(n_targets, t->targets, sizeof(*n_targets) * n);
	}

	memset(n_highs + n, -1, sizeof(*n_highs) * (num - n));
	vfree(t->highs);

	t->num_allocated = num;
	t->highs = n_highs;
	t->targets = n_targets;

	return 0;
}

206
int dm_table_create(struct dm_table **result, int mode, unsigned num_targets)
207
{
208
	struct dm_table *t = kmalloc(sizeof(*t), GFP_KERNEL);
209 210 211 212 213 214 215 216

	if (!t)
		return -ENOMEM;

	memset(t, 0, sizeof(*t));
	INIT_LIST_HEAD(&t->devices);
	atomic_set(&t->holders, 1);

217 218 219 220 221 222
	if (!num_targets)
		num_targets = KEYS_PER_NODE;

	num_targets = dm_round_up(num_targets, KEYS_PER_NODE);

	if (alloc_targets(t, num_targets)) {
223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245
		kfree(t);
		t = NULL;
		return -ENOMEM;
	}

	t->mode = mode;
	*result = t;
	return 0;
}

static void free_devices(struct list_head *devices)
{
	struct list_head *tmp, *next;

	for (tmp = devices->next; tmp != devices; tmp = next) {
		struct dm_dev *dd = list_entry(tmp, struct dm_dev, list);
		next = tmp->next;
		kfree(dd);
	}
}

void table_destroy(struct dm_table *t)
{
246
	unsigned int i;
247 248 249 250 251 252 253

	/* free the indexes (see dm_table_complete) */
	if (t->depth >= 2)
		vfree(t->index[t->depth - 2]);

	/* free the targets */
	for (i = 0; i < t->num_targets; i++) {
254
		struct dm_target *tgt = t->targets + i;
255 256 257

		if (tgt->type->dtr)
			tgt->type->dtr(tgt);
258 259

		dm_put_target_type(tgt->type);
260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281
	}

	vfree(t->highs);

	/* free the device list */
	if (t->devices.next != &t->devices) {
		DMWARN("devices still present during destroy: "
		       "dm_table_remove_device calls missing");

		free_devices(&t->devices);
	}

	kfree(t);
}

void dm_table_get(struct dm_table *t)
{
	atomic_inc(&t->holders);
}

void dm_table_put(struct dm_table *t)
{
282 283 284
	if (!t)
		return;

285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318
	if (atomic_dec_and_test(&t->holders))
		table_destroy(t);
}

/*
 * Checks to see if we need to extend highs or targets.
 */
static inline int check_space(struct dm_table *t)
{
	if (t->num_targets >= t->num_allocated)
		return alloc_targets(t, t->num_allocated * 2);

	return 0;
}

/*
 * Convert a device path to a dev_t.
 */
static int lookup_device(const char *path, dev_t *dev)
{
	int r;
	struct nameidata nd;
	struct inode *inode;

	if ((r = path_lookup(path, LOOKUP_FOLLOW, &nd)))
		return r;

	inode = nd.dentry->d_inode;
	if (!inode) {
		r = -ENOENT;
		goto out;
	}

	if (!S_ISBLK(inode->i_mode)) {
Joe Thornber's avatar
Joe Thornber committed
319
		r = -ENOTBLK;
320 321 322
		goto out;
	}

323
	*dev = inode->i_rdev;
324 325 326 327 328 329 330 331 332 333 334

 out:
	path_release(&nd);
	return r;
}

/*
 * See if we've already got a device in the list.
 */
static struct dm_dev *find_device(struct list_head *l, dev_t dev)
{
335
	struct dm_dev *dd;
336

337
	list_for_each_entry (dd, l, list)
338 339 340 341 342 343 344 345 346 347 348 349
		if (dd->bdev->bd_dev == dev)
			return dd;

	return NULL;
}

/*
 * Open a device so we can use it as a map destination.
 */
static int open_dev(struct dm_dev *d, dev_t dev)
{
	static char *_claim_ptr = "I belong to device-mapper";
Alexander Viro's avatar
Alexander Viro committed
350
	struct block_device *bdev;
351 352 353 354 355 356

	int r;

	if (d->bdev)
		BUG();

357
	bdev = open_by_devnum(dev, d->mode);
Alexander Viro's avatar
Alexander Viro committed
358 359 360
	if (IS_ERR(bdev))
		return PTR_ERR(bdev);
	r = bd_claim(bdev, _claim_ptr);
361
	if (r)
362
		blkdev_put(bdev);
Alexander Viro's avatar
Alexander Viro committed
363 364
	else
		d->bdev = bdev;
365 366 367 368 369 370 371 372 373 374 375 376
	return r;
}

/*
 * Close a device that we've been using.
 */
static void close_dev(struct dm_dev *d)
{
	if (!d->bdev)
		return;

	bd_release(d->bdev);
377
	blkdev_put(d->bdev);
378 379 380 381 382 383 384 385 386 387
	d->bdev = NULL;
}

/*
 * If possible (ie. blk_size[major] is set), this checks an area
 * of a destination device is valid.
 */
static int check_device_area(struct dm_dev *dd, sector_t start, sector_t len)
{
	sector_t dev_size;
388
	dev_size = dd->bdev->bd_inode->i_size >> SECTOR_SHIFT;
389 390 391 392 393 394 395 396 397 398 399 400 401 402
	return ((start < dev_size) && (len <= (dev_size - start)));
}

/*
 * This upgrades the mode on an already open dm_dev.  Being
 * careful to leave things as they were if we fail to reopen the
 * device.
 */
static int upgrade_mode(struct dm_dev *dd, int new_mode)
{
	int r;
	struct dm_dev dd_copy;
	dev_t dev = dd->bdev->bd_dev;

403
	dd_copy = *dd;
404 405 406 407 408 409 410

	dd->mode |= new_mode;
	dd->bdev = NULL;
	r = open_dev(dd, dev);
	if (!r)
		close_dev(&dd_copy);
	else
411
		*dd = dd_copy;
412 413 414 415 416 417 418 419 420 421 422 423 424 425 426

	return r;
}

/*
 * Add a device to the list, or just increment the usage count if
 * it's already present.
 */
static int __table_get_device(struct dm_table *t, struct dm_target *ti,
			      const char *path, sector_t start, sector_t len,
			      int mode, struct dm_dev **result)
{
	int r;
	dev_t dev;
	struct dm_dev *dd;
427
	unsigned int major, minor;
428 429 430 431

	if (!t)
		BUG();

432
	if (sscanf(path, "%u:%u", &major, &minor) == 2) {
433 434
		/* Extract the major/minor numbers */
		dev = MKDEV(major, minor);
435 436
		if (MAJOR(dev) != major || MINOR(dev) != minor)
			return -EOVERFLOW;
437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487
	} else {
		/* convert the path to a device */
		if ((r = lookup_device(path, &dev)))
			return r;
	}

	dd = find_device(&t->devices, dev);
	if (!dd) {
		dd = kmalloc(sizeof(*dd), GFP_KERNEL);
		if (!dd)
			return -ENOMEM;

		dd->mode = mode;
		dd->bdev = NULL;

		if ((r = open_dev(dd, dev))) {
			kfree(dd);
			return r;
		}

		atomic_set(&dd->count, 0);
		list_add(&dd->list, &t->devices);

	} else if (dd->mode != (mode | dd->mode)) {
		r = upgrade_mode(dd, mode);
		if (r)
			return r;
	}
	atomic_inc(&dd->count);

	if (!check_device_area(dd, start, len)) {
		DMWARN("device %s too small for target", path);
		dm_put_device(ti, dd);
		return -EINVAL;
	}

	*result = dd;

	return 0;
}


int dm_get_device(struct dm_target *ti, const char *path, sector_t start,
		  sector_t len, int mode, struct dm_dev **result)
{
	int r = __table_get_device(ti->table, ti, path,
				   start, len, mode, result);
	if (!r) {
		request_queue_t *q = bdev_get_queue((*result)->bdev);
		struct io_restrictions *rs = &ti->limits;

488 489 490 491 492 493 494 495 496 497
		/*
		 * Combine the device limits low.
		 *
		 * FIXME: if we move an io_restriction struct
		 *        into q this would just be a call to
		 *        combine_restrictions_low()
		 */
		rs->max_sectors =
			min_not_zero(rs->max_sectors, q->max_sectors);

Kevin Corry's avatar
Kevin Corry committed
498 499 500 501 502 503 504 505 506 507 508 509
		/* FIXME: Device-Mapper on top of RAID-0 breaks because DM
		 *        currently doesn't honor MD's merge_bvec_fn routine.
		 *        In this case, we'll force DM to use PAGE_SIZE or
		 *        smaller I/O, just to be safe. A better fix is in the
		 *        works, but add this for the time being so it will at
		 *        least operate correctly.
		 */
		if (q->merge_bvec_fn)
			rs->max_sectors =
				min_not_zero(rs->max_sectors,
					     (unsigned short)(PAGE_SIZE >> 9));

510 511 512 513 514 515 516 517 518 519 520 521 522 523 524
		rs->max_phys_segments =
			min_not_zero(rs->max_phys_segments,
				     q->max_phys_segments);

		rs->max_hw_segments =
			min_not_zero(rs->max_hw_segments, q->max_hw_segments);

		rs->hardsect_size = max(rs->hardsect_size, q->hardsect_size);

		rs->max_segment_size =
			min_not_zero(rs->max_segment_size, q->max_segment_size);

		rs->seg_boundary_mask =
			min_not_zero(rs->seg_boundary_mask,
				     q->seg_boundary_mask);
525 526 527 528 529 530
	}

	return r;
}

/*
531
 * Decrement a devices use count and remove it if necessary.
532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555
 */
void dm_put_device(struct dm_target *ti, struct dm_dev *dd)
{
	if (atomic_dec_and_test(&dd->count)) {
		close_dev(dd);
		list_del(&dd->list);
		kfree(dd);
	}
}

/*
 * Checks to see if the target joins onto the end of the table.
 */
static int adjoin(struct dm_table *table, struct dm_target *ti)
{
	struct dm_target *prev;

	if (!table->num_targets)
		return !ti->begin;

	prev = &table->targets[table->num_targets - 1];
	return (ti->begin == (prev->begin + prev->len));
}

556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574
/*
 * Used to dynamically allocate the arg array.
 */
static char **realloc_argv(unsigned *array_size, char **old_argv)
{
	char **argv;
	unsigned new_size;

	new_size = *array_size ? *array_size * 2 : 64;
	argv = kmalloc(new_size * sizeof(*argv), GFP_KERNEL);
	if (argv) {
		memcpy(argv, old_argv, *array_size * sizeof(*argv));
		*array_size = new_size;
	}

	kfree(old_argv);
	return argv;
}

575 576 577
/*
 * Destructively splits up the argument list to pass to ctr.
 */
578
static int split_args(int *argc, char ***argvp, char *input)
579
{
580 581 582
	char *start, *end = input, *out, **argv = NULL;
	unsigned array_size = 0;

583
	*argc = 0;
584 585 586
	argv = realloc_argv(&array_size, argv);
	if (!argv)
		return -ENOMEM;
587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614

	while (1) {
		start = end;

		/* Skip whitespace */
		while (*start && isspace(*start))
			start++;

		if (!*start)
			break;	/* success, we hit the end */

		/* 'out' is used to remove any back-quotes */
		end = out = start;
		while (*end) {
			/* Everything apart from '\0' can be quoted */
			if (*end == '\\' && *(end + 1)) {
				*out++ = *(end + 1);
				end += 2;
				continue;
			}

			if (isspace(*end))
				break;	/* end of token */

			*out++ = *end++;
		}

		/* have we already filled the array ? */
615 616 617 618 619
		if ((*argc + 1) > array_size) {
			argv = realloc_argv(&array_size, argv);
			if (!argv)
				return -ENOMEM;
		}
620 621 622 623 624 625 626 627 628 629 630

		/* we know this is whitespace */
		if (*end)
			end++;

		/* terminate the string and put it in the array */
		*out = '\0';
		argv[*argc] = start;
		(*argc)++;
	}

631
	*argvp = argv;
632 633 634
	return 0;
}

635 636 637 638 639 640 641 642 643 644 645 646 647 648
static void check_for_valid_limits(struct io_restrictions *rs)
{
	if (!rs->max_sectors)
		rs->max_sectors = MAX_SECTORS;
	if (!rs->max_phys_segments)
		rs->max_phys_segments = MAX_PHYS_SEGMENTS;
	if (!rs->max_hw_segments)
		rs->max_hw_segments = MAX_HW_SEGMENTS;
	if (!rs->hardsect_size)
		rs->hardsect_size = 1 << SECTOR_SHIFT;
	if (!rs->max_segment_size)
		rs->max_segment_size = MAX_SEGMENT_SIZE;
	if (!rs->seg_boundary_mask)
		rs->seg_boundary_mask = -1;
649 650
}

651 652 653
int dm_table_add_target(struct dm_table *t, const char *type,
			sector_t start, sector_t len, char *params)
{
654
	int r = -EINVAL, argc;
655
	char **argv;
656 657 658 659 660 661 662 663
	struct dm_target *tgt;

	if ((r = check_space(t)))
		return r;

	tgt = t->targets + t->num_targets;
	memset(tgt, 0, sizeof(*tgt));

664 665
	if (!len) {
		tgt->error = "zero-length target";
666
		DMERR(": %s\n", tgt->error);
667 668 669
		return -EINVAL;
	}

670 671
	tgt->type = dm_get_target_type(type);
	if (!tgt->type) {
672
		tgt->error = "unknown target type";
673
		DMERR(": %s\n", tgt->error);
674
		return -EINVAL;
675 676 677 678 679 680 681 682 683 684 685
	}

	tgt->table = t;
	tgt->begin = start;
	tgt->len = len;
	tgt->error = "Unknown error";

	/*
	 * Does this target adjoin the previous one ?
	 */
	if (!adjoin(t, tgt)) {
686
		tgt->error = "Gap in table";
687
		r = -EINVAL;
688
		goto bad;
689 690
	}

691
	r = split_args(&argc, &argv, params);
692
	if (r) {
693
		tgt->error = "couldn't split parameters (insufficient memory)";
694
		goto bad;
695 696
	}

697
	r = tgt->type->ctr(tgt, argc, argv);
698
	kfree(argv);
699 700
	if (r)
		goto bad;
701 702 703 704 705 706 707

	t->highs[t->num_targets++] = tgt->begin + tgt->len - 1;

	/* FIXME: the plan is to combine high here and then have
	 * the merge fn apply the target level restrictions. */
	combine_restrictions_low(&t->limits, &tgt->limits);
	return 0;
708 709

 bad:
710
	DMERR(": %s\n", tgt->error);
711 712
	dm_put_target_type(tgt->type);
	return r;
713 714 715 716
}

static int setup_indexes(struct dm_table *t)
{
717 718
	int i;
	unsigned int total = 0;
719 720 721 722
	sector_t *indexes;

	/* allocate the space for *all* the indexes */
	for (i = t->depth - 2; i >= 0; i--) {
Joe Thornber's avatar
Joe Thornber committed
723
		t->counts[i] = dm_div_up(t->counts[i + 1], CHILDREN_PER_NODE);
724 725 726
		total += t->counts[i];
	}

727
	indexes = (sector_t *) dm_vcalloc(total, (unsigned long) NODE_SIZE);
728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745
	if (!indexes)
		return -ENOMEM;

	/* set up internal nodes, bottom-up */
	for (i = t->depth - 2, total = 0; i >= 0; i--) {
		t->index[i] = indexes;
		indexes += (KEYS_PER_NODE * t->counts[i]);
		setup_btree_index(i, t);
	}

	return 0;
}

/*
 * Builds the btree to index the map.
 */
int dm_table_complete(struct dm_table *t)
{
746 747
	int r = 0;
	unsigned int leaf_nodes;
748

749 750
	check_for_valid_limits(&t->limits);

751
	/* how many indexes will the btree have ? */
Joe Thornber's avatar
Joe Thornber committed
752
	leaf_nodes = dm_div_up(t->num_targets, KEYS_PER_NODE);
753 754 755 756 757 758 759 760 761 762 763 764
	t->depth = 1 + int_log(leaf_nodes, CHILDREN_PER_NODE);

	/* leaf layer has already been set up */
	t->counts[t->depth - 1] = leaf_nodes;
	t->index[t->depth - 1] = t->highs;

	if (t->depth >= 2)
		r = setup_indexes(t);

	return r;
}

765
static DECLARE_MUTEX(_event_lock);
766 767 768
void dm_table_event_callback(struct dm_table *t,
			     void (*fn)(void *), void *context)
{
769
	down(&_event_lock);
770 771
	t->event_fn = fn;
	t->event_context = context;
772
	up(&_event_lock);
773 774
}

775 776
void dm_table_event(struct dm_table *t)
{
777 778 779 780 781 782 783
	/*
	 * You can no longer call dm_table_event() from interrupt
	 * context, use a bottom half instead.
	 */
	BUG_ON(in_interrupt());

	down(&_event_lock);
784 785
	if (t->event_fn)
		t->event_fn(t->event_context);
786
	up(&_event_lock);
787 788 789 790 791 792 793
}

sector_t dm_table_get_size(struct dm_table *t)
{
	return t->num_targets ? (t->highs[t->num_targets - 1] + 1) : 0;
}

794
struct dm_target *dm_table_get_target(struct dm_table *t, unsigned int index)
795 796 797 798 799 800 801 802 803 804 805 806
{
	if (index > t->num_targets)
		return NULL;

	return t->targets + index;
}

/*
 * Search the btree for the correct target.
 */
struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector)
{
807
	unsigned int l, n = 0, k = 0;
808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850
	sector_t *node;

	for (l = 0; l < t->depth; l++) {
		n = get_child(n, k);
		node = get_node(t, l, n);

		for (k = 0; k < KEYS_PER_NODE; k++)
			if (node[k] >= sector)
				break;
	}

	return &t->targets[(KEYS_PER_NODE * n) + k];
}

void dm_table_set_restrictions(struct dm_table *t, struct request_queue *q)
{
	/*
	 * Make sure we obey the optimistic sub devices
	 * restrictions.
	 */
	q->max_sectors = t->limits.max_sectors;
	q->max_phys_segments = t->limits.max_phys_segments;
	q->max_hw_segments = t->limits.max_hw_segments;
	q->hardsect_size = t->limits.hardsect_size;
	q->max_segment_size = t->limits.max_segment_size;
	q->seg_boundary_mask = t->limits.seg_boundary_mask;
}

unsigned int dm_table_get_num_targets(struct dm_table *t)
{
	return t->num_targets;
}

struct list_head *dm_table_get_devices(struct dm_table *t)
{
	return &t->devices;
}

int dm_table_get_mode(struct dm_table *t)
{
	return t->mode;
}

851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874
void dm_table_suspend_targets(struct dm_table *t)
{
	int i;

	for (i = 0; i < t->num_targets; i++) {
		struct dm_target *ti = t->targets + i;

		if (ti->type->suspend)
			ti->type->suspend(ti);
	}
}

void dm_table_resume_targets(struct dm_table *t)
{
	int i;

	for (i = 0; i < t->num_targets; i++) {
		struct dm_target *ti = t->targets + i;

		if (ti->type->resume)
			ti->type->resume(ti);
	}
}

875 876 877 878 879 880 881 882 883 884 885 886 887 888
int dm_table_any_congested(struct dm_table *t, int bdi_bits)
{
	struct list_head *d, *devices;
	int r = 0;

	devices = dm_table_get_devices(t);
	for (d = devices->next; d != devices; d = d->next) {
		struct dm_dev *dd = list_entry(d, struct dm_dev, list);
		request_queue_t *q = bdev_get_queue(dd->bdev);
		r |= bdi_congested(&q->backing_dev_info, bdi_bits);
	}

	return r;
}
889

890 891 892 893 894 895 896 897 898 899 900 901 902
void dm_table_unplug_all(struct dm_table *t)
{
	struct list_head *d, *devices = dm_table_get_devices(t);

	for (d = devices->next; d != devices; d = d->next) {
		struct dm_dev *dd = list_entry(d, struct dm_dev, list);
		request_queue_t *q = bdev_get_queue(dd->bdev);

		if (q->unplug_fn)
			q->unplug_fn(q);
	}
}

903
EXPORT_SYMBOL(dm_vcalloc);
904 905 906
EXPORT_SYMBOL(dm_get_device);
EXPORT_SYMBOL(dm_put_device);
EXPORT_SYMBOL(dm_table_event);
907
EXPORT_SYMBOL(dm_table_get_mode);
908 909 910
EXPORT_SYMBOL(dm_table_put);
EXPORT_SYMBOL(dm_table_get);
EXPORT_SYMBOL(dm_table_unplug_all);