css.c 13.8 KB
Newer Older
Linus Torvalds's avatar
Linus Torvalds committed
1
2
3
/*
 *  drivers/s390/cio/css.c
 *  driver for channel subsystem
4
 *   $Revision: 1.96 $
Linus Torvalds's avatar
Linus Torvalds committed
5
6
7
8
 *
 *    Copyright (C) 2002 IBM Deutschland Entwicklung GmbH,
 *			 IBM Corporation
 *    Author(s): Arnd Bergmann (arndb@de.ibm.com)
9
 *		 Cornelia Huck (cornelia.huck@de.ibm.com)
Linus Torvalds's avatar
Linus Torvalds committed
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
 */
#include <linux/module.h>
#include <linux/init.h>
#include <linux/device.h>
#include <linux/slab.h>
#include <linux/errno.h>
#include <linux/list.h>

#include "css.h"
#include "cio.h"
#include "cio_debug.h"
#include "ioasm.h"
#include "chsc.h"

int need_rescan = 0;
int css_init_done = 0;
26
static int max_ssid = 0;
Linus Torvalds's avatar
Linus Torvalds committed
27

28
struct channel_subsystem *css[__MAX_CSSID + 1];
Linus Torvalds's avatar
Linus Torvalds committed
29

30
int css_characteristics_avail = 0;
Linus Torvalds's avatar
Linus Torvalds committed
31

32
33
34
35
36
37
38
39
40
inline int
for_each_subchannel(int(*fn)(struct subchannel_id, void *), void *data)
{
	struct subchannel_id schid;
	int ret;

	init_subchannel_id(&schid);
	ret = -ENODEV;
	do {
41
42
43
44
45
46
47
		do {
			ret = fn(schid, data);
			if (ret)
				break;
		} while (schid.sch_no++ < __MAX_SUBCHANNEL);
		schid.sch_no = 0;
	} while (schid.ssid++ < max_ssid);
48
49
50
	return ret;
}

Linus Torvalds's avatar
Linus Torvalds committed
51
static struct subchannel *
52
css_alloc_subchannel(struct subchannel_id schid)
Linus Torvalds's avatar
Linus Torvalds committed
53
54
55
56
57
58
59
{
	struct subchannel *sch;
	int ret;

	sch = kmalloc (sizeof (*sch), GFP_KERNEL | GFP_DMA);
	if (sch == NULL)
		return ERR_PTR(-ENOMEM);
60
	ret = cio_validate_subchannel (sch, schid);
Linus Torvalds's avatar
Linus Torvalds committed
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
	if (ret < 0) {
		kfree(sch);
		return ERR_PTR(ret);
	}

	if (sch->st != SUBCHANNEL_TYPE_IO) {
		/* For now we ignore all non-io subchannels. */
		kfree(sch);
		return ERR_PTR(-EINVAL);
	}

	/* 
	 * Set intparm to subchannel address.
	 * This is fine even on 64bit since the subchannel is always located
	 * under 2G.
	 */
	sch->schib.pmcw.intparm = (__u32)(unsigned long)sch;
	ret = cio_modify(sch);
	if (ret) {
		kfree(sch);
		return ERR_PTR(ret);
	}
	return sch;
}

static void
css_free_subchannel(struct subchannel *sch)
{
	if (sch) {
		/* Reset intparm to zeroes. */
		sch->schib.pmcw.intparm = 0;
		cio_modify(sch);
		kfree(sch);
	}
	
}

static void
css_subchannel_release(struct device *dev)
{
	struct subchannel *sch;

	sch = to_subchannel(dev);
104
	if (!cio_is_console(sch->schid))
Linus Torvalds's avatar
Linus Torvalds committed
105
106
107
108
109
110
111
112
113
114
115
		kfree(sch);
}

extern int css_get_ssd_info(struct subchannel *sch);

static int
css_register_subchannel(struct subchannel *sch)
{
	int ret;

	/* Initialize the subchannel structure */
116
	sch->dev.parent = &css[0]->device;
Linus Torvalds's avatar
Linus Torvalds committed
117
118
119
120
121
122
123
124
125
126
127
128
129
130
	sch->dev.bus = &css_bus_type;
	sch->dev.release = &css_subchannel_release;
	
	/* make it known to the system */
	ret = device_register(&sch->dev);
	if (ret)
		printk (KERN_WARNING "%s: could not register %s\n",
			__func__, sch->dev.bus_id);
	else
		css_get_ssd_info(sch);
	return ret;
}

int
131
css_probe_device(struct subchannel_id schid)
Linus Torvalds's avatar
Linus Torvalds committed
132
133
134
135
{
	int ret;
	struct subchannel *sch;

136
	sch = css_alloc_subchannel(schid);
Linus Torvalds's avatar
Linus Torvalds committed
137
138
139
140
141
142
143
144
	if (IS_ERR(sch))
		return PTR_ERR(sch);
	ret = css_register_subchannel(sch);
	if (ret)
		css_free_subchannel(sch);
	return ret;
}

145
146
147
148
static int
check_subchannel(struct device * dev, void * data)
{
	struct subchannel *sch;
149
	struct subchannel_id *schid = data;
150
151

	sch = to_subchannel(dev);
152
	return schid_equal(&sch->schid, schid);
153
154
}

Linus Torvalds's avatar
Linus Torvalds committed
155
struct subchannel *
156
get_subchannel_by_schid(struct subchannel_id schid)
Linus Torvalds's avatar
Linus Torvalds committed
157
158
159
{
	struct device *dev;

160
	dev = bus_find_device(&css_bus_type, NULL,
161
			      (void *)&schid, check_subchannel);
Linus Torvalds's avatar
Linus Torvalds committed
162

163
	return dev ? to_subchannel(dev) : NULL;
Linus Torvalds's avatar
Linus Torvalds committed
164
165
}

166

Linus Torvalds's avatar
Linus Torvalds committed
167
static inline int
168
css_get_subchannel_status(struct subchannel *sch, struct subchannel_id schid)
Linus Torvalds's avatar
Linus Torvalds committed
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
{
	struct schib schib;
	int cc;

	cc = stsch(schid, &schib);
	if (cc)
		return CIO_GONE;
	if (!schib.pmcw.dnv)
		return CIO_GONE;
	if (sch && sch->schib.pmcw.dnv &&
	    (schib.pmcw.dev != sch->schib.pmcw.dev))
		return CIO_REVALIDATE;
	if (sch && !sch->lpm)
		return CIO_NO_PATH;
	return CIO_OPER;
}
	
static int
187
css_evaluate_subchannel(struct subchannel_id schid, int slow)
Linus Torvalds's avatar
Linus Torvalds committed
188
189
190
191
192
{
	int event, ret, disc;
	struct subchannel *sch;
	unsigned long flags;

193
	sch = get_subchannel_by_schid(schid);
Linus Torvalds's avatar
Linus Torvalds committed
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
	disc = sch ? device_is_disconnected(sch) : 0;
	if (disc && slow) {
		if (sch)
			put_device(&sch->dev);
		return 0; /* Already processed. */
	}
	/*
	 * We've got a machine check, so running I/O won't get an interrupt.
	 * Kill any pending timers.
	 */
	if (sch)
		device_kill_pending_timer(sch);
	if (!disc && !slow) {
		if (sch)
			put_device(&sch->dev);
		return -EAGAIN; /* Will be done on the slow path. */
	}
211
	event = css_get_subchannel_status(sch, schid);
212
213
	CIO_MSG_EVENT(4, "Evaluating schid 0.%x.%04x, event %d, %s, %s path.\n",
		      schid.ssid, schid.sch_no, event,
214
		      sch?(disc?"disconnected":"normal"):"unknown",
Linus Torvalds's avatar
Linus Torvalds committed
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
		      slow?"slow":"fast");
	switch (event) {
	case CIO_NO_PATH:
	case CIO_GONE:
		if (!sch) {
			/* Never used this subchannel. Ignore. */
			ret = 0;
			break;
		}
		if (disc && (event == CIO_NO_PATH)) {
			/*
			 * Uargh, hack again. Because we don't get a machine
			 * check on configure on, our path bookkeeping can
			 * be out of date here (it's fine while we only do
			 * logical varying or get chsc machine checks). We
			 * need to force reprobing or we might miss devices
			 * coming operational again. It won't do harm in real
			 * no path situations.
			 */
			spin_lock_irqsave(&sch->lock, flags);
			device_trigger_reprobe(sch);
			spin_unlock_irqrestore(&sch->lock, flags);
			ret = 0;
			break;
		}
		if (sch->driver && sch->driver->notify &&
		    sch->driver->notify(&sch->dev, event)) {
			cio_disable_subchannel(sch);
			device_set_disconnected(sch);
			ret = 0;
			break;
		}
		/*
		 * Unregister subchannel.
		 * The device will be killed automatically.
		 */
		cio_disable_subchannel(sch);
		device_unregister(&sch->dev);
		/* Reset intparm to zeroes. */
		sch->schib.pmcw.intparm = 0;
		cio_modify(sch);
		put_device(&sch->dev);
		ret = 0;
		break;
	case CIO_REVALIDATE:
		/* 
		 * Revalidation machine check. Sick.
		 * We don't notify the driver since we have to throw the device
		 * away in any case.
		 */
		if (!disc) {
			device_unregister(&sch->dev);
			/* Reset intparm to zeroes. */
			sch->schib.pmcw.intparm = 0;
			cio_modify(sch);
			put_device(&sch->dev);
271
			ret = css_probe_device(schid);
Linus Torvalds's avatar
Linus Torvalds committed
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
		} else {
			/*
			 * We can't immediately deregister the disconnected
			 * device since it might block.
			 */
			spin_lock_irqsave(&sch->lock, flags);
			device_trigger_reprobe(sch);
			spin_unlock_irqrestore(&sch->lock, flags);
			ret = 0;
		}
		break;
	case CIO_OPER:
		if (disc) {
			spin_lock_irqsave(&sch->lock, flags);
			/* Get device operational again. */
			device_trigger_reprobe(sch);
			spin_unlock_irqrestore(&sch->lock, flags);
		}
290
		ret = sch ? 0 : css_probe_device(schid);
Linus Torvalds's avatar
Linus Torvalds committed
291
292
293
294
295
296
297
298
		break;
	default:
		BUG();
		ret = 0;
	}
	return ret;
}

299
300
static int
css_rescan_devices(struct subchannel_id schid, void *data)
Linus Torvalds's avatar
Linus Torvalds committed
301
{
302
	return css_evaluate_subchannel(schid, 1);
Linus Torvalds's avatar
Linus Torvalds committed
303
304
305
306
}

struct slow_subchannel {
	struct list_head slow_list;
307
	struct subchannel_id schid;
Linus Torvalds's avatar
Linus Torvalds committed
308
309
310
311
312
313
314
315
316
317
318
319
};

static LIST_HEAD(slow_subchannels_head);
static DEFINE_SPINLOCK(slow_subchannel_lock);

static void
css_trigger_slow_path(void)
{
	CIO_TRACE_EVENT(4, "slowpath");

	if (need_rescan) {
		need_rescan = 0;
320
		for_each_subchannel(css_rescan_devices, NULL);
Linus Torvalds's avatar
Linus Torvalds committed
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
		return;
	}

	spin_lock_irq(&slow_subchannel_lock);
	while (!list_empty(&slow_subchannels_head)) {
		struct slow_subchannel *slow_sch =
			list_entry(slow_subchannels_head.next,
				   struct slow_subchannel, slow_list);

		list_del_init(slow_subchannels_head.next);
		spin_unlock_irq(&slow_subchannel_lock);
		css_evaluate_subchannel(slow_sch->schid, 1);
		spin_lock_irq(&slow_subchannel_lock);
		kfree(slow_sch);
	}
	spin_unlock_irq(&slow_subchannel_lock);
}

typedef void (*workfunc)(void *);
DECLARE_WORK(slow_path_work, (workfunc)css_trigger_slow_path, NULL);
struct workqueue_struct *slow_path_wq;

/*
 * Rescan for new devices. FIXME: This is slow.
 * This function is called when we have lost CRWs due to overflows and we have
 * to do subchannel housekeeping.
 */
void
css_reiterate_subchannels(void)
{
	css_clear_subchannel_slow_list();
	need_rescan = 1;
}

/*
 * Called from the machine check handler for subchannel report words.
 */
int
359
css_process_crw(int rsid1, int rsid2)
Linus Torvalds's avatar
Linus Torvalds committed
360
361
{
	int ret;
362
	struct subchannel_id mchk_schid;
Linus Torvalds's avatar
Linus Torvalds committed
363

364
365
	CIO_CRW_EVENT(2, "source is subchannel %04X, subsystem id %x\n",
		      rsid1, rsid2);
Linus Torvalds's avatar
Linus Torvalds committed
366
367
368
369

	if (need_rescan)
		/* We need to iterate all subchannels anyway. */
		return -EAGAIN;
370
371

	init_subchannel_id(&mchk_schid);
372
373
374
375
	mchk_schid.sch_no = rsid1;
	if (rsid2 != 0)
		mchk_schid.ssid = (rsid2 >> 8) & 3;

Linus Torvalds's avatar
Linus Torvalds committed
376
377
378
379
380
	/* 
	 * Since we are always presented with IPI in the CRW, we have to
	 * use stsch() to find out if the subchannel in question has come
	 * or gone.
	 */
381
	ret = css_evaluate_subchannel(mchk_schid, 0);
Linus Torvalds's avatar
Linus Torvalds committed
382
	if (ret == -EAGAIN) {
383
		if (css_enqueue_subchannel_slow(mchk_schid)) {
Linus Torvalds's avatar
Linus Torvalds committed
384
385
386
387
388
389
390
			css_clear_subchannel_slow_list();
			need_rescan = 1;
		}
	}
	return ret;
}

391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
static int __init
__init_channel_subsystem(struct subchannel_id schid, void *data)
{
	struct subchannel *sch;
	int ret;

	if (cio_is_console(schid))
		sch = cio_get_console_subchannel();
	else {
		sch = css_alloc_subchannel(schid);
		if (IS_ERR(sch))
			ret = PTR_ERR(sch);
		else
			ret = 0;
		switch (ret) {
		case 0:
			break;
		case -ENOMEM:
			panic("Out of memory in init_channel_subsystem\n");
		/* -ENXIO: no more subchannels. */
		case -ENXIO:
			return ret;
		default:
			return 0;
		}
	}
	/*
	 * We register ALL valid subchannels in ioinfo, even those
	 * that have been present before init_channel_subsystem.
	 * These subchannels can't have been registered yet (kmalloc
	 * not working) so we do it now. This is true e.g. for the
	 * console subchannel.
	 */
	css_register_subchannel(sch);
	return 0;
}

Linus Torvalds's avatar
Linus Torvalds committed
428
static void __init
429
css_generate_pgid(struct channel_subsystem *css, u32 tod_high)
Linus Torvalds's avatar
Linus Torvalds committed
430
{
431
432
433
434
	if (css_characteristics_avail && css_general_characteristics.mcss) {
		css->global_pgid.pgid_high.ext_cssid.version = 0x80;
		css->global_pgid.pgid_high.ext_cssid.cssid = css->cssid;
	} else {
Linus Torvalds's avatar
Linus Torvalds committed
435
#ifdef CONFIG_SMP
436
		css->global_pgid.pgid_high.cpu_addr = hard_smp_processor_id();
Linus Torvalds's avatar
Linus Torvalds committed
437
#else
438
		css->global_pgid.pgid_high.cpu_addr = 0;
Linus Torvalds's avatar
Linus Torvalds committed
439
440
#endif
	}
441
442
443
444
445
446
	css->global_pgid.cpu_id = ((cpuid_t *) __LC_CPUID)->ident;
	css->global_pgid.cpu_model = ((cpuid_t *) __LC_CPUID)->machine;
	css->global_pgid.tod_high = tod_high;

}

447
448
449
450
451
452
453
454
455
static void
channel_subsystem_release(struct device *dev)
{
	struct channel_subsystem *css;

	css = to_css(dev);
	kfree(css);
}

456
457
458
459
460
461
462
463
464
static inline void __init
setup_css(int nr)
{
	u32 tod_high;

	memset(css[nr], 0, sizeof(struct channel_subsystem));
	css[nr]->valid = 1;
	css[nr]->cssid = nr;
	sprintf(css[nr]->device.bus_id, "css%x", nr);
465
	css[nr]->device.release = channel_subsystem_release;
466
467
	tod_high = (u32) (get_clock() >> 32);
	css_generate_pgid(css[nr], tod_high);
Linus Torvalds's avatar
Linus Torvalds committed
468
469
470
471
472
473
474
475
476
477
}

/*
 * Now that the driver core is running, we can setup our channel subsystem.
 * The struct subchannel's are created during probing (except for the
 * static console subchannel).
 */
static int __init
init_channel_subsystem (void)
{
478
	int ret, i;
Linus Torvalds's avatar
Linus Torvalds committed
479
480
481
482
483
484
485

	if (chsc_determine_css_characteristics() == 0)
		css_characteristics_avail = 1;

	if ((ret = bus_register(&css_bus_type)))
		goto out;

486
487
488
489
490
491
492
493
494
495
496
	/* Try to enable MSS. */
	ret = chsc_enable_facility(CHSC_SDA_OC_MSS);
	switch (ret) {
	case 0: /* Success. */
		max_ssid = __MAX_SSID;
		break;
	case -ENOMEM:
		goto out_bus;
	default:
		max_ssid = 0;
	}
497
498
499
500
501
	/* Setup css structure. */
	for (i = 0; i <= __MAX_CSSID; i++) {
		css[i] = kmalloc(sizeof(struct channel_subsystem), GFP_KERNEL);
		if (!css[i]) {
			ret = -ENOMEM;
502
			goto out_unregister;
503
504
505
506
507
508
		}
		setup_css(i);
		ret = device_register(&css[i]->device);
		if (ret)
			goto out_free;
	}
Linus Torvalds's avatar
Linus Torvalds committed
509
510
511
512
	css_init_done = 1;

	ctl_set_bit(6, 28);

513
	for_each_subchannel(__init_channel_subsystem, NULL);
Linus Torvalds's avatar
Linus Torvalds committed
514
	return 0;
515
516
out_free:
	kfree(css[i]);
517
out_unregister:
518
519
520
521
	while (i > 0) {
		i--;
		device_unregister(&css[i]->device);
	}
522
out_bus:
Linus Torvalds's avatar
Linus Torvalds committed
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
	bus_unregister(&css_bus_type);
out:
	return ret;
}

/*
 * find a driver for a subchannel. They identify by the subchannel
 * type with the exception that the console subchannel driver has its own
 * subchannel type although the device is an i/o subchannel
 */
static int
css_bus_match (struct device *dev, struct device_driver *drv)
{
	struct subchannel *sch = container_of (dev, struct subchannel, dev);
	struct css_driver *driver = container_of (drv, struct css_driver, drv);

	if (sch->st == driver->subchannel_type)
		return 1;

	return 0;
}

545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
static int
css_probe (struct device *dev)
{
	struct subchannel *sch;

	sch = to_subchannel(dev);
	sch->driver = container_of (dev->driver, struct css_driver, drv);
	return (sch->driver->probe ? sch->driver->probe(sch) : 0);
}

static int
css_remove (struct device *dev)
{
	struct subchannel *sch;

	sch = to_subchannel(dev);
	return (sch->driver->remove ? sch->driver->remove(sch) : 0);
}

static void
css_shutdown (struct device *dev)
{
	struct subchannel *sch;

	sch = to_subchannel(dev);
	if (sch->driver->shutdown)
		sch->driver->shutdown(sch);
}

Linus Torvalds's avatar
Linus Torvalds committed
574
struct bus_type css_bus_type = {
575
576
577
578
579
	.name     = "css",
	.match    = css_bus_match,
	.probe    = css_probe,
	.remove   = css_remove,
	.shutdown = css_shutdown,
Linus Torvalds's avatar
Linus Torvalds committed
580
581
582
583
584
};

subsys_initcall(init_channel_subsystem);

int
585
css_enqueue_subchannel_slow(struct subchannel_id schid)
Linus Torvalds's avatar
Linus Torvalds committed
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
{
	struct slow_subchannel *new_slow_sch;
	unsigned long flags;

	new_slow_sch = kmalloc(sizeof(struct slow_subchannel), GFP_ATOMIC);
	if (!new_slow_sch)
		return -ENOMEM;
	memset(new_slow_sch, 0, sizeof(struct slow_subchannel));
	new_slow_sch->schid = schid;
	spin_lock_irqsave(&slow_subchannel_lock, flags);
	list_add_tail(&new_slow_sch->slow_list, &slow_subchannels_head);
	spin_unlock_irqrestore(&slow_subchannel_lock, flags);
	return 0;
}

void
css_clear_subchannel_slow_list(void)
{
	unsigned long flags;

	spin_lock_irqsave(&slow_subchannel_lock, flags);
	while (!list_empty(&slow_subchannels_head)) {
		struct slow_subchannel *slow_sch =
			list_entry(slow_subchannels_head.next,
				   struct slow_subchannel, slow_list);

		list_del_init(slow_subchannels_head.next);
		kfree(slow_sch);
	}
	spin_unlock_irqrestore(&slow_subchannel_lock, flags);
}



int
css_slow_subchannels_exist(void)
{
	return (!list_empty(&slow_subchannels_head));
}

MODULE_LICENSE("GPL");
EXPORT_SYMBOL(css_bus_type);
EXPORT_SYMBOL_GPL(css_characteristics_avail);