Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
awilliam
GitHub Repository: awilliam/linux-vfio
Path: blob/master/arch/alpha/kernel/core_wildfire.c
10817 views
1
/*
2
* linux/arch/alpha/kernel/core_wildfire.c
3
*
4
* Wildfire support.
5
*
6
* Copyright (C) 2000 Andrea Arcangeli <[email protected]> SuSE
7
*/
8
9
#define __EXTERN_INLINE inline
10
#include <asm/io.h>
11
#include <asm/core_wildfire.h>
12
#undef __EXTERN_INLINE
13
14
#include <linux/types.h>
15
#include <linux/pci.h>
16
#include <linux/sched.h>
17
#include <linux/init.h>
18
19
#include <asm/ptrace.h>
20
#include <asm/smp.h>
21
22
#include "proto.h"
23
#include "pci_impl.h"
24
25
#define DEBUG_CONFIG 0
26
#define DEBUG_DUMP_REGS 0
27
#define DEBUG_DUMP_CONFIG 1
28
29
#if DEBUG_CONFIG
30
# define DBG_CFG(args) printk args
31
#else
32
# define DBG_CFG(args)
33
#endif
34
35
#if DEBUG_DUMP_REGS
36
static void wildfire_dump_pci_regs(int qbbno, int hoseno);
37
static void wildfire_dump_pca_regs(int qbbno, int pcano);
38
static void wildfire_dump_qsa_regs(int qbbno);
39
static void wildfire_dump_qsd_regs(int qbbno);
40
static void wildfire_dump_iop_regs(int qbbno);
41
static void wildfire_dump_gp_regs(int qbbno);
42
#endif
43
#if DEBUG_DUMP_CONFIG
44
static void wildfire_dump_hardware_config(void);
45
#endif
46
47
unsigned char wildfire_hard_qbb_map[WILDFIRE_MAX_QBB];
48
unsigned char wildfire_soft_qbb_map[WILDFIRE_MAX_QBB];
49
#define QBB_MAP_EMPTY 0xff
50
51
unsigned long wildfire_hard_qbb_mask;
52
unsigned long wildfire_soft_qbb_mask;
53
unsigned long wildfire_gp_mask;
54
unsigned long wildfire_hs_mask;
55
unsigned long wildfire_iop_mask;
56
unsigned long wildfire_ior_mask;
57
unsigned long wildfire_pca_mask;
58
unsigned long wildfire_cpu_mask;
59
unsigned long wildfire_mem_mask;
60
61
void __init
62
wildfire_init_hose(int qbbno, int hoseno)
63
{
64
struct pci_controller *hose;
65
wildfire_pci *pci;
66
67
hose = alloc_pci_controller();
68
hose->io_space = alloc_resource();
69
hose->mem_space = alloc_resource();
70
71
/* This is for userland consumption. */
72
hose->sparse_mem_base = 0;
73
hose->sparse_io_base = 0;
74
hose->dense_mem_base = WILDFIRE_MEM(qbbno, hoseno);
75
hose->dense_io_base = WILDFIRE_IO(qbbno, hoseno);
76
77
hose->config_space_base = WILDFIRE_CONF(qbbno, hoseno);
78
hose->index = (qbbno << 3) + hoseno;
79
80
hose->io_space->start = WILDFIRE_IO(qbbno, hoseno) - WILDFIRE_IO_BIAS;
81
hose->io_space->end = hose->io_space->start + WILDFIRE_IO_SPACE - 1;
82
hose->io_space->name = pci_io_names[hoseno];
83
hose->io_space->flags = IORESOURCE_IO;
84
85
hose->mem_space->start = WILDFIRE_MEM(qbbno, hoseno)-WILDFIRE_MEM_BIAS;
86
hose->mem_space->end = hose->mem_space->start + 0xffffffff;
87
hose->mem_space->name = pci_mem_names[hoseno];
88
hose->mem_space->flags = IORESOURCE_MEM;
89
90
if (request_resource(&ioport_resource, hose->io_space) < 0)
91
printk(KERN_ERR "Failed to request IO on qbb %d hose %d\n",
92
qbbno, hoseno);
93
if (request_resource(&iomem_resource, hose->mem_space) < 0)
94
printk(KERN_ERR "Failed to request MEM on qbb %d hose %d\n",
95
qbbno, hoseno);
96
97
#if DEBUG_DUMP_REGS
98
wildfire_dump_pci_regs(qbbno, hoseno);
99
#endif
100
101
/*
102
* Set up the PCI to main memory translation windows.
103
*
104
* Note: Window 3 is scatter-gather only
105
*
106
* Window 0 is scatter-gather 8MB at 8MB (for isa)
107
* Window 1 is direct access 1GB at 1GB
108
* Window 2 is direct access 1GB at 2GB
109
* Window 3 is scatter-gather 128MB at 3GB
110
* ??? We ought to scale window 3 memory.
111
*
112
*/
113
hose->sg_isa = iommu_arena_new(hose, 0x00800000, 0x00800000, 0);
114
hose->sg_pci = iommu_arena_new(hose, 0xc0000000, 0x08000000, 0);
115
116
pci = WILDFIRE_pci(qbbno, hoseno);
117
118
pci->pci_window[0].wbase.csr = hose->sg_isa->dma_base | 3;
119
pci->pci_window[0].wmask.csr = (hose->sg_isa->size - 1) & 0xfff00000;
120
pci->pci_window[0].tbase.csr = virt_to_phys(hose->sg_isa->ptes);
121
122
pci->pci_window[1].wbase.csr = 0x40000000 | 1;
123
pci->pci_window[1].wmask.csr = (0x40000000 -1) & 0xfff00000;
124
pci->pci_window[1].tbase.csr = 0;
125
126
pci->pci_window[2].wbase.csr = 0x80000000 | 1;
127
pci->pci_window[2].wmask.csr = (0x40000000 -1) & 0xfff00000;
128
pci->pci_window[2].tbase.csr = 0x40000000;
129
130
pci->pci_window[3].wbase.csr = hose->sg_pci->dma_base | 3;
131
pci->pci_window[3].wmask.csr = (hose->sg_pci->size - 1) & 0xfff00000;
132
pci->pci_window[3].tbase.csr = virt_to_phys(hose->sg_pci->ptes);
133
134
wildfire_pci_tbi(hose, 0, 0); /* Flush TLB at the end. */
135
}
136
137
void __init
138
wildfire_init_pca(int qbbno, int pcano)
139
{
140
141
/* Test for PCA existence first. */
142
if (!WILDFIRE_PCA_EXISTS(qbbno, pcano))
143
return;
144
145
#if DEBUG_DUMP_REGS
146
wildfire_dump_pca_regs(qbbno, pcano);
147
#endif
148
149
/* Do both hoses of the PCA. */
150
wildfire_init_hose(qbbno, (pcano << 1) + 0);
151
wildfire_init_hose(qbbno, (pcano << 1) + 1);
152
}
153
154
void __init
155
wildfire_init_qbb(int qbbno)
156
{
157
int pcano;
158
159
/* Test for QBB existence first. */
160
if (!WILDFIRE_QBB_EXISTS(qbbno))
161
return;
162
163
#if DEBUG_DUMP_REGS
164
wildfire_dump_qsa_regs(qbbno);
165
wildfire_dump_qsd_regs(qbbno);
166
wildfire_dump_iop_regs(qbbno);
167
wildfire_dump_gp_regs(qbbno);
168
#endif
169
170
/* Init all PCAs here. */
171
for (pcano = 0; pcano < WILDFIRE_PCA_PER_QBB; pcano++) {
172
wildfire_init_pca(qbbno, pcano);
173
}
174
}
175
176
void __init
177
wildfire_hardware_probe(void)
178
{
179
unsigned long temp;
180
unsigned int hard_qbb, soft_qbb;
181
wildfire_fast_qsd *fast = WILDFIRE_fast_qsd();
182
wildfire_qsd *qsd;
183
wildfire_qsa *qsa;
184
wildfire_iop *iop;
185
wildfire_gp *gp;
186
wildfire_ne *ne;
187
wildfire_fe *fe;
188
int i;
189
190
temp = fast->qsd_whami.csr;
191
#if 0
192
printk(KERN_ERR "fast QSD_WHAMI at base %p is 0x%lx\n", fast, temp);
193
#endif
194
195
hard_qbb = (temp >> 8) & 7;
196
soft_qbb = (temp >> 4) & 7;
197
198
/* Init the HW configuration variables. */
199
wildfire_hard_qbb_mask = (1 << hard_qbb);
200
wildfire_soft_qbb_mask = (1 << soft_qbb);
201
202
wildfire_gp_mask = 0;
203
wildfire_hs_mask = 0;
204
wildfire_iop_mask = 0;
205
wildfire_ior_mask = 0;
206
wildfire_pca_mask = 0;
207
208
wildfire_cpu_mask = 0;
209
wildfire_mem_mask = 0;
210
211
memset(wildfire_hard_qbb_map, QBB_MAP_EMPTY, WILDFIRE_MAX_QBB);
212
memset(wildfire_soft_qbb_map, QBB_MAP_EMPTY, WILDFIRE_MAX_QBB);
213
214
/* First, determine which QBBs are present. */
215
qsa = WILDFIRE_qsa(soft_qbb);
216
217
temp = qsa->qsa_qbb_id.csr;
218
#if 0
219
printk(KERN_ERR "QSA_QBB_ID at base %p is 0x%lx\n", qsa, temp);
220
#endif
221
222
if (temp & 0x40) /* Is there an HS? */
223
wildfire_hs_mask = 1;
224
225
if (temp & 0x20) { /* Is there a GP? */
226
gp = WILDFIRE_gp(soft_qbb);
227
temp = 0;
228
for (i = 0; i < 4; i++) {
229
temp |= gp->gpa_qbb_map[i].csr << (i * 8);
230
#if 0
231
printk(KERN_ERR "GPA_QBB_MAP[%d] at base %p is 0x%lx\n",
232
i, gp, temp);
233
#endif
234
}
235
236
for (hard_qbb = 0; hard_qbb < WILDFIRE_MAX_QBB; hard_qbb++) {
237
if (temp & 8) { /* Is there a QBB? */
238
soft_qbb = temp & 7;
239
wildfire_hard_qbb_mask |= (1 << hard_qbb);
240
wildfire_soft_qbb_mask |= (1 << soft_qbb);
241
}
242
temp >>= 4;
243
}
244
wildfire_gp_mask = wildfire_soft_qbb_mask;
245
}
246
247
/* Next determine each QBBs resources. */
248
for (soft_qbb = 0; soft_qbb < WILDFIRE_MAX_QBB; soft_qbb++) {
249
if (WILDFIRE_QBB_EXISTS(soft_qbb)) {
250
qsd = WILDFIRE_qsd(soft_qbb);
251
temp = qsd->qsd_whami.csr;
252
#if 0
253
printk(KERN_ERR "QSD_WHAMI at base %p is 0x%lx\n", qsd, temp);
254
#endif
255
hard_qbb = (temp >> 8) & 7;
256
wildfire_hard_qbb_map[hard_qbb] = soft_qbb;
257
wildfire_soft_qbb_map[soft_qbb] = hard_qbb;
258
259
qsa = WILDFIRE_qsa(soft_qbb);
260
temp = qsa->qsa_qbb_pop[0].csr;
261
#if 0
262
printk(KERN_ERR "QSA_QBB_POP_0 at base %p is 0x%lx\n", qsa, temp);
263
#endif
264
wildfire_cpu_mask |= ((temp >> 0) & 0xf) << (soft_qbb << 2);
265
wildfire_mem_mask |= ((temp >> 4) & 0xf) << (soft_qbb << 2);
266
267
temp = qsa->qsa_qbb_pop[1].csr;
268
#if 0
269
printk(KERN_ERR "QSA_QBB_POP_1 at base %p is 0x%lx\n", qsa, temp);
270
#endif
271
wildfire_iop_mask |= (1 << soft_qbb);
272
wildfire_ior_mask |= ((temp >> 4) & 0xf) << (soft_qbb << 2);
273
274
temp = qsa->qsa_qbb_id.csr;
275
#if 0
276
printk(KERN_ERR "QSA_QBB_ID at %p is 0x%lx\n", qsa, temp);
277
#endif
278
if (temp & 0x20)
279
wildfire_gp_mask |= (1 << soft_qbb);
280
281
/* Probe for PCA existence here. */
282
for (i = 0; i < WILDFIRE_PCA_PER_QBB; i++) {
283
iop = WILDFIRE_iop(soft_qbb);
284
ne = WILDFIRE_ne(soft_qbb, i);
285
fe = WILDFIRE_fe(soft_qbb, i);
286
287
if ((iop->iop_hose[i].init.csr & 1) == 1 &&
288
((ne->ne_what_am_i.csr & 0xf00000300UL) == 0x100000300UL) &&
289
((fe->fe_what_am_i.csr & 0xf00000300UL) == 0x100000200UL))
290
{
291
wildfire_pca_mask |= 1 << ((soft_qbb << 2) + i);
292
}
293
}
294
295
}
296
}
297
#if DEBUG_DUMP_CONFIG
298
wildfire_dump_hardware_config();
299
#endif
300
}
301
302
void __init
303
wildfire_init_arch(void)
304
{
305
int qbbno;
306
307
/* With multiple PCI buses, we play with I/O as physical addrs. */
308
ioport_resource.end = ~0UL;
309
310
311
/* Probe the hardware for info about configuration. */
312
wildfire_hardware_probe();
313
314
/* Now init all the found QBBs. */
315
for (qbbno = 0; qbbno < WILDFIRE_MAX_QBB; qbbno++) {
316
wildfire_init_qbb(qbbno);
317
}
318
319
/* Normal direct PCI DMA mapping. */
320
__direct_map_base = 0x40000000UL;
321
__direct_map_size = 0x80000000UL;
322
}
323
324
void
325
wildfire_machine_check(unsigned long vector, unsigned long la_ptr)
326
{
327
mb();
328
mb(); /* magic */
329
draina();
330
/* FIXME: clear pci errors */
331
wrmces(0x7);
332
mb();
333
334
process_mcheck_info(vector, la_ptr, "WILDFIRE",
335
mcheck_expected(smp_processor_id()));
336
}
337
338
void
339
wildfire_kill_arch(int mode)
340
{
341
}
342
343
void
344
wildfire_pci_tbi(struct pci_controller *hose, dma_addr_t start, dma_addr_t end)
345
{
346
int qbbno = hose->index >> 3;
347
int hoseno = hose->index & 7;
348
wildfire_pci *pci = WILDFIRE_pci(qbbno, hoseno);
349
350
mb();
351
pci->pci_flush_tlb.csr; /* reading does the trick */
352
}
353
354
static int
355
mk_conf_addr(struct pci_bus *pbus, unsigned int device_fn, int where,
356
unsigned long *pci_addr, unsigned char *type1)
357
{
358
struct pci_controller *hose = pbus->sysdata;
359
unsigned long addr;
360
u8 bus = pbus->number;
361
362
DBG_CFG(("mk_conf_addr(bus=%d ,device_fn=0x%x, where=0x%x, "
363
"pci_addr=0x%p, type1=0x%p)\n",
364
bus, device_fn, where, pci_addr, type1));
365
366
if (!pbus->parent) /* No parent means peer PCI bus. */
367
bus = 0;
368
*type1 = (bus != 0);
369
370
addr = (bus << 16) | (device_fn << 8) | where;
371
addr |= hose->config_space_base;
372
373
*pci_addr = addr;
374
DBG_CFG(("mk_conf_addr: returning pci_addr 0x%lx\n", addr));
375
return 0;
376
}
377
378
static int
379
wildfire_read_config(struct pci_bus *bus, unsigned int devfn, int where,
380
int size, u32 *value)
381
{
382
unsigned long addr;
383
unsigned char type1;
384
385
if (mk_conf_addr(bus, devfn, where, &addr, &type1))
386
return PCIBIOS_DEVICE_NOT_FOUND;
387
388
switch (size) {
389
case 1:
390
*value = __kernel_ldbu(*(vucp)addr);
391
break;
392
case 2:
393
*value = __kernel_ldwu(*(vusp)addr);
394
break;
395
case 4:
396
*value = *(vuip)addr;
397
break;
398
}
399
400
return PCIBIOS_SUCCESSFUL;
401
}
402
403
static int
404
wildfire_write_config(struct pci_bus *bus, unsigned int devfn, int where,
405
int size, u32 value)
406
{
407
unsigned long addr;
408
unsigned char type1;
409
410
if (mk_conf_addr(bus, devfn, where, &addr, &type1))
411
return PCIBIOS_DEVICE_NOT_FOUND;
412
413
switch (size) {
414
case 1:
415
__kernel_stb(value, *(vucp)addr);
416
mb();
417
__kernel_ldbu(*(vucp)addr);
418
break;
419
case 2:
420
__kernel_stw(value, *(vusp)addr);
421
mb();
422
__kernel_ldwu(*(vusp)addr);
423
break;
424
case 4:
425
*(vuip)addr = value;
426
mb();
427
*(vuip)addr;
428
break;
429
}
430
431
return PCIBIOS_SUCCESSFUL;
432
}
433
434
struct pci_ops wildfire_pci_ops =
435
{
436
.read = wildfire_read_config,
437
.write = wildfire_write_config,
438
};
439
440
441
/*
442
* NUMA Support
443
*/
444
int wildfire_pa_to_nid(unsigned long pa)
445
{
446
return pa >> 36;
447
}
448
449
int wildfire_cpuid_to_nid(int cpuid)
450
{
451
/* assume 4 CPUs per node */
452
return cpuid >> 2;
453
}
454
455
unsigned long wildfire_node_mem_start(int nid)
456
{
457
/* 64GB per node */
458
return (unsigned long)nid * (64UL * 1024 * 1024 * 1024);
459
}
460
461
unsigned long wildfire_node_mem_size(int nid)
462
{
463
/* 64GB per node */
464
return 64UL * 1024 * 1024 * 1024;
465
}
466
467
#if DEBUG_DUMP_REGS
468
469
static void __init
470
wildfire_dump_pci_regs(int qbbno, int hoseno)
471
{
472
wildfire_pci *pci = WILDFIRE_pci(qbbno, hoseno);
473
int i;
474
475
printk(KERN_ERR "PCI registers for QBB %d hose %d (%p)\n",
476
qbbno, hoseno, pci);
477
478
printk(KERN_ERR " PCI_IO_ADDR_EXT: 0x%16lx\n",
479
pci->pci_io_addr_ext.csr);
480
printk(KERN_ERR " PCI_CTRL: 0x%16lx\n", pci->pci_ctrl.csr);
481
printk(KERN_ERR " PCI_ERR_SUM: 0x%16lx\n", pci->pci_err_sum.csr);
482
printk(KERN_ERR " PCI_ERR_ADDR: 0x%16lx\n", pci->pci_err_addr.csr);
483
printk(KERN_ERR " PCI_STALL_CNT: 0x%16lx\n", pci->pci_stall_cnt.csr);
484
printk(KERN_ERR " PCI_PEND_INT: 0x%16lx\n", pci->pci_pend_int.csr);
485
printk(KERN_ERR " PCI_SENT_INT: 0x%16lx\n", pci->pci_sent_int.csr);
486
487
printk(KERN_ERR " DMA window registers for QBB %d hose %d (%p)\n",
488
qbbno, hoseno, pci);
489
for (i = 0; i < 4; i++) {
490
printk(KERN_ERR " window %d: 0x%16lx 0x%16lx 0x%16lx\n", i,
491
pci->pci_window[i].wbase.csr,
492
pci->pci_window[i].wmask.csr,
493
pci->pci_window[i].tbase.csr);
494
}
495
printk(KERN_ERR "\n");
496
}
497
498
static void __init
499
wildfire_dump_pca_regs(int qbbno, int pcano)
500
{
501
wildfire_pca *pca = WILDFIRE_pca(qbbno, pcano);
502
int i;
503
504
printk(KERN_ERR "PCA registers for QBB %d PCA %d (%p)\n",
505
qbbno, pcano, pca);
506
507
printk(KERN_ERR " PCA_WHAT_AM_I: 0x%16lx\n", pca->pca_what_am_i.csr);
508
printk(KERN_ERR " PCA_ERR_SUM: 0x%16lx\n", pca->pca_err_sum.csr);
509
printk(KERN_ERR " PCA_PEND_INT: 0x%16lx\n", pca->pca_pend_int.csr);
510
printk(KERN_ERR " PCA_SENT_INT: 0x%16lx\n", pca->pca_sent_int.csr);
511
printk(KERN_ERR " PCA_STDIO_EL: 0x%16lx\n",
512
pca->pca_stdio_edge_level.csr);
513
514
printk(KERN_ERR " PCA target registers for QBB %d PCA %d (%p)\n",
515
qbbno, pcano, pca);
516
for (i = 0; i < 4; i++) {
517
printk(KERN_ERR " target %d: 0x%16lx 0x%16lx\n", i,
518
pca->pca_int[i].target.csr,
519
pca->pca_int[i].enable.csr);
520
}
521
522
printk(KERN_ERR "\n");
523
}
524
525
static void __init
526
wildfire_dump_qsa_regs(int qbbno)
527
{
528
wildfire_qsa *qsa = WILDFIRE_qsa(qbbno);
529
int i;
530
531
printk(KERN_ERR "QSA registers for QBB %d (%p)\n", qbbno, qsa);
532
533
printk(KERN_ERR " QSA_QBB_ID: 0x%16lx\n", qsa->qsa_qbb_id.csr);
534
printk(KERN_ERR " QSA_PORT_ENA: 0x%16lx\n", qsa->qsa_port_ena.csr);
535
printk(KERN_ERR " QSA_REF_INT: 0x%16lx\n", qsa->qsa_ref_int.csr);
536
537
for (i = 0; i < 5; i++)
538
printk(KERN_ERR " QSA_CONFIG_%d: 0x%16lx\n",
539
i, qsa->qsa_config[i].csr);
540
541
for (i = 0; i < 2; i++)
542
printk(KERN_ERR " QSA_QBB_POP_%d: 0x%16lx\n",
543
i, qsa->qsa_qbb_pop[0].csr);
544
545
printk(KERN_ERR "\n");
546
}
547
548
static void __init
549
wildfire_dump_qsd_regs(int qbbno)
550
{
551
wildfire_qsd *qsd = WILDFIRE_qsd(qbbno);
552
553
printk(KERN_ERR "QSD registers for QBB %d (%p)\n", qbbno, qsd);
554
555
printk(KERN_ERR " QSD_WHAMI: 0x%16lx\n", qsd->qsd_whami.csr);
556
printk(KERN_ERR " QSD_REV: 0x%16lx\n", qsd->qsd_rev.csr);
557
printk(KERN_ERR " QSD_PORT_PRESENT: 0x%16lx\n",
558
qsd->qsd_port_present.csr);
559
printk(KERN_ERR " QSD_PORT_ACTUVE: 0x%16lx\n",
560
qsd->qsd_port_active.csr);
561
printk(KERN_ERR " QSD_FAULT_ENA: 0x%16lx\n",
562
qsd->qsd_fault_ena.csr);
563
printk(KERN_ERR " QSD_CPU_INT_ENA: 0x%16lx\n",
564
qsd->qsd_cpu_int_ena.csr);
565
printk(KERN_ERR " QSD_MEM_CONFIG: 0x%16lx\n",
566
qsd->qsd_mem_config.csr);
567
printk(KERN_ERR " QSD_ERR_SUM: 0x%16lx\n",
568
qsd->qsd_err_sum.csr);
569
570
printk(KERN_ERR "\n");
571
}
572
573
static void __init
574
wildfire_dump_iop_regs(int qbbno)
575
{
576
wildfire_iop *iop = WILDFIRE_iop(qbbno);
577
int i;
578
579
printk(KERN_ERR "IOP registers for QBB %d (%p)\n", qbbno, iop);
580
581
printk(KERN_ERR " IOA_CONFIG: 0x%16lx\n", iop->ioa_config.csr);
582
printk(KERN_ERR " IOD_CONFIG: 0x%16lx\n", iop->iod_config.csr);
583
printk(KERN_ERR " IOP_SWITCH_CREDITS: 0x%16lx\n",
584
iop->iop_switch_credits.csr);
585
printk(KERN_ERR " IOP_HOSE_CREDITS: 0x%16lx\n",
586
iop->iop_hose_credits.csr);
587
588
for (i = 0; i < 4; i++)
589
printk(KERN_ERR " IOP_HOSE_%d_INIT: 0x%16lx\n",
590
i, iop->iop_hose[i].init.csr);
591
for (i = 0; i < 4; i++)
592
printk(KERN_ERR " IOP_DEV_INT_TARGET_%d: 0x%16lx\n",
593
i, iop->iop_dev_int[i].target.csr);
594
595
printk(KERN_ERR "\n");
596
}
597
598
static void __init
599
wildfire_dump_gp_regs(int qbbno)
600
{
601
wildfire_gp *gp = WILDFIRE_gp(qbbno);
602
int i;
603
604
printk(KERN_ERR "GP registers for QBB %d (%p)\n", qbbno, gp);
605
for (i = 0; i < 4; i++)
606
printk(KERN_ERR " GPA_QBB_MAP_%d: 0x%16lx\n",
607
i, gp->gpa_qbb_map[i].csr);
608
609
printk(KERN_ERR " GPA_MEM_POP_MAP: 0x%16lx\n",
610
gp->gpa_mem_pop_map.csr);
611
printk(KERN_ERR " GPA_SCRATCH: 0x%16lx\n", gp->gpa_scratch.csr);
612
printk(KERN_ERR " GPA_DIAG: 0x%16lx\n", gp->gpa_diag.csr);
613
printk(KERN_ERR " GPA_CONFIG_0: 0x%16lx\n", gp->gpa_config_0.csr);
614
printk(KERN_ERR " GPA_INIT_ID: 0x%16lx\n", gp->gpa_init_id.csr);
615
printk(KERN_ERR " GPA_CONFIG_2: 0x%16lx\n", gp->gpa_config_2.csr);
616
617
printk(KERN_ERR "\n");
618
}
619
#endif /* DUMP_REGS */
620
621
#if DEBUG_DUMP_CONFIG
622
static void __init
623
wildfire_dump_hardware_config(void)
624
{
625
int i;
626
627
printk(KERN_ERR "Probed Hardware Configuration\n");
628
629
printk(KERN_ERR " hard_qbb_mask: 0x%16lx\n", wildfire_hard_qbb_mask);
630
printk(KERN_ERR " soft_qbb_mask: 0x%16lx\n", wildfire_soft_qbb_mask);
631
632
printk(KERN_ERR " gp_mask: 0x%16lx\n", wildfire_gp_mask);
633
printk(KERN_ERR " hs_mask: 0x%16lx\n", wildfire_hs_mask);
634
printk(KERN_ERR " iop_mask: 0x%16lx\n", wildfire_iop_mask);
635
printk(KERN_ERR " ior_mask: 0x%16lx\n", wildfire_ior_mask);
636
printk(KERN_ERR " pca_mask: 0x%16lx\n", wildfire_pca_mask);
637
638
printk(KERN_ERR " cpu_mask: 0x%16lx\n", wildfire_cpu_mask);
639
printk(KERN_ERR " mem_mask: 0x%16lx\n", wildfire_mem_mask);
640
641
printk(" hard_qbb_map: ");
642
for (i = 0; i < WILDFIRE_MAX_QBB; i++)
643
if (wildfire_hard_qbb_map[i] == QBB_MAP_EMPTY)
644
printk("--- ");
645
else
646
printk("%3d ", wildfire_hard_qbb_map[i]);
647
printk("\n");
648
649
printk(" soft_qbb_map: ");
650
for (i = 0; i < WILDFIRE_MAX_QBB; i++)
651
if (wildfire_soft_qbb_map[i] == QBB_MAP_EMPTY)
652
printk("--- ");
653
else
654
printk("%3d ", wildfire_soft_qbb_map[i]);
655
printk("\n");
656
}
657
#endif /* DUMP_CONFIG */
658
659