Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/drivers/edac/e7xxx_edac.c
26278 views
1
/*
2
* Intel e7xxx Memory Controller kernel module
3
* (C) 2003 Linux Networx (http://lnxi.com)
4
* This file may be distributed under the terms of the
5
* GNU General Public License.
6
*
7
* See "enum e7xxx_chips" below for supported chipsets
8
*
9
* Written by Thayne Harbaugh
10
* Based on work by Dan Hollis <goemon at anime dot net> and others.
11
* http://www.anime.net/~goemon/linux-ecc/
12
*
13
* Datasheet:
14
* http://www.intel.com/content/www/us/en/chipsets/e7501-chipset-memory-controller-hub-datasheet.html
15
*
16
* Contributors:
17
* Eric Biederman (Linux Networx)
18
* Tom Zimmerman (Linux Networx)
19
* Jim Garlick (Lawrence Livermore National Labs)
20
* Dave Peterson (Lawrence Livermore National Labs)
21
* That One Guy (Some other place)
22
* Wang Zhenyu (intel.com)
23
*
24
* $Id: edac_e7xxx.c,v 1.5.2.9 2005/10/05 00:43:44 dsp_llnl Exp $
25
*
26
*/
27
28
#include <linux/module.h>
29
#include <linux/init.h>
30
#include <linux/pci.h>
31
#include <linux/pci_ids.h>
32
#include <linux/edac.h>
33
#include "edac_module.h"
34
35
#define EDAC_MOD_STR "e7xxx_edac"
36
37
#define e7xxx_printk(level, fmt, arg...) \
38
edac_printk(level, "e7xxx", fmt, ##arg)
39
40
#define e7xxx_mc_printk(mci, level, fmt, arg...) \
41
edac_mc_chipset_printk(mci, level, "e7xxx", fmt, ##arg)
42
43
#ifndef PCI_DEVICE_ID_INTEL_7205_0
44
#define PCI_DEVICE_ID_INTEL_7205_0 0x255d
45
#endif /* PCI_DEVICE_ID_INTEL_7205_0 */
46
47
#ifndef PCI_DEVICE_ID_INTEL_7205_1_ERR
48
#define PCI_DEVICE_ID_INTEL_7205_1_ERR 0x2551
49
#endif /* PCI_DEVICE_ID_INTEL_7205_1_ERR */
50
51
#ifndef PCI_DEVICE_ID_INTEL_7500_0
52
#define PCI_DEVICE_ID_INTEL_7500_0 0x2540
53
#endif /* PCI_DEVICE_ID_INTEL_7500_0 */
54
55
#ifndef PCI_DEVICE_ID_INTEL_7500_1_ERR
56
#define PCI_DEVICE_ID_INTEL_7500_1_ERR 0x2541
57
#endif /* PCI_DEVICE_ID_INTEL_7500_1_ERR */
58
59
#ifndef PCI_DEVICE_ID_INTEL_7501_0
60
#define PCI_DEVICE_ID_INTEL_7501_0 0x254c
61
#endif /* PCI_DEVICE_ID_INTEL_7501_0 */
62
63
#ifndef PCI_DEVICE_ID_INTEL_7501_1_ERR
64
#define PCI_DEVICE_ID_INTEL_7501_1_ERR 0x2541
65
#endif /* PCI_DEVICE_ID_INTEL_7501_1_ERR */
66
67
#ifndef PCI_DEVICE_ID_INTEL_7505_0
68
#define PCI_DEVICE_ID_INTEL_7505_0 0x2550
69
#endif /* PCI_DEVICE_ID_INTEL_7505_0 */
70
71
#ifndef PCI_DEVICE_ID_INTEL_7505_1_ERR
72
#define PCI_DEVICE_ID_INTEL_7505_1_ERR 0x2551
73
#endif /* PCI_DEVICE_ID_INTEL_7505_1_ERR */
74
75
#define E7XXX_NR_CSROWS 8 /* number of csrows */
76
#define E7XXX_NR_DIMMS 8 /* 2 channels, 4 dimms/channel */
77
78
/* E7XXX register addresses - device 0 function 0 */
79
#define E7XXX_DRB 0x60 /* DRAM row boundary register (8b) */
80
#define E7XXX_DRA 0x70 /* DRAM row attribute register (8b) */
81
/*
82
* 31 Device width row 7 0=x8 1=x4
83
* 27 Device width row 6
84
* 23 Device width row 5
85
* 19 Device width row 4
86
* 15 Device width row 3
87
* 11 Device width row 2
88
* 7 Device width row 1
89
* 3 Device width row 0
90
*/
91
#define E7XXX_DRC 0x7C /* DRAM controller mode reg (32b) */
92
/*
93
* 22 Number channels 0=1,1=2
94
* 19:18 DRB Granularity 32/64MB
95
*/
96
#define E7XXX_TOLM 0xC4 /* DRAM top of low memory reg (16b) */
97
#define E7XXX_REMAPBASE 0xC6 /* DRAM remap base address reg (16b) */
98
#define E7XXX_REMAPLIMIT 0xC8 /* DRAM remap limit address reg (16b) */
99
100
/* E7XXX register addresses - device 0 function 1 */
101
#define E7XXX_DRAM_FERR 0x80 /* DRAM first error register (8b) */
102
#define E7XXX_DRAM_NERR 0x82 /* DRAM next error register (8b) */
103
#define E7XXX_DRAM_CELOG_ADD 0xA0 /* DRAM first correctable memory */
104
/* error address register (32b) */
105
/*
106
* 31:28 Reserved
107
* 27:6 CE address (4k block 33:12)
108
* 5:0 Reserved
109
*/
110
#define E7XXX_DRAM_UELOG_ADD 0xB0 /* DRAM first uncorrectable memory */
111
/* error address register (32b) */
112
/*
113
* 31:28 Reserved
114
* 27:6 CE address (4k block 33:12)
115
* 5:0 Reserved
116
*/
117
#define E7XXX_DRAM_CELOG_SYNDROME 0xD0 /* DRAM first correctable memory */
118
/* error syndrome register (16b) */
119
120
enum e7xxx_chips {
121
E7500 = 0,
122
E7501,
123
E7505,
124
E7205,
125
};
126
127
struct e7xxx_pvt {
128
struct pci_dev *bridge_ck;
129
u32 tolm;
130
u32 remapbase;
131
u32 remaplimit;
132
const struct e7xxx_dev_info *dev_info;
133
};
134
135
struct e7xxx_dev_info {
136
u16 err_dev;
137
const char *ctl_name;
138
};
139
140
struct e7xxx_error_info {
141
u8 dram_ferr;
142
u8 dram_nerr;
143
u32 dram_celog_add;
144
u16 dram_celog_syndrome;
145
u32 dram_uelog_add;
146
};
147
148
static struct edac_pci_ctl_info *e7xxx_pci;
149
150
static const struct e7xxx_dev_info e7xxx_devs[] = {
151
[E7500] = {
152
.err_dev = PCI_DEVICE_ID_INTEL_7500_1_ERR,
153
.ctl_name = "E7500"},
154
[E7501] = {
155
.err_dev = PCI_DEVICE_ID_INTEL_7501_1_ERR,
156
.ctl_name = "E7501"},
157
[E7505] = {
158
.err_dev = PCI_DEVICE_ID_INTEL_7505_1_ERR,
159
.ctl_name = "E7505"},
160
[E7205] = {
161
.err_dev = PCI_DEVICE_ID_INTEL_7205_1_ERR,
162
.ctl_name = "E7205"},
163
};
164
165
/* FIXME - is this valid for both SECDED and S4ECD4ED? */
166
static inline int e7xxx_find_channel(u16 syndrome)
167
{
168
edac_dbg(3, "\n");
169
170
if ((syndrome & 0xff00) == 0)
171
return 0;
172
173
if ((syndrome & 0x00ff) == 0)
174
return 1;
175
176
if ((syndrome & 0xf000) == 0 || (syndrome & 0x0f00) == 0)
177
return 0;
178
179
return 1;
180
}
181
182
static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci,
183
unsigned long page)
184
{
185
u32 remap;
186
struct e7xxx_pvt *pvt = (struct e7xxx_pvt *)mci->pvt_info;
187
188
edac_dbg(3, "\n");
189
190
if ((page < pvt->tolm) ||
191
((page >= 0x100000) && (page < pvt->remapbase)))
192
return page;
193
194
remap = (page - pvt->tolm) + pvt->remapbase;
195
196
if (remap < pvt->remaplimit)
197
return remap;
198
199
e7xxx_printk(KERN_ERR, "Invalid page %lx - out of range\n", page);
200
return pvt->tolm - 1;
201
}
202
203
static void process_ce(struct mem_ctl_info *mci, struct e7xxx_error_info *info)
204
{
205
u32 error_1b, page;
206
u16 syndrome;
207
int row;
208
int channel;
209
210
edac_dbg(3, "\n");
211
/* read the error address */
212
error_1b = info->dram_celog_add;
213
/* FIXME - should use PAGE_SHIFT */
214
page = error_1b >> 6; /* convert the address to 4k page */
215
/* read the syndrome */
216
syndrome = info->dram_celog_syndrome;
217
/* FIXME - check for -1 */
218
row = edac_mc_find_csrow_by_page(mci, page);
219
/* convert syndrome to channel */
220
channel = e7xxx_find_channel(syndrome);
221
edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1, page, 0, syndrome,
222
row, channel, -1, "e7xxx CE", "");
223
}
224
225
static void process_ce_no_info(struct mem_ctl_info *mci)
226
{
227
edac_dbg(3, "\n");
228
edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1, 0, 0, 0, -1, -1, -1,
229
"e7xxx CE log register overflow", "");
230
}
231
232
static void process_ue(struct mem_ctl_info *mci, struct e7xxx_error_info *info)
233
{
234
u32 error_2b, block_page;
235
int row;
236
237
edac_dbg(3, "\n");
238
/* read the error address */
239
error_2b = info->dram_uelog_add;
240
/* FIXME - should use PAGE_SHIFT */
241
block_page = error_2b >> 6; /* convert to 4k address */
242
row = edac_mc_find_csrow_by_page(mci, block_page);
243
244
edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, block_page, 0, 0,
245
row, -1, -1, "e7xxx UE", "");
246
}
247
248
static void process_ue_no_info(struct mem_ctl_info *mci)
249
{
250
edac_dbg(3, "\n");
251
252
edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, 0, 0, 0, -1, -1, -1,
253
"e7xxx UE log register overflow", "");
254
}
255
256
static void e7xxx_get_error_info(struct mem_ctl_info *mci,
257
struct e7xxx_error_info *info)
258
{
259
struct e7xxx_pvt *pvt;
260
261
pvt = (struct e7xxx_pvt *)mci->pvt_info;
262
pci_read_config_byte(pvt->bridge_ck, E7XXX_DRAM_FERR, &info->dram_ferr);
263
pci_read_config_byte(pvt->bridge_ck, E7XXX_DRAM_NERR, &info->dram_nerr);
264
265
if ((info->dram_ferr & 1) || (info->dram_nerr & 1)) {
266
pci_read_config_dword(pvt->bridge_ck, E7XXX_DRAM_CELOG_ADD,
267
&info->dram_celog_add);
268
pci_read_config_word(pvt->bridge_ck,
269
E7XXX_DRAM_CELOG_SYNDROME,
270
&info->dram_celog_syndrome);
271
}
272
273
if ((info->dram_ferr & 2) || (info->dram_nerr & 2))
274
pci_read_config_dword(pvt->bridge_ck, E7XXX_DRAM_UELOG_ADD,
275
&info->dram_uelog_add);
276
277
if (info->dram_ferr & 3)
278
pci_write_bits8(pvt->bridge_ck, E7XXX_DRAM_FERR, 0x03, 0x03);
279
280
if (info->dram_nerr & 3)
281
pci_write_bits8(pvt->bridge_ck, E7XXX_DRAM_NERR, 0x03, 0x03);
282
}
283
284
static int e7xxx_process_error_info(struct mem_ctl_info *mci,
285
struct e7xxx_error_info *info,
286
int handle_errors)
287
{
288
int error_found;
289
290
error_found = 0;
291
292
/* decode and report errors */
293
if (info->dram_ferr & 1) { /* check first error correctable */
294
error_found = 1;
295
296
if (handle_errors)
297
process_ce(mci, info);
298
}
299
300
if (info->dram_ferr & 2) { /* check first error uncorrectable */
301
error_found = 1;
302
303
if (handle_errors)
304
process_ue(mci, info);
305
}
306
307
if (info->dram_nerr & 1) { /* check next error correctable */
308
error_found = 1;
309
310
if (handle_errors) {
311
if (info->dram_ferr & 1)
312
process_ce_no_info(mci);
313
else
314
process_ce(mci, info);
315
}
316
}
317
318
if (info->dram_nerr & 2) { /* check next error uncorrectable */
319
error_found = 1;
320
321
if (handle_errors) {
322
if (info->dram_ferr & 2)
323
process_ue_no_info(mci);
324
else
325
process_ue(mci, info);
326
}
327
}
328
329
return error_found;
330
}
331
332
static void e7xxx_check(struct mem_ctl_info *mci)
333
{
334
struct e7xxx_error_info info;
335
336
e7xxx_get_error_info(mci, &info);
337
e7xxx_process_error_info(mci, &info, 1);
338
}
339
340
/* Return 1 if dual channel mode is active. Else return 0. */
341
static inline int dual_channel_active(u32 drc, int dev_idx)
342
{
343
return (dev_idx == E7501) ? ((drc >> 22) & 0x1) : 1;
344
}
345
346
/* Return DRB granularity (0=32mb, 1=64mb). */
347
static inline int drb_granularity(u32 drc, int dev_idx)
348
{
349
/* only e7501 can be single channel */
350
return (dev_idx == E7501) ? ((drc >> 18) & 0x3) : 1;
351
}
352
353
static void e7xxx_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev,
354
int dev_idx, u32 drc)
355
{
356
unsigned long last_cumul_size;
357
int index, j;
358
u8 value;
359
u32 dra, cumul_size, nr_pages;
360
int drc_chan, drc_drbg, drc_ddim, mem_dev;
361
struct csrow_info *csrow;
362
struct dimm_info *dimm;
363
enum edac_type edac_mode;
364
365
pci_read_config_dword(pdev, E7XXX_DRA, &dra);
366
drc_chan = dual_channel_active(drc, dev_idx);
367
drc_drbg = drb_granularity(drc, dev_idx);
368
drc_ddim = (drc >> 20) & 0x3;
369
last_cumul_size = 0;
370
371
/* The dram row boundary (DRB) reg values are boundary address
372
* for each DRAM row with a granularity of 32 or 64MB (single/dual
373
* channel operation). DRB regs are cumulative; therefore DRB7 will
374
* contain the total memory contained in all eight rows.
375
*/
376
for (index = 0; index < mci->nr_csrows; index++) {
377
/* mem_dev 0=x8, 1=x4 */
378
mem_dev = (dra >> (index * 4 + 3)) & 0x1;
379
csrow = mci->csrows[index];
380
381
pci_read_config_byte(pdev, E7XXX_DRB + index, &value);
382
/* convert a 64 or 32 MiB DRB to a page size. */
383
cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
384
edac_dbg(3, "(%d) cumul_size 0x%x\n", index, cumul_size);
385
if (cumul_size == last_cumul_size)
386
continue; /* not populated */
387
388
csrow->first_page = last_cumul_size;
389
csrow->last_page = cumul_size - 1;
390
nr_pages = cumul_size - last_cumul_size;
391
last_cumul_size = cumul_size;
392
393
/*
394
* if single channel or x8 devices then SECDED
395
* if dual channel and x4 then S4ECD4ED
396
*/
397
if (drc_ddim) {
398
if (drc_chan && mem_dev) {
399
edac_mode = EDAC_S4ECD4ED;
400
mci->edac_cap |= EDAC_FLAG_S4ECD4ED;
401
} else {
402
edac_mode = EDAC_SECDED;
403
mci->edac_cap |= EDAC_FLAG_SECDED;
404
}
405
} else
406
edac_mode = EDAC_NONE;
407
408
for (j = 0; j < drc_chan + 1; j++) {
409
dimm = csrow->channels[j]->dimm;
410
411
dimm->nr_pages = nr_pages / (drc_chan + 1);
412
dimm->grain = 1 << 12; /* 4KiB - resolution of CELOG */
413
dimm->mtype = MEM_RDDR; /* only one type supported */
414
dimm->dtype = mem_dev ? DEV_X4 : DEV_X8;
415
dimm->edac_mode = edac_mode;
416
}
417
}
418
}
419
420
static int e7xxx_probe1(struct pci_dev *pdev, int dev_idx)
421
{
422
u16 pci_data;
423
struct mem_ctl_info *mci = NULL;
424
struct edac_mc_layer layers[2];
425
struct e7xxx_pvt *pvt = NULL;
426
u32 drc;
427
int drc_chan;
428
struct e7xxx_error_info discard;
429
430
edac_dbg(0, "mci\n");
431
432
pci_read_config_dword(pdev, E7XXX_DRC, &drc);
433
434
drc_chan = dual_channel_active(drc, dev_idx);
435
/*
436
* According with the datasheet, this device has a maximum of
437
* 4 DIMMS per channel, either single-rank or dual-rank. So, the
438
* total amount of dimms is 8 (E7XXX_NR_DIMMS).
439
* That means that the DIMM is mapped as CSROWs, and the channel
440
* will map the rank. So, an error to either channel should be
441
* attributed to the same dimm.
442
*/
443
layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
444
layers[0].size = E7XXX_NR_CSROWS;
445
layers[0].is_virt_csrow = true;
446
layers[1].type = EDAC_MC_LAYER_CHANNEL;
447
layers[1].size = drc_chan + 1;
448
layers[1].is_virt_csrow = false;
449
mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt));
450
if (mci == NULL)
451
return -ENOMEM;
452
453
edac_dbg(3, "init mci\n");
454
mci->mtype_cap = MEM_FLAG_RDDR;
455
mci->edac_ctl_cap = EDAC_FLAG_NONE | EDAC_FLAG_SECDED |
456
EDAC_FLAG_S4ECD4ED;
457
/* FIXME - what if different memory types are in different csrows? */
458
mci->mod_name = EDAC_MOD_STR;
459
mci->pdev = &pdev->dev;
460
edac_dbg(3, "init pvt\n");
461
pvt = (struct e7xxx_pvt *)mci->pvt_info;
462
pvt->dev_info = &e7xxx_devs[dev_idx];
463
pvt->bridge_ck = pci_get_device(PCI_VENDOR_ID_INTEL,
464
pvt->dev_info->err_dev, pvt->bridge_ck);
465
466
if (!pvt->bridge_ck) {
467
e7xxx_printk(KERN_ERR, "error reporting device not found:"
468
"vendor %x device 0x%x (broken BIOS?)\n",
469
PCI_VENDOR_ID_INTEL, e7xxx_devs[dev_idx].err_dev);
470
goto fail0;
471
}
472
473
edac_dbg(3, "more mci init\n");
474
mci->ctl_name = pvt->dev_info->ctl_name;
475
mci->dev_name = pci_name(pdev);
476
mci->edac_check = e7xxx_check;
477
mci->ctl_page_to_phys = ctl_page_to_phys;
478
e7xxx_init_csrows(mci, pdev, dev_idx, drc);
479
mci->edac_cap |= EDAC_FLAG_NONE;
480
edac_dbg(3, "tolm, remapbase, remaplimit\n");
481
/* load the top of low memory, remap base, and remap limit vars */
482
pci_read_config_word(pdev, E7XXX_TOLM, &pci_data);
483
pvt->tolm = ((u32) pci_data) << 4;
484
pci_read_config_word(pdev, E7XXX_REMAPBASE, &pci_data);
485
pvt->remapbase = ((u32) pci_data) << 14;
486
pci_read_config_word(pdev, E7XXX_REMAPLIMIT, &pci_data);
487
pvt->remaplimit = ((u32) pci_data) << 14;
488
e7xxx_printk(KERN_INFO,
489
"tolm = %x, remapbase = %x, remaplimit = %x\n", pvt->tolm,
490
pvt->remapbase, pvt->remaplimit);
491
492
/* clear any pending errors, or initial state bits */
493
e7xxx_get_error_info(mci, &discard);
494
495
/* Here we assume that we will never see multiple instances of this
496
* type of memory controller. The ID is therefore hardcoded to 0.
497
*/
498
if (edac_mc_add_mc(mci)) {
499
edac_dbg(3, "failed edac_mc_add_mc()\n");
500
goto fail1;
501
}
502
503
/* allocating generic PCI control info */
504
e7xxx_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
505
if (!e7xxx_pci) {
506
printk(KERN_WARNING
507
"%s(): Unable to create PCI control\n",
508
__func__);
509
printk(KERN_WARNING
510
"%s(): PCI error report via EDAC not setup\n",
511
__func__);
512
}
513
514
/* get this far and it's successful */
515
edac_dbg(3, "success\n");
516
return 0;
517
518
fail1:
519
pci_dev_put(pvt->bridge_ck);
520
521
fail0:
522
edac_mc_free(mci);
523
524
return -ENODEV;
525
}
526
527
/* returns count (>= 0), or negative on error */
528
static int e7xxx_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
529
{
530
edac_dbg(0, "\n");
531
532
/* wake up and enable device */
533
return pci_enable_device(pdev) ?
534
-EIO : e7xxx_probe1(pdev, ent->driver_data);
535
}
536
537
static void e7xxx_remove_one(struct pci_dev *pdev)
538
{
539
struct mem_ctl_info *mci;
540
struct e7xxx_pvt *pvt;
541
542
edac_dbg(0, "\n");
543
544
if (e7xxx_pci)
545
edac_pci_release_generic_ctl(e7xxx_pci);
546
547
if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL)
548
return;
549
550
pvt = (struct e7xxx_pvt *)mci->pvt_info;
551
pci_dev_put(pvt->bridge_ck);
552
edac_mc_free(mci);
553
}
554
555
static const struct pci_device_id e7xxx_pci_tbl[] = {
556
{
557
PCI_VEND_DEV(INTEL, 7205_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
558
E7205},
559
{
560
PCI_VEND_DEV(INTEL, 7500_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
561
E7500},
562
{
563
PCI_VEND_DEV(INTEL, 7501_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
564
E7501},
565
{
566
PCI_VEND_DEV(INTEL, 7505_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
567
E7505},
568
{
569
0,
570
} /* 0 terminated list. */
571
};
572
573
MODULE_DEVICE_TABLE(pci, e7xxx_pci_tbl);
574
575
static struct pci_driver e7xxx_driver = {
576
.name = EDAC_MOD_STR,
577
.probe = e7xxx_init_one,
578
.remove = e7xxx_remove_one,
579
.id_table = e7xxx_pci_tbl,
580
};
581
582
static int __init e7xxx_init(void)
583
{
584
/* Ensure that the OPSTATE is set correctly for POLL or NMI */
585
opstate_init();
586
587
return pci_register_driver(&e7xxx_driver);
588
}
589
590
static void __exit e7xxx_exit(void)
591
{
592
pci_unregister_driver(&e7xxx_driver);
593
}
594
595
module_init(e7xxx_init);
596
module_exit(e7xxx_exit);
597
598
MODULE_LICENSE("GPL");
599
MODULE_AUTHOR("Linux Networx (http://lnxi.com) Thayne Harbaugh et al");
600
MODULE_DESCRIPTION("MC support for Intel e7xxx memory controllers");
601
module_param(edac_op_state, int, 0444);
602
MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
603
604