Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
awilliam
GitHub Repository: awilliam/linux-vfio
Path: blob/master/drivers/ata/pata_bf54x.c
15109 views
1
/*
2
* File: drivers/ata/pata_bf54x.c
3
* Author: Sonic Zhang <[email protected]>
4
*
5
* Created:
6
* Description: PATA Driver for blackfin 54x
7
*
8
* Modified:
9
* Copyright 2007 Analog Devices Inc.
10
*
11
* Bugs: Enter bugs at http://blackfin.uclinux.org/
12
*
13
* This program is free software; you can redistribute it and/or modify
14
* it under the terms of the GNU General Public License as published by
15
* the Free Software Foundation; either version 2 of the License, or
16
* (at your option) any later version.
17
*
18
* This program is distributed in the hope that it will be useful,
19
* but WITHOUT ANY WARRANTY; without even the implied warranty of
20
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
21
* GNU General Public License for more details.
22
*
23
* You should have received a copy of the GNU General Public License
24
* along with this program; if not, see the file COPYING, or write
25
* to the Free Software Foundation, Inc.,
26
* 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
27
*/
28
29
#include <linux/kernel.h>
30
#include <linux/module.h>
31
#include <linux/pci.h>
32
#include <linux/init.h>
33
#include <linux/blkdev.h>
34
#include <linux/delay.h>
35
#include <linux/device.h>
36
#include <scsi/scsi_host.h>
37
#include <linux/libata.h>
38
#include <linux/platform_device.h>
39
#include <asm/dma.h>
40
#include <asm/gpio.h>
41
#include <asm/portmux.h>
42
43
#define DRV_NAME "pata-bf54x"
44
#define DRV_VERSION "0.9"
45
46
#define ATA_REG_CTRL 0x0E
47
#define ATA_REG_ALTSTATUS ATA_REG_CTRL
48
49
/* These are the offset of the controller's registers */
50
#define ATAPI_OFFSET_CONTROL 0x00
51
#define ATAPI_OFFSET_STATUS 0x04
52
#define ATAPI_OFFSET_DEV_ADDR 0x08
53
#define ATAPI_OFFSET_DEV_TXBUF 0x0c
54
#define ATAPI_OFFSET_DEV_RXBUF 0x10
55
#define ATAPI_OFFSET_INT_MASK 0x14
56
#define ATAPI_OFFSET_INT_STATUS 0x18
57
#define ATAPI_OFFSET_XFER_LEN 0x1c
58
#define ATAPI_OFFSET_LINE_STATUS 0x20
59
#define ATAPI_OFFSET_SM_STATE 0x24
60
#define ATAPI_OFFSET_TERMINATE 0x28
61
#define ATAPI_OFFSET_PIO_TFRCNT 0x2c
62
#define ATAPI_OFFSET_DMA_TFRCNT 0x30
63
#define ATAPI_OFFSET_UMAIN_TFRCNT 0x34
64
#define ATAPI_OFFSET_UDMAOUT_TFRCNT 0x38
65
#define ATAPI_OFFSET_REG_TIM_0 0x40
66
#define ATAPI_OFFSET_PIO_TIM_0 0x44
67
#define ATAPI_OFFSET_PIO_TIM_1 0x48
68
#define ATAPI_OFFSET_MULTI_TIM_0 0x50
69
#define ATAPI_OFFSET_MULTI_TIM_1 0x54
70
#define ATAPI_OFFSET_MULTI_TIM_2 0x58
71
#define ATAPI_OFFSET_ULTRA_TIM_0 0x60
72
#define ATAPI_OFFSET_ULTRA_TIM_1 0x64
73
#define ATAPI_OFFSET_ULTRA_TIM_2 0x68
74
#define ATAPI_OFFSET_ULTRA_TIM_3 0x6c
75
76
77
#define ATAPI_GET_CONTROL(base)\
78
bfin_read16(base + ATAPI_OFFSET_CONTROL)
79
#define ATAPI_SET_CONTROL(base, val)\
80
bfin_write16(base + ATAPI_OFFSET_CONTROL, val)
81
#define ATAPI_GET_STATUS(base)\
82
bfin_read16(base + ATAPI_OFFSET_STATUS)
83
#define ATAPI_GET_DEV_ADDR(base)\
84
bfin_read16(base + ATAPI_OFFSET_DEV_ADDR)
85
#define ATAPI_SET_DEV_ADDR(base, val)\
86
bfin_write16(base + ATAPI_OFFSET_DEV_ADDR, val)
87
#define ATAPI_GET_DEV_TXBUF(base)\
88
bfin_read16(base + ATAPI_OFFSET_DEV_TXBUF)
89
#define ATAPI_SET_DEV_TXBUF(base, val)\
90
bfin_write16(base + ATAPI_OFFSET_DEV_TXBUF, val)
91
#define ATAPI_GET_DEV_RXBUF(base)\
92
bfin_read16(base + ATAPI_OFFSET_DEV_RXBUF)
93
#define ATAPI_SET_DEV_RXBUF(base, val)\
94
bfin_write16(base + ATAPI_OFFSET_DEV_RXBUF, val)
95
#define ATAPI_GET_INT_MASK(base)\
96
bfin_read16(base + ATAPI_OFFSET_INT_MASK)
97
#define ATAPI_SET_INT_MASK(base, val)\
98
bfin_write16(base + ATAPI_OFFSET_INT_MASK, val)
99
#define ATAPI_GET_INT_STATUS(base)\
100
bfin_read16(base + ATAPI_OFFSET_INT_STATUS)
101
#define ATAPI_SET_INT_STATUS(base, val)\
102
bfin_write16(base + ATAPI_OFFSET_INT_STATUS, val)
103
#define ATAPI_GET_XFER_LEN(base)\
104
bfin_read16(base + ATAPI_OFFSET_XFER_LEN)
105
#define ATAPI_SET_XFER_LEN(base, val)\
106
bfin_write16(base + ATAPI_OFFSET_XFER_LEN, val)
107
#define ATAPI_GET_LINE_STATUS(base)\
108
bfin_read16(base + ATAPI_OFFSET_LINE_STATUS)
109
#define ATAPI_GET_SM_STATE(base)\
110
bfin_read16(base + ATAPI_OFFSET_SM_STATE)
111
#define ATAPI_GET_TERMINATE(base)\
112
bfin_read16(base + ATAPI_OFFSET_TERMINATE)
113
#define ATAPI_SET_TERMINATE(base, val)\
114
bfin_write16(base + ATAPI_OFFSET_TERMINATE, val)
115
#define ATAPI_GET_PIO_TFRCNT(base)\
116
bfin_read16(base + ATAPI_OFFSET_PIO_TFRCNT)
117
#define ATAPI_GET_DMA_TFRCNT(base)\
118
bfin_read16(base + ATAPI_OFFSET_DMA_TFRCNT)
119
#define ATAPI_GET_UMAIN_TFRCNT(base)\
120
bfin_read16(base + ATAPI_OFFSET_UMAIN_TFRCNT)
121
#define ATAPI_GET_UDMAOUT_TFRCNT(base)\
122
bfin_read16(base + ATAPI_OFFSET_UDMAOUT_TFRCNT)
123
#define ATAPI_GET_REG_TIM_0(base)\
124
bfin_read16(base + ATAPI_OFFSET_REG_TIM_0)
125
#define ATAPI_SET_REG_TIM_0(base, val)\
126
bfin_write16(base + ATAPI_OFFSET_REG_TIM_0, val)
127
#define ATAPI_GET_PIO_TIM_0(base)\
128
bfin_read16(base + ATAPI_OFFSET_PIO_TIM_0)
129
#define ATAPI_SET_PIO_TIM_0(base, val)\
130
bfin_write16(base + ATAPI_OFFSET_PIO_TIM_0, val)
131
#define ATAPI_GET_PIO_TIM_1(base)\
132
bfin_read16(base + ATAPI_OFFSET_PIO_TIM_1)
133
#define ATAPI_SET_PIO_TIM_1(base, val)\
134
bfin_write16(base + ATAPI_OFFSET_PIO_TIM_1, val)
135
#define ATAPI_GET_MULTI_TIM_0(base)\
136
bfin_read16(base + ATAPI_OFFSET_MULTI_TIM_0)
137
#define ATAPI_SET_MULTI_TIM_0(base, val)\
138
bfin_write16(base + ATAPI_OFFSET_MULTI_TIM_0, val)
139
#define ATAPI_GET_MULTI_TIM_1(base)\
140
bfin_read16(base + ATAPI_OFFSET_MULTI_TIM_1)
141
#define ATAPI_SET_MULTI_TIM_1(base, val)\
142
bfin_write16(base + ATAPI_OFFSET_MULTI_TIM_1, val)
143
#define ATAPI_GET_MULTI_TIM_2(base)\
144
bfin_read16(base + ATAPI_OFFSET_MULTI_TIM_2)
145
#define ATAPI_SET_MULTI_TIM_2(base, val)\
146
bfin_write16(base + ATAPI_OFFSET_MULTI_TIM_2, val)
147
#define ATAPI_GET_ULTRA_TIM_0(base)\
148
bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_0)
149
#define ATAPI_SET_ULTRA_TIM_0(base, val)\
150
bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_0, val)
151
#define ATAPI_GET_ULTRA_TIM_1(base)\
152
bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_1)
153
#define ATAPI_SET_ULTRA_TIM_1(base, val)\
154
bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_1, val)
155
#define ATAPI_GET_ULTRA_TIM_2(base)\
156
bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_2)
157
#define ATAPI_SET_ULTRA_TIM_2(base, val)\
158
bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_2, val)
159
#define ATAPI_GET_ULTRA_TIM_3(base)\
160
bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_3)
161
#define ATAPI_SET_ULTRA_TIM_3(base, val)\
162
bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_3, val)
163
164
/**
165
* PIO Mode - Frequency compatibility
166
*/
167
/* mode: 0 1 2 3 4 */
168
static const u32 pio_fsclk[] =
169
{ 33333333, 33333333, 33333333, 33333333, 33333333 };
170
171
/**
172
* MDMA Mode - Frequency compatibility
173
*/
174
/* mode: 0 1 2 */
175
static const u32 mdma_fsclk[] = { 33333333, 33333333, 33333333 };
176
177
/**
178
* UDMA Mode - Frequency compatibility
179
*
180
* UDMA5 - 100 MB/s - SCLK = 133 MHz
181
* UDMA4 - 66 MB/s - SCLK >= 80 MHz
182
* UDMA3 - 44.4 MB/s - SCLK >= 50 MHz
183
* UDMA2 - 33 MB/s - SCLK >= 40 MHz
184
*/
185
/* mode: 0 1 2 3 4 5 */
186
static const u32 udma_fsclk[] =
187
{ 33333333, 33333333, 40000000, 50000000, 80000000, 133333333 };
188
189
/**
190
* Register transfer timing table
191
*/
192
/* mode: 0 1 2 3 4 */
193
/* Cycle Time */
194
static const u32 reg_t0min[] = { 600, 383, 330, 180, 120 };
195
/* DIOR/DIOW to end cycle */
196
static const u32 reg_t2min[] = { 290, 290, 290, 70, 25 };
197
/* DIOR/DIOW asserted pulse width */
198
static const u32 reg_teocmin[] = { 290, 290, 290, 80, 70 };
199
200
/**
201
* PIO timing table
202
*/
203
/* mode: 0 1 2 3 4 */
204
/* Cycle Time */
205
static const u32 pio_t0min[] = { 600, 383, 240, 180, 120 };
206
/* Address valid to DIOR/DIORW */
207
static const u32 pio_t1min[] = { 70, 50, 30, 30, 25 };
208
/* DIOR/DIOW to end cycle */
209
static const u32 pio_t2min[] = { 165, 125, 100, 80, 70 };
210
/* DIOR/DIOW asserted pulse width */
211
static const u32 pio_teocmin[] = { 165, 125, 100, 70, 25 };
212
/* DIOW data hold */
213
static const u32 pio_t4min[] = { 30, 20, 15, 10, 10 };
214
215
/* ******************************************************************
216
* Multiword DMA timing table
217
* ******************************************************************
218
*/
219
/* mode: 0 1 2 */
220
/* Cycle Time */
221
static const u32 mdma_t0min[] = { 480, 150, 120 };
222
/* DIOR/DIOW asserted pulse width */
223
static const u32 mdma_tdmin[] = { 215, 80, 70 };
224
/* DMACK to read data released */
225
static const u32 mdma_thmin[] = { 20, 15, 10 };
226
/* DIOR/DIOW to DMACK hold */
227
static const u32 mdma_tjmin[] = { 20, 5, 5 };
228
/* DIOR negated pulse width */
229
static const u32 mdma_tkrmin[] = { 50, 50, 25 };
230
/* DIOR negated pulse width */
231
static const u32 mdma_tkwmin[] = { 215, 50, 25 };
232
/* CS[1:0] valid to DIOR/DIOW */
233
static const u32 mdma_tmmin[] = { 50, 30, 25 };
234
/* DMACK to read data released */
235
static const u32 mdma_tzmax[] = { 20, 25, 25 };
236
237
/**
238
* Ultra DMA timing table
239
*/
240
/* mode: 0 1 2 3 4 5 */
241
static const u32 udma_tcycmin[] = { 112, 73, 54, 39, 25, 17 };
242
static const u32 udma_tdvsmin[] = { 70, 48, 31, 20, 7, 5 };
243
static const u32 udma_tenvmax[] = { 70, 70, 70, 55, 55, 50 };
244
static const u32 udma_trpmin[] = { 160, 125, 100, 100, 100, 85 };
245
static const u32 udma_tmin[] = { 5, 5, 5, 5, 3, 3 };
246
247
248
static const u32 udma_tmlimin = 20;
249
static const u32 udma_tzahmin = 20;
250
static const u32 udma_tenvmin = 20;
251
static const u32 udma_tackmin = 20;
252
static const u32 udma_tssmin = 50;
253
254
/**
255
*
256
* Function: num_clocks_min
257
*
258
* Description:
259
* calculate number of SCLK cycles to meet minimum timing
260
*/
261
static unsigned short num_clocks_min(unsigned long tmin,
262
unsigned long fsclk)
263
{
264
unsigned long tmp ;
265
unsigned short result;
266
267
tmp = tmin * (fsclk/1000/1000) / 1000;
268
result = (unsigned short)tmp;
269
if ((tmp*1000*1000) < (tmin*(fsclk/1000))) {
270
result++;
271
}
272
273
return result;
274
}
275
276
/**
277
* bfin_set_piomode - Initialize host controller PATA PIO timings
278
* @ap: Port whose timings we are configuring
279
* @adev: um
280
*
281
* Set PIO mode for device.
282
*
283
* LOCKING:
284
* None (inherited from caller).
285
*/
286
287
static void bfin_set_piomode(struct ata_port *ap, struct ata_device *adev)
288
{
289
int mode = adev->pio_mode - XFER_PIO_0;
290
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
291
unsigned int fsclk = get_sclk();
292
unsigned short teoc_reg, t2_reg, teoc_pio;
293
unsigned short t4_reg, t2_pio, t1_reg;
294
unsigned short n0, n6, t6min = 5;
295
296
/* the most restrictive timing value is t6 and tc, the DIOW - data hold
297
* If one SCLK pulse is longer than this minimum value then register
298
* transfers cannot be supported at this frequency.
299
*/
300
n6 = num_clocks_min(t6min, fsclk);
301
if (mode >= 0 && mode <= 4 && n6 >= 1) {
302
dev_dbg(adev->link->ap->dev, "set piomode: mode=%d, fsclk=%ud\n", mode, fsclk);
303
/* calculate the timing values for register transfers. */
304
while (mode > 0 && pio_fsclk[mode] > fsclk)
305
mode--;
306
307
/* DIOR/DIOW to end cycle time */
308
t2_reg = num_clocks_min(reg_t2min[mode], fsclk);
309
/* DIOR/DIOW asserted pulse width */
310
teoc_reg = num_clocks_min(reg_teocmin[mode], fsclk);
311
/* Cycle Time */
312
n0 = num_clocks_min(reg_t0min[mode], fsclk);
313
314
/* increase t2 until we meed the minimum cycle length */
315
if (t2_reg + teoc_reg < n0)
316
t2_reg = n0 - teoc_reg;
317
318
/* calculate the timing values for pio transfers. */
319
320
/* DIOR/DIOW to end cycle time */
321
t2_pio = num_clocks_min(pio_t2min[mode], fsclk);
322
/* DIOR/DIOW asserted pulse width */
323
teoc_pio = num_clocks_min(pio_teocmin[mode], fsclk);
324
/* Cycle Time */
325
n0 = num_clocks_min(pio_t0min[mode], fsclk);
326
327
/* increase t2 until we meed the minimum cycle length */
328
if (t2_pio + teoc_pio < n0)
329
t2_pio = n0 - teoc_pio;
330
331
/* Address valid to DIOR/DIORW */
332
t1_reg = num_clocks_min(pio_t1min[mode], fsclk);
333
334
/* DIOW data hold */
335
t4_reg = num_clocks_min(pio_t4min[mode], fsclk);
336
337
ATAPI_SET_REG_TIM_0(base, (teoc_reg<<8 | t2_reg));
338
ATAPI_SET_PIO_TIM_0(base, (t4_reg<<12 | t2_pio<<4 | t1_reg));
339
ATAPI_SET_PIO_TIM_1(base, teoc_pio);
340
if (mode > 2) {
341
ATAPI_SET_CONTROL(base,
342
ATAPI_GET_CONTROL(base) | IORDY_EN);
343
} else {
344
ATAPI_SET_CONTROL(base,
345
ATAPI_GET_CONTROL(base) & ~IORDY_EN);
346
}
347
348
/* Disable host ATAPI PIO interrupts */
349
ATAPI_SET_INT_MASK(base, ATAPI_GET_INT_MASK(base)
350
& ~(PIO_DONE_MASK | HOST_TERM_XFER_MASK));
351
SSYNC();
352
}
353
}
354
355
/**
356
* bfin_set_dmamode - Initialize host controller PATA DMA timings
357
* @ap: Port whose timings we are configuring
358
* @adev: um
359
*
360
* Set UDMA mode for device.
361
*
362
* LOCKING:
363
* None (inherited from caller).
364
*/
365
366
static void bfin_set_dmamode(struct ata_port *ap, struct ata_device *adev)
367
{
368
int mode;
369
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
370
unsigned long fsclk = get_sclk();
371
unsigned short tenv, tack, tcyc_tdvs, tdvs, tmli, tss, trp, tzah;
372
unsigned short tm, td, tkr, tkw, teoc, th;
373
unsigned short n0, nf, tfmin = 5;
374
unsigned short nmin, tcyc;
375
376
mode = adev->dma_mode - XFER_UDMA_0;
377
if (mode >= 0 && mode <= 5) {
378
dev_dbg(adev->link->ap->dev, "set udmamode: mode=%d\n", mode);
379
/* the most restrictive timing value is t6 and tc,
380
* the DIOW - data hold. If one SCLK pulse is longer
381
* than this minimum value then register
382
* transfers cannot be supported at this frequency.
383
*/
384
while (mode > 0 && udma_fsclk[mode] > fsclk)
385
mode--;
386
387
nmin = num_clocks_min(udma_tmin[mode], fsclk);
388
if (nmin >= 1) {
389
/* calculate the timing values for Ultra DMA. */
390
tdvs = num_clocks_min(udma_tdvsmin[mode], fsclk);
391
tcyc = num_clocks_min(udma_tcycmin[mode], fsclk);
392
tcyc_tdvs = 2;
393
394
/* increase tcyc - tdvs (tcyc_tdvs) until we meed
395
* the minimum cycle length
396
*/
397
if (tdvs + tcyc_tdvs < tcyc)
398
tcyc_tdvs = tcyc - tdvs;
399
400
/* Mow assign the values required for the timing
401
* registers
402
*/
403
if (tcyc_tdvs < 2)
404
tcyc_tdvs = 2;
405
406
if (tdvs < 2)
407
tdvs = 2;
408
409
tack = num_clocks_min(udma_tackmin, fsclk);
410
tss = num_clocks_min(udma_tssmin, fsclk);
411
tmli = num_clocks_min(udma_tmlimin, fsclk);
412
tzah = num_clocks_min(udma_tzahmin, fsclk);
413
trp = num_clocks_min(udma_trpmin[mode], fsclk);
414
tenv = num_clocks_min(udma_tenvmin, fsclk);
415
if (tenv <= udma_tenvmax[mode]) {
416
ATAPI_SET_ULTRA_TIM_0(base, (tenv<<8 | tack));
417
ATAPI_SET_ULTRA_TIM_1(base,
418
(tcyc_tdvs<<8 | tdvs));
419
ATAPI_SET_ULTRA_TIM_2(base, (tmli<<8 | tss));
420
ATAPI_SET_ULTRA_TIM_3(base, (trp<<8 | tzah));
421
422
/* Enable host ATAPI Untra DMA interrupts */
423
ATAPI_SET_INT_MASK(base,
424
ATAPI_GET_INT_MASK(base)
425
| UDMAIN_DONE_MASK
426
| UDMAOUT_DONE_MASK
427
| UDMAIN_TERM_MASK
428
| UDMAOUT_TERM_MASK);
429
}
430
}
431
}
432
433
mode = adev->dma_mode - XFER_MW_DMA_0;
434
if (mode >= 0 && mode <= 2) {
435
dev_dbg(adev->link->ap->dev, "set mdmamode: mode=%d\n", mode);
436
/* the most restrictive timing value is tf, the DMACK to
437
* read data released. If one SCLK pulse is longer than
438
* this maximum value then the MDMA mode
439
* cannot be supported at this frequency.
440
*/
441
while (mode > 0 && mdma_fsclk[mode] > fsclk)
442
mode--;
443
444
nf = num_clocks_min(tfmin, fsclk);
445
if (nf >= 1) {
446
/* calculate the timing values for Multi-word DMA. */
447
448
/* DIOR/DIOW asserted pulse width */
449
td = num_clocks_min(mdma_tdmin[mode], fsclk);
450
451
/* DIOR negated pulse width */
452
tkw = num_clocks_min(mdma_tkwmin[mode], fsclk);
453
454
/* Cycle Time */
455
n0 = num_clocks_min(mdma_t0min[mode], fsclk);
456
457
/* increase tk until we meed the minimum cycle length */
458
if (tkw + td < n0)
459
tkw = n0 - td;
460
461
/* DIOR negated pulse width - read */
462
tkr = num_clocks_min(mdma_tkrmin[mode], fsclk);
463
/* CS{1:0] valid to DIOR/DIOW */
464
tm = num_clocks_min(mdma_tmmin[mode], fsclk);
465
/* DIOR/DIOW to DMACK hold */
466
teoc = num_clocks_min(mdma_tjmin[mode], fsclk);
467
/* DIOW Data hold */
468
th = num_clocks_min(mdma_thmin[mode], fsclk);
469
470
ATAPI_SET_MULTI_TIM_0(base, (tm<<8 | td));
471
ATAPI_SET_MULTI_TIM_1(base, (tkr<<8 | tkw));
472
ATAPI_SET_MULTI_TIM_2(base, (teoc<<8 | th));
473
474
/* Enable host ATAPI Multi DMA interrupts */
475
ATAPI_SET_INT_MASK(base, ATAPI_GET_INT_MASK(base)
476
| MULTI_DONE_MASK | MULTI_TERM_MASK);
477
SSYNC();
478
}
479
}
480
return;
481
}
482
483
/**
484
*
485
* Function: wait_complete
486
*
487
* Description: Waits the interrupt from device
488
*
489
*/
490
static inline void wait_complete(void __iomem *base, unsigned short mask)
491
{
492
unsigned short status;
493
unsigned int i = 0;
494
495
#define PATA_BF54X_WAIT_TIMEOUT 10000
496
497
for (i = 0; i < PATA_BF54X_WAIT_TIMEOUT; i++) {
498
status = ATAPI_GET_INT_STATUS(base) & mask;
499
if (status)
500
break;
501
}
502
503
ATAPI_SET_INT_STATUS(base, mask);
504
}
505
506
/**
507
*
508
* Function: write_atapi_register
509
*
510
* Description: Writes to ATA Device Resgister
511
*
512
*/
513
514
static void write_atapi_register(void __iomem *base,
515
unsigned long ata_reg, unsigned short value)
516
{
517
/* Program the ATA_DEV_TXBUF register with write data (to be
518
* written into the device).
519
*/
520
ATAPI_SET_DEV_TXBUF(base, value);
521
522
/* Program the ATA_DEV_ADDR register with address of the
523
* device register (0x01 to 0x0F).
524
*/
525
ATAPI_SET_DEV_ADDR(base, ata_reg);
526
527
/* Program the ATA_CTRL register with dir set to write (1)
528
*/
529
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) | XFER_DIR));
530
531
/* ensure PIO DMA is not set */
532
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~PIO_USE_DMA));
533
534
/* and start the transfer */
535
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) | PIO_START));
536
537
/* Wait for the interrupt to indicate the end of the transfer.
538
* (We need to wait on and clear rhe ATA_DEV_INT interrupt status)
539
*/
540
wait_complete(base, PIO_DONE_INT);
541
}
542
543
/**
544
*
545
* Function: read_atapi_register
546
*
547
*Description: Reads from ATA Device Resgister
548
*
549
*/
550
551
static unsigned short read_atapi_register(void __iomem *base,
552
unsigned long ata_reg)
553
{
554
/* Program the ATA_DEV_ADDR register with address of the
555
* device register (0x01 to 0x0F).
556
*/
557
ATAPI_SET_DEV_ADDR(base, ata_reg);
558
559
/* Program the ATA_CTRL register with dir set to read (0) and
560
*/
561
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~XFER_DIR));
562
563
/* ensure PIO DMA is not set */
564
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~PIO_USE_DMA));
565
566
/* and start the transfer */
567
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) | PIO_START));
568
569
/* Wait for the interrupt to indicate the end of the transfer.
570
* (PIO_DONE interrupt is set and it doesn't seem to matter
571
* that we don't clear it)
572
*/
573
wait_complete(base, PIO_DONE_INT);
574
575
/* Read the ATA_DEV_RXBUF register with write data (to be
576
* written into the device).
577
*/
578
return ATAPI_GET_DEV_RXBUF(base);
579
}
580
581
/**
582
*
583
* Function: write_atapi_register_data
584
*
585
* Description: Writes to ATA Device Resgister
586
*
587
*/
588
589
static void write_atapi_data(void __iomem *base,
590
int len, unsigned short *buf)
591
{
592
int i;
593
594
/* Set transfer length to 1 */
595
ATAPI_SET_XFER_LEN(base, 1);
596
597
/* Program the ATA_DEV_ADDR register with address of the
598
* ATA_REG_DATA
599
*/
600
ATAPI_SET_DEV_ADDR(base, ATA_REG_DATA);
601
602
/* Program the ATA_CTRL register with dir set to write (1)
603
*/
604
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) | XFER_DIR));
605
606
/* ensure PIO DMA is not set */
607
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~PIO_USE_DMA));
608
609
for (i = 0; i < len; i++) {
610
/* Program the ATA_DEV_TXBUF register with write data (to be
611
* written into the device).
612
*/
613
ATAPI_SET_DEV_TXBUF(base, buf[i]);
614
615
/* and start the transfer */
616
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) | PIO_START));
617
618
/* Wait for the interrupt to indicate the end of the transfer.
619
* (We need to wait on and clear rhe ATA_DEV_INT
620
* interrupt status)
621
*/
622
wait_complete(base, PIO_DONE_INT);
623
}
624
}
625
626
/**
627
*
628
* Function: read_atapi_register_data
629
*
630
* Description: Reads from ATA Device Resgister
631
*
632
*/
633
634
static void read_atapi_data(void __iomem *base,
635
int len, unsigned short *buf)
636
{
637
int i;
638
639
/* Set transfer length to 1 */
640
ATAPI_SET_XFER_LEN(base, 1);
641
642
/* Program the ATA_DEV_ADDR register with address of the
643
* ATA_REG_DATA
644
*/
645
ATAPI_SET_DEV_ADDR(base, ATA_REG_DATA);
646
647
/* Program the ATA_CTRL register with dir set to read (0) and
648
*/
649
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~XFER_DIR));
650
651
/* ensure PIO DMA is not set */
652
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~PIO_USE_DMA));
653
654
for (i = 0; i < len; i++) {
655
/* and start the transfer */
656
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) | PIO_START));
657
658
/* Wait for the interrupt to indicate the end of the transfer.
659
* (PIO_DONE interrupt is set and it doesn't seem to matter
660
* that we don't clear it)
661
*/
662
wait_complete(base, PIO_DONE_INT);
663
664
/* Read the ATA_DEV_RXBUF register with write data (to be
665
* written into the device).
666
*/
667
buf[i] = ATAPI_GET_DEV_RXBUF(base);
668
}
669
}
670
671
/**
672
* bfin_tf_load - send taskfile registers to host controller
673
* @ap: Port to which output is sent
674
* @tf: ATA taskfile register set
675
*
676
* Note: Original code is ata_sff_tf_load().
677
*/
678
679
static void bfin_tf_load(struct ata_port *ap, const struct ata_taskfile *tf)
680
{
681
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
682
unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR;
683
684
if (tf->ctl != ap->last_ctl) {
685
write_atapi_register(base, ATA_REG_CTRL, tf->ctl);
686
ap->last_ctl = tf->ctl;
687
ata_wait_idle(ap);
688
}
689
690
if (is_addr) {
691
if (tf->flags & ATA_TFLAG_LBA48) {
692
write_atapi_register(base, ATA_REG_FEATURE,
693
tf->hob_feature);
694
write_atapi_register(base, ATA_REG_NSECT,
695
tf->hob_nsect);
696
write_atapi_register(base, ATA_REG_LBAL, tf->hob_lbal);
697
write_atapi_register(base, ATA_REG_LBAM, tf->hob_lbam);
698
write_atapi_register(base, ATA_REG_LBAH, tf->hob_lbah);
699
dev_dbg(ap->dev, "hob: feat 0x%X nsect 0x%X, lba 0x%X "
700
"0x%X 0x%X\n",
701
tf->hob_feature,
702
tf->hob_nsect,
703
tf->hob_lbal,
704
tf->hob_lbam,
705
tf->hob_lbah);
706
}
707
708
write_atapi_register(base, ATA_REG_FEATURE, tf->feature);
709
write_atapi_register(base, ATA_REG_NSECT, tf->nsect);
710
write_atapi_register(base, ATA_REG_LBAL, tf->lbal);
711
write_atapi_register(base, ATA_REG_LBAM, tf->lbam);
712
write_atapi_register(base, ATA_REG_LBAH, tf->lbah);
713
dev_dbg(ap->dev, "feat 0x%X nsect 0x%X lba 0x%X 0x%X 0x%X\n",
714
tf->feature,
715
tf->nsect,
716
tf->lbal,
717
tf->lbam,
718
tf->lbah);
719
}
720
721
if (tf->flags & ATA_TFLAG_DEVICE) {
722
write_atapi_register(base, ATA_REG_DEVICE, tf->device);
723
dev_dbg(ap->dev, "device 0x%X\n", tf->device);
724
}
725
726
ata_wait_idle(ap);
727
}
728
729
/**
730
* bfin_check_status - Read device status reg & clear interrupt
731
* @ap: port where the device is
732
*
733
* Note: Original code is ata_check_status().
734
*/
735
736
static u8 bfin_check_status(struct ata_port *ap)
737
{
738
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
739
return read_atapi_register(base, ATA_REG_STATUS);
740
}
741
742
/**
743
* bfin_tf_read - input device's ATA taskfile shadow registers
744
* @ap: Port from which input is read
745
* @tf: ATA taskfile register set for storing input
746
*
747
* Note: Original code is ata_sff_tf_read().
748
*/
749
750
static void bfin_tf_read(struct ata_port *ap, struct ata_taskfile *tf)
751
{
752
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
753
754
tf->command = bfin_check_status(ap);
755
tf->feature = read_atapi_register(base, ATA_REG_ERR);
756
tf->nsect = read_atapi_register(base, ATA_REG_NSECT);
757
tf->lbal = read_atapi_register(base, ATA_REG_LBAL);
758
tf->lbam = read_atapi_register(base, ATA_REG_LBAM);
759
tf->lbah = read_atapi_register(base, ATA_REG_LBAH);
760
tf->device = read_atapi_register(base, ATA_REG_DEVICE);
761
762
if (tf->flags & ATA_TFLAG_LBA48) {
763
write_atapi_register(base, ATA_REG_CTRL, tf->ctl | ATA_HOB);
764
tf->hob_feature = read_atapi_register(base, ATA_REG_ERR);
765
tf->hob_nsect = read_atapi_register(base, ATA_REG_NSECT);
766
tf->hob_lbal = read_atapi_register(base, ATA_REG_LBAL);
767
tf->hob_lbam = read_atapi_register(base, ATA_REG_LBAM);
768
tf->hob_lbah = read_atapi_register(base, ATA_REG_LBAH);
769
}
770
}
771
772
/**
773
* bfin_exec_command - issue ATA command to host controller
774
* @ap: port to which command is being issued
775
* @tf: ATA taskfile register set
776
*
777
* Note: Original code is ata_sff_exec_command().
778
*/
779
780
static void bfin_exec_command(struct ata_port *ap,
781
const struct ata_taskfile *tf)
782
{
783
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
784
dev_dbg(ap->dev, "ata%u: cmd 0x%X\n", ap->print_id, tf->command);
785
786
write_atapi_register(base, ATA_REG_CMD, tf->command);
787
ata_sff_pause(ap);
788
}
789
790
/**
791
* bfin_check_altstatus - Read device alternate status reg
792
* @ap: port where the device is
793
*/
794
795
static u8 bfin_check_altstatus(struct ata_port *ap)
796
{
797
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
798
return read_atapi_register(base, ATA_REG_ALTSTATUS);
799
}
800
801
/**
802
* bfin_dev_select - Select device 0/1 on ATA bus
803
* @ap: ATA channel to manipulate
804
* @device: ATA device (numbered from zero) to select
805
*
806
* Note: Original code is ata_sff_dev_select().
807
*/
808
809
static void bfin_dev_select(struct ata_port *ap, unsigned int device)
810
{
811
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
812
u8 tmp;
813
814
if (device == 0)
815
tmp = ATA_DEVICE_OBS;
816
else
817
tmp = ATA_DEVICE_OBS | ATA_DEV1;
818
819
write_atapi_register(base, ATA_REG_DEVICE, tmp);
820
ata_sff_pause(ap);
821
}
822
823
/**
824
* bfin_set_devctl - Write device control reg
825
* @ap: port where the device is
826
* @ctl: value to write
827
*/
828
829
static void bfin_set_devctl(struct ata_port *ap, u8 ctl)
830
{
831
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
832
write_atapi_register(base, ATA_REG_CTRL, ctl);
833
}
834
835
/**
836
* bfin_bmdma_setup - Set up IDE DMA transaction
837
* @qc: Info associated with this ATA transaction.
838
*
839
* Note: Original code is ata_bmdma_setup().
840
*/
841
842
static void bfin_bmdma_setup(struct ata_queued_cmd *qc)
843
{
844
unsigned short config = WDSIZE_16;
845
struct scatterlist *sg;
846
unsigned int si;
847
848
dev_dbg(qc->ap->dev, "in atapi dma setup\n");
849
/* Program the ATA_CTRL register with dir */
850
if (qc->tf.flags & ATA_TFLAG_WRITE) {
851
/* fill the ATAPI DMA controller */
852
set_dma_config(CH_ATAPI_TX, config);
853
set_dma_x_modify(CH_ATAPI_TX, 2);
854
for_each_sg(qc->sg, sg, qc->n_elem, si) {
855
set_dma_start_addr(CH_ATAPI_TX, sg_dma_address(sg));
856
set_dma_x_count(CH_ATAPI_TX, sg_dma_len(sg) >> 1);
857
}
858
} else {
859
config |= WNR;
860
/* fill the ATAPI DMA controller */
861
set_dma_config(CH_ATAPI_RX, config);
862
set_dma_x_modify(CH_ATAPI_RX, 2);
863
for_each_sg(qc->sg, sg, qc->n_elem, si) {
864
set_dma_start_addr(CH_ATAPI_RX, sg_dma_address(sg));
865
set_dma_x_count(CH_ATAPI_RX, sg_dma_len(sg) >> 1);
866
}
867
}
868
}
869
870
/**
871
* bfin_bmdma_start - Start an IDE DMA transaction
872
* @qc: Info associated with this ATA transaction.
873
*
874
* Note: Original code is ata_bmdma_start().
875
*/
876
877
static void bfin_bmdma_start(struct ata_queued_cmd *qc)
878
{
879
struct ata_port *ap = qc->ap;
880
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
881
struct scatterlist *sg;
882
unsigned int si;
883
884
dev_dbg(qc->ap->dev, "in atapi dma start\n");
885
if (!(ap->udma_mask || ap->mwdma_mask))
886
return;
887
888
/* start ATAPI DMA controller*/
889
if (qc->tf.flags & ATA_TFLAG_WRITE) {
890
/*
891
* On blackfin arch, uncacheable memory is not
892
* allocated with flag GFP_DMA. DMA buffer from
893
* common kenel code should be flushed if WB
894
* data cache is enabled. Otherwise, this loop
895
* is an empty loop and optimized out.
896
*/
897
for_each_sg(qc->sg, sg, qc->n_elem, si) {
898
flush_dcache_range(sg_dma_address(sg),
899
sg_dma_address(sg) + sg_dma_len(sg));
900
}
901
enable_dma(CH_ATAPI_TX);
902
dev_dbg(qc->ap->dev, "enable udma write\n");
903
904
/* Send ATA DMA write command */
905
bfin_exec_command(ap, &qc->tf);
906
907
/* set ATA DMA write direction */
908
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base)
909
| XFER_DIR));
910
} else {
911
enable_dma(CH_ATAPI_RX);
912
dev_dbg(qc->ap->dev, "enable udma read\n");
913
914
/* Send ATA DMA read command */
915
bfin_exec_command(ap, &qc->tf);
916
917
/* set ATA DMA read direction */
918
ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base)
919
& ~XFER_DIR));
920
}
921
922
/* Reset all transfer count */
923
ATAPI_SET_CONTROL(base, ATAPI_GET_CONTROL(base) | TFRCNT_RST);
924
925
/* Set ATAPI state machine contorl in terminate sequence */
926
ATAPI_SET_CONTROL(base, ATAPI_GET_CONTROL(base) | END_ON_TERM);
927
928
/* Set transfer length to buffer len */
929
for_each_sg(qc->sg, sg, qc->n_elem, si) {
930
ATAPI_SET_XFER_LEN(base, (sg_dma_len(sg) >> 1));
931
}
932
933
/* Enable ATA DMA operation*/
934
if (ap->udma_mask)
935
ATAPI_SET_CONTROL(base, ATAPI_GET_CONTROL(base)
936
| ULTRA_START);
937
else
938
ATAPI_SET_CONTROL(base, ATAPI_GET_CONTROL(base)
939
| MULTI_START);
940
}
941
942
/**
943
* bfin_bmdma_stop - Stop IDE DMA transfer
944
* @qc: Command we are ending DMA for
945
*/
946
947
static void bfin_bmdma_stop(struct ata_queued_cmd *qc)
948
{
949
struct ata_port *ap = qc->ap;
950
struct scatterlist *sg;
951
unsigned int si;
952
953
dev_dbg(qc->ap->dev, "in atapi dma stop\n");
954
if (!(ap->udma_mask || ap->mwdma_mask))
955
return;
956
957
/* stop ATAPI DMA controller*/
958
if (qc->tf.flags & ATA_TFLAG_WRITE)
959
disable_dma(CH_ATAPI_TX);
960
else {
961
disable_dma(CH_ATAPI_RX);
962
if (ap->hsm_task_state & HSM_ST_LAST) {
963
/*
964
* On blackfin arch, uncacheable memory is not
965
* allocated with flag GFP_DMA. DMA buffer from
966
* common kenel code should be invalidated if
967
* data cache is enabled. Otherwise, this loop
968
* is an empty loop and optimized out.
969
*/
970
for_each_sg(qc->sg, sg, qc->n_elem, si) {
971
invalidate_dcache_range(
972
sg_dma_address(sg),
973
sg_dma_address(sg)
974
+ sg_dma_len(sg));
975
}
976
}
977
}
978
}
979
980
/**
981
* bfin_devchk - PATA device presence detection
982
* @ap: ATA channel to examine
983
* @device: Device to examine (starting at zero)
984
*
985
* Note: Original code is ata_devchk().
986
*/
987
988
static unsigned int bfin_devchk(struct ata_port *ap,
989
unsigned int device)
990
{
991
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
992
u8 nsect, lbal;
993
994
bfin_dev_select(ap, device);
995
996
write_atapi_register(base, ATA_REG_NSECT, 0x55);
997
write_atapi_register(base, ATA_REG_LBAL, 0xaa);
998
999
write_atapi_register(base, ATA_REG_NSECT, 0xaa);
1000
write_atapi_register(base, ATA_REG_LBAL, 0x55);
1001
1002
write_atapi_register(base, ATA_REG_NSECT, 0x55);
1003
write_atapi_register(base, ATA_REG_LBAL, 0xaa);
1004
1005
nsect = read_atapi_register(base, ATA_REG_NSECT);
1006
lbal = read_atapi_register(base, ATA_REG_LBAL);
1007
1008
if ((nsect == 0x55) && (lbal == 0xaa))
1009
return 1; /* we found a device */
1010
1011
return 0; /* nothing found */
1012
}
1013
1014
/**
1015
* bfin_bus_post_reset - PATA device post reset
1016
*
1017
* Note: Original code is ata_bus_post_reset().
1018
*/
1019
1020
static void bfin_bus_post_reset(struct ata_port *ap, unsigned int devmask)
1021
{
1022
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1023
unsigned int dev0 = devmask & (1 << 0);
1024
unsigned int dev1 = devmask & (1 << 1);
1025
unsigned long deadline;
1026
1027
/* if device 0 was found in ata_devchk, wait for its
1028
* BSY bit to clear
1029
*/
1030
if (dev0)
1031
ata_sff_busy_sleep(ap, ATA_TMOUT_BOOT_QUICK, ATA_TMOUT_BOOT);
1032
1033
/* if device 1 was found in ata_devchk, wait for
1034
* register access, then wait for BSY to clear
1035
*/
1036
deadline = ata_deadline(jiffies, ATA_TMOUT_BOOT);
1037
while (dev1) {
1038
u8 nsect, lbal;
1039
1040
bfin_dev_select(ap, 1);
1041
nsect = read_atapi_register(base, ATA_REG_NSECT);
1042
lbal = read_atapi_register(base, ATA_REG_LBAL);
1043
if ((nsect == 1) && (lbal == 1))
1044
break;
1045
if (time_after(jiffies, deadline)) {
1046
dev1 = 0;
1047
break;
1048
}
1049
ata_msleep(ap, 50); /* give drive a breather */
1050
}
1051
if (dev1)
1052
ata_sff_busy_sleep(ap, ATA_TMOUT_BOOT_QUICK, ATA_TMOUT_BOOT);
1053
1054
/* is all this really necessary? */
1055
bfin_dev_select(ap, 0);
1056
if (dev1)
1057
bfin_dev_select(ap, 1);
1058
if (dev0)
1059
bfin_dev_select(ap, 0);
1060
}
1061
1062
/**
1063
* bfin_bus_softreset - PATA device software reset
1064
*
1065
* Note: Original code is ata_bus_softreset().
1066
*/
1067
1068
static unsigned int bfin_bus_softreset(struct ata_port *ap,
1069
unsigned int devmask)
1070
{
1071
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1072
1073
/* software reset. causes dev0 to be selected */
1074
write_atapi_register(base, ATA_REG_CTRL, ap->ctl);
1075
udelay(20);
1076
write_atapi_register(base, ATA_REG_CTRL, ap->ctl | ATA_SRST);
1077
udelay(20);
1078
write_atapi_register(base, ATA_REG_CTRL, ap->ctl);
1079
1080
/* spec mandates ">= 2ms" before checking status.
1081
* We wait 150ms, because that was the magic delay used for
1082
* ATAPI devices in Hale Landis's ATADRVR, for the period of time
1083
* between when the ATA command register is written, and then
1084
* status is checked. Because waiting for "a while" before
1085
* checking status is fine, post SRST, we perform this magic
1086
* delay here as well.
1087
*
1088
* Old drivers/ide uses the 2mS rule and then waits for ready
1089
*/
1090
ata_msleep(ap, 150);
1091
1092
/* Before we perform post reset processing we want to see if
1093
* the bus shows 0xFF because the odd clown forgets the D7
1094
* pulldown resistor.
1095
*/
1096
if (bfin_check_status(ap) == 0xFF)
1097
return 0;
1098
1099
bfin_bus_post_reset(ap, devmask);
1100
1101
return 0;
1102
}
1103
1104
/**
1105
* bfin_softreset - reset host port via ATA SRST
1106
* @ap: port to reset
1107
* @classes: resulting classes of attached devices
1108
*
1109
* Note: Original code is ata_sff_softreset().
1110
*/
1111
1112
static int bfin_softreset(struct ata_link *link, unsigned int *classes,
1113
unsigned long deadline)
1114
{
1115
struct ata_port *ap = link->ap;
1116
unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS;
1117
unsigned int devmask = 0, err_mask;
1118
u8 err;
1119
1120
/* determine if device 0/1 are present */
1121
if (bfin_devchk(ap, 0))
1122
devmask |= (1 << 0);
1123
if (slave_possible && bfin_devchk(ap, 1))
1124
devmask |= (1 << 1);
1125
1126
/* select device 0 again */
1127
bfin_dev_select(ap, 0);
1128
1129
/* issue bus reset */
1130
err_mask = bfin_bus_softreset(ap, devmask);
1131
if (err_mask) {
1132
ata_port_printk(ap, KERN_ERR, "SRST failed (err_mask=0x%x)\n",
1133
err_mask);
1134
return -EIO;
1135
}
1136
1137
/* determine by signature whether we have ATA or ATAPI devices */
1138
classes[0] = ata_sff_dev_classify(&ap->link.device[0],
1139
devmask & (1 << 0), &err);
1140
if (slave_possible && err != 0x81)
1141
classes[1] = ata_sff_dev_classify(&ap->link.device[1],
1142
devmask & (1 << 1), &err);
1143
1144
return 0;
1145
}
1146
1147
/**
1148
* bfin_bmdma_status - Read IDE DMA status
1149
* @ap: Port associated with this ATA transaction.
1150
*/
1151
1152
static unsigned char bfin_bmdma_status(struct ata_port *ap)
1153
{
1154
unsigned char host_stat = 0;
1155
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1156
unsigned short int_status = ATAPI_GET_INT_STATUS(base);
1157
1158
if (ATAPI_GET_STATUS(base) & (MULTI_XFER_ON|ULTRA_XFER_ON))
1159
host_stat |= ATA_DMA_ACTIVE;
1160
if (int_status & (MULTI_DONE_INT|UDMAIN_DONE_INT|UDMAOUT_DONE_INT|
1161
ATAPI_DEV_INT))
1162
host_stat |= ATA_DMA_INTR;
1163
if (int_status & (MULTI_TERM_INT|UDMAIN_TERM_INT|UDMAOUT_TERM_INT))
1164
host_stat |= ATA_DMA_ERR|ATA_DMA_INTR;
1165
1166
dev_dbg(ap->dev, "ATAPI: host_stat=0x%x\n", host_stat);
1167
1168
return host_stat;
1169
}
1170
1171
/**
1172
* bfin_data_xfer - Transfer data by PIO
1173
* @adev: device for this I/O
1174
* @buf: data buffer
1175
* @buflen: buffer length
1176
* @write_data: read/write
1177
*
1178
* Note: Original code is ata_sff_data_xfer().
1179
*/
1180
1181
static unsigned int bfin_data_xfer(struct ata_device *dev, unsigned char *buf,
1182
unsigned int buflen, int rw)
1183
{
1184
struct ata_port *ap = dev->link->ap;
1185
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1186
unsigned int words = buflen >> 1;
1187
unsigned short *buf16 = (u16 *)buf;
1188
1189
/* Transfer multiple of 2 bytes */
1190
if (rw == READ)
1191
read_atapi_data(base, words, buf16);
1192
else
1193
write_atapi_data(base, words, buf16);
1194
1195
/* Transfer trailing 1 byte, if any. */
1196
if (unlikely(buflen & 0x01)) {
1197
unsigned short align_buf[1] = { 0 };
1198
unsigned char *trailing_buf = buf + buflen - 1;
1199
1200
if (rw == READ) {
1201
read_atapi_data(base, 1, align_buf);
1202
memcpy(trailing_buf, align_buf, 1);
1203
} else {
1204
memcpy(align_buf, trailing_buf, 1);
1205
write_atapi_data(base, 1, align_buf);
1206
}
1207
words++;
1208
}
1209
1210
return words << 1;
1211
}
1212
1213
/**
1214
* bfin_irq_clear - Clear ATAPI interrupt.
1215
* @ap: Port associated with this ATA transaction.
1216
*
1217
* Note: Original code is ata_bmdma_irq_clear().
1218
*/
1219
1220
static void bfin_irq_clear(struct ata_port *ap)
1221
{
1222
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1223
1224
dev_dbg(ap->dev, "in atapi irq clear\n");
1225
ATAPI_SET_INT_STATUS(base, ATAPI_GET_INT_STATUS(base)|ATAPI_DEV_INT
1226
| MULTI_DONE_INT | UDMAIN_DONE_INT | UDMAOUT_DONE_INT
1227
| MULTI_TERM_INT | UDMAIN_TERM_INT | UDMAOUT_TERM_INT);
1228
}
1229
1230
/**
1231
* bfin_thaw - Thaw DMA controller port
1232
* @ap: port to thaw
1233
*
1234
* Note: Original code is ata_sff_thaw().
1235
*/
1236
1237
void bfin_thaw(struct ata_port *ap)
1238
{
1239
dev_dbg(ap->dev, "in atapi dma thaw\n");
1240
bfin_check_status(ap);
1241
ata_sff_irq_on(ap);
1242
}
1243
1244
/**
1245
* bfin_postreset - standard postreset callback
1246
* @ap: the target ata_port
1247
* @classes: classes of attached devices
1248
*
1249
* Note: Original code is ata_sff_postreset().
1250
*/
1251
1252
static void bfin_postreset(struct ata_link *link, unsigned int *classes)
1253
{
1254
struct ata_port *ap = link->ap;
1255
void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1256
1257
/* re-enable interrupts */
1258
ata_sff_irq_on(ap);
1259
1260
/* is double-select really necessary? */
1261
if (classes[0] != ATA_DEV_NONE)
1262
bfin_dev_select(ap, 1);
1263
if (classes[1] != ATA_DEV_NONE)
1264
bfin_dev_select(ap, 0);
1265
1266
/* bail out if no device is present */
1267
if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) {
1268
return;
1269
}
1270
1271
/* set up device control */
1272
write_atapi_register(base, ATA_REG_CTRL, ap->ctl);
1273
}
1274
1275
static void bfin_port_stop(struct ata_port *ap)
1276
{
1277
dev_dbg(ap->dev, "in atapi port stop\n");
1278
if (ap->udma_mask != 0 || ap->mwdma_mask != 0) {
1279
free_dma(CH_ATAPI_RX);
1280
free_dma(CH_ATAPI_TX);
1281
}
1282
}
1283
1284
static int bfin_port_start(struct ata_port *ap)
1285
{
1286
dev_dbg(ap->dev, "in atapi port start\n");
1287
if (!(ap->udma_mask || ap->mwdma_mask))
1288
return 0;
1289
1290
if (request_dma(CH_ATAPI_RX, "BFIN ATAPI RX DMA") >= 0) {
1291
if (request_dma(CH_ATAPI_TX,
1292
"BFIN ATAPI TX DMA") >= 0)
1293
return 0;
1294
1295
free_dma(CH_ATAPI_RX);
1296
}
1297
1298
ap->udma_mask = 0;
1299
ap->mwdma_mask = 0;
1300
dev_err(ap->dev, "Unable to request ATAPI DMA!"
1301
" Continue in PIO mode.\n");
1302
1303
return 0;
1304
}
1305
1306
static unsigned int bfin_ata_host_intr(struct ata_port *ap,
1307
struct ata_queued_cmd *qc)
1308
{
1309
struct ata_eh_info *ehi = &ap->link.eh_info;
1310
u8 status, host_stat = 0;
1311
1312
VPRINTK("ata%u: protocol %d task_state %d\n",
1313
ap->print_id, qc->tf.protocol, ap->hsm_task_state);
1314
1315
/* Check whether we are expecting interrupt in this state */
1316
switch (ap->hsm_task_state) {
1317
case HSM_ST_FIRST:
1318
/* Some pre-ATAPI-4 devices assert INTRQ
1319
* at this state when ready to receive CDB.
1320
*/
1321
1322
/* Check the ATA_DFLAG_CDB_INTR flag is enough here.
1323
* The flag was turned on only for atapi devices.
1324
* No need to check is_atapi_taskfile(&qc->tf) again.
1325
*/
1326
if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR))
1327
goto idle_irq;
1328
break;
1329
case HSM_ST_LAST:
1330
if (qc->tf.protocol == ATA_PROT_DMA ||
1331
qc->tf.protocol == ATAPI_PROT_DMA) {
1332
/* check status of DMA engine */
1333
host_stat = ap->ops->bmdma_status(ap);
1334
VPRINTK("ata%u: host_stat 0x%X\n",
1335
ap->print_id, host_stat);
1336
1337
/* if it's not our irq... */
1338
if (!(host_stat & ATA_DMA_INTR))
1339
goto idle_irq;
1340
1341
/* before we do anything else, clear DMA-Start bit */
1342
ap->ops->bmdma_stop(qc);
1343
1344
if (unlikely(host_stat & ATA_DMA_ERR)) {
1345
/* error when transferring data to/from memory */
1346
qc->err_mask |= AC_ERR_HOST_BUS;
1347
ap->hsm_task_state = HSM_ST_ERR;
1348
}
1349
}
1350
break;
1351
case HSM_ST:
1352
break;
1353
default:
1354
goto idle_irq;
1355
}
1356
1357
/* check altstatus */
1358
status = ap->ops->sff_check_altstatus(ap);
1359
if (status & ATA_BUSY)
1360
goto busy_ata;
1361
1362
/* check main status, clearing INTRQ */
1363
status = ap->ops->sff_check_status(ap);
1364
if (unlikely(status & ATA_BUSY))
1365
goto busy_ata;
1366
1367
/* ack bmdma irq events */
1368
ap->ops->sff_irq_clear(ap);
1369
1370
ata_sff_hsm_move(ap, qc, status, 0);
1371
1372
if (unlikely(qc->err_mask) && (qc->tf.protocol == ATA_PROT_DMA ||
1373
qc->tf.protocol == ATAPI_PROT_DMA))
1374
ata_ehi_push_desc(ehi, "BMDMA stat 0x%x", host_stat);
1375
1376
busy_ata:
1377
return 1; /* irq handled */
1378
1379
idle_irq:
1380
ap->stats.idle_irq++;
1381
1382
#ifdef ATA_IRQ_TRAP
1383
if ((ap->stats.idle_irq % 1000) == 0) {
1384
ap->ops->irq_ack(ap, 0); /* debug trap */
1385
ata_port_printk(ap, KERN_WARNING, "irq trap\n");
1386
return 1;
1387
}
1388
#endif
1389
return 0; /* irq not handled */
1390
}
1391
1392
static irqreturn_t bfin_ata_interrupt(int irq, void *dev_instance)
1393
{
1394
struct ata_host *host = dev_instance;
1395
unsigned int i;
1396
unsigned int handled = 0;
1397
unsigned long flags;
1398
1399
/* TODO: make _irqsave conditional on x86 PCI IDE legacy mode */
1400
spin_lock_irqsave(&host->lock, flags);
1401
1402
for (i = 0; i < host->n_ports; i++) {
1403
struct ata_port *ap = host->ports[i];
1404
struct ata_queued_cmd *qc;
1405
1406
qc = ata_qc_from_tag(ap, ap->link.active_tag);
1407
if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING)))
1408
handled |= bfin_ata_host_intr(ap, qc);
1409
}
1410
1411
spin_unlock_irqrestore(&host->lock, flags);
1412
1413
return IRQ_RETVAL(handled);
1414
}
1415
1416
1417
static struct scsi_host_template bfin_sht = {
1418
ATA_BASE_SHT(DRV_NAME),
1419
.sg_tablesize = SG_NONE,
1420
.dma_boundary = ATA_DMA_BOUNDARY,
1421
};
1422
1423
static struct ata_port_operations bfin_pata_ops = {
1424
.inherits = &ata_bmdma_port_ops,
1425
1426
.set_piomode = bfin_set_piomode,
1427
.set_dmamode = bfin_set_dmamode,
1428
1429
.sff_tf_load = bfin_tf_load,
1430
.sff_tf_read = bfin_tf_read,
1431
.sff_exec_command = bfin_exec_command,
1432
.sff_check_status = bfin_check_status,
1433
.sff_check_altstatus = bfin_check_altstatus,
1434
.sff_dev_select = bfin_dev_select,
1435
.sff_set_devctl = bfin_set_devctl,
1436
1437
.bmdma_setup = bfin_bmdma_setup,
1438
.bmdma_start = bfin_bmdma_start,
1439
.bmdma_stop = bfin_bmdma_stop,
1440
.bmdma_status = bfin_bmdma_status,
1441
.sff_data_xfer = bfin_data_xfer,
1442
1443
.qc_prep = ata_noop_qc_prep,
1444
1445
.thaw = bfin_thaw,
1446
.softreset = bfin_softreset,
1447
.postreset = bfin_postreset,
1448
1449
.sff_irq_clear = bfin_irq_clear,
1450
1451
.port_start = bfin_port_start,
1452
.port_stop = bfin_port_stop,
1453
};
1454
1455
static struct ata_port_info bfin_port_info[] = {
1456
{
1457
.flags = ATA_FLAG_SLAVE_POSS,
1458
.pio_mask = ATA_PIO4,
1459
.mwdma_mask = 0,
1460
.udma_mask = 0,
1461
.port_ops = &bfin_pata_ops,
1462
},
1463
};
1464
1465
/**
1466
* bfin_reset_controller - initialize BF54x ATAPI controller.
1467
*/
1468
1469
static int bfin_reset_controller(struct ata_host *host)
1470
{
1471
void __iomem *base = (void __iomem *)host->ports[0]->ioaddr.ctl_addr;
1472
int count;
1473
unsigned short status;
1474
1475
/* Disable all ATAPI interrupts */
1476
ATAPI_SET_INT_MASK(base, 0);
1477
SSYNC();
1478
1479
/* Assert the RESET signal 25us*/
1480
ATAPI_SET_CONTROL(base, ATAPI_GET_CONTROL(base) | DEV_RST);
1481
udelay(30);
1482
1483
/* Negate the RESET signal for 2ms*/
1484
ATAPI_SET_CONTROL(base, ATAPI_GET_CONTROL(base) & ~DEV_RST);
1485
msleep(2);
1486
1487
/* Wait on Busy flag to clear */
1488
count = 10000000;
1489
do {
1490
status = read_atapi_register(base, ATA_REG_STATUS);
1491
} while (--count && (status & ATA_BUSY));
1492
1493
/* Enable only ATAPI Device interrupt */
1494
ATAPI_SET_INT_MASK(base, 1);
1495
SSYNC();
1496
1497
return (!count);
1498
}
1499
1500
/**
1501
* atapi_io_port - define atapi peripheral port pins.
1502
*/
1503
static unsigned short atapi_io_port[] = {
1504
P_ATAPI_RESET,
1505
P_ATAPI_DIOR,
1506
P_ATAPI_DIOW,
1507
P_ATAPI_CS0,
1508
P_ATAPI_CS1,
1509
P_ATAPI_DMACK,
1510
P_ATAPI_DMARQ,
1511
P_ATAPI_INTRQ,
1512
P_ATAPI_IORDY,
1513
P_ATAPI_D0A,
1514
P_ATAPI_D1A,
1515
P_ATAPI_D2A,
1516
P_ATAPI_D3A,
1517
P_ATAPI_D4A,
1518
P_ATAPI_D5A,
1519
P_ATAPI_D6A,
1520
P_ATAPI_D7A,
1521
P_ATAPI_D8A,
1522
P_ATAPI_D9A,
1523
P_ATAPI_D10A,
1524
P_ATAPI_D11A,
1525
P_ATAPI_D12A,
1526
P_ATAPI_D13A,
1527
P_ATAPI_D14A,
1528
P_ATAPI_D15A,
1529
P_ATAPI_A0A,
1530
P_ATAPI_A1A,
1531
P_ATAPI_A2A,
1532
0
1533
};
1534
1535
/**
1536
* bfin_atapi_probe - attach a bfin atapi interface
1537
* @pdev: platform device
1538
*
1539
* Register a bfin atapi interface.
1540
*
1541
*
1542
* Platform devices are expected to contain 2 resources per port:
1543
*
1544
* - I/O Base (IORESOURCE_IO)
1545
* - IRQ (IORESOURCE_IRQ)
1546
*
1547
*/
1548
static int __devinit bfin_atapi_probe(struct platform_device *pdev)
1549
{
1550
int board_idx = 0;
1551
struct resource *res;
1552
struct ata_host *host;
1553
unsigned int fsclk = get_sclk();
1554
int udma_mode = 5;
1555
const struct ata_port_info *ppi[] =
1556
{ &bfin_port_info[board_idx], NULL };
1557
1558
/*
1559
* Simple resource validation ..
1560
*/
1561
if (unlikely(pdev->num_resources != 2)) {
1562
dev_err(&pdev->dev, "invalid number of resources\n");
1563
return -EINVAL;
1564
}
1565
1566
/*
1567
* Get the register base first
1568
*/
1569
res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1570
if (res == NULL)
1571
return -EINVAL;
1572
1573
while (bfin_port_info[board_idx].udma_mask > 0 &&
1574
udma_fsclk[udma_mode] > fsclk) {
1575
udma_mode--;
1576
bfin_port_info[board_idx].udma_mask >>= 1;
1577
}
1578
1579
/*
1580
* Now that that's out of the way, wire up the port..
1581
*/
1582
host = ata_host_alloc_pinfo(&pdev->dev, ppi, 1);
1583
if (!host)
1584
return -ENOMEM;
1585
1586
host->ports[0]->ioaddr.ctl_addr = (void *)res->start;
1587
1588
if (peripheral_request_list(atapi_io_port, "atapi-io-port")) {
1589
dev_err(&pdev->dev, "Requesting Peripherals failed\n");
1590
return -EFAULT;
1591
}
1592
1593
if (bfin_reset_controller(host)) {
1594
peripheral_free_list(atapi_io_port);
1595
dev_err(&pdev->dev, "Fail to reset ATAPI device\n");
1596
return -EFAULT;
1597
}
1598
1599
if (ata_host_activate(host, platform_get_irq(pdev, 0),
1600
bfin_ata_interrupt, IRQF_SHARED, &bfin_sht) != 0) {
1601
peripheral_free_list(atapi_io_port);
1602
dev_err(&pdev->dev, "Fail to attach ATAPI device\n");
1603
return -ENODEV;
1604
}
1605
1606
dev_set_drvdata(&pdev->dev, host);
1607
1608
return 0;
1609
}
1610
1611
/**
1612
* bfin_atapi_remove - unplug a bfin atapi interface
1613
* @pdev: platform device
1614
*
1615
* A bfin atapi device has been unplugged. Perform the needed
1616
* cleanup. Also called on module unload for any active devices.
1617
*/
1618
static int __devexit bfin_atapi_remove(struct platform_device *pdev)
1619
{
1620
struct device *dev = &pdev->dev;
1621
struct ata_host *host = dev_get_drvdata(dev);
1622
1623
ata_host_detach(host);
1624
dev_set_drvdata(&pdev->dev, NULL);
1625
1626
peripheral_free_list(atapi_io_port);
1627
1628
return 0;
1629
}
1630
1631
#ifdef CONFIG_PM
1632
static int bfin_atapi_suspend(struct platform_device *pdev, pm_message_t state)
1633
{
1634
struct ata_host *host = dev_get_drvdata(&pdev->dev);
1635
if (host)
1636
return ata_host_suspend(host, state);
1637
else
1638
return 0;
1639
}
1640
1641
static int bfin_atapi_resume(struct platform_device *pdev)
1642
{
1643
struct ata_host *host = dev_get_drvdata(&pdev->dev);
1644
int ret;
1645
1646
if (host) {
1647
ret = bfin_reset_controller(host);
1648
if (ret) {
1649
printk(KERN_ERR DRV_NAME ": Error during HW init\n");
1650
return ret;
1651
}
1652
ata_host_resume(host);
1653
}
1654
1655
return 0;
1656
}
1657
#else
1658
#define bfin_atapi_suspend NULL
1659
#define bfin_atapi_resume NULL
1660
#endif
1661
1662
static struct platform_driver bfin_atapi_driver = {
1663
.probe = bfin_atapi_probe,
1664
.remove = __devexit_p(bfin_atapi_remove),
1665
.suspend = bfin_atapi_suspend,
1666
.resume = bfin_atapi_resume,
1667
.driver = {
1668
.name = DRV_NAME,
1669
.owner = THIS_MODULE,
1670
},
1671
};
1672
1673
#define ATAPI_MODE_SIZE 10
1674
static char bfin_atapi_mode[ATAPI_MODE_SIZE];
1675
1676
static int __init bfin_atapi_init(void)
1677
{
1678
pr_info("register bfin atapi driver\n");
1679
1680
switch(bfin_atapi_mode[0]) {
1681
case 'p':
1682
case 'P':
1683
break;
1684
case 'm':
1685
case 'M':
1686
bfin_port_info[0].mwdma_mask = ATA_MWDMA2;
1687
break;
1688
default:
1689
bfin_port_info[0].udma_mask = ATA_UDMA5;
1690
};
1691
1692
return platform_driver_register(&bfin_atapi_driver);
1693
}
1694
1695
static void __exit bfin_atapi_exit(void)
1696
{
1697
platform_driver_unregister(&bfin_atapi_driver);
1698
}
1699
1700
module_init(bfin_atapi_init);
1701
module_exit(bfin_atapi_exit);
1702
/*
1703
* ATAPI mode:
1704
* pio/PIO
1705
* udma/UDMA (default)
1706
* mwdma/MWDMA
1707
*/
1708
module_param_string(bfin_atapi_mode, bfin_atapi_mode, ATAPI_MODE_SIZE, 0);
1709
1710
MODULE_AUTHOR("Sonic Zhang <[email protected]>");
1711
MODULE_DESCRIPTION("PATA driver for blackfin 54x ATAPI controller");
1712
MODULE_LICENSE("GPL");
1713
MODULE_VERSION(DRV_VERSION);
1714
MODULE_ALIAS("platform:" DRV_NAME);
1715
1716