Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
nexedi
linux
Commits
0f8b375f
Commit
0f8b375f
authored
Feb 19, 2004
by
Linus Torvalds
Browse files
Options
Browse Files
Download
Plain Diff
Merge
bk://gkernel.bkbits.net/net-drivers-2.5
into ppc970.osdl.org:/home/torvalds/v2.5/linux
parents
1f0bb3eb
6bac5a4d
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
40 additions
and
24 deletions
+40
-24
drivers/ide/ppc/pmac.c
drivers/ide/ppc/pmac.c
+28
-23
drivers/net/sungem.c
drivers/net/sungem.c
+12
-1
No files found.
drivers/ide/ppc/pmac.c
View file @
0f8b375f
...
...
@@ -55,7 +55,7 @@ extern void ide_do_request(ide_hwgroup_t *hwgroup, int masked_irq);
#define IDE_PMAC_DEBUG
#define DMA_WAIT_TIMEOUT
10
0
#define DMA_WAIT_TIMEOUT
5
0
typedef
struct
pmac_ide_hwif
{
unsigned
long
regbase
;
...
...
@@ -2026,8 +2026,11 @@ pmac_ide_dma_end (ide_drive_t *drive)
dstat
=
readl
(
&
dma
->
status
);
writel
(((
RUN
|
WAKE
|
DEAD
)
<<
16
),
&
dma
->
control
);
pmac_ide_destroy_dmatable
(
drive
);
/* verify good dma status */
return
(
dstat
&
(
RUN
|
DEAD
|
ACTIVE
))
!=
RUN
;
/* verify good dma status. we don't check for ACTIVE beeing 0. We should...
* in theory, but with ATAPI decices doing buffer underruns, that would
* cause us to disable DMA, which isn't what we want
*/
return
(
dstat
&
(
RUN
|
DEAD
))
!=
RUN
;
}
/*
...
...
@@ -2041,7 +2044,7 @@ pmac_ide_dma_test_irq (ide_drive_t *drive)
{
pmac_ide_hwif_t
*
pmif
=
(
pmac_ide_hwif_t
*
)
HWIF
(
drive
)
->
hwif_data
;
volatile
struct
dbdma_regs
*
dma
;
unsigned
long
status
;
unsigned
long
status
,
timeout
;
if
(
pmif
==
NULL
)
return
0
;
...
...
@@ -2057,15 +2060,6 @@ pmac_ide_dma_test_irq (ide_drive_t *drive)
* - The dbdma fifo hasn't yet finished flushing to
* to system memory when the disk interrupt occurs.
*
* The trick here is to increment drive->waiting_for_dma,
* and return as if no interrupt occurred. If the counter
* reach a certain timeout value, we then return 1. If
* we really got the interrupt, it will happen right away
* again.
* Apple's solution here may be more elegant. They issue
* a DMA channel interrupt (a separate irq line) via a DBDMA
* NOP command just before the STOP, and wait for both the
* disk and DBDMA interrupts to have completed.
*/
/* If ACTIVE is cleared, the STOP command have passed and
...
...
@@ -2079,15 +2073,26 @@ pmac_ide_dma_test_irq (ide_drive_t *drive)
called while not waiting
\n
"
,
HWIF
(
drive
)
->
index
);
/* If dbdma didn't execute the STOP command yet, the
* active bit is still set */
drive
->
waiting_for_dma
++
;
if
(
drive
->
waiting_for_dma
>=
DMA_WAIT_TIMEOUT
)
{
printk
(
KERN_WARNING
"ide%d, timeout waiting \
for dbdma command stop
\n
"
,
HWIF
(
drive
)
->
index
);
return
1
;
* active bit is still set. We consider that we aren't
* sharing interrupts (which is hopefully the case with
* those controllers) and so we just try to flush the
* channel for pending data in the fifo
*/
udelay
(
1
);
writel
((
FLUSH
<<
16
)
|
FLUSH
,
&
dma
->
control
);
timeout
=
0
;
for
(;;)
{
udelay
(
1
);
status
=
readl
(
&
dma
->
status
);
if
((
status
&
FLUSH
)
==
0
)
break
;
if
(
++
timeout
>
100
)
{
printk
(
KERN_WARNING
"ide%d, ide_dma_test_irq \
timeout flushing channel
\n
"
,
HWIF
(
drive
)
->
index
);
break
;
}
udelay
(
5
);
return
0
;
}
return
1
;
}
static
int
__pmac
...
...
drivers/net/sungem.c
View file @
0f8b375f
...
...
@@ -654,6 +654,7 @@ static __inline__ void gem_post_rxds(struct gem *gp, int limit)
cluster_start
=
curr
=
(
gp
->
rx_new
&
~
(
4
-
1
));
count
=
0
;
kick
=
-
1
;
wmb
();
while
(
curr
!=
limit
)
{
curr
=
NEXT_RX
(
curr
);
if
(
++
count
==
4
)
{
...
...
@@ -670,8 +671,10 @@ static __inline__ void gem_post_rxds(struct gem *gp, int limit)
count
=
0
;
}
}
if
(
kick
>=
0
)
if
(
kick
>=
0
)
{
mb
();
writel
(
kick
,
gp
->
regs
+
RXDMA_KICK
);
}
}
static
void
gem_rx
(
struct
gem
*
gp
)
...
...
@@ -884,6 +887,7 @@ static int gem_start_xmit(struct sk_buff *skb, struct net_device *dev)
if
(
gem_intme
(
entry
))
ctrl
|=
TXDCTRL_INTME
;
txd
->
buffer
=
cpu_to_le64
(
mapping
);
wmb
();
txd
->
control_word
=
cpu_to_le64
(
ctrl
);
entry
=
NEXT_TX
(
entry
);
}
else
{
...
...
@@ -923,6 +927,7 @@ static int gem_start_xmit(struct sk_buff *skb, struct net_device *dev)
txd
=
&
gp
->
init_block
->
txd
[
entry
];
txd
->
buffer
=
cpu_to_le64
(
mapping
);
wmb
();
txd
->
control_word
=
cpu_to_le64
(
this_ctrl
|
len
);
if
(
gem_intme
(
entry
))
...
...
@@ -932,6 +937,7 @@ static int gem_start_xmit(struct sk_buff *skb, struct net_device *dev)
}
txd
=
&
gp
->
init_block
->
txd
[
first_entry
];
txd
->
buffer
=
cpu_to_le64
(
first_mapping
);
wmb
();
txd
->
control_word
=
cpu_to_le64
(
ctrl
|
TXDCTRL_SOF
|
intme
|
first_len
);
}
...
...
@@ -943,6 +949,7 @@ static int gem_start_xmit(struct sk_buff *skb, struct net_device *dev)
if
(
netif_msg_tx_queued
(
gp
))
printk
(
KERN_DEBUG
"%s: tx queued, slot %d, skblen %d
\n
"
,
dev
->
name
,
entry
,
skb
->
len
);
mb
();
writel
(
gp
->
tx_new
,
gp
->
regs
+
TXDMA_KICK
);
spin_unlock_irq
(
&
gp
->
lock
);
...
...
@@ -1418,6 +1425,7 @@ static void gem_clean_rings(struct gem *gp)
gp
->
rx_skbs
[
i
]
=
NULL
;
}
rxd
->
status_word
=
0
;
wmb
();
rxd
->
buffer
=
0
;
}
...
...
@@ -1478,6 +1486,7 @@ static void gem_init_rings(struct gem *gp)
RX_BUF_ALLOC_SIZE
(
gp
),
PCI_DMA_FROMDEVICE
);
rxd
->
buffer
=
cpu_to_le64
(
dma_addr
);
wmb
();
rxd
->
status_word
=
cpu_to_le64
(
RXDCTRL_FRESH
(
gp
));
skb_reserve
(
skb
,
RX_OFFSET
);
}
...
...
@@ -1486,8 +1495,10 @@ static void gem_init_rings(struct gem *gp)
struct
gem_txd
*
txd
=
&
gb
->
txd
[
i
];
txd
->
control_word
=
0
;
wmb
();
txd
->
buffer
=
0
;
}
wmb
();
}
/* Must be invoked under gp->lock. */
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment