Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
5ed26a4f
Commit
5ed26a4f
authored
May 05, 2003
by
Steven Cole
Committed by
David Mosberger
May 05, 2003
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[PATCH] ia64: spelling fixes
parent
39d4a447
Changes
13
Hide whitespace changes
Inline
Side-by-side
Showing
13 changed files
with
15 additions
and
15 deletions
+15
-15
arch/ia64/hp/common/sba_iommu.c
arch/ia64/hp/common/sba_iommu.c
+1
-1
arch/ia64/hp/sim/simeth.c
arch/ia64/hp/sim/simeth.c
+1
-1
arch/ia64/ia32/ia32_traps.c
arch/ia64/ia32/ia32_traps.c
+1
-1
arch/ia64/kernel/irq.c
arch/ia64/kernel/irq.c
+2
-2
arch/ia64/kernel/mca.c
arch/ia64/kernel/mca.c
+1
-1
arch/ia64/kernel/sal.c
arch/ia64/kernel/sal.c
+1
-1
arch/ia64/kernel/time.c
arch/ia64/kernel/time.c
+1
-1
arch/ia64/kernel/unaligned.c
arch/ia64/kernel/unaligned.c
+1
-1
arch/ia64/lib/copy_user.S
arch/ia64/lib/copy_user.S
+1
-1
arch/ia64/lib/do_csum.S
arch/ia64/lib/do_csum.S
+1
-1
arch/ia64/lib/swiotlb.c
arch/ia64/lib/swiotlb.c
+2
-2
include/asm-ia64/sal.h
include/asm-ia64/sal.h
+1
-1
include/asm-ia64/uaccess.h
include/asm-ia64/uaccess.h
+1
-1
No files found.
arch/ia64/hp/common/sba_iommu.c
View file @
5ed26a4f
...
...
@@ -122,7 +122,7 @@
#endif
/*
** The number of pdir entries to "free" before issu
e
ing
** The number of pdir entries to "free" before issuing
** a read to PCOM register to flush out PCOM writes.
** Interacts with allocation granularity (ie 4 or 8 entries
** allocated and free'd/purged at a time might make this
...
...
arch/ia64/hp/sim/simeth.c
View file @
5ed26a4f
...
...
@@ -505,7 +505,7 @@ simeth_interrupt(int irq, void *dev_id, struct pt_regs * regs)
}
/*
* very simple loop because we get interrupts only when receving
* very simple loop because we get interrupts only when rece
i
ving
*/
while
(
simeth_rx
(
dev
));
}
...
...
arch/ia64/ia32/ia32_traps.c
View file @
5ed26a4f
...
...
@@ -103,7 +103,7 @@ ia32_exception (struct pt_regs *regs, unsigned long isr)
* C1 reg you need in case of a stack fault, 0x040 is the stack
* fault bit. We should only be taking one exception at a time,
* so if this combination doesn't produce any single exception,
* then we have a bad program that isn't syncronizing its FPU usage
* then we have a bad program that isn't sync
h
ronizing its FPU usage
* and it will suffer the consequences since we won't be able to
* fully reproduce the context of the exception
*/
...
...
arch/ia64/kernel/irq.c
View file @
5ed26a4f
...
...
@@ -50,7 +50,7 @@
* Linux has a controller-independent x86 interrupt architecture.
* every controller has a 'controller-template', that is used
* by the main code to do the right thing. Each driver-visible
* interrupt source is transparently wired to the apropriate
* interrupt source is transparently wired to the ap
p
ropriate
* controller. Thus drivers need not be aware of the
* interrupt-controller.
*
...
...
@@ -705,7 +705,7 @@ unsigned int probe_irq_mask(unsigned long val)
* The interrupt probe logic state is returned to its previous
* value.
*
* BUGS: When used in a module (which arguably shouldnt happen)
* BUGS: When used in a module (which arguably shouldn
'
t happen)
* nothing prevents two IRQ probe callers from overlapping. The
* results of this are non-optimal.
*/
...
...
arch/ia64/kernel/mca.c
View file @
5ed26a4f
...
...
@@ -1138,7 +1138,7 @@ ia64_mca_cpe_int_caller(void *dummy)
* ia64_mca_cpe_poll
*
* Poll for Corrected Platform Errors (CPEs), dynamically adjust
* polling interval based on occur
a
nce of an event.
* polling interval based on occur
re
nce of an event.
*
* Inputs : dummy(unused)
* Outputs : None
...
...
arch/ia64/kernel/sal.c
View file @
5ed26a4f
...
...
@@ -116,7 +116,7 @@ ia64_sal_init (struct ia64_sal_systab *systab)
p
=
(
char
*
)
(
systab
+
1
);
for
(
i
=
0
;
i
<
systab
->
entry_count
;
i
++
)
{
/*
* The first byte of each entry type contains the type desciptor.
* The first byte of each entry type contains the type desc
r
iptor.
*/
switch
(
*
p
)
{
case
SAL_DESC_ENTRY_POINT
:
...
...
arch/ia64/kernel/time.c
View file @
5ed26a4f
...
...
@@ -221,7 +221,7 @@ timer_interrupt(int irq, void *dev_id, struct pt_regs *regs)
do
{
/*
* If we're too close to the next clock tick for comfort, we increase the
* saf
te
y margin by intentionally dropping the next tick(s). We do NOT update
* saf
et
y margin by intentionally dropping the next tick(s). We do NOT update
* itm.next because that would force us to call do_timer() which in turn would
* let our clock run too fast (with the potentially devastating effect of
* losing monotony of time).
...
...
arch/ia64/kernel/unaligned.c
View file @
5ed26a4f
...
...
@@ -789,7 +789,7 @@ emulate_load_int (unsigned long ifa, load_store_t ld, struct pt_regs *regs)
*
* ldX.a (advanced load):
* - suppose ldX.a r1=[r3]. If we get to the unaligned trap it's because the
* address doesn't match requested size align
e
ment. This means that we would
* address doesn't match requested size alignment. This means that we would
* possibly need more than one load to get the result.
*
* The load part can be handled just like a normal load, however the difficult
...
...
arch/ia64/lib/copy_user.S
View file @
5ed26a4f
...
...
@@ -316,7 +316,7 @@ GLOBAL_ENTRY(__copy_user)
//
Beginning
of
long
mempcy
(
i
.
e
.
>
16
bytes
)
//
.
long_copy_user
:
tbit.nz
p6
,
p7
=
src1
,
0
//
odd
align
e
ment
tbit.nz
p6
,
p7
=
src1
,
0
//
odd
alignment
and
tmp
=
7
,
tmp
;;
cmp.eq
p10
,
p8
=
r0
,
tmp
...
...
arch/ia64/lib/do_csum.S
View file @
5ed26a4f
...
...
@@ -137,7 +137,7 @@ GLOBAL_ENTRY(do_csum)
mov
saved_pr
=
pr
//
preserve
predicates
(
rotation
)
(
p6
)
br.ret.spnt.many
rp
//
return
if
zero
or
negative
length
mov
hmask
=-
1
//
intialize
head
mask
mov
hmask
=-
1
//
in
i
tialize
head
mask
tbit.nz
p15
,
p0
=
buf
,
0
//
is
buf
an
odd
address
?
and
first1
=-
8
,
buf
//
8
-
byte
align
down
address
of
first1
element
...
...
arch/ia64/lib/swiotlb.c
View file @
5ed26a4f
...
...
@@ -239,7 +239,7 @@ unmap_single (struct pci_dev *hwdev, char *dma_addr, size_t size, int direction)
for
(
i
=
index
+
nslots
-
1
;
i
>=
index
;
i
--
)
io_tlb_list
[
i
]
=
++
count
;
/*
* Step 2: merge the returned slots with the prece
e
ding slots, if
* Step 2: merge the returned slots with the preceding slots, if
* available (non zero)
*/
for
(
i
=
index
-
1
;
(
OFFSET
(
i
,
IO_TLB_SEGSIZE
)
!=
IO_TLB_SEGSIZE
-
1
)
&&
...
...
@@ -399,7 +399,7 @@ swiotlb_sync_single (struct pci_dev *hwdev, dma_addr_t pci_addr, size_t size, in
/*
* Map a set of buffers described by scatterlist in streaming mode for DMA. This is the
* scat
h
er-gather version of the above swiotlb_map_single interface. Here the scatter
* scat
t
er-gather version of the above swiotlb_map_single interface. Here the scatter
* gather list elements are each tagged with the appropriate dma address and length. They
* are obtained via sg_dma_{address,length}(SG).
*
...
...
include/asm-ia64/sal.h
View file @
5ed26a4f
...
...
@@ -226,7 +226,7 @@ enum {
/* Encodings for machine check parameter types */
enum
{
SAL_MC_PARAM_RENDEZ_INT
=
1
,
/* Rendez
e
vous interrupt */
SAL_MC_PARAM_RENDEZ_INT
=
1
,
/* Rendezvous interrupt */
SAL_MC_PARAM_RENDEZ_WAKEUP
=
2
,
/* Wakeup */
SAL_MC_PARAM_CPE_INT
=
3
/* Corrected Platform Error Int */
};
...
...
include/asm-ia64/uaccess.h
View file @
5ed26a4f
...
...
@@ -8,7 +8,7 @@
* addresses. Thus, we need to be careful not to let the user to
* trick us into accessing kernel memory that would normally be
* inaccessible. This code is also fairly performance sensitive,
* so we want to spend as little time doing saf
te
y checks as
* so we want to spend as little time doing saf
et
y checks as
* possible.
*
* To make matters a bit more interesting, these macros sometimes also
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment