Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
nexedi
linux
Commits
fb8e569c
Commit
fb8e569c
authored
Nov 21, 2007
by
Paul Mundt
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
sh: Fix up user_fpu_struct typo for SH-5.
Signed-off-by:
Paul Mundt
<
lethal@linux-sh.org
>
parent
5db141a9
Changes
4
Hide whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
1 addition
and
220 deletions
+1
-220
arch/sh64/kernel/asm-offsets.c
arch/sh64/kernel/asm-offsets.c
+0
-33
arch/sh64/kernel/init_task.c
arch/sh64/kernel/init_task.c
+0
-46
arch/sh64/kernel/semaphore.c
arch/sh64/kernel/semaphore.c
+0
-140
include/asm-sh/user.h
include/asm-sh/user.h
+1
-1
No files found.
arch/sh64/kernel/asm-offsets.c
deleted
100644 → 0
View file @
5db141a9
/*
* This program is used to generate definitions needed by
* assembly language modules.
*
* We use the technique used in the OSF Mach kernel code:
* generate asm statements containing #defines,
* compile this file to assembler, and then extract the
* #defines from the assembly-language output.
*/
#include <linux/stddef.h>
#include <linux/types.h>
#include <linux/mm.h>
#include <asm/thread_info.h>
#define DEFINE(sym, val) \
asm volatile("\n->" #sym " %0 " #val : : "i" (val))
#define BLANK() asm volatile("\n->" : : )
int
main
(
void
)
{
/* offsets into the thread_info struct */
DEFINE
(
TI_TASK
,
offsetof
(
struct
thread_info
,
task
));
DEFINE
(
TI_EXEC_DOMAIN
,
offsetof
(
struct
thread_info
,
exec_domain
));
DEFINE
(
TI_FLAGS
,
offsetof
(
struct
thread_info
,
flags
));
DEFINE
(
TI_PRE_COUNT
,
offsetof
(
struct
thread_info
,
preempt_count
));
DEFINE
(
TI_CPU
,
offsetof
(
struct
thread_info
,
cpu
));
DEFINE
(
TI_ADDR_LIMIT
,
offsetof
(
struct
thread_info
,
addr_limit
));
DEFINE
(
TI_RESTART_BLOCK
,
offsetof
(
struct
thread_info
,
restart_block
));
return
0
;
}
arch/sh64/kernel/init_task.c
deleted
100644 → 0
View file @
5db141a9
/*
* This file is subject to the terms and conditions of the GNU General Public
* License. See the file "COPYING" in the main directory of this archive
* for more details.
*
* arch/sh64/kernel/init_task.c
*
* Copyright (C) 2000, 2001 Paolo Alberelli
* Copyright (C) 2003 Paul Mundt
*
*/
#include <linux/rwsem.h>
#include <linux/mm.h>
#include <linux/sched.h>
#include <linux/init_task.h>
#include <linux/mqueue.h>
#include <linux/fs.h>
#include <asm/uaccess.h>
#include <asm/pgtable.h>
static
struct
fs_struct
init_fs
=
INIT_FS
;
static
struct
files_struct
init_files
=
INIT_FILES
;
static
struct
signal_struct
init_signals
=
INIT_SIGNALS
(
init_signals
);
static
struct
sighand_struct
init_sighand
=
INIT_SIGHAND
(
init_sighand
);
struct
mm_struct
init_mm
=
INIT_MM
(
init_mm
);
struct
pt_regs
fake_swapper_regs
;
/*
* Initial thread structure.
*
* We need to make sure that this is THREAD_SIZE-byte aligned due
* to the way process stacks are handled. This is done by having a
* special "init_task" linker map entry..
*/
union
thread_union
init_thread_union
__attribute__
((
__section__
(
".data.init_task"
)))
=
{
INIT_THREAD_INFO
(
init_task
)
};
/*
* Initial task structure.
*
* All other task structs will be allocated on slabs in fork.c
*/
struct
task_struct
init_task
=
INIT_TASK
(
init_task
);
arch/sh64/kernel/semaphore.c
deleted
100644 → 0
View file @
5db141a9
/*
* Just taken from alpha implementation.
* This can't work well, perhaps.
*/
/*
* Generic semaphore code. Buyer beware. Do your own
* specific changes in <asm/semaphore-helper.h>
*/
#include <linux/errno.h>
#include <linux/rwsem.h>
#include <linux/sched.h>
#include <linux/wait.h>
#include <linux/init.h>
#include <asm/semaphore.h>
#include <asm/semaphore-helper.h>
spinlock_t
semaphore_wake_lock
;
/*
* Semaphores are implemented using a two-way counter:
* The "count" variable is decremented for each process
* that tries to sleep, while the "waking" variable is
* incremented when the "up()" code goes to wake up waiting
* processes.
*
* Notably, the inline "up()" and "down()" functions can
* efficiently test if they need to do any extra work (up
* needs to do something only if count was negative before
* the increment operation.
*
* waking_non_zero() (from asm/semaphore.h) must execute
* atomically.
*
* When __up() is called, the count was negative before
* incrementing it, and we need to wake up somebody.
*
* This routine adds one to the count of processes that need to
* wake up and exit. ALL waiting processes actually wake up but
* only the one that gets to the "waking" field first will gate
* through and acquire the semaphore. The others will go back
* to sleep.
*
* Note that these functions are only called when there is
* contention on the lock, and as such all this is the
* "non-critical" part of the whole semaphore business. The
* critical part is the inline stuff in <asm/semaphore.h>
* where we want to avoid any extra jumps and calls.
*/
void
__up
(
struct
semaphore
*
sem
)
{
wake_one_more
(
sem
);
wake_up
(
&
sem
->
wait
);
}
/*
* Perform the "down" function. Return zero for semaphore acquired,
* return negative for signalled out of the function.
*
* If called from __down, the return is ignored and the wait loop is
* not interruptible. This means that a task waiting on a semaphore
* using "down()" cannot be killed until someone does an "up()" on
* the semaphore.
*
* If called from __down_interruptible, the return value gets checked
* upon return. If the return value is negative then the task continues
* with the negative value in the return register (it can be tested by
* the caller).
*
* Either form may be used in conjunction with "up()".
*
*/
#define DOWN_VAR \
struct task_struct *tsk = current; \
wait_queue_t wait; \
init_waitqueue_entry(&wait, tsk);
#define DOWN_HEAD(task_state) \
\
\
tsk->state = (task_state); \
add_wait_queue(&sem->wait, &wait); \
\
/* \
* Ok, we're set up. sem->count is known to be less than zero \
* so we must wait. \
* \
* We can let go the lock for purposes of waiting. \
* We re-acquire it after awaking so as to protect \
* all semaphore operations. \
* \
* If "up()" is called before we call waking_non_zero() then \
* we will catch it right away. If it is called later then \
* we will have to go through a wakeup cycle to catch it. \
* \
* Multiple waiters contend for the semaphore lock to see \
* who gets to gate through and who has to wait some more. \
*/
\
for (;;) {
#define DOWN_TAIL(task_state) \
tsk->state = (task_state); \
} \
tsk->state = TASK_RUNNING; \
remove_wait_queue(&sem->wait, &wait);
void
__sched
__down
(
struct
semaphore
*
sem
)
{
DOWN_VAR
DOWN_HEAD
(
TASK_UNINTERRUPTIBLE
)
if
(
waking_non_zero
(
sem
))
break
;
schedule
();
DOWN_TAIL
(
TASK_UNINTERRUPTIBLE
)
}
int
__sched
__down_interruptible
(
struct
semaphore
*
sem
)
{
int
ret
=
0
;
DOWN_VAR
DOWN_HEAD
(
TASK_INTERRUPTIBLE
)
ret
=
waking_non_zero_interruptible
(
sem
,
tsk
);
if
(
ret
)
{
if
(
ret
==
1
)
/* ret != 0 only if we get interrupted -arca */
ret
=
0
;
break
;
}
schedule
();
DOWN_TAIL
(
TASK_INTERRUPTIBLE
)
return
ret
;
}
int
__down_trylock
(
struct
semaphore
*
sem
)
{
return
waking_non_zero_trylock
(
sem
);
}
include/asm-sh/user.h
View file @
fb8e569c
...
...
@@ -28,7 +28,7 @@
*/
#if defined(__SH5__) || defined(CONFIG_CPU_SH5)
struct
user
fpu_struct
{
struct
user
_
fpu_struct
{
unsigned
long
fp_regs
[
32
];
unsigned
int
fpscr
;
};
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment