Merge branch 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6
* 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6: [IA64] make pfm_get_task work with virtual pids [IA64] honor notify_die() returning NOTIFY_STOP [IA64] remove dead code: __cpu_{down,die} from !HOTPLUG_CPU [IA64] Appoint kvm/ia64 Maintainers [IA64] ia64_set_psr should use srlz.i [IA64] Export three symbols for module use [IA64] mca style cleanup [IA64] sn_hwperf semaphore to mutex [IA64] generalize attribute of fsyscall_gtod_data [IA64] efi.c Add /* never reached */ annotation [IA64] efi.c Spelling/punctuation fixes [IA64] Make efi.c mostly fit in 80 columns [IA64] aliasing-test: fix gcc warnings on non-ia64 [IA64] Slim-down __clear_bit_unlock [IA64] Fix the order of atomic operations in restore_previous_kprobes on ia64 [IA64] constify function pointer tables [IA64] fix userspace compile error in gcc_intrin.h
This commit is contained in:
commit
21511abd0a
@ -16,6 +16,7 @@
|
|||||||
#include <fcntl.h>
|
#include <fcntl.h>
|
||||||
#include <fnmatch.h>
|
#include <fnmatch.h>
|
||||||
#include <string.h>
|
#include <string.h>
|
||||||
|
#include <sys/ioctl.h>
|
||||||
#include <sys/mman.h>
|
#include <sys/mman.h>
|
||||||
#include <sys/stat.h>
|
#include <sys/stat.h>
|
||||||
#include <unistd.h>
|
#include <unistd.h>
|
||||||
@ -65,7 +66,7 @@ int scan_tree(char *path, char *file, off_t offset, size_t length, int touch)
|
|||||||
{
|
{
|
||||||
struct dirent **namelist;
|
struct dirent **namelist;
|
||||||
char *name, *path2;
|
char *name, *path2;
|
||||||
int i, n, r, rc, result = 0;
|
int i, n, r, rc = 0, result = 0;
|
||||||
struct stat buf;
|
struct stat buf;
|
||||||
|
|
||||||
n = scandir(path, &namelist, 0, alphasort);
|
n = scandir(path, &namelist, 0, alphasort);
|
||||||
@ -113,7 +114,7 @@ skip:
|
|||||||
free(namelist[i]);
|
free(namelist[i]);
|
||||||
}
|
}
|
||||||
free(namelist);
|
free(namelist);
|
||||||
return rc;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
char buf[1024];
|
char buf[1024];
|
||||||
@ -149,7 +150,7 @@ int scan_rom(char *path, char *file)
|
|||||||
{
|
{
|
||||||
struct dirent **namelist;
|
struct dirent **namelist;
|
||||||
char *name, *path2;
|
char *name, *path2;
|
||||||
int i, n, r, rc, result = 0;
|
int i, n, r, rc = 0, result = 0;
|
||||||
struct stat buf;
|
struct stat buf;
|
||||||
|
|
||||||
n = scandir(path, &namelist, 0, alphasort);
|
n = scandir(path, &namelist, 0, alphasort);
|
||||||
@ -180,7 +181,7 @@ int scan_rom(char *path, char *file)
|
|||||||
* important thing is that no MCA happened.
|
* important thing is that no MCA happened.
|
||||||
*/
|
*/
|
||||||
if (rc > 0)
|
if (rc > 0)
|
||||||
fprintf(stderr, "PASS: %s read %ld bytes\n", path2, rc);
|
fprintf(stderr, "PASS: %s read %d bytes\n", path2, rc);
|
||||||
else {
|
else {
|
||||||
fprintf(stderr, "PASS: %s not readable\n", path2);
|
fprintf(stderr, "PASS: %s not readable\n", path2);
|
||||||
return rc;
|
return rc;
|
||||||
@ -201,10 +202,10 @@ skip:
|
|||||||
free(namelist[i]);
|
free(namelist[i]);
|
||||||
}
|
}
|
||||||
free(namelist);
|
free(namelist);
|
||||||
return rc;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
int main()
|
int main(void)
|
||||||
{
|
{
|
||||||
int rc;
|
int rc;
|
||||||
|
|
||||||
@ -256,4 +257,6 @@ int main()
|
|||||||
scan_tree("/proc/bus/pci", "??.?", 0xA0000, 0x20000, 0);
|
scan_tree("/proc/bus/pci", "??.?", 0xA0000, 0x20000, 0);
|
||||||
scan_tree("/proc/bus/pci", "??.?", 0xC0000, 0x40000, 1);
|
scan_tree("/proc/bus/pci", "??.?", 0xC0000, 0x40000, 1);
|
||||||
scan_tree("/proc/bus/pci", "??.?", 0, 1024*1024, 0);
|
scan_tree("/proc/bus/pci", "??.?", 0, 1024*1024, 0);
|
||||||
|
|
||||||
|
return rc;
|
||||||
}
|
}
|
||||||
|
@ -2249,6 +2249,15 @@ L: kvm-devel@lists.sourceforge.net
|
|||||||
W: kvm.sourceforge.net
|
W: kvm.sourceforge.net
|
||||||
S: Supported
|
S: Supported
|
||||||
|
|
||||||
|
KERNEL VIRTUAL MACHINE For Itanium(KVM/IA64)
|
||||||
|
P: Anthony Xu
|
||||||
|
M: anthony.xu@intel.com
|
||||||
|
P: Xiantao Zhang
|
||||||
|
M: xiantao.zhang@intel.com
|
||||||
|
L: kvm-ia64-devel@lists.sourceforge.net
|
||||||
|
W: kvm.sourceforge.net
|
||||||
|
S: Supported
|
||||||
|
|
||||||
KEXEC
|
KEXEC
|
||||||
P: Eric Biederman
|
P: Eric Biederman
|
||||||
M: ebiederm@xmission.com
|
M: ebiederm@xmission.com
|
||||||
|
@ -1875,7 +1875,7 @@ ioc_show(struct seq_file *s, void *v)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct seq_operations ioc_seq_ops = {
|
static const struct seq_operations ioc_seq_ops = {
|
||||||
.start = ioc_start,
|
.start = ioc_start,
|
||||||
.next = ioc_next,
|
.next = ioc_next,
|
||||||
.stop = ioc_stop,
|
.stop = ioc_stop,
|
||||||
|
@ -27,7 +27,7 @@
|
|||||||
|
|
||||||
#include "ia32priv.h"
|
#include "ia32priv.h"
|
||||||
|
|
||||||
extern void die_if_kernel (char *str, struct pt_regs *regs, long err);
|
extern int die_if_kernel (char *str, struct pt_regs *regs, long err);
|
||||||
|
|
||||||
struct exec_domain ia32_exec_domain;
|
struct exec_domain ia32_exec_domain;
|
||||||
struct page *ia32_shared_page[NR_CPUS];
|
struct page *ia32_shared_page[NR_CPUS];
|
||||||
@ -217,7 +217,8 @@ ia32_bad_interrupt (unsigned long int_num, struct pt_regs *regs)
|
|||||||
{
|
{
|
||||||
siginfo_t siginfo;
|
siginfo_t siginfo;
|
||||||
|
|
||||||
die_if_kernel("Bad IA-32 interrupt", regs, int_num);
|
if (die_if_kernel("Bad IA-32 interrupt", regs, int_num))
|
||||||
|
return;
|
||||||
|
|
||||||
siginfo.si_signo = SIGTRAP;
|
siginfo.si_signo = SIGTRAP;
|
||||||
siginfo.si_errno = int_num; /* XXX is it OK to abuse si_errno like this? */
|
siginfo.si_errno = int_num; /* XXX is it OK to abuse si_errno like this? */
|
||||||
|
@ -1,7 +1,8 @@
|
|||||||
/*
|
/*
|
||||||
* Extensible Firmware Interface
|
* Extensible Firmware Interface
|
||||||
*
|
*
|
||||||
* Based on Extensible Firmware Interface Specification version 0.9 April 30, 1999
|
* Based on Extensible Firmware Interface Specification version 0.9
|
||||||
|
* April 30, 1999
|
||||||
*
|
*
|
||||||
* Copyright (C) 1999 VA Linux Systems
|
* Copyright (C) 1999 VA Linux Systems
|
||||||
* Copyright (C) 1999 Walt Drummond <drummond@valinux.com>
|
* Copyright (C) 1999 Walt Drummond <drummond@valinux.com>
|
||||||
@ -48,145 +49,157 @@ static unsigned long mem_limit = ~0UL, max_addr = ~0UL, min_addr = 0UL;
|
|||||||
|
|
||||||
#define efi_call_virt(f, args...) (*(f))(args)
|
#define efi_call_virt(f, args...) (*(f))(args)
|
||||||
|
|
||||||
#define STUB_GET_TIME(prefix, adjust_arg) \
|
#define STUB_GET_TIME(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_get_time (efi_time_t *tm, efi_time_cap_t *tc) \
|
prefix##_get_time (efi_time_t *tm, efi_time_cap_t *tc) \
|
||||||
{ \
|
{ \
|
||||||
struct ia64_fpreg fr[6]; \
|
struct ia64_fpreg fr[6]; \
|
||||||
efi_time_cap_t *atc = NULL; \
|
efi_time_cap_t *atc = NULL; \
|
||||||
efi_status_t ret; \
|
efi_status_t ret; \
|
||||||
\
|
\
|
||||||
if (tc) \
|
if (tc) \
|
||||||
atc = adjust_arg(tc); \
|
atc = adjust_arg(tc); \
|
||||||
ia64_save_scratch_fpregs(fr); \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
ret = efi_call_##prefix((efi_get_time_t *) __va(runtime->get_time), adjust_arg(tm), atc); \
|
ret = efi_call_##prefix((efi_get_time_t *) __va(runtime->get_time), \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
adjust_arg(tm), atc); \
|
||||||
return ret; \
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_SET_TIME(prefix, adjust_arg) \
|
#define STUB_SET_TIME(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_set_time (efi_time_t *tm) \
|
prefix##_set_time (efi_time_t *tm) \
|
||||||
{ \
|
{ \
|
||||||
struct ia64_fpreg fr[6]; \
|
struct ia64_fpreg fr[6]; \
|
||||||
efi_status_t ret; \
|
efi_status_t ret; \
|
||||||
\
|
\
|
||||||
ia64_save_scratch_fpregs(fr); \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
ret = efi_call_##prefix((efi_set_time_t *) __va(runtime->set_time), adjust_arg(tm)); \
|
ret = efi_call_##prefix((efi_set_time_t *) __va(runtime->set_time), \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
adjust_arg(tm)); \
|
||||||
return ret; \
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_GET_WAKEUP_TIME(prefix, adjust_arg) \
|
#define STUB_GET_WAKEUP_TIME(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_get_wakeup_time (efi_bool_t *enabled, efi_bool_t *pending, efi_time_t *tm) \
|
prefix##_get_wakeup_time (efi_bool_t *enabled, efi_bool_t *pending, \
|
||||||
{ \
|
efi_time_t *tm) \
|
||||||
struct ia64_fpreg fr[6]; \
|
{ \
|
||||||
efi_status_t ret; \
|
struct ia64_fpreg fr[6]; \
|
||||||
\
|
efi_status_t ret; \
|
||||||
ia64_save_scratch_fpregs(fr); \
|
\
|
||||||
ret = efi_call_##prefix((efi_get_wakeup_time_t *) __va(runtime->get_wakeup_time), \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
adjust_arg(enabled), adjust_arg(pending), adjust_arg(tm)); \
|
ret = efi_call_##prefix( \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
(efi_get_wakeup_time_t *) __va(runtime->get_wakeup_time), \
|
||||||
return ret; \
|
adjust_arg(enabled), adjust_arg(pending), adjust_arg(tm)); \
|
||||||
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_SET_WAKEUP_TIME(prefix, adjust_arg) \
|
#define STUB_SET_WAKEUP_TIME(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_set_wakeup_time (efi_bool_t enabled, efi_time_t *tm) \
|
prefix##_set_wakeup_time (efi_bool_t enabled, efi_time_t *tm) \
|
||||||
{ \
|
{ \
|
||||||
struct ia64_fpreg fr[6]; \
|
struct ia64_fpreg fr[6]; \
|
||||||
efi_time_t *atm = NULL; \
|
efi_time_t *atm = NULL; \
|
||||||
efi_status_t ret; \
|
efi_status_t ret; \
|
||||||
\
|
\
|
||||||
if (tm) \
|
if (tm) \
|
||||||
atm = adjust_arg(tm); \
|
atm = adjust_arg(tm); \
|
||||||
ia64_save_scratch_fpregs(fr); \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
ret = efi_call_##prefix((efi_set_wakeup_time_t *) __va(runtime->set_wakeup_time), \
|
ret = efi_call_##prefix( \
|
||||||
enabled, atm); \
|
(efi_set_wakeup_time_t *) __va(runtime->set_wakeup_time), \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
enabled, atm); \
|
||||||
return ret; \
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_GET_VARIABLE(prefix, adjust_arg) \
|
#define STUB_GET_VARIABLE(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_get_variable (efi_char16_t *name, efi_guid_t *vendor, u32 *attr, \
|
prefix##_get_variable (efi_char16_t *name, efi_guid_t *vendor, u32 *attr, \
|
||||||
unsigned long *data_size, void *data) \
|
unsigned long *data_size, void *data) \
|
||||||
{ \
|
{ \
|
||||||
struct ia64_fpreg fr[6]; \
|
struct ia64_fpreg fr[6]; \
|
||||||
u32 *aattr = NULL; \
|
u32 *aattr = NULL; \
|
||||||
efi_status_t ret; \
|
efi_status_t ret; \
|
||||||
\
|
\
|
||||||
if (attr) \
|
if (attr) \
|
||||||
aattr = adjust_arg(attr); \
|
aattr = adjust_arg(attr); \
|
||||||
ia64_save_scratch_fpregs(fr); \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
ret = efi_call_##prefix((efi_get_variable_t *) __va(runtime->get_variable), \
|
ret = efi_call_##prefix( \
|
||||||
adjust_arg(name), adjust_arg(vendor), aattr, \
|
(efi_get_variable_t *) __va(runtime->get_variable), \
|
||||||
adjust_arg(data_size), adjust_arg(data)); \
|
adjust_arg(name), adjust_arg(vendor), aattr, \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
adjust_arg(data_size), adjust_arg(data)); \
|
||||||
return ret; \
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_GET_NEXT_VARIABLE(prefix, adjust_arg) \
|
#define STUB_GET_NEXT_VARIABLE(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_get_next_variable (unsigned long *name_size, efi_char16_t *name, efi_guid_t *vendor) \
|
prefix##_get_next_variable (unsigned long *name_size, efi_char16_t *name, \
|
||||||
{ \
|
efi_guid_t *vendor) \
|
||||||
struct ia64_fpreg fr[6]; \
|
{ \
|
||||||
efi_status_t ret; \
|
struct ia64_fpreg fr[6]; \
|
||||||
\
|
efi_status_t ret; \
|
||||||
ia64_save_scratch_fpregs(fr); \
|
\
|
||||||
ret = efi_call_##prefix((efi_get_next_variable_t *) __va(runtime->get_next_variable), \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
adjust_arg(name_size), adjust_arg(name), adjust_arg(vendor)); \
|
ret = efi_call_##prefix( \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
(efi_get_next_variable_t *) __va(runtime->get_next_variable), \
|
||||||
return ret; \
|
adjust_arg(name_size), adjust_arg(name), adjust_arg(vendor)); \
|
||||||
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_SET_VARIABLE(prefix, adjust_arg) \
|
#define STUB_SET_VARIABLE(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_set_variable (efi_char16_t *name, efi_guid_t *vendor, unsigned long attr, \
|
prefix##_set_variable (efi_char16_t *name, efi_guid_t *vendor, \
|
||||||
unsigned long data_size, void *data) \
|
unsigned long attr, unsigned long data_size, \
|
||||||
{ \
|
void *data) \
|
||||||
struct ia64_fpreg fr[6]; \
|
{ \
|
||||||
efi_status_t ret; \
|
struct ia64_fpreg fr[6]; \
|
||||||
\
|
efi_status_t ret; \
|
||||||
ia64_save_scratch_fpregs(fr); \
|
\
|
||||||
ret = efi_call_##prefix((efi_set_variable_t *) __va(runtime->set_variable), \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
adjust_arg(name), adjust_arg(vendor), attr, data_size, \
|
ret = efi_call_##prefix( \
|
||||||
adjust_arg(data)); \
|
(efi_set_variable_t *) __va(runtime->set_variable), \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
adjust_arg(name), adjust_arg(vendor), attr, data_size, \
|
||||||
return ret; \
|
adjust_arg(data)); \
|
||||||
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_GET_NEXT_HIGH_MONO_COUNT(prefix, adjust_arg) \
|
#define STUB_GET_NEXT_HIGH_MONO_COUNT(prefix, adjust_arg) \
|
||||||
static efi_status_t \
|
static efi_status_t \
|
||||||
prefix##_get_next_high_mono_count (u32 *count) \
|
prefix##_get_next_high_mono_count (u32 *count) \
|
||||||
{ \
|
{ \
|
||||||
struct ia64_fpreg fr[6]; \
|
struct ia64_fpreg fr[6]; \
|
||||||
efi_status_t ret; \
|
efi_status_t ret; \
|
||||||
\
|
\
|
||||||
ia64_save_scratch_fpregs(fr); \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
ret = efi_call_##prefix((efi_get_next_high_mono_count_t *) \
|
ret = efi_call_##prefix((efi_get_next_high_mono_count_t *) \
|
||||||
__va(runtime->get_next_high_mono_count), adjust_arg(count)); \
|
__va(runtime->get_next_high_mono_count), \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
adjust_arg(count)); \
|
||||||
return ret; \
|
ia64_load_scratch_fpregs(fr); \
|
||||||
|
return ret; \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define STUB_RESET_SYSTEM(prefix, adjust_arg) \
|
#define STUB_RESET_SYSTEM(prefix, adjust_arg) \
|
||||||
static void \
|
static void \
|
||||||
prefix##_reset_system (int reset_type, efi_status_t status, \
|
prefix##_reset_system (int reset_type, efi_status_t status, \
|
||||||
unsigned long data_size, efi_char16_t *data) \
|
unsigned long data_size, efi_char16_t *data) \
|
||||||
{ \
|
{ \
|
||||||
struct ia64_fpreg fr[6]; \
|
struct ia64_fpreg fr[6]; \
|
||||||
efi_char16_t *adata = NULL; \
|
efi_char16_t *adata = NULL; \
|
||||||
\
|
\
|
||||||
if (data) \
|
if (data) \
|
||||||
adata = adjust_arg(data); \
|
adata = adjust_arg(data); \
|
||||||
\
|
\
|
||||||
ia64_save_scratch_fpregs(fr); \
|
ia64_save_scratch_fpregs(fr); \
|
||||||
efi_call_##prefix((efi_reset_system_t *) __va(runtime->reset_system), \
|
efi_call_##prefix( \
|
||||||
reset_type, status, data_size, adata); \
|
(efi_reset_system_t *) __va(runtime->reset_system), \
|
||||||
/* should not return, but just in case... */ \
|
reset_type, status, data_size, adata); \
|
||||||
ia64_load_scratch_fpregs(fr); \
|
/* should not return, but just in case... */ \
|
||||||
|
ia64_load_scratch_fpregs(fr); \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define phys_ptr(arg) ((__typeof__(arg)) ia64_tpa(arg))
|
#define phys_ptr(arg) ((__typeof__(arg)) ia64_tpa(arg))
|
||||||
@ -223,7 +236,8 @@ efi_gettimeofday (struct timespec *ts)
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
ts->tv_sec = mktime(tm.year, tm.month, tm.day, tm.hour, tm.minute, tm.second);
|
ts->tv_sec = mktime(tm.year, tm.month, tm.day,
|
||||||
|
tm.hour, tm.minute, tm.second);
|
||||||
ts->tv_nsec = tm.nanosecond;
|
ts->tv_nsec = tm.nanosecond;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -297,8 +311,8 @@ walk (efi_freemem_callback_t callback, void *arg, u64 attr)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Walks the EFI memory map and calls CALLBACK once for each EFI memory descriptor that
|
* Walk the EFI memory map and call CALLBACK once for each EFI memory
|
||||||
* has memory that is available for OS use.
|
* descriptor that has memory that is available for OS use.
|
||||||
*/
|
*/
|
||||||
void
|
void
|
||||||
efi_memmap_walk (efi_freemem_callback_t callback, void *arg)
|
efi_memmap_walk (efi_freemem_callback_t callback, void *arg)
|
||||||
@ -307,8 +321,8 @@ efi_memmap_walk (efi_freemem_callback_t callback, void *arg)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Walks the EFI memory map and calls CALLBACK once for each EFI memory descriptor that
|
* Walk the EFI memory map and call CALLBACK once for each EFI memory
|
||||||
* has memory that is available for uncached allocator.
|
* descriptor that has memory that is available for uncached allocator.
|
||||||
*/
|
*/
|
||||||
void
|
void
|
||||||
efi_memmap_walk_uc (efi_freemem_callback_t callback, void *arg)
|
efi_memmap_walk_uc (efi_freemem_callback_t callback, void *arg)
|
||||||
@ -317,11 +331,10 @@ efi_memmap_walk_uc (efi_freemem_callback_t callback, void *arg)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Look for the PAL_CODE region reported by EFI and maps it using an
|
* Look for the PAL_CODE region reported by EFI and map it using an
|
||||||
* ITR to enable safe PAL calls in virtual mode. See IA-64 Processor
|
* ITR to enable safe PAL calls in virtual mode. See IA-64 Processor
|
||||||
* Abstraction Layer chapter 11 in ADAG
|
* Abstraction Layer chapter 11 in ADAG
|
||||||
*/
|
*/
|
||||||
|
|
||||||
void *
|
void *
|
||||||
efi_get_pal_addr (void)
|
efi_get_pal_addr (void)
|
||||||
{
|
{
|
||||||
@ -341,45 +354,47 @@ efi_get_pal_addr (void)
|
|||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (++pal_code_count > 1) {
|
if (++pal_code_count > 1) {
|
||||||
printk(KERN_ERR "Too many EFI Pal Code memory ranges, dropped @ %lx\n",
|
printk(KERN_ERR "Too many EFI Pal Code memory ranges, "
|
||||||
md->phys_addr);
|
"dropped @ %lx\n", md->phys_addr);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
/*
|
/*
|
||||||
* The only ITLB entry in region 7 that is used is the one installed by
|
* The only ITLB entry in region 7 that is used is the one
|
||||||
* __start(). That entry covers a 64MB range.
|
* installed by __start(). That entry covers a 64MB range.
|
||||||
*/
|
*/
|
||||||
mask = ~((1 << KERNEL_TR_PAGE_SHIFT) - 1);
|
mask = ~((1 << KERNEL_TR_PAGE_SHIFT) - 1);
|
||||||
vaddr = PAGE_OFFSET + md->phys_addr;
|
vaddr = PAGE_OFFSET + md->phys_addr;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We must check that the PAL mapping won't overlap with the kernel
|
* We must check that the PAL mapping won't overlap with the
|
||||||
* mapping.
|
* kernel mapping.
|
||||||
*
|
*
|
||||||
* PAL code is guaranteed to be aligned on a power of 2 between 4k and
|
* PAL code is guaranteed to be aligned on a power of 2 between
|
||||||
* 256KB and that only one ITR is needed to map it. This implies that the
|
* 4k and 256KB and that only one ITR is needed to map it. This
|
||||||
* PAL code is always aligned on its size, i.e., the closest matching page
|
* implies that the PAL code is always aligned on its size,
|
||||||
* size supported by the TLB. Therefore PAL code is guaranteed never to
|
* i.e., the closest matching page size supported by the TLB.
|
||||||
* cross a 64MB unless it is bigger than 64MB (very unlikely!). So for
|
* Therefore PAL code is guaranteed never to cross a 64MB unless
|
||||||
* now the following test is enough to determine whether or not we need a
|
* it is bigger than 64MB (very unlikely!). So for now the
|
||||||
* dedicated ITR for the PAL code.
|
* following test is enough to determine whether or not we need
|
||||||
|
* a dedicated ITR for the PAL code.
|
||||||
*/
|
*/
|
||||||
if ((vaddr & mask) == (KERNEL_START & mask)) {
|
if ((vaddr & mask) == (KERNEL_START & mask)) {
|
||||||
printk(KERN_INFO "%s: no need to install ITR for PAL code\n",
|
printk(KERN_INFO "%s: no need to install ITR for "
|
||||||
__FUNCTION__);
|
"PAL code\n", __FUNCTION__);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (efi_md_size(md) > IA64_GRANULE_SIZE)
|
if (efi_md_size(md) > IA64_GRANULE_SIZE)
|
||||||
panic("Woah! PAL code size bigger than a granule!");
|
panic("Whoa! PAL code size bigger than a granule!");
|
||||||
|
|
||||||
#if EFI_DEBUG
|
#if EFI_DEBUG
|
||||||
mask = ~((1 << IA64_GRANULE_SHIFT) - 1);
|
mask = ~((1 << IA64_GRANULE_SHIFT) - 1);
|
||||||
|
|
||||||
printk(KERN_INFO "CPU %d: mapping PAL code [0x%lx-0x%lx) into [0x%lx-0x%lx)\n",
|
printk(KERN_INFO "CPU %d: mapping PAL code "
|
||||||
smp_processor_id(), md->phys_addr,
|
"[0x%lx-0x%lx) into [0x%lx-0x%lx)\n",
|
||||||
md->phys_addr + efi_md_size(md),
|
smp_processor_id(), md->phys_addr,
|
||||||
vaddr & mask, (vaddr & mask) + IA64_GRANULE_SIZE);
|
md->phys_addr + efi_md_size(md),
|
||||||
|
vaddr & mask, (vaddr & mask) + IA64_GRANULE_SIZE);
|
||||||
#endif
|
#endif
|
||||||
return __va(md->phys_addr);
|
return __va(md->phys_addr);
|
||||||
}
|
}
|
||||||
@ -401,11 +416,11 @@ efi_map_pal_code (void)
|
|||||||
* Cannot write to CRx with PSR.ic=1
|
* Cannot write to CRx with PSR.ic=1
|
||||||
*/
|
*/
|
||||||
psr = ia64_clear_ic();
|
psr = ia64_clear_ic();
|
||||||
ia64_itr(0x1, IA64_TR_PALCODE, GRANULEROUNDDOWN((unsigned long) pal_vaddr),
|
ia64_itr(0x1, IA64_TR_PALCODE,
|
||||||
|
GRANULEROUNDDOWN((unsigned long) pal_vaddr),
|
||||||
pte_val(pfn_pte(__pa(pal_vaddr) >> PAGE_SHIFT, PAGE_KERNEL)),
|
pte_val(pfn_pte(__pa(pal_vaddr) >> PAGE_SHIFT, PAGE_KERNEL)),
|
||||||
IA64_GRANULE_SHIFT);
|
IA64_GRANULE_SHIFT);
|
||||||
ia64_set_psr(psr); /* restore psr */
|
ia64_set_psr(psr); /* restore psr */
|
||||||
ia64_srlz_i();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void __init
|
void __init
|
||||||
@ -418,7 +433,10 @@ efi_init (void)
|
|||||||
char *cp, vendor[100] = "unknown";
|
char *cp, vendor[100] = "unknown";
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
/* it's too early to be able to use the standard kernel command line support... */
|
/*
|
||||||
|
* It's too early to be able to use the standard kernel command line
|
||||||
|
* support...
|
||||||
|
*/
|
||||||
for (cp = boot_command_line; *cp; ) {
|
for (cp = boot_command_line; *cp; ) {
|
||||||
if (memcmp(cp, "mem=", 4) == 0) {
|
if (memcmp(cp, "mem=", 4) == 0) {
|
||||||
mem_limit = memparse(cp + 4, &cp);
|
mem_limit = memparse(cp + 4, &cp);
|
||||||
@ -434,9 +452,11 @@ efi_init (void)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (min_addr != 0UL)
|
if (min_addr != 0UL)
|
||||||
printk(KERN_INFO "Ignoring memory below %luMB\n", min_addr >> 20);
|
printk(KERN_INFO "Ignoring memory below %luMB\n",
|
||||||
|
min_addr >> 20);
|
||||||
if (max_addr != ~0UL)
|
if (max_addr != ~0UL)
|
||||||
printk(KERN_INFO "Ignoring memory above %luMB\n", max_addr >> 20);
|
printk(KERN_INFO "Ignoring memory above %luMB\n",
|
||||||
|
max_addr >> 20);
|
||||||
|
|
||||||
efi.systab = __va(ia64_boot_param->efi_systab);
|
efi.systab = __va(ia64_boot_param->efi_systab);
|
||||||
|
|
||||||
@ -444,9 +464,9 @@ efi_init (void)
|
|||||||
* Verify the EFI Table
|
* Verify the EFI Table
|
||||||
*/
|
*/
|
||||||
if (efi.systab == NULL)
|
if (efi.systab == NULL)
|
||||||
panic("Woah! Can't find EFI system table.\n");
|
panic("Whoa! Can't find EFI system table.\n");
|
||||||
if (efi.systab->hdr.signature != EFI_SYSTEM_TABLE_SIGNATURE)
|
if (efi.systab->hdr.signature != EFI_SYSTEM_TABLE_SIGNATURE)
|
||||||
panic("Woah! EFI system table signature incorrect\n");
|
panic("Whoa! EFI system table signature incorrect\n");
|
||||||
if ((efi.systab->hdr.revision >> 16) == 0)
|
if ((efi.systab->hdr.revision >> 16) == 0)
|
||||||
printk(KERN_WARNING "Warning: EFI system table version "
|
printk(KERN_WARNING "Warning: EFI system table version "
|
||||||
"%d.%02d, expected 1.00 or greater\n",
|
"%d.%02d, expected 1.00 or greater\n",
|
||||||
@ -464,7 +484,8 @@ efi_init (void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
printk(KERN_INFO "EFI v%u.%.02u by %s:",
|
printk(KERN_INFO "EFI v%u.%.02u by %s:",
|
||||||
efi.systab->hdr.revision >> 16, efi.systab->hdr.revision & 0xffff, vendor);
|
efi.systab->hdr.revision >> 16,
|
||||||
|
efi.systab->hdr.revision & 0xffff, vendor);
|
||||||
|
|
||||||
efi.mps = EFI_INVALID_TABLE_ADDR;
|
efi.mps = EFI_INVALID_TABLE_ADDR;
|
||||||
efi.acpi = EFI_INVALID_TABLE_ADDR;
|
efi.acpi = EFI_INVALID_TABLE_ADDR;
|
||||||
@ -519,9 +540,12 @@ efi_init (void)
|
|||||||
efi_memory_desc_t *md;
|
efi_memory_desc_t *md;
|
||||||
void *p;
|
void *p;
|
||||||
|
|
||||||
for (i = 0, p = efi_map_start; p < efi_map_end; ++i, p += efi_desc_size) {
|
for (i = 0, p = efi_map_start; p < efi_map_end;
|
||||||
|
++i, p += efi_desc_size)
|
||||||
|
{
|
||||||
md = p;
|
md = p;
|
||||||
printk("mem%02u: type=%u, attr=0x%lx, range=[0x%016lx-0x%016lx) (%luMB)\n",
|
printk("mem%02u: type=%u, attr=0x%lx, "
|
||||||
|
"range=[0x%016lx-0x%016lx) (%luMB)\n",
|
||||||
i, md->type, md->attribute, md->phys_addr,
|
i, md->type, md->attribute, md->phys_addr,
|
||||||
md->phys_addr + efi_md_size(md),
|
md->phys_addr + efi_md_size(md),
|
||||||
md->num_pages >> (20 - EFI_PAGE_SHIFT));
|
md->num_pages >> (20 - EFI_PAGE_SHIFT));
|
||||||
@ -549,8 +573,8 @@ efi_enter_virtual_mode (void)
|
|||||||
md = p;
|
md = p;
|
||||||
if (md->attribute & EFI_MEMORY_RUNTIME) {
|
if (md->attribute & EFI_MEMORY_RUNTIME) {
|
||||||
/*
|
/*
|
||||||
* Some descriptors have multiple bits set, so the order of
|
* Some descriptors have multiple bits set, so the
|
||||||
* the tests is relevant.
|
* order of the tests is relevant.
|
||||||
*/
|
*/
|
||||||
if (md->attribute & EFI_MEMORY_WB) {
|
if (md->attribute & EFI_MEMORY_WB) {
|
||||||
md->virt_addr = (u64) __va(md->phys_addr);
|
md->virt_addr = (u64) __va(md->phys_addr);
|
||||||
@ -558,21 +582,26 @@ efi_enter_virtual_mode (void)
|
|||||||
md->virt_addr = (u64) ioremap(md->phys_addr, 0);
|
md->virt_addr = (u64) ioremap(md->phys_addr, 0);
|
||||||
} else if (md->attribute & EFI_MEMORY_WC) {
|
} else if (md->attribute & EFI_MEMORY_WC) {
|
||||||
#if 0
|
#if 0
|
||||||
md->virt_addr = ia64_remap(md->phys_addr, (_PAGE_A | _PAGE_P
|
md->virt_addr = ia64_remap(md->phys_addr,
|
||||||
| _PAGE_D
|
(_PAGE_A |
|
||||||
| _PAGE_MA_WC
|
_PAGE_P |
|
||||||
| _PAGE_PL_0
|
_PAGE_D |
|
||||||
| _PAGE_AR_RW));
|
_PAGE_MA_WC |
|
||||||
|
_PAGE_PL_0 |
|
||||||
|
_PAGE_AR_RW));
|
||||||
#else
|
#else
|
||||||
printk(KERN_INFO "EFI_MEMORY_WC mapping\n");
|
printk(KERN_INFO "EFI_MEMORY_WC mapping\n");
|
||||||
md->virt_addr = (u64) ioremap(md->phys_addr, 0);
|
md->virt_addr = (u64) ioremap(md->phys_addr, 0);
|
||||||
#endif
|
#endif
|
||||||
} else if (md->attribute & EFI_MEMORY_WT) {
|
} else if (md->attribute & EFI_MEMORY_WT) {
|
||||||
#if 0
|
#if 0
|
||||||
md->virt_addr = ia64_remap(md->phys_addr, (_PAGE_A | _PAGE_P
|
md->virt_addr = ia64_remap(md->phys_addr,
|
||||||
| _PAGE_D | _PAGE_MA_WT
|
(_PAGE_A |
|
||||||
| _PAGE_PL_0
|
_PAGE_P |
|
||||||
| _PAGE_AR_RW));
|
_PAGE_D |
|
||||||
|
_PAGE_MA_WT |
|
||||||
|
_PAGE_PL_0 |
|
||||||
|
_PAGE_AR_RW));
|
||||||
#else
|
#else
|
||||||
printk(KERN_INFO "EFI_MEMORY_WT mapping\n");
|
printk(KERN_INFO "EFI_MEMORY_WT mapping\n");
|
||||||
md->virt_addr = (u64) ioremap(md->phys_addr, 0);
|
md->virt_addr = (u64) ioremap(md->phys_addr, 0);
|
||||||
@ -583,16 +612,18 @@ efi_enter_virtual_mode (void)
|
|||||||
|
|
||||||
status = efi_call_phys(__va(runtime->set_virtual_address_map),
|
status = efi_call_phys(__va(runtime->set_virtual_address_map),
|
||||||
ia64_boot_param->efi_memmap_size,
|
ia64_boot_param->efi_memmap_size,
|
||||||
efi_desc_size, ia64_boot_param->efi_memdesc_version,
|
efi_desc_size,
|
||||||
|
ia64_boot_param->efi_memdesc_version,
|
||||||
ia64_boot_param->efi_memmap);
|
ia64_boot_param->efi_memmap);
|
||||||
if (status != EFI_SUCCESS) {
|
if (status != EFI_SUCCESS) {
|
||||||
printk(KERN_WARNING "warning: unable to switch EFI into virtual mode "
|
printk(KERN_WARNING "warning: unable to switch EFI into "
|
||||||
"(status=%lu)\n", status);
|
"virtual mode (status=%lu)\n", status);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Now that EFI is in virtual mode, we call the EFI functions more efficiently:
|
* Now that EFI is in virtual mode, we call the EFI functions more
|
||||||
|
* efficiently:
|
||||||
*/
|
*/
|
||||||
efi.get_time = virt_get_time;
|
efi.get_time = virt_get_time;
|
||||||
efi.set_time = virt_set_time;
|
efi.set_time = virt_set_time;
|
||||||
@ -606,8 +637,8 @@ efi_enter_virtual_mode (void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Walk the EFI memory map looking for the I/O port range. There can only be one entry of
|
* Walk the EFI memory map looking for the I/O port range. There can only be
|
||||||
* this type, other I/O port ranges should be described via ACPI.
|
* one entry of this type, other I/O port ranges should be described via ACPI.
|
||||||
*/
|
*/
|
||||||
u64
|
u64
|
||||||
efi_get_iobase (void)
|
efi_get_iobase (void)
|
||||||
@ -678,7 +709,6 @@ efi_memmap_intersects (unsigned long phys_addr, unsigned long size)
|
|||||||
|
|
||||||
for (p = efi_map_start; p < efi_map_end; p += efi_desc_size) {
|
for (p = efi_map_start; p < efi_map_end; p += efi_desc_size) {
|
||||||
md = p;
|
md = p;
|
||||||
|
|
||||||
if (md->phys_addr < end && efi_md_end(md) > phys_addr)
|
if (md->phys_addr < end && efi_md_end(md) > phys_addr)
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
@ -731,7 +761,7 @@ efi_mem_attribute (unsigned long phys_addr, unsigned long size)
|
|||||||
if (!md || (md->attribute & ~EFI_MEMORY_RUNTIME) != attr)
|
if (!md || (md->attribute & ~EFI_MEMORY_RUNTIME) != attr)
|
||||||
return 0;
|
return 0;
|
||||||
} while (md);
|
} while (md);
|
||||||
return 0;
|
return 0; /* never reached */
|
||||||
}
|
}
|
||||||
|
|
||||||
u64
|
u64
|
||||||
@ -767,7 +797,7 @@ kern_mem_attribute (unsigned long phys_addr, unsigned long size)
|
|||||||
if (!md || md->attribute != attr)
|
if (!md || md->attribute != attr)
|
||||||
return 0;
|
return 0;
|
||||||
} while (md);
|
} while (md);
|
||||||
return 0;
|
return 0; /* never reached */
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(kern_mem_attribute);
|
EXPORT_SYMBOL(kern_mem_attribute);
|
||||||
|
|
||||||
@ -883,7 +913,7 @@ efi_uart_console_only(void)
|
|||||||
return 1;
|
return 1;
|
||||||
uart = 0;
|
uart = 0;
|
||||||
}
|
}
|
||||||
hdr = (struct efi_generic_dev_path *) ((u8 *) hdr + hdr->length);
|
hdr = (struct efi_generic_dev_path *)((u8 *) hdr + hdr->length);
|
||||||
}
|
}
|
||||||
printk(KERN_ERR "Malformed %s value\n", name);
|
printk(KERN_ERR "Malformed %s value\n", name);
|
||||||
return 0;
|
return 0;
|
||||||
@ -921,10 +951,12 @@ find_memmap_space (void)
|
|||||||
if (!efi_wb(md)) {
|
if (!efi_wb(md)) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (pmd == NULL || !efi_wb(pmd) || efi_md_end(pmd) != md->phys_addr) {
|
if (pmd == NULL || !efi_wb(pmd) ||
|
||||||
|
efi_md_end(pmd) != md->phys_addr) {
|
||||||
contig_low = GRANULEROUNDUP(md->phys_addr);
|
contig_low = GRANULEROUNDUP(md->phys_addr);
|
||||||
contig_high = efi_md_end(md);
|
contig_high = efi_md_end(md);
|
||||||
for (q = p + efi_desc_size; q < efi_map_end; q += efi_desc_size) {
|
for (q = p + efi_desc_size; q < efi_map_end;
|
||||||
|
q += efi_desc_size) {
|
||||||
check_md = q;
|
check_md = q;
|
||||||
if (!efi_wb(check_md))
|
if (!efi_wb(check_md))
|
||||||
break;
|
break;
|
||||||
@ -988,8 +1020,9 @@ efi_memmap_init(unsigned long *s, unsigned long *e)
|
|||||||
for (p = efi_map_start; p < efi_map_end; pmd = md, p += efi_desc_size) {
|
for (p = efi_map_start; p < efi_map_end; pmd = md, p += efi_desc_size) {
|
||||||
md = p;
|
md = p;
|
||||||
if (!efi_wb(md)) {
|
if (!efi_wb(md)) {
|
||||||
if (efi_uc(md) && (md->type == EFI_CONVENTIONAL_MEMORY ||
|
if (efi_uc(md) &&
|
||||||
md->type == EFI_BOOT_SERVICES_DATA)) {
|
(md->type == EFI_CONVENTIONAL_MEMORY ||
|
||||||
|
md->type == EFI_BOOT_SERVICES_DATA)) {
|
||||||
k->attribute = EFI_MEMORY_UC;
|
k->attribute = EFI_MEMORY_UC;
|
||||||
k->start = md->phys_addr;
|
k->start = md->phys_addr;
|
||||||
k->num_pages = md->num_pages;
|
k->num_pages = md->num_pages;
|
||||||
@ -997,10 +1030,12 @@ efi_memmap_init(unsigned long *s, unsigned long *e)
|
|||||||
}
|
}
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (pmd == NULL || !efi_wb(pmd) || efi_md_end(pmd) != md->phys_addr) {
|
if (pmd == NULL || !efi_wb(pmd) ||
|
||||||
|
efi_md_end(pmd) != md->phys_addr) {
|
||||||
contig_low = GRANULEROUNDUP(md->phys_addr);
|
contig_low = GRANULEROUNDUP(md->phys_addr);
|
||||||
contig_high = efi_md_end(md);
|
contig_high = efi_md_end(md);
|
||||||
for (q = p + efi_desc_size; q < efi_map_end; q += efi_desc_size) {
|
for (q = p + efi_desc_size; q < efi_map_end;
|
||||||
|
q += efi_desc_size) {
|
||||||
check_md = q;
|
check_md = q;
|
||||||
if (!efi_wb(check_md))
|
if (!efi_wb(check_md))
|
||||||
break;
|
break;
|
||||||
@ -1025,13 +1060,17 @@ efi_memmap_init(unsigned long *s, unsigned long *e)
|
|||||||
if (md->phys_addr < contig_low) {
|
if (md->phys_addr < contig_low) {
|
||||||
lim = min(efi_md_end(md), contig_low);
|
lim = min(efi_md_end(md), contig_low);
|
||||||
if (efi_uc(md)) {
|
if (efi_uc(md)) {
|
||||||
if (k > kern_memmap && (k-1)->attribute == EFI_MEMORY_UC &&
|
if (k > kern_memmap &&
|
||||||
|
(k-1)->attribute == EFI_MEMORY_UC &&
|
||||||
kmd_end(k-1) == md->phys_addr) {
|
kmd_end(k-1) == md->phys_addr) {
|
||||||
(k-1)->num_pages += (lim - md->phys_addr) >> EFI_PAGE_SHIFT;
|
(k-1)->num_pages +=
|
||||||
|
(lim - md->phys_addr)
|
||||||
|
>> EFI_PAGE_SHIFT;
|
||||||
} else {
|
} else {
|
||||||
k->attribute = EFI_MEMORY_UC;
|
k->attribute = EFI_MEMORY_UC;
|
||||||
k->start = md->phys_addr;
|
k->start = md->phys_addr;
|
||||||
k->num_pages = (lim - md->phys_addr) >> EFI_PAGE_SHIFT;
|
k->num_pages = (lim - md->phys_addr)
|
||||||
|
>> EFI_PAGE_SHIFT;
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1049,7 +1088,8 @@ efi_memmap_init(unsigned long *s, unsigned long *e)
|
|||||||
} else {
|
} else {
|
||||||
k->attribute = EFI_MEMORY_UC;
|
k->attribute = EFI_MEMORY_UC;
|
||||||
k->start = lim;
|
k->start = lim;
|
||||||
k->num_pages = (efi_md_end(md) - lim) >> EFI_PAGE_SHIFT;
|
k->num_pages = (efi_md_end(md) - lim)
|
||||||
|
>> EFI_PAGE_SHIFT;
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1151,8 +1191,10 @@ efi_initialize_iomem_resources(struct resource *code_resource,
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
if ((res = kzalloc(sizeof(struct resource), GFP_KERNEL)) == NULL) {
|
if ((res = kzalloc(sizeof(struct resource),
|
||||||
printk(KERN_ERR "failed to alocate resource for iomem\n");
|
GFP_KERNEL)) == NULL) {
|
||||||
|
printk(KERN_ERR
|
||||||
|
"failed to allocate resource for iomem\n");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1187,44 +1229,44 @@ efi_initialize_iomem_resources(struct resource *code_resource,
|
|||||||
rsvd_regions are sorted
|
rsvd_regions are sorted
|
||||||
*/
|
*/
|
||||||
unsigned long __init
|
unsigned long __init
|
||||||
kdump_find_rsvd_region (unsigned long size,
|
kdump_find_rsvd_region (unsigned long size, struct rsvd_region *r, int n)
|
||||||
struct rsvd_region *r, int n)
|
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
u64 start, end;
|
u64 start, end;
|
||||||
u64 alignment = 1UL << _PAGE_SIZE_64M;
|
u64 alignment = 1UL << _PAGE_SIZE_64M;
|
||||||
void *efi_map_start, *efi_map_end, *p;
|
void *efi_map_start, *efi_map_end, *p;
|
||||||
efi_memory_desc_t *md;
|
efi_memory_desc_t *md;
|
||||||
u64 efi_desc_size;
|
u64 efi_desc_size;
|
||||||
|
|
||||||
efi_map_start = __va(ia64_boot_param->efi_memmap);
|
efi_map_start = __va(ia64_boot_param->efi_memmap);
|
||||||
efi_map_end = efi_map_start + ia64_boot_param->efi_memmap_size;
|
efi_map_end = efi_map_start + ia64_boot_param->efi_memmap_size;
|
||||||
efi_desc_size = ia64_boot_param->efi_memdesc_size;
|
efi_desc_size = ia64_boot_param->efi_memdesc_size;
|
||||||
|
|
||||||
for (p = efi_map_start; p < efi_map_end; p += efi_desc_size) {
|
for (p = efi_map_start; p < efi_map_end; p += efi_desc_size) {
|
||||||
md = p;
|
md = p;
|
||||||
if (!efi_wb(md))
|
if (!efi_wb(md))
|
||||||
continue;
|
continue;
|
||||||
start = ALIGN(md->phys_addr, alignment);
|
start = ALIGN(md->phys_addr, alignment);
|
||||||
end = efi_md_end(md);
|
end = efi_md_end(md);
|
||||||
for (i = 0; i < n; i++) {
|
for (i = 0; i < n; i++) {
|
||||||
if (__pa(r[i].start) >= start && __pa(r[i].end) < end) {
|
if (__pa(r[i].start) >= start && __pa(r[i].end) < end) {
|
||||||
if (__pa(r[i].start) > start + size)
|
if (__pa(r[i].start) > start + size)
|
||||||
return start;
|
return start;
|
||||||
start = ALIGN(__pa(r[i].end), alignment);
|
start = ALIGN(__pa(r[i].end), alignment);
|
||||||
if (i < n-1 && __pa(r[i+1].start) < start + size)
|
if (i < n-1 &&
|
||||||
continue;
|
__pa(r[i+1].start) < start + size)
|
||||||
else
|
continue;
|
||||||
break;
|
else
|
||||||
|
break;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
if (end > start + size)
|
||||||
if (end > start + size)
|
return start;
|
||||||
return start;
|
}
|
||||||
}
|
|
||||||
|
|
||||||
printk(KERN_WARNING "Cannot reserve 0x%lx byte of memory for crashdump\n",
|
printk(KERN_WARNING
|
||||||
size);
|
"Cannot reserve 0x%lx byte of memory for crashdump\n", size);
|
||||||
return ~0UL;
|
return ~0UL;
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -14,10 +14,10 @@ struct fsyscall_gtod_data_t {
|
|||||||
u32 clk_shift;
|
u32 clk_shift;
|
||||||
void *clk_fsys_mmio;
|
void *clk_fsys_mmio;
|
||||||
cycle_t clk_cycle_last;
|
cycle_t clk_cycle_last;
|
||||||
} __attribute__ ((aligned (L1_CACHE_BYTES)));
|
} ____cacheline_aligned;
|
||||||
|
|
||||||
struct itc_jitter_data_t {
|
struct itc_jitter_data_t {
|
||||||
int itc_jitter;
|
int itc_jitter;
|
||||||
cycle_t itc_lastcycle;
|
cycle_t itc_lastcycle;
|
||||||
} __attribute__ ((aligned (L1_CACHE_BYTES)));
|
} ____cacheline_aligned;
|
||||||
|
|
||||||
|
@ -12,6 +12,9 @@ EXPORT_SYMBOL(memset);
|
|||||||
EXPORT_SYMBOL(memcpy);
|
EXPORT_SYMBOL(memcpy);
|
||||||
EXPORT_SYMBOL(strlen);
|
EXPORT_SYMBOL(strlen);
|
||||||
|
|
||||||
|
#include<asm/pgtable.h>
|
||||||
|
EXPORT_SYMBOL_GPL(empty_zero_page);
|
||||||
|
|
||||||
#include <asm/checksum.h>
|
#include <asm/checksum.h>
|
||||||
EXPORT_SYMBOL(ip_fast_csum); /* hand-coded assembly */
|
EXPORT_SYMBOL(ip_fast_csum); /* hand-coded assembly */
|
||||||
EXPORT_SYMBOL(csum_ipv6_magic);
|
EXPORT_SYMBOL(csum_ipv6_magic);
|
||||||
|
@ -381,9 +381,10 @@ static void __kprobes save_previous_kprobe(struct kprobe_ctlblk *kcb)
|
|||||||
static void __kprobes restore_previous_kprobe(struct kprobe_ctlblk *kcb)
|
static void __kprobes restore_previous_kprobe(struct kprobe_ctlblk *kcb)
|
||||||
{
|
{
|
||||||
unsigned int i;
|
unsigned int i;
|
||||||
i = atomic_sub_return(1, &kcb->prev_kprobe_index);
|
i = atomic_read(&kcb->prev_kprobe_index);
|
||||||
__get_cpu_var(current_kprobe) = kcb->prev_kprobe[i].kp;
|
__get_cpu_var(current_kprobe) = kcb->prev_kprobe[i-1].kp;
|
||||||
kcb->kprobe_status = kcb->prev_kprobe[i].status;
|
kcb->kprobe_status = kcb->prev_kprobe[i-1].status;
|
||||||
|
atomic_sub(1, &kcb->prev_kprobe_index);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __kprobes set_current_kprobe(struct kprobe *p,
|
static void __kprobes set_current_kprobe(struct kprobe *p,
|
||||||
|
@ -2,61 +2,69 @@
|
|||||||
* File: mca.c
|
* File: mca.c
|
||||||
* Purpose: Generic MCA handling layer
|
* Purpose: Generic MCA handling layer
|
||||||
*
|
*
|
||||||
* Updated for latest kernel
|
|
||||||
* Copyright (C) 2003 Hewlett-Packard Co
|
* Copyright (C) 2003 Hewlett-Packard Co
|
||||||
* David Mosberger-Tang <davidm@hpl.hp.com>
|
* David Mosberger-Tang <davidm@hpl.hp.com>
|
||||||
*
|
*
|
||||||
* Copyright (C) 2002 Dell Inc.
|
* Copyright (C) 2002 Dell Inc.
|
||||||
* Copyright (C) Matt Domsch (Matt_Domsch@dell.com)
|
* Copyright (C) Matt Domsch <Matt_Domsch@dell.com>
|
||||||
*
|
*
|
||||||
* Copyright (C) 2002 Intel
|
* Copyright (C) 2002 Intel
|
||||||
* Copyright (C) Jenna Hall (jenna.s.hall@intel.com)
|
* Copyright (C) Jenna Hall <jenna.s.hall@intel.com>
|
||||||
*
|
*
|
||||||
* Copyright (C) 2001 Intel
|
* Copyright (C) 2001 Intel
|
||||||
* Copyright (C) Fred Lewis (frederick.v.lewis@intel.com)
|
* Copyright (C) Fred Lewis <frederick.v.lewis@intel.com>
|
||||||
*
|
*
|
||||||
* Copyright (C) 2000 Intel
|
* Copyright (C) 2000 Intel
|
||||||
* Copyright (C) Chuck Fleckenstein (cfleck@co.intel.com)
|
* Copyright (C) Chuck Fleckenstein <cfleck@co.intel.com>
|
||||||
*
|
*
|
||||||
* Copyright (C) 1999, 2004 Silicon Graphics, Inc.
|
* Copyright (C) 1999, 2004 Silicon Graphics, Inc.
|
||||||
* Copyright (C) Vijay Chander(vijay@engr.sgi.com)
|
* Copyright (C) Vijay Chander <vijay@engr.sgi.com>
|
||||||
*
|
*
|
||||||
* 03/04/15 D. Mosberger Added INIT backtrace support.
|
* Copyright (C) 2006 FUJITSU LIMITED
|
||||||
* 02/03/25 M. Domsch GUID cleanups
|
* Copyright (C) Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com>
|
||||||
*
|
*
|
||||||
* 02/01/04 J. Hall Aligned MCA stack to 16 bytes, added platform vs. CPU
|
* 2000-03-29 Chuck Fleckenstein <cfleck@co.intel.com>
|
||||||
* error flag, set SAL default return values, changed
|
* Fixed PAL/SAL update issues, began MCA bug fixes, logging issues,
|
||||||
* error record structure to linked list, added init call
|
* added min save state dump, added INIT handler.
|
||||||
* to sal_get_state_info_size().
|
|
||||||
*
|
*
|
||||||
* 01/01/03 F. Lewis Added setup of CMCI and CPEI IRQs, logging of corrected
|
* 2001-01-03 Fred Lewis <frederick.v.lewis@intel.com>
|
||||||
* platform errors, completed code for logging of
|
* Added setup of CMCI and CPEI IRQs, logging of corrected platform
|
||||||
* corrected & uncorrected machine check errors, and
|
* errors, completed code for logging of corrected & uncorrected
|
||||||
* updated for conformance with Nov. 2000 revision of the
|
* machine check errors, and updated for conformance with Nov. 2000
|
||||||
* SAL 3.0 spec.
|
* revision of the SAL 3.0 spec.
|
||||||
* 00/03/29 C. Fleckenstein Fixed PAL/SAL update issues, began MCA bug fixes, logging issues,
|
*
|
||||||
* added min save state dump, added INIT handler.
|
* 2002-01-04 Jenna Hall <jenna.s.hall@intel.com>
|
||||||
|
* Aligned MCA stack to 16 bytes, added platform vs. CPU error flag,
|
||||||
|
* set SAL default return values, changed error record structure to
|
||||||
|
* linked list, added init call to sal_get_state_info_size().
|
||||||
|
*
|
||||||
|
* 2002-03-25 Matt Domsch <Matt_Domsch@dell.com>
|
||||||
|
* GUID cleanups.
|
||||||
|
*
|
||||||
|
* 2003-04-15 David Mosberger-Tang <davidm@hpl.hp.com>
|
||||||
|
* Added INIT backtrace support.
|
||||||
*
|
*
|
||||||
* 2003-12-08 Keith Owens <kaos@sgi.com>
|
* 2003-12-08 Keith Owens <kaos@sgi.com>
|
||||||
* smp_call_function() must not be called from interrupt context (can
|
* smp_call_function() must not be called from interrupt context
|
||||||
* deadlock on tasklist_lock). Use keventd to call smp_call_function().
|
* (can deadlock on tasklist_lock).
|
||||||
|
* Use keventd to call smp_call_function().
|
||||||
*
|
*
|
||||||
* 2004-02-01 Keith Owens <kaos@sgi.com>
|
* 2004-02-01 Keith Owens <kaos@sgi.com>
|
||||||
* Avoid deadlock when using printk() for MCA and INIT records.
|
* Avoid deadlock when using printk() for MCA and INIT records.
|
||||||
* Delete all record printing code, moved to salinfo_decode in user space.
|
* Delete all record printing code, moved to salinfo_decode in user
|
||||||
* Mark variables and functions static where possible.
|
* space. Mark variables and functions static where possible.
|
||||||
* Delete dead variables and functions.
|
* Delete dead variables and functions. Reorder to remove the need
|
||||||
* Reorder to remove the need for forward declarations and to consolidate
|
* for forward declarations and to consolidate related code.
|
||||||
* related code.
|
|
||||||
*
|
*
|
||||||
* 2005-08-12 Keith Owens <kaos@sgi.com>
|
* 2005-08-12 Keith Owens <kaos@sgi.com>
|
||||||
* Convert MCA/INIT handlers to use per event stacks and SAL/OS state.
|
* Convert MCA/INIT handlers to use per event stacks and SAL/OS
|
||||||
|
* state.
|
||||||
*
|
*
|
||||||
* 2005-10-07 Keith Owens <kaos@sgi.com>
|
* 2005-10-07 Keith Owens <kaos@sgi.com>
|
||||||
* Add notify_die() hooks.
|
* Add notify_die() hooks.
|
||||||
*
|
*
|
||||||
* 2006-09-15 Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com>
|
* 2006-09-15 Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com>
|
||||||
* Add printing support for MCA/INIT.
|
* Add printing support for MCA/INIT.
|
||||||
*
|
*
|
||||||
* 2007-04-27 Russ Anderson <rja@sgi.com>
|
* 2007-04-27 Russ Anderson <rja@sgi.com>
|
||||||
* Support multiple cpus going through OS_MCA in the same event.
|
* Support multiple cpus going through OS_MCA in the same event.
|
||||||
|
@ -1,24 +1,28 @@
|
|||||||
//
|
/*
|
||||||
// assembly portion of the IA64 MCA handling
|
* File: mca_asm.S
|
||||||
//
|
* Purpose: assembly portion of the IA64 MCA handling
|
||||||
// Mods by cfleck to integrate into kernel build
|
*
|
||||||
// 00/03/15 davidm Added various stop bits to get a clean compile
|
* Mods by cfleck to integrate into kernel build
|
||||||
//
|
*
|
||||||
// 00/03/29 cfleck Added code to save INIT handoff state in pt_regs format, switch to temp
|
* 2000-03-15 David Mosberger-Tang <davidm@hpl.hp.com>
|
||||||
// kstack, switch modes, jump to C INIT handler
|
* Added various stop bits to get a clean compile
|
||||||
//
|
*
|
||||||
// 02/01/04 J.Hall <jenna.s.hall@intel.com>
|
* 2000-03-29 Chuck Fleckenstein <cfleck@co.intel.com>
|
||||||
// Before entering virtual mode code:
|
* Added code to save INIT handoff state in pt_regs format,
|
||||||
// 1. Check for TLB CPU error
|
* switch to temp kstack, switch modes, jump to C INIT handler
|
||||||
// 2. Restore current thread pointer to kr6
|
*
|
||||||
// 3. Move stack ptr 16 bytes to conform to C calling convention
|
* 2002-01-04 J.Hall <jenna.s.hall@intel.com>
|
||||||
//
|
* Before entering virtual mode code:
|
||||||
// 04/11/12 Russ Anderson <rja@sgi.com>
|
* 1. Check for TLB CPU error
|
||||||
// Added per cpu MCA/INIT stack save areas.
|
* 2. Restore current thread pointer to kr6
|
||||||
//
|
* 3. Move stack ptr 16 bytes to conform to C calling convention
|
||||||
// 12/08/05 Keith Owens <kaos@sgi.com>
|
*
|
||||||
// Use per cpu MCA/INIT stacks for all data.
|
* 2004-11-12 Russ Anderson <rja@sgi.com>
|
||||||
//
|
* Added per cpu MCA/INIT stack save areas.
|
||||||
|
*
|
||||||
|
* 2005-12-08 Keith Owens <kaos@sgi.com>
|
||||||
|
* Use per cpu MCA/INIT stacks for all data.
|
||||||
|
*/
|
||||||
#include <linux/threads.h>
|
#include <linux/threads.h>
|
||||||
|
|
||||||
#include <asm/asmmacro.h>
|
#include <asm/asmmacro.h>
|
||||||
|
@ -3,7 +3,7 @@
|
|||||||
* Purpose: Generic MCA handling layer
|
* Purpose: Generic MCA handling layer
|
||||||
*
|
*
|
||||||
* Copyright (C) 2004 FUJITSU LIMITED
|
* Copyright (C) 2004 FUJITSU LIMITED
|
||||||
* Copyright (C) Hidetoshi Seto (seto.hidetoshi@jp.fujitsu.com)
|
* Copyright (C) 2004 Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com>
|
||||||
* Copyright (C) 2005 Silicon Graphics, Inc
|
* Copyright (C) 2005 Silicon Graphics, Inc
|
||||||
* Copyright (C) 2005 Keith Owens <kaos@sgi.com>
|
* Copyright (C) 2005 Keith Owens <kaos@sgi.com>
|
||||||
* Copyright (C) 2006 Russ Anderson <rja@sgi.com>
|
* Copyright (C) 2006 Russ Anderson <rja@sgi.com>
|
||||||
|
@ -3,7 +3,7 @@
|
|||||||
* Purpose: Define helpers for Generic MCA handling
|
* Purpose: Define helpers for Generic MCA handling
|
||||||
*
|
*
|
||||||
* Copyright (C) 2004 FUJITSU LIMITED
|
* Copyright (C) 2004 FUJITSU LIMITED
|
||||||
* Copyright (C) Hidetoshi Seto (seto.hidetoshi@jp.fujitsu.com)
|
* Copyright (C) 2004 Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com>
|
||||||
*/
|
*/
|
||||||
/*
|
/*
|
||||||
* Processor error section:
|
* Processor error section:
|
||||||
|
@ -3,7 +3,7 @@
|
|||||||
* Purpose: Assembly portion of Generic MCA handling
|
* Purpose: Assembly portion of Generic MCA handling
|
||||||
*
|
*
|
||||||
* Copyright (C) 2004 FUJITSU LIMITED
|
* Copyright (C) 2004 FUJITSU LIMITED
|
||||||
* Copyright (C) Hidetoshi Seto (seto.hidetoshi@jp.fujitsu.com)
|
* Copyright (C) 2004 Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com>
|
||||||
*/
|
*/
|
||||||
#include <linux/threads.h>
|
#include <linux/threads.h>
|
||||||
|
|
||||||
|
@ -2654,11 +2654,11 @@ pfm_get_task(pfm_context_t *ctx, pid_t pid, struct task_struct **task)
|
|||||||
/* XXX: need to add more checks here */
|
/* XXX: need to add more checks here */
|
||||||
if (pid < 2) return -EPERM;
|
if (pid < 2) return -EPERM;
|
||||||
|
|
||||||
if (pid != current->pid) {
|
if (pid != task_pid_vnr(current)) {
|
||||||
|
|
||||||
read_lock(&tasklist_lock);
|
read_lock(&tasklist_lock);
|
||||||
|
|
||||||
p = find_task_by_pid(pid);
|
p = find_task_by_vpid(pid);
|
||||||
|
|
||||||
/* make sure task cannot go away while we operate on it */
|
/* make sure task cannot go away while we operate on it */
|
||||||
if (p) get_task_struct(p);
|
if (p) get_task_struct(p);
|
||||||
@ -5795,7 +5795,7 @@ pfm_proc_show(struct seq_file *m, void *v)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
struct seq_operations pfm_seq_ops = {
|
const struct seq_operations pfm_seq_ops = {
|
||||||
.start = pfm_proc_start,
|
.start = pfm_proc_start,
|
||||||
.next = pfm_proc_next,
|
.next = pfm_proc_next,
|
||||||
.stop = pfm_proc_stop,
|
.stop = pfm_proc_stop,
|
||||||
|
@ -284,6 +284,7 @@ ia64_sal_cache_flush (u64 cache_type)
|
|||||||
SAL_CALL(isrv, SAL_CACHE_FLUSH, cache_type, 0, 0, 0, 0, 0, 0);
|
SAL_CALL(isrv, SAL_CACHE_FLUSH, cache_type, 0, 0, 0, 0, 0, 0);
|
||||||
return isrv.status;
|
return isrv.status;
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(ia64_sal_cache_flush);
|
||||||
|
|
||||||
void __init
|
void __init
|
||||||
ia64_sal_init (struct ia64_sal_systab *systab)
|
ia64_sal_init (struct ia64_sal_systab *systab)
|
||||||
@ -372,3 +373,16 @@ ia64_sal_oemcall_reentrant(struct ia64_sal_retval *isrvp, u64 oemfunc,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(ia64_sal_oemcall_reentrant);
|
EXPORT_SYMBOL(ia64_sal_oemcall_reentrant);
|
||||||
|
|
||||||
|
long
|
||||||
|
ia64_sal_freq_base (unsigned long which, unsigned long *ticks_per_second,
|
||||||
|
unsigned long *drift_info)
|
||||||
|
{
|
||||||
|
struct ia64_sal_retval isrv;
|
||||||
|
|
||||||
|
SAL_CALL(isrv, SAL_FREQ_BASE, which, 0, 0, 0, 0, 0, 0);
|
||||||
|
*ticks_per_second = isrv.v0;
|
||||||
|
*drift_info = isrv.v1;
|
||||||
|
return isrv.status;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(ia64_sal_freq_base);
|
||||||
|
@ -654,7 +654,7 @@ c_stop (struct seq_file *m, void *v)
|
|||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
struct seq_operations cpuinfo_op = {
|
const struct seq_operations cpuinfo_op = {
|
||||||
.start = c_start,
|
.start = c_start,
|
||||||
.next = c_next,
|
.next = c_next,
|
||||||
.stop = c_stop,
|
.stop = c_stop,
|
||||||
|
@ -767,17 +767,6 @@ void __cpu_die(unsigned int cpu)
|
|||||||
}
|
}
|
||||||
printk(KERN_ERR "CPU %u didn't die...\n", cpu);
|
printk(KERN_ERR "CPU %u didn't die...\n", cpu);
|
||||||
}
|
}
|
||||||
#else /* !CONFIG_HOTPLUG_CPU */
|
|
||||||
int __cpu_disable(void)
|
|
||||||
{
|
|
||||||
return -ENOSYS;
|
|
||||||
}
|
|
||||||
|
|
||||||
void __cpu_die(unsigned int cpu)
|
|
||||||
{
|
|
||||||
/* We said "no" in __cpu_disable */
|
|
||||||
BUG();
|
|
||||||
}
|
|
||||||
#endif /* CONFIG_HOTPLUG_CPU */
|
#endif /* CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
void
|
void
|
||||||
|
@ -35,7 +35,7 @@ trap_init (void)
|
|||||||
fpswa_interface = __va(ia64_boot_param->fpswa);
|
fpswa_interface = __va(ia64_boot_param->fpswa);
|
||||||
}
|
}
|
||||||
|
|
||||||
void
|
int
|
||||||
die (const char *str, struct pt_regs *regs, long err)
|
die (const char *str, struct pt_regs *regs, long err)
|
||||||
{
|
{
|
||||||
static struct {
|
static struct {
|
||||||
@ -62,8 +62,11 @@ die (const char *str, struct pt_regs *regs, long err)
|
|||||||
if (++die.lock_owner_depth < 3) {
|
if (++die.lock_owner_depth < 3) {
|
||||||
printk("%s[%d]: %s %ld [%d]\n",
|
printk("%s[%d]: %s %ld [%d]\n",
|
||||||
current->comm, task_pid_nr(current), str, err, ++die_counter);
|
current->comm, task_pid_nr(current), str, err, ++die_counter);
|
||||||
(void) notify_die(DIE_OOPS, (char *)str, regs, err, 255, SIGSEGV);
|
if (notify_die(DIE_OOPS, str, regs, err, 255, SIGSEGV)
|
||||||
show_regs(regs);
|
!= NOTIFY_STOP)
|
||||||
|
show_regs(regs);
|
||||||
|
else
|
||||||
|
regs = NULL;
|
||||||
} else
|
} else
|
||||||
printk(KERN_ERR "Recursive die() failure, output suppressed\n");
|
printk(KERN_ERR "Recursive die() failure, output suppressed\n");
|
||||||
|
|
||||||
@ -72,17 +75,22 @@ die (const char *str, struct pt_regs *regs, long err)
|
|||||||
add_taint(TAINT_DIE);
|
add_taint(TAINT_DIE);
|
||||||
spin_unlock_irq(&die.lock);
|
spin_unlock_irq(&die.lock);
|
||||||
|
|
||||||
|
if (!regs)
|
||||||
|
return 1;
|
||||||
|
|
||||||
if (panic_on_oops)
|
if (panic_on_oops)
|
||||||
panic("Fatal exception");
|
panic("Fatal exception");
|
||||||
|
|
||||||
do_exit(SIGSEGV);
|
do_exit(SIGSEGV);
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
void
|
int
|
||||||
die_if_kernel (char *str, struct pt_regs *regs, long err)
|
die_if_kernel (char *str, struct pt_regs *regs, long err)
|
||||||
{
|
{
|
||||||
if (!user_mode(regs))
|
if (!user_mode(regs))
|
||||||
die(str, regs, err);
|
return die(str, regs, err);
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
void
|
void
|
||||||
@ -102,7 +110,8 @@ __kprobes ia64_bad_break (unsigned long break_num, struct pt_regs *regs)
|
|||||||
if (notify_die(DIE_BREAK, "break 0", regs, break_num, TRAP_BRKPT, SIGTRAP)
|
if (notify_die(DIE_BREAK, "break 0", regs, break_num, TRAP_BRKPT, SIGTRAP)
|
||||||
== NOTIFY_STOP)
|
== NOTIFY_STOP)
|
||||||
return;
|
return;
|
||||||
die_if_kernel("bugcheck!", regs, break_num);
|
if (die_if_kernel("bugcheck!", regs, break_num))
|
||||||
|
return;
|
||||||
sig = SIGILL; code = ILL_ILLOPC;
|
sig = SIGILL; code = ILL_ILLOPC;
|
||||||
break;
|
break;
|
||||||
|
|
||||||
@ -155,8 +164,9 @@ __kprobes ia64_bad_break (unsigned long break_num, struct pt_regs *regs)
|
|||||||
break;
|
break;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
if (break_num < 0x40000 || break_num > 0x100000)
|
if ((break_num < 0x40000 || break_num > 0x100000)
|
||||||
die_if_kernel("Bad break", regs, break_num);
|
&& die_if_kernel("Bad break", regs, break_num))
|
||||||
|
return;
|
||||||
|
|
||||||
if (break_num < 0x80000) {
|
if (break_num < 0x80000) {
|
||||||
sig = SIGILL; code = __ILL_BREAK;
|
sig = SIGILL; code = __ILL_BREAK;
|
||||||
@ -402,14 +412,15 @@ ia64_illegal_op_fault (unsigned long ec, long arg1, long arg2, long arg3,
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
sprintf(buf, "IA-64 Illegal operation fault");
|
sprintf(buf, "IA-64 Illegal operation fault");
|
||||||
die_if_kernel(buf, ®s, 0);
|
rv.fkt = 0;
|
||||||
|
if (die_if_kernel(buf, ®s, 0))
|
||||||
|
return rv;
|
||||||
|
|
||||||
memset(&si, 0, sizeof(si));
|
memset(&si, 0, sizeof(si));
|
||||||
si.si_signo = SIGILL;
|
si.si_signo = SIGILL;
|
||||||
si.si_code = ILL_ILLOPC;
|
si.si_code = ILL_ILLOPC;
|
||||||
si.si_addr = (void __user *) (regs.cr_iip + ia64_psr(®s)->ri);
|
si.si_addr = (void __user *) (regs.cr_iip + ia64_psr(®s)->ri);
|
||||||
force_sig_info(SIGILL, &si, current);
|
force_sig_info(SIGILL, &si, current);
|
||||||
rv.fkt = 0;
|
|
||||||
return rv;
|
return rv;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -644,6 +655,6 @@ ia64_fault (unsigned long vector, unsigned long isr, unsigned long ifa,
|
|||||||
sprintf(buf, "Fault %lu", vector);
|
sprintf(buf, "Fault %lu", vector);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
die_if_kernel(buf, ®s, error);
|
if (!die_if_kernel(buf, ®s, error))
|
||||||
force_sig(SIGILL, current);
|
force_sig(SIGILL, current);
|
||||||
}
|
}
|
||||||
|
@ -23,7 +23,7 @@
|
|||||||
#include <asm/uaccess.h>
|
#include <asm/uaccess.h>
|
||||||
#include <asm/unaligned.h>
|
#include <asm/unaligned.h>
|
||||||
|
|
||||||
extern void die_if_kernel(char *str, struct pt_regs *regs, long err);
|
extern int die_if_kernel(char *str, struct pt_regs *regs, long err);
|
||||||
|
|
||||||
#undef DEBUG_UNALIGNED_TRAP
|
#undef DEBUG_UNALIGNED_TRAP
|
||||||
|
|
||||||
@ -675,8 +675,9 @@ emulate_load_updates (update_t type, load_store_t ld, struct pt_regs *regs, unsi
|
|||||||
*/
|
*/
|
||||||
if (ld.x6_op == 1 || ld.x6_op == 3) {
|
if (ld.x6_op == 1 || ld.x6_op == 3) {
|
||||||
printk(KERN_ERR "%s: register update on speculative load, error\n", __FUNCTION__);
|
printk(KERN_ERR "%s: register update on speculative load, error\n", __FUNCTION__);
|
||||||
die_if_kernel("unaligned reference on speculative load with register update\n",
|
if (die_if_kernel("unaligned reference on speculative load with register update\n",
|
||||||
regs, 30);
|
regs, 30))
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -1317,7 +1318,8 @@ ia64_handle_unaligned (unsigned long ifa, struct pt_regs *regs)
|
|||||||
|
|
||||||
if (ia64_psr(regs)->be) {
|
if (ia64_psr(regs)->be) {
|
||||||
/* we don't support big-endian accesses */
|
/* we don't support big-endian accesses */
|
||||||
die_if_kernel("big-endian unaligned accesses are not supported", regs, 0);
|
if (die_if_kernel("big-endian unaligned accesses are not supported", regs, 0))
|
||||||
|
return;
|
||||||
goto force_sigbus;
|
goto force_sigbus;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1534,7 +1536,8 @@ ia64_handle_unaligned (unsigned long ifa, struct pt_regs *regs)
|
|||||||
ia64_handle_exception(regs, eh);
|
ia64_handle_exception(regs, eh);
|
||||||
goto done;
|
goto done;
|
||||||
}
|
}
|
||||||
die_if_kernel("error during unaligned kernel access\n", regs, ret);
|
if (die_if_kernel("error during unaligned kernel access\n", regs, ret))
|
||||||
|
return;
|
||||||
/* NOT_REACHED */
|
/* NOT_REACHED */
|
||||||
}
|
}
|
||||||
force_sigbus:
|
force_sigbus:
|
||||||
|
@ -16,7 +16,7 @@
|
|||||||
#include <asm/system.h>
|
#include <asm/system.h>
|
||||||
#include <asm/uaccess.h>
|
#include <asm/uaccess.h>
|
||||||
|
|
||||||
extern void die (char *, struct pt_regs *, long);
|
extern int die(char *, struct pt_regs *, long);
|
||||||
|
|
||||||
#ifdef CONFIG_KPROBES
|
#ifdef CONFIG_KPROBES
|
||||||
static inline int notify_page_fault(struct pt_regs *regs, int trap)
|
static inline int notify_page_fault(struct pt_regs *regs, int trap)
|
||||||
@ -267,9 +267,11 @@ ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *re
|
|||||||
else
|
else
|
||||||
printk(KERN_ALERT "Unable to handle kernel paging request at "
|
printk(KERN_ALERT "Unable to handle kernel paging request at "
|
||||||
"virtual address %016lx\n", address);
|
"virtual address %016lx\n", address);
|
||||||
die("Oops", regs, isr);
|
if (die("Oops", regs, isr))
|
||||||
|
regs = NULL;
|
||||||
bust_spinlocks(0);
|
bust_spinlocks(0);
|
||||||
do_exit(SIGKILL);
|
if (regs)
|
||||||
|
do_exit(SIGKILL);
|
||||||
return;
|
return;
|
||||||
|
|
||||||
out_of_memory:
|
out_of_memory:
|
||||||
|
@ -523,7 +523,7 @@ static ssize_t sn2_ptc_proc_write(struct file *file, const char __user *user, si
|
|||||||
return count;
|
return count;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct seq_operations sn2_ptc_seq_ops = {
|
static const struct seq_operations sn2_ptc_seq_ops = {
|
||||||
.start = sn2_ptc_seq_start,
|
.start = sn2_ptc_seq_start,
|
||||||
.next = sn2_ptc_seq_next,
|
.next = sn2_ptc_seq_next,
|
||||||
.stop = sn2_ptc_seq_stop,
|
.stop = sn2_ptc_seq_stop,
|
||||||
|
@ -33,6 +33,7 @@
|
|||||||
#include <linux/smp_lock.h>
|
#include <linux/smp_lock.h>
|
||||||
#include <linux/nodemask.h>
|
#include <linux/nodemask.h>
|
||||||
#include <linux/smp.h>
|
#include <linux/smp.h>
|
||||||
|
#include <linux/mutex.h>
|
||||||
|
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
#include <asm/topology.h>
|
#include <asm/topology.h>
|
||||||
@ -50,7 +51,7 @@ static void *sn_hwperf_salheap = NULL;
|
|||||||
static int sn_hwperf_obj_cnt = 0;
|
static int sn_hwperf_obj_cnt = 0;
|
||||||
static nasid_t sn_hwperf_master_nasid = INVALID_NASID;
|
static nasid_t sn_hwperf_master_nasid = INVALID_NASID;
|
||||||
static int sn_hwperf_init(void);
|
static int sn_hwperf_init(void);
|
||||||
static DECLARE_MUTEX(sn_hwperf_init_mutex);
|
static DEFINE_MUTEX(sn_hwperf_init_mutex);
|
||||||
|
|
||||||
#define cnode_possible(n) ((n) < num_cnodes)
|
#define cnode_possible(n) ((n) < num_cnodes)
|
||||||
|
|
||||||
@ -577,7 +578,7 @@ static void sn_topology_stop(struct seq_file *m, void *v)
|
|||||||
/*
|
/*
|
||||||
* /proc/sgi_sn/sn_topology, read-only using seq_file
|
* /proc/sgi_sn/sn_topology, read-only using seq_file
|
||||||
*/
|
*/
|
||||||
static struct seq_operations sn_topology_seq_ops = {
|
static const struct seq_operations sn_topology_seq_ops = {
|
||||||
.start = sn_topology_start,
|
.start = sn_topology_start,
|
||||||
.next = sn_topology_next,
|
.next = sn_topology_next,
|
||||||
.stop = sn_topology_stop,
|
.stop = sn_topology_stop,
|
||||||
@ -884,10 +885,10 @@ static int sn_hwperf_init(void)
|
|||||||
int e = 0;
|
int e = 0;
|
||||||
|
|
||||||
/* single threaded, once-only initialization */
|
/* single threaded, once-only initialization */
|
||||||
down(&sn_hwperf_init_mutex);
|
mutex_lock(&sn_hwperf_init_mutex);
|
||||||
|
|
||||||
if (sn_hwperf_salheap) {
|
if (sn_hwperf_salheap) {
|
||||||
up(&sn_hwperf_init_mutex);
|
mutex_unlock(&sn_hwperf_init_mutex);
|
||||||
return e;
|
return e;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -936,7 +937,7 @@ out:
|
|||||||
sn_hwperf_salheap = NULL;
|
sn_hwperf_salheap = NULL;
|
||||||
sn_hwperf_obj_cnt = 0;
|
sn_hwperf_obj_cnt = 0;
|
||||||
}
|
}
|
||||||
up(&sn_hwperf_init_mutex);
|
mutex_unlock(&sn_hwperf_init_mutex);
|
||||||
return e;
|
return e;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -122,38 +122,40 @@ clear_bit_unlock (int nr, volatile void *addr)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* __clear_bit_unlock - Non-atomically clear a bit with release
|
* __clear_bit_unlock - Non-atomically clears a bit in memory with release
|
||||||
|
* @nr: Bit to clear
|
||||||
|
* @addr: Address to start counting from
|
||||||
*
|
*
|
||||||
* This is like clear_bit_unlock, but the implementation uses a store
|
* Similarly to clear_bit_unlock, the implementation uses a store
|
||||||
* with release semantics. See also __raw_spin_unlock().
|
* with release semantics. See also __raw_spin_unlock().
|
||||||
*/
|
*/
|
||||||
static __inline__ void
|
static __inline__ void
|
||||||
__clear_bit_unlock(int nr, volatile void *addr)
|
__clear_bit_unlock(int nr, void *addr)
|
||||||
{
|
{
|
||||||
__u32 mask, new;
|
__u32 * const m = (__u32 *) addr + (nr >> 5);
|
||||||
volatile __u32 *m;
|
__u32 const new = *m & ~(1 << (nr & 31));
|
||||||
|
|
||||||
m = (volatile __u32 *)addr + (nr >> 5);
|
|
||||||
mask = ~(1 << (nr & 31));
|
|
||||||
new = *m & mask;
|
|
||||||
barrier();
|
|
||||||
ia64_st4_rel_nta(m, new);
|
ia64_st4_rel_nta(m, new);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* __clear_bit - Clears a bit in memory (non-atomic version)
|
* __clear_bit - Clears a bit in memory (non-atomic version)
|
||||||
|
* @nr: the bit to clear
|
||||||
|
* @addr: the address to start counting from
|
||||||
|
*
|
||||||
|
* Unlike clear_bit(), this function is non-atomic and may be reordered.
|
||||||
|
* If it's called on the same region of memory simultaneously, the effect
|
||||||
|
* may be that only one operation succeeds.
|
||||||
*/
|
*/
|
||||||
static __inline__ void
|
static __inline__ void
|
||||||
__clear_bit (int nr, volatile void *addr)
|
__clear_bit (int nr, volatile void *addr)
|
||||||
{
|
{
|
||||||
volatile __u32 *p = (__u32 *) addr + (nr >> 5);
|
*((__u32 *) addr + (nr >> 5)) &= ~(1 << (nr & 31));
|
||||||
__u32 m = 1 << (nr & 31);
|
|
||||||
*p &= ~m;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* change_bit - Toggle a bit in memory
|
* change_bit - Toggle a bit in memory
|
||||||
* @nr: Bit to clear
|
* @nr: Bit to toggle
|
||||||
* @addr: Address to start counting from
|
* @addr: Address to start counting from
|
||||||
*
|
*
|
||||||
* change_bit() is atomic and may not be reordered.
|
* change_bit() is atomic and may not be reordered.
|
||||||
@ -178,7 +180,7 @@ change_bit (int nr, volatile void *addr)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* __change_bit - Toggle a bit in memory
|
* __change_bit - Toggle a bit in memory
|
||||||
* @nr: the bit to set
|
* @nr: the bit to toggle
|
||||||
* @addr: the address to start counting from
|
* @addr: the address to start counting from
|
||||||
*
|
*
|
||||||
* Unlike change_bit(), this function is non-atomic and may be reordered.
|
* Unlike change_bit(), this function is non-atomic and may be reordered.
|
||||||
@ -197,7 +199,7 @@ __change_bit (int nr, volatile void *addr)
|
|||||||
* @addr: Address to count from
|
* @addr: Address to count from
|
||||||
*
|
*
|
||||||
* This operation is atomic and cannot be reordered.
|
* This operation is atomic and cannot be reordered.
|
||||||
* It also implies a memory barrier.
|
* It also implies the acquisition side of the memory barrier.
|
||||||
*/
|
*/
|
||||||
static __inline__ int
|
static __inline__ int
|
||||||
test_and_set_bit (int nr, volatile void *addr)
|
test_and_set_bit (int nr, volatile void *addr)
|
||||||
@ -247,11 +249,11 @@ __test_and_set_bit (int nr, volatile void *addr)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* test_and_clear_bit - Clear a bit and return its old value
|
* test_and_clear_bit - Clear a bit and return its old value
|
||||||
* @nr: Bit to set
|
* @nr: Bit to clear
|
||||||
* @addr: Address to count from
|
* @addr: Address to count from
|
||||||
*
|
*
|
||||||
* This operation is atomic and cannot be reordered.
|
* This operation is atomic and cannot be reordered.
|
||||||
* It also implies a memory barrier.
|
* It also implies the acquisition side of the memory barrier.
|
||||||
*/
|
*/
|
||||||
static __inline__ int
|
static __inline__ int
|
||||||
test_and_clear_bit (int nr, volatile void *addr)
|
test_and_clear_bit (int nr, volatile void *addr)
|
||||||
@ -272,7 +274,7 @@ test_and_clear_bit (int nr, volatile void *addr)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* __test_and_clear_bit - Clear a bit and return its old value
|
* __test_and_clear_bit - Clear a bit and return its old value
|
||||||
* @nr: Bit to set
|
* @nr: Bit to clear
|
||||||
* @addr: Address to count from
|
* @addr: Address to count from
|
||||||
*
|
*
|
||||||
* This operation is non-atomic and can be reordered.
|
* This operation is non-atomic and can be reordered.
|
||||||
@ -292,11 +294,11 @@ __test_and_clear_bit(int nr, volatile void * addr)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* test_and_change_bit - Change a bit and return its old value
|
* test_and_change_bit - Change a bit and return its old value
|
||||||
* @nr: Bit to set
|
* @nr: Bit to change
|
||||||
* @addr: Address to count from
|
* @addr: Address to count from
|
||||||
*
|
*
|
||||||
* This operation is atomic and cannot be reordered.
|
* This operation is atomic and cannot be reordered.
|
||||||
* It also implies a memory barrier.
|
* It also implies the acquisition side of the memory barrier.
|
||||||
*/
|
*/
|
||||||
static __inline__ int
|
static __inline__ int
|
||||||
test_and_change_bit (int nr, volatile void *addr)
|
test_and_change_bit (int nr, volatile void *addr)
|
||||||
@ -315,8 +317,12 @@ test_and_change_bit (int nr, volatile void *addr)
|
|||||||
return (old & bit) != 0;
|
return (old & bit) != 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/**
|
||||||
* WARNING: non atomic version.
|
* __test_and_change_bit - Change a bit and return its old value
|
||||||
|
* @nr: Bit to change
|
||||||
|
* @addr: Address to count from
|
||||||
|
*
|
||||||
|
* This operation is non-atomic and can be reordered.
|
||||||
*/
|
*/
|
||||||
static __inline__ int
|
static __inline__ int
|
||||||
__test_and_change_bit (int nr, void *addr)
|
__test_and_change_bit (int nr, void *addr)
|
||||||
|
@ -24,7 +24,9 @@
|
|||||||
extern void ia64_bad_param_for_setreg (void);
|
extern void ia64_bad_param_for_setreg (void);
|
||||||
extern void ia64_bad_param_for_getreg (void);
|
extern void ia64_bad_param_for_getreg (void);
|
||||||
|
|
||||||
|
#ifdef __KERNEL__
|
||||||
register unsigned long ia64_r13 asm ("r13") __used;
|
register unsigned long ia64_r13 asm ("r13") __used;
|
||||||
|
#endif
|
||||||
|
|
||||||
#define ia64_setreg(regnum, val) \
|
#define ia64_setreg(regnum, val) \
|
||||||
({ \
|
({ \
|
||||||
|
@ -3,9 +3,9 @@
|
|||||||
* Purpose: Machine check handling specific defines
|
* Purpose: Machine check handling specific defines
|
||||||
*
|
*
|
||||||
* Copyright (C) 1999, 2004 Silicon Graphics, Inc.
|
* Copyright (C) 1999, 2004 Silicon Graphics, Inc.
|
||||||
* Copyright (C) Vijay Chander (vijay@engr.sgi.com)
|
* Copyright (C) Vijay Chander <vijay@engr.sgi.com>
|
||||||
* Copyright (C) Srinivasa Thirumalachar (sprasad@engr.sgi.com)
|
* Copyright (C) Srinivasa Thirumalachar <sprasad@engr.sgi.com>
|
||||||
* Copyright (C) Russ Anderson (rja@sgi.com)
|
* Copyright (C) Russ Anderson <rja@sgi.com>
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#ifndef _ASM_IA64_MCA_H
|
#ifndef _ASM_IA64_MCA_H
|
||||||
|
@ -1,8 +1,9 @@
|
|||||||
/*
|
/*
|
||||||
* File: mca_asm.h
|
* File: mca_asm.h
|
||||||
|
* Purpose: Machine check handling specific defines
|
||||||
*
|
*
|
||||||
* Copyright (C) 1999 Silicon Graphics, Inc.
|
* Copyright (C) 1999 Silicon Graphics, Inc.
|
||||||
* Copyright (C) Vijay Chander (vijay@engr.sgi.com)
|
* Copyright (C) Vijay Chander <vijay@engr.sgi.com>
|
||||||
* Copyright (C) Srinivasa Thirumalachar <sprasad@engr.sgi.com>
|
* Copyright (C) Srinivasa Thirumalachar <sprasad@engr.sgi.com>
|
||||||
* Copyright (C) 2000 Hewlett-Packard Co.
|
* Copyright (C) 2000 Hewlett-Packard Co.
|
||||||
* Copyright (C) 2000 David Mosberger-Tang <davidm@hpl.hp.com>
|
* Copyright (C) 2000 David Mosberger-Tang <davidm@hpl.hp.com>
|
||||||
|
@ -473,7 +473,7 @@ ia64_set_psr (__u64 psr)
|
|||||||
{
|
{
|
||||||
ia64_stop();
|
ia64_stop();
|
||||||
ia64_setreg(_IA64_REG_PSR_L, psr);
|
ia64_setreg(_IA64_REG_PSR_L, psr);
|
||||||
ia64_srlz_d();
|
ia64_srlz_i();
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -649,17 +649,6 @@ typedef struct err_rec {
|
|||||||
* Now define a couple of inline functions for improved type checking
|
* Now define a couple of inline functions for improved type checking
|
||||||
* and convenience.
|
* and convenience.
|
||||||
*/
|
*/
|
||||||
static inline long
|
|
||||||
ia64_sal_freq_base (unsigned long which, unsigned long *ticks_per_second,
|
|
||||||
unsigned long *drift_info)
|
|
||||||
{
|
|
||||||
struct ia64_sal_retval isrv;
|
|
||||||
|
|
||||||
SAL_CALL(isrv, SAL_FREQ_BASE, which, 0, 0, 0, 0, 0, 0);
|
|
||||||
*ticks_per_second = isrv.v0;
|
|
||||||
*drift_info = isrv.v1;
|
|
||||||
return isrv.status;
|
|
||||||
}
|
|
||||||
|
|
||||||
extern s64 ia64_sal_cache_flush (u64 cache_type);
|
extern s64 ia64_sal_cache_flush (u64 cache_type);
|
||||||
extern void __init check_sal_cache_flush (void);
|
extern void __init check_sal_cache_flush (void);
|
||||||
@ -841,6 +830,9 @@ extern int ia64_sal_oemcall_nolock(struct ia64_sal_retval *, u64, u64, u64,
|
|||||||
u64, u64, u64, u64, u64);
|
u64, u64, u64, u64, u64);
|
||||||
extern int ia64_sal_oemcall_reentrant(struct ia64_sal_retval *, u64, u64, u64,
|
extern int ia64_sal_oemcall_reentrant(struct ia64_sal_retval *, u64, u64, u64,
|
||||||
u64, u64, u64, u64, u64);
|
u64, u64, u64, u64, u64);
|
||||||
|
extern long
|
||||||
|
ia64_sal_freq_base (unsigned long which, unsigned long *ticks_per_second,
|
||||||
|
unsigned long *drift_info);
|
||||||
#ifdef CONFIG_HOTPLUG_CPU
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
/*
|
/*
|
||||||
* System Abstraction Layer Specification
|
* System Abstraction Layer Specification
|
||||||
|
Loading…
Reference in New Issue
Block a user