Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
L
linux-davinci
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Redmine
Redmine
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Operations
Operations
Metrics
Environments
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
linux
linux-davinci
Commits
d1751388
Commit
d1751388
authored
Sep 29, 2009
by
David S. Miller
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
sparc64: Cache per-cpu %pcr register value in perf code.
Signed-off-by:
David S. Miller
<
davem@davemloft.net
>
parent
6e804251
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
42 additions
and
17 deletions
+42
-17
arch/sparc/kernel/perf_event.c
arch/sparc/kernel/perf_event.c
+42
-17
No files found.
arch/sparc/kernel/perf_event.c
View file @
d1751388
...
@@ -56,7 +56,8 @@ struct cpu_hw_events {
...
@@ -56,7 +56,8 @@ struct cpu_hw_events {
struct
perf_event
*
events
[
MAX_HWEVENTS
];
struct
perf_event
*
events
[
MAX_HWEVENTS
];
unsigned
long
used_mask
[
BITS_TO_LONGS
(
MAX_HWEVENTS
)];
unsigned
long
used_mask
[
BITS_TO_LONGS
(
MAX_HWEVENTS
)];
unsigned
long
active_mask
[
BITS_TO_LONGS
(
MAX_HWEVENTS
)];
unsigned
long
active_mask
[
BITS_TO_LONGS
(
MAX_HWEVENTS
)];
int
enabled
;
u64
pcr
;
int
enabled
;
};
};
DEFINE_PER_CPU
(
struct
cpu_hw_events
,
cpu_hw_events
)
=
{
.
enabled
=
1
,
};
DEFINE_PER_CPU
(
struct
cpu_hw_events
,
cpu_hw_events
)
=
{
.
enabled
=
1
,
};
...
@@ -464,21 +465,30 @@ static u64 nop_for_index(int idx)
...
@@ -464,21 +465,30 @@ static u64 nop_for_index(int idx)
sparc_pmu
->
lower_nop
,
idx
);
sparc_pmu
->
lower_nop
,
idx
);
}
}
static
inline
void
sparc_pmu_enable_event
(
struct
hw_perf_event
*
hwc
,
int
idx
)
static
inline
void
sparc_pmu_enable_event
(
struct
cpu_hw_events
*
cpuc
,
struct
hw_perf_event
*
hwc
,
int
idx
)
{
{
u64
val
,
mask
=
mask_for_index
(
idx
);
u64
val
,
mask
=
mask_for_index
(
idx
);
val
=
pcr_ops
->
read
();
val
=
cpuc
->
pcr
;
pcr_ops
->
write
((
val
&
~
mask
)
|
hwc
->
config
);
val
&=
~
mask
;
val
|=
hwc
->
config
;
cpuc
->
pcr
=
val
;
pcr_ops
->
write
(
cpuc
->
pcr
);
}
}
static
inline
void
sparc_pmu_disable_event
(
struct
hw_perf_event
*
hwc
,
int
idx
)
static
inline
void
sparc_pmu_disable_event
(
struct
cpu_hw_events
*
cpuc
,
struct
hw_perf_event
*
hwc
,
int
idx
)
{
{
u64
mask
=
mask_for_index
(
idx
);
u64
mask
=
mask_for_index
(
idx
);
u64
nop
=
nop_for_index
(
idx
);
u64
nop
=
nop_for_index
(
idx
);
u64
val
=
pcr_ops
->
read
()
;
u64
val
;
pcr_ops
->
write
((
val
&
~
mask
)
|
nop
);
val
=
cpuc
->
pcr
;
val
&=
~
mask
;
val
|=
nop
;
cpuc
->
pcr
=
val
;
pcr_ops
->
write
(
cpuc
->
pcr
);
}
}
void
hw_perf_enable
(
void
)
void
hw_perf_enable
(
void
)
...
@@ -493,7 +503,7 @@ void hw_perf_enable(void)
...
@@ -493,7 +503,7 @@ void hw_perf_enable(void)
cpuc
->
enabled
=
1
;
cpuc
->
enabled
=
1
;
barrier
();
barrier
();
val
=
pcr_ops
->
read
()
;
val
=
cpuc
->
pcr
;
for
(
i
=
0
;
i
<
MAX_HWEVENTS
;
i
++
)
{
for
(
i
=
0
;
i
<
MAX_HWEVENTS
;
i
++
)
{
struct
perf_event
*
cp
=
cpuc
->
events
[
i
];
struct
perf_event
*
cp
=
cpuc
->
events
[
i
];
...
@@ -505,7 +515,9 @@ void hw_perf_enable(void)
...
@@ -505,7 +515,9 @@ void hw_perf_enable(void)
val
|=
hwc
->
config_base
;
val
|=
hwc
->
config_base
;
}
}
pcr_ops
->
write
(
val
);
cpuc
->
pcr
=
val
;
pcr_ops
->
write
(
cpuc
->
pcr
);
}
}
void
hw_perf_disable
(
void
)
void
hw_perf_disable
(
void
)
...
@@ -518,10 +530,12 @@ void hw_perf_disable(void)
...
@@ -518,10 +530,12 @@ void hw_perf_disable(void)
cpuc
->
enabled
=
0
;
cpuc
->
enabled
=
0
;
val
=
pcr_ops
->
read
()
;
val
=
cpuc
->
pcr
;
val
&=
~
(
PCR_UTRACE
|
PCR_STRACE
|
val
&=
~
(
PCR_UTRACE
|
PCR_STRACE
|
sparc_pmu
->
hv_bit
|
sparc_pmu
->
irq_bit
);
sparc_pmu
->
hv_bit
|
sparc_pmu
->
irq_bit
);
pcr_ops
->
write
(
val
);
cpuc
->
pcr
=
val
;
pcr_ops
->
write
(
cpuc
->
pcr
);
}
}
static
u32
read_pmc
(
int
idx
)
static
u32
read_pmc
(
int
idx
)
...
@@ -593,13 +607,13 @@ static int sparc_pmu_enable(struct perf_event *event)
...
@@ -593,13 +607,13 @@ static int sparc_pmu_enable(struct perf_event *event)
if
(
test_and_set_bit
(
idx
,
cpuc
->
used_mask
))
if
(
test_and_set_bit
(
idx
,
cpuc
->
used_mask
))
return
-
EAGAIN
;
return
-
EAGAIN
;
sparc_pmu_disable_event
(
hwc
,
idx
);
sparc_pmu_disable_event
(
cpuc
,
hwc
,
idx
);
cpuc
->
events
[
idx
]
=
event
;
cpuc
->
events
[
idx
]
=
event
;
set_bit
(
idx
,
cpuc
->
active_mask
);
set_bit
(
idx
,
cpuc
->
active_mask
);
sparc_perf_event_set_period
(
event
,
hwc
,
idx
);
sparc_perf_event_set_period
(
event
,
hwc
,
idx
);
sparc_pmu_enable_event
(
hwc
,
idx
);
sparc_pmu_enable_event
(
cpuc
,
hwc
,
idx
);
perf_event_update_userpage
(
event
);
perf_event_update_userpage
(
event
);
return
0
;
return
0
;
}
}
...
@@ -635,7 +649,7 @@ static void sparc_pmu_disable(struct perf_event *event)
...
@@ -635,7 +649,7 @@ static void sparc_pmu_disable(struct perf_event *event)
int
idx
=
hwc
->
idx
;
int
idx
=
hwc
->
idx
;
clear_bit
(
idx
,
cpuc
->
active_mask
);
clear_bit
(
idx
,
cpuc
->
active_mask
);
sparc_pmu_disable_event
(
hwc
,
idx
);
sparc_pmu_disable_event
(
cpuc
,
hwc
,
idx
);
barrier
();
barrier
();
...
@@ -649,18 +663,29 @@ static void sparc_pmu_disable(struct perf_event *event)
...
@@ -649,18 +663,29 @@ static void sparc_pmu_disable(struct perf_event *event)
static
void
sparc_pmu_read
(
struct
perf_event
*
event
)
static
void
sparc_pmu_read
(
struct
perf_event
*
event
)
{
{
struct
hw_perf_event
*
hwc
=
&
event
->
hw
;
struct
hw_perf_event
*
hwc
=
&
event
->
hw
;
sparc_perf_event_update
(
event
,
hwc
,
hwc
->
idx
);
sparc_perf_event_update
(
event
,
hwc
,
hwc
->
idx
);
}
}
static
void
sparc_pmu_unthrottle
(
struct
perf_event
*
event
)
static
void
sparc_pmu_unthrottle
(
struct
perf_event
*
event
)
{
{
struct
cpu_hw_events
*
cpuc
=
&
__get_cpu_var
(
cpu_hw_events
);
struct
hw_perf_event
*
hwc
=
&
event
->
hw
;
struct
hw_perf_event
*
hwc
=
&
event
->
hw
;
sparc_pmu_enable_event
(
hwc
,
hwc
->
idx
);
sparc_pmu_enable_event
(
cpuc
,
hwc
,
hwc
->
idx
);
}
}
static
atomic_t
active_events
=
ATOMIC_INIT
(
0
);
static
atomic_t
active_events
=
ATOMIC_INIT
(
0
);
static
DEFINE_MUTEX
(
pmc_grab_mutex
);
static
DEFINE_MUTEX
(
pmc_grab_mutex
);
static
void
perf_stop_nmi_watchdog
(
void
*
unused
)
{
struct
cpu_hw_events
*
cpuc
=
&
__get_cpu_var
(
cpu_hw_events
);
stop_nmi_watchdog
(
NULL
);
cpuc
->
pcr
=
pcr_ops
->
read
();
}
void
perf_event_grab_pmc
(
void
)
void
perf_event_grab_pmc
(
void
)
{
{
if
(
atomic_inc_not_zero
(
&
active_events
))
if
(
atomic_inc_not_zero
(
&
active_events
))
...
@@ -669,7 +694,7 @@ void perf_event_grab_pmc(void)
...
@@ -669,7 +694,7 @@ void perf_event_grab_pmc(void)
mutex_lock
(
&
pmc_grab_mutex
);
mutex_lock
(
&
pmc_grab_mutex
);
if
(
atomic_read
(
&
active_events
)
==
0
)
{
if
(
atomic_read
(
&
active_events
)
==
0
)
{
if
(
atomic_read
(
&
nmi_active
)
>
0
)
{
if
(
atomic_read
(
&
nmi_active
)
>
0
)
{
on_each_cpu
(
stop_nmi_watchdog
,
NULL
,
1
);
on_each_cpu
(
perf_
stop_nmi_watchdog
,
NULL
,
1
);
BUG_ON
(
atomic_read
(
&
nmi_active
)
!=
0
);
BUG_ON
(
atomic_read
(
&
nmi_active
)
!=
0
);
}
}
atomic_inc
(
&
active_events
);
atomic_inc
(
&
active_events
);
...
@@ -978,7 +1003,7 @@ static int __kprobes perf_event_nmi_handler(struct notifier_block *self,
...
@@ -978,7 +1003,7 @@ static int __kprobes perf_event_nmi_handler(struct notifier_block *self,
continue
;
continue
;
if
(
perf_event_overflow
(
event
,
1
,
&
data
,
regs
))
if
(
perf_event_overflow
(
event
,
1
,
&
data
,
regs
))
sparc_pmu_disable_event
(
hwc
,
idx
);
sparc_pmu_disable_event
(
cpuc
,
hwc
,
idx
);
}
}
return
NOTIFY_STOP
;
return
NOTIFY_STOP
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment