Kill the ever confusing typedef and use u64.
NOT FOR INCLUSION - Must be regenerated at some point via coccinelle
Not-Signed-off-by: Thomas Gleixner <[email protected]>
---
arch/alpha/kernel/time.c | 4 -
arch/arc/kernel/time.c | 12 ++--
arch/arm/mach-davinci/time.c | 2
arch/arm/mach-ep93xx/timer-ep93xx.c | 4 -
arch/arm/mach-footbridge/dc21285-timer.c | 2
arch/arm/mach-ixp4xx/common.c | 2
arch/arm/mach-mmp/time.c | 2
arch/arm/mach-omap2/timer.c | 4 -
arch/arm/plat-iop/time.c | 2
arch/avr32/kernel/time.c | 4 -
arch/blackfin/kernel/time-ts.c | 4 -
arch/c6x/kernel/time.c | 2
arch/hexagon/kernel/time.c | 4 -
arch/ia64/kernel/cyclone.c | 4 -
arch/ia64/kernel/fsyscall_gtod_data.h | 6 +-
arch/ia64/kernel/time.c | 6 +-
arch/ia64/sn/kernel/sn2/timer.c | 4 -
arch/m68k/68000/timers.c | 2
arch/m68k/coldfire/dma_timer.c | 2
arch/m68k/coldfire/pit.c | 2
arch/m68k/coldfire/sltimers.c | 2
arch/m68k/coldfire/timers.c | 2
arch/microblaze/kernel/timer.c | 6 +-
arch/mips/alchemy/common/time.c | 2
arch/mips/cavium-octeon/csrc-octeon.c | 2
arch/mips/jz4740/time.c | 2
arch/mips/kernel/cevt-txx9.c | 2
arch/mips/kernel/csrc-bcm1480.c | 4 -
arch/mips/kernel/csrc-ioasic.c | 2
arch/mips/kernel/csrc-r4k.c | 2
arch/mips/kernel/csrc-sb1250.c | 4 -
arch/mips/loongson32/common/time.c | 4 -
arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c | 4 -
arch/mips/loongson64/loongson-3/hpet.c | 4 -
arch/mips/mti-malta/malta-time.c | 2
arch/mips/netlogic/common/time.c | 4 -
arch/mips/sgi-ip27/ip27-timer.c | 2
arch/mn10300/kernel/csrc-mn10300.c | 2
arch/nios2/kernel/time.c | 2
arch/openrisc/kernel/time.c | 4 -
arch/parisc/kernel/time.c | 2
arch/powerpc/kernel/time.c | 14 ++---
arch/s390/kernel/time.c | 2
arch/sparc/kernel/time_32.c | 2
arch/sparc/kernel/time_64.c | 2
arch/um/kernel/time.c | 2
arch/unicore32/kernel/time.c | 2
arch/x86/entry/vdso/vclock_gettime.c | 8 +--
arch/x86/include/asm/kvm_host.h | 2
arch/x86/include/asm/pvclock.h | 6 +-
arch/x86/include/asm/tsc.h | 2
arch/x86/include/asm/vgtod.h | 4 -
arch/x86/kernel/apb_timer.c | 4 -
arch/x86/kernel/cpu/mshyperv.c | 4 -
arch/x86/kernel/hpet.c | 14 ++---
arch/x86/kernel/kvmclock.c | 10 ++--
arch/x86/kernel/pvclock.c | 4 -
arch/x86/kernel/tsc.c | 6 +-
arch/x86/kvm/x86.c | 14 ++---
arch/x86/lguest/boot.c | 2
arch/x86/platform/uv/uv_time.c | 8 +--
arch/x86/xen/time.c | 6 +-
arch/x86/xen/xen-ops.h | 2
arch/xtensa/kernel/time.c | 4 -
drivers/char/hpet.c | 4 -
drivers/clocksource/acpi_pm.c | 14 ++---
drivers/clocksource/arm_arch_timer.c | 4 -
drivers/clocksource/arm_global_timer.c | 2
drivers/clocksource/cadence_ttc_timer.c | 4 -
drivers/clocksource/clksrc-dbx500-prcmu.c | 2
drivers/clocksource/dw_apb_timer.c | 8 +--
drivers/clocksource/em_sti.c | 12 ++--
drivers/clocksource/exynos_mct.c | 6 +-
drivers/clocksource/h8300_timer16.c | 2
drivers/clocksource/h8300_tpu.c | 2
drivers/clocksource/i8253.c | 4 -
drivers/clocksource/jcore-pit.c | 2
drivers/clocksource/metag_generic.c | 2
drivers/clocksource/mips-gic-timer.c | 2
drivers/clocksource/mmio.c | 18 +++----
drivers/clocksource/mxs_timer.c | 2
drivers/clocksource/qcom-timer.c | 2
drivers/clocksource/samsung_pwm_timer.c | 2
drivers/clocksource/scx200_hrt.c | 4 -
drivers/clocksource/sh_cmt.c | 2
drivers/clocksource/sh_tmu.c | 2
drivers/clocksource/tcb_clksrc.c | 4 -
drivers/clocksource/time-pistachio.c | 4 -
drivers/clocksource/timer-atlas7.c | 2
drivers/clocksource/timer-atmel-pit.c | 2
drivers/clocksource/timer-atmel-st.c | 2
drivers/clocksource/timer-nps.c | 4 -
drivers/clocksource/timer-prima2.c | 2
drivers/clocksource/timer-sun5i.c | 2
drivers/clocksource/timer-ti-32k.c | 4 -
drivers/clocksource/vt8500_timer.c | 4 -
drivers/hv/hv.c | 8 +--
drivers/irqchip/irq-mips-gic.c | 16 +++---
drivers/net/ethernet/amd/xgbe/xgbe-ptp.c | 2
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c | 2
drivers/net/ethernet/freescale/fec_ptp.c | 2
drivers/net/ethernet/intel/e1000e/netdev.c | 18 +++----
drivers/net/ethernet/intel/e1000e/ptp.c | 4 -
drivers/net/ethernet/intel/igb/igb_ptp.c | 4 -
drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c | 4 -
drivers/net/ethernet/mellanox/mlx4/en_clock.c | 2
drivers/net/ethernet/mellanox/mlx4/main.c | 4 -
drivers/net/ethernet/mellanox/mlx5/core/en_clock.c | 2
drivers/net/ethernet/mellanox/mlx5/core/main.c | 4 -
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h | 2
drivers/net/ethernet/ti/cpts.c | 2
include/kvm/arm_arch_timer.h | 4 -
include/linux/clocksource.h | 22 ++++----
include/linux/dw_apb_timer.h | 2
include/linux/irqchip/mips-gic.h | 8 +--
include/linux/mlx4/device.h | 2
include/linux/timecounter.h | 12 ++--
include/linux/timekeeper_internal.h | 10 ++--
include/linux/timekeeping.h | 4 -
include/linux/types.h | 3 -
kernel/time/clocksource.c | 2
kernel/time/jiffies.c | 4 -
kernel/time/timecounter.c | 6 +-
kernel/time/timekeeping.c | 50 ++++++++++----------
kernel/time/timekeeping_internal.h | 6 +-
kernel/trace/ftrace.c | 4 -
kernel/trace/trace.c | 6 +-
kernel/trace/trace.h | 8 +--
kernel/trace/trace_irqsoff.c | 4 -
kernel/trace/trace_sched_wakeup.c | 4 -
sound/hda/hdac_stream.c | 6 +-
virt/kvm/arm/arch_timer.c | 6 +-
132 files changed, 318 insertions(+), 321 deletions(-)
--- a/arch/alpha/kernel/time.c
+++ b/arch/alpha/kernel/time.c
@@ -133,7 +133,7 @@ init_rtc_clockevent(void)
* The QEMU clock as a clocksource primitive.
*/
-static cycle_t
+static u64
qemu_cs_read(struct clocksource *cs)
{
return qemu_get_vmtime();
@@ -260,7 +260,7 @@ common_init_rtc(void)
* use this method when WTINT is in use.
*/
-static cycle_t read_rpcc(struct clocksource *cs)
+static u64 read_rpcc(struct clocksource *cs)
{
return rpcc();
}
--- a/arch/arc/kernel/time.c
+++ b/arch/arc/kernel/time.c
@@ -83,7 +83,7 @@ static int noinline arc_get_timer_clk(st
#ifdef CONFIG_ARC_HAS_GFRC
-static cycle_t arc_read_gfrc(struct clocksource *cs)
+static u64 arc_read_gfrc(struct clocksource *cs)
{
unsigned long flags;
union {
@@ -92,7 +92,7 @@ static cycle_t arc_read_gfrc(struct cloc
#else
struct { u32 l, h; };
#endif
- cycle_t full;
+ u64 full;
} stamp;
local_irq_save(flags);
@@ -140,7 +140,7 @@ CLOCKSOURCE_OF_DECLARE(arc_gfrc, "snps,a
#define AUX_RTC_LOW 0x104
#define AUX_RTC_HIGH 0x105
-static cycle_t arc_read_rtc(struct clocksource *cs)
+static u64 arc_read_rtc(struct clocksource *cs)
{
unsigned long status;
union {
@@ -149,7 +149,7 @@ static cycle_t arc_read_rtc(struct clock
#else
struct { u32 low, high; };
#endif
- cycle_t full;
+ u64 full;
} stamp;
/*
@@ -203,9 +203,9 @@ CLOCKSOURCE_OF_DECLARE(arc_rtc, "snps,ar
* 32bit TIMER1 to keep counting monotonically and wraparound
*/
-static cycle_t arc_read_timer1(struct clocksource *cs)
+static u64 arc_read_timer1(struct clocksource *cs)
{
- return (cycle_t) read_aux_reg(ARC_REG_TIMER1_CNT);
+ return (u64) read_aux_reg(ARC_REG_TIMER1_CNT);
}
static struct clocksource arc_counter_timer1 = {
--- a/arch/arm/mach-davinci/time.c
+++ b/arch/arm/mach-davinci/time.c
@@ -268,7 +268,7 @@ static void __init timer_init(void)
/*
* clocksource
*/
-static cycle_t read_cycles(struct clocksource *cs)
+static u64 read_cycles(struct clocksource *cs)
{
struct timer_s *t = &timers[TID_CLOCKSOURCE];
--- a/arch/arm/mach-ep93xx/timer-ep93xx.c
+++ b/arch/arm/mach-ep93xx/timer-ep93xx.c
@@ -59,13 +59,13 @@ static u64 notrace ep93xx_read_sched_clo
return ret;
}
-cycle_t ep93xx_clocksource_read(struct clocksource *c)
+u64 ep93xx_clocksource_read(struct clocksource *c)
{
u64 ret;
ret = readl(EP93XX_TIMER4_VALUE_LOW);
ret |= ((u64) (readl(EP93XX_TIMER4_VALUE_HIGH) & 0xff) << 32);
- return (cycle_t) ret;
+ return ret;
}
static int ep93xx_clkevt_set_next_event(unsigned long next,
--- a/arch/arm/mach-footbridge/dc21285-timer.c
+++ b/arch/arm/mach-footbridge/dc21285-timer.c
@@ -19,7 +19,7 @@
#include "common.h"
-static cycle_t cksrc_dc21285_read(struct clocksource *cs)
+static u64 cksrc_dc21285_read(struct clocksource *cs)
{
return cs->mask - *CSR_TIMER2_VALUE;
}
--- a/arch/arm/mach-ixp4xx/common.c
+++ b/arch/arm/mach-ixp4xx/common.c
@@ -493,7 +493,7 @@ static u64 notrace ixp4xx_read_sched_clo
* clocksource
*/
-static cycle_t ixp4xx_clocksource_read(struct clocksource *c)
+static u64 ixp4xx_clocksource_read(struct clocksource *c)
{
return *IXP4XX_OSTS;
}
--- a/arch/arm/mach-mmp/time.c
+++ b/arch/arm/mach-mmp/time.c
@@ -144,7 +144,7 @@ static struct clock_event_device ckevt =
.set_state_oneshot = timer_set_shutdown,
};
-static cycle_t clksrc_read(struct clocksource *cs)
+static u64 clksrc_read(struct clocksource *cs)
{
return timer_read();
}
--- a/arch/arm/mach-omap2/timer.c
+++ b/arch/arm/mach-omap2/timer.c
@@ -369,9 +369,9 @@ static bool use_gptimer_clksrc __initdat
/*
* clocksource
*/
-static cycle_t clocksource_read_cycles(struct clocksource *cs)
+static u64 clocksource_read_cycles(struct clocksource *cs)
{
- return (cycle_t)__omap_dm_timer_read_counter(&clksrc,
+ return (u64)__omap_dm_timer_read_counter(&clksrc,
OMAP_TIMER_NONPOSTED);
}
--- a/arch/arm/plat-iop/time.c
+++ b/arch/arm/plat-iop/time.c
@@ -38,7 +38,7 @@
/*
* IOP clocksource (free-running timer 1).
*/
-static cycle_t notrace iop_clocksource_read(struct clocksource *unused)
+static u64 notrace iop_clocksource_read(struct clocksource *unused)
{
return 0xffffffffu - read_tcr1();
}
--- a/arch/avr32/kernel/time.c
+++ b/arch/avr32/kernel/time.c
@@ -20,9 +20,9 @@
static bool disable_cpu_idle_poll;
-static cycle_t read_cycle_count(struct clocksource *cs)
+static u64 read_cycle_count(struct clocksource *cs)
{
- return (cycle_t)sysreg_read(COUNT);
+ return (u64)sysreg_read(COUNT);
}
/*
--- a/arch/blackfin/kernel/time-ts.c
+++ b/arch/blackfin/kernel/time-ts.c
@@ -26,7 +26,7 @@
#if defined(CONFIG_CYCLES_CLOCKSOURCE)
-static notrace cycle_t bfin_read_cycles(struct clocksource *cs)
+static notrace u64 bfin_read_cycles(struct clocksource *cs)
{
#ifdef CONFIG_CPU_FREQ
return __bfin_cycles_off + (get_cycles() << __bfin_cycles_mod);
@@ -80,7 +80,7 @@ void __init setup_gptimer0(void)
enable_gptimers(TIMER0bit);
}
-static cycle_t bfin_read_gptimer0(struct clocksource *cs)
+static u64 bfin_read_gptimer0(struct clocksource *cs)
{
return bfin_read_TIMER0_COUNTER();
}
--- a/arch/c6x/kernel/time.c
+++ b/arch/c6x/kernel/time.c
@@ -26,7 +26,7 @@
static u32 sched_clock_multiplier;
#define SCHED_CLOCK_SHIFT 16
-static cycle_t tsc_read(struct clocksource *cs)
+static u64 tsc_read(struct clocksource *cs)
{
return get_cycles();
}
--- a/arch/hexagon/kernel/time.c
+++ b/arch/hexagon/kernel/time.c
@@ -72,9 +72,9 @@ struct adsp_hw_timer_struct {
/* Look for "TCX0" for related constants. */
static __iomem struct adsp_hw_timer_struct *rtos_timer;
-static cycle_t timer_get_cycles(struct clocksource *cs)
+static u64 timer_get_cycles(struct clocksource *cs)
{
- return (cycle_t) __vmgettime();
+ return (u64) __vmgettime();
}
static struct clocksource hexagon_clocksource = {
--- a/arch/ia64/kernel/cyclone.c
+++ b/arch/ia64/kernel/cyclone.c
@@ -21,9 +21,9 @@ void __init cyclone_setup(void)
static void __iomem *cyclone_mc;
-static cycle_t read_cyclone(struct clocksource *cs)
+static u64 read_cyclone(struct clocksource *cs)
{
- return (cycle_t)readq((void __iomem *)cyclone_mc);
+ return (u64)readq((void __iomem *)cyclone_mc);
}
static struct clocksource clocksource_cyclone = {
--- a/arch/ia64/kernel/fsyscall_gtod_data.h
+++ b/arch/ia64/kernel/fsyscall_gtod_data.h
@@ -9,15 +9,15 @@ struct fsyscall_gtod_data_t {
seqcount_t seq;
struct timespec wall_time;
struct timespec monotonic_time;
- cycle_t clk_mask;
+ u64 clk_mask;
u32 clk_mult;
u32 clk_shift;
void *clk_fsys_mmio;
- cycle_t clk_cycle_last;
+ u64 clk_cycle_last;
} ____cacheline_aligned;
struct itc_jitter_data_t {
int itc_jitter;
- cycle_t itc_lastcycle;
+ u64 itc_lastcycle;
} ____cacheline_aligned;
--- a/arch/ia64/kernel/time.c
+++ b/arch/ia64/kernel/time.c
@@ -31,7 +31,7 @@
#include "fsyscall_gtod_data.h"
-static cycle_t itc_get_cycles(struct clocksource *cs);
+static u64 itc_get_cycles(struct clocksource *cs);
struct fsyscall_gtod_data_t fsyscall_gtod_data;
@@ -323,7 +323,7 @@ void ia64_init_itm(void)
}
}
-static cycle_t itc_get_cycles(struct clocksource *cs)
+static u64 itc_get_cycles(struct clocksource *cs)
{
unsigned long lcycle, now, ret;
@@ -397,7 +397,7 @@ void update_vsyscall_tz(void)
}
void update_vsyscall_old(struct timespec *wall, struct timespec *wtm,
- struct clocksource *c, u32 mult, cycle_t cycle_last)
+ struct clocksource *c, u32 mult, u64 cycle_last)
{
write_seqcount_begin(&fsyscall_gtod_data.seq);
--- a/arch/ia64/sn/kernel/sn2/timer.c
+++ b/arch/ia64/sn/kernel/sn2/timer.c
@@ -22,9 +22,9 @@
extern unsigned long sn_rtc_cycles_per_second;
-static cycle_t read_sn2(struct clocksource *cs)
+static u64 read_sn2(struct clocksource *cs)
{
- return (cycle_t)readq(RTC_COUNTER_ADDR);
+ return (u64)readq(RTC_COUNTER_ADDR);
}
static struct clocksource clocksource_sn2 = {
--- a/arch/m68k/68000/timers.c
+++ b/arch/m68k/68000/timers.c
@@ -76,7 +76,7 @@ static struct irqaction m68328_timer_irq
/***************************************************************************/
-static cycle_t m68328_read_clk(struct clocksource *cs)
+static u64 m68328_read_clk(struct clocksource *cs)
{
unsigned long flags;
u32 cycles;
--- a/arch/m68k/coldfire/dma_timer.c
+++ b/arch/m68k/coldfire/dma_timer.c
@@ -34,7 +34,7 @@
#define DMA_DTMR_CLK_DIV_16 (2 << 1)
#define DMA_DTMR_ENABLE (1 << 0)
-static cycle_t cf_dt_get_cycles(struct clocksource *cs)
+static u64 cf_dt_get_cycles(struct clocksource *cs)
{
return __raw_readl(DTCN0);
}
--- a/arch/m68k/coldfire/pit.c
+++ b/arch/m68k/coldfire/pit.c
@@ -118,7 +118,7 @@ static struct irqaction pit_irq = {
/***************************************************************************/
-static cycle_t pit_read_clk(struct clocksource *cs)
+static u64 pit_read_clk(struct clocksource *cs)
{
unsigned long flags;
u32 cycles;
--- a/arch/m68k/coldfire/sltimers.c
+++ b/arch/m68k/coldfire/sltimers.c
@@ -97,7 +97,7 @@ static struct irqaction mcfslt_timer_irq
.handler = mcfslt_tick,
};
-static cycle_t mcfslt_read_clk(struct clocksource *cs)
+static u64 mcfslt_read_clk(struct clocksource *cs)
{
unsigned long flags;
u32 cycles, scnt;
--- a/arch/m68k/coldfire/timers.c
+++ b/arch/m68k/coldfire/timers.c
@@ -89,7 +89,7 @@ static struct irqaction mcftmr_timer_irq
/***************************************************************************/
-static cycle_t mcftmr_read_clk(struct clocksource *cs)
+static u64 mcftmr_read_clk(struct clocksource *cs)
{
unsigned long flags;
u32 cycles;
--- a/arch/microblaze/kernel/timer.c
+++ b/arch/microblaze/kernel/timer.c
@@ -190,17 +190,17 @@ static u64 xilinx_clock_read(void)
return read_fn(timer_baseaddr + TCR1);
}
-static cycle_t xilinx_read(struct clocksource *cs)
+static u64 xilinx_read(struct clocksource *cs)
{
/* reading actual value of timer 1 */
- return (cycle_t)xilinx_clock_read();
+ return (u64)xilinx_clock_read();
}
static struct timecounter xilinx_tc = {
.cc = NULL,
};
-static cycle_t xilinx_cc_read(const struct cyclecounter *cc)
+static u64 xilinx_cc_read(const struct cyclecounter *cc)
{
return xilinx_read(NULL);
}
--- a/arch/mips/alchemy/common/time.c
+++ b/arch/mips/alchemy/common/time.c
@@ -44,7 +44,7 @@
/* 32kHz clock enabled and detected */
#define CNTR_OK (SYS_CNTRL_E0 | SYS_CNTRL_32S)
-static cycle_t au1x_counter1_read(struct clocksource *cs)
+static u64 au1x_counter1_read(struct clocksource *cs)
{
return alchemy_rdsys(AU1000_SYS_RTCREAD);
}
--- a/arch/mips/cavium-octeon/csrc-octeon.c
+++ b/arch/mips/cavium-octeon/csrc-octeon.c
@@ -98,7 +98,7 @@ void octeon_init_cvmcount(void)
local_irq_restore(flags);
}
-static cycle_t octeon_cvmcount_read(struct clocksource *cs)
+static u64 octeon_cvmcount_read(struct clocksource *cs)
{
return read_c0_cvmcount();
}
--- a/arch/mips/jz4740/time.c
+++ b/arch/mips/jz4740/time.c
@@ -34,7 +34,7 @@
static uint16_t jz4740_jiffies_per_tick;
-static cycle_t jz4740_clocksource_read(struct clocksource *cs)
+static u64 jz4740_clocksource_read(struct clocksource *cs)
{
return jz4740_timer_get_count(TIMER_CLOCKSOURCE);
}
--- a/arch/mips/kernel/cevt-txx9.c
+++ b/arch/mips/kernel/cevt-txx9.c
@@ -27,7 +27,7 @@ struct txx9_clocksource {
struct txx9_tmr_reg __iomem *tmrptr;
};
-static cycle_t txx9_cs_read(struct clocksource *cs)
+static u64 txx9_cs_read(struct clocksource *cs)
{
struct txx9_clocksource *txx9_cs =
container_of(cs, struct txx9_clocksource, cs);
--- a/arch/mips/kernel/csrc-bcm1480.c
+++ b/arch/mips/kernel/csrc-bcm1480.c
@@ -25,9 +25,9 @@
#include <asm/sibyte/sb1250.h>
-static cycle_t bcm1480_hpt_read(struct clocksource *cs)
+static u64 bcm1480_hpt_read(struct clocksource *cs)
{
- return (cycle_t) __raw_readq(IOADDR(A_SCD_ZBBUS_CYCLE_COUNT));
+ return (u64) __raw_readq(IOADDR(A_SCD_ZBBUS_CYCLE_COUNT));
}
struct clocksource bcm1480_clocksource = {
--- a/arch/mips/kernel/csrc-ioasic.c
+++ b/arch/mips/kernel/csrc-ioasic.c
@@ -22,7 +22,7 @@
#include <asm/dec/ioasic.h>
#include <asm/dec/ioasic_addrs.h>
-static cycle_t dec_ioasic_hpt_read(struct clocksource *cs)
+static u64 dec_ioasic_hpt_read(struct clocksource *cs)
{
return ioasic_read(IO_REG_FCTR);
}
--- a/arch/mips/kernel/csrc-r4k.c
+++ b/arch/mips/kernel/csrc-r4k.c
@@ -11,7 +11,7 @@
#include <asm/time.h>
-static cycle_t c0_hpt_read(struct clocksource *cs)
+static u64 c0_hpt_read(struct clocksource *cs)
{
return read_c0_count();
}
--- a/arch/mips/kernel/csrc-sb1250.c
+++ b/arch/mips/kernel/csrc-sb1250.c
@@ -30,7 +30,7 @@
* The HPT is free running from SB1250_HPT_VALUE down to 0 then starts over
* again.
*/
-static inline cycle_t sb1250_hpt_get_cycles(void)
+static inline u64 sb1250_hpt_get_cycles(void)
{
unsigned int count;
void __iomem *addr;
@@ -41,7 +41,7 @@ static inline cycle_t sb1250_hpt_get_cyc
return SB1250_HPT_VALUE - count;
}
-static cycle_t sb1250_hpt_read(struct clocksource *cs)
+static u64 sb1250_hpt_read(struct clocksource *cs)
{
return sb1250_hpt_get_cycles();
}
--- a/arch/mips/loongson32/common/time.c
+++ b/arch/mips/loongson32/common/time.c
@@ -63,7 +63,7 @@ void __init ls1x_pwmtimer_init(void)
ls1x_pwmtimer_restart();
}
-static cycle_t ls1x_clocksource_read(struct clocksource *cs)
+static u64 ls1x_clocksource_read(struct clocksource *cs)
{
unsigned long flags;
int count;
@@ -107,7 +107,7 @@ static cycle_t ls1x_clocksource_read(str
raw_spin_unlock_irqrestore(&ls1x_timer_lock, flags);
- return (cycle_t) (jifs * ls1x_jiffies_per_tick) + count;
+ return (u64) (jifs * ls1x_jiffies_per_tick) + count;
}
static struct clocksource ls1x_clocksource = {
--- a/arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c
+++ b/arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c
@@ -144,7 +144,7 @@ void __init setup_mfgpt0_timer(void)
* to just read by itself. So use jiffies to emulate a free
* running counter:
*/
-static cycle_t mfgpt_read(struct clocksource *cs)
+static u64 mfgpt_read(struct clocksource *cs)
{
unsigned long flags;
int count;
@@ -188,7 +188,7 @@ static cycle_t mfgpt_read(struct clockso
raw_spin_unlock_irqrestore(&mfgpt_lock, flags);
- return (cycle_t) (jifs * COMPARE) + count;
+ return (u64) (jifs * COMPARE) + count;
}
static struct clocksource clocksource_mfgpt = {
--- a/arch/mips/loongson64/loongson-3/hpet.c
+++ b/arch/mips/loongson64/loongson-3/hpet.c
@@ -248,9 +248,9 @@ void __init setup_hpet_timer(void)
pr_info("hpet clock event device register\n");
}
-static cycle_t hpet_read_counter(struct clocksource *cs)
+static u64 hpet_read_counter(struct clocksource *cs)
{
- return (cycle_t)hpet_read(HPET_COUNTER);
+ return (u64)hpet_read(HPET_COUNTER);
}
static void hpet_suspend(struct clocksource *cs)
--- a/arch/mips/mti-malta/malta-time.c
+++ b/arch/mips/mti-malta/malta-time.c
@@ -75,7 +75,7 @@ static void __init estimate_frequencies(
unsigned int count, start;
unsigned char secs1, secs2, ctrl;
int secs;
- cycle_t giccount = 0, gicstart = 0;
+ u64 giccount = 0, gicstart = 0;
#if defined(CONFIG_KVM_GUEST) && CONFIG_KVM_GUEST_TIMER_FREQ
mips_hpt_frequency = CONFIG_KVM_GUEST_TIMER_FREQ * 1000000;
--- a/arch/mips/netlogic/common/time.c
+++ b/arch/mips/netlogic/common/time.c
@@ -59,14 +59,14 @@ unsigned int get_c0_compare_int(void)
return IRQ_TIMER;
}
-static cycle_t nlm_get_pic_timer(struct clocksource *cs)
+static u64 nlm_get_pic_timer(struct clocksource *cs)
{
uint64_t picbase = nlm_get_node(0)->picbase;
return ~nlm_pic_read_timer(picbase, PIC_CLOCK_TIMER);
}
-static cycle_t nlm_get_pic_timer32(struct clocksource *cs)
+static u64 nlm_get_pic_timer32(struct clocksource *cs)
{
uint64_t picbase = nlm_get_node(0)->picbase;
--- a/arch/mips/sgi-ip27/ip27-timer.c
+++ b/arch/mips/sgi-ip27/ip27-timer.c
@@ -140,7 +140,7 @@ static void __init hub_rt_clock_event_gl
setup_irq(irq, &hub_rt_irqaction);
}
-static cycle_t hub_rt_read(struct clocksource *cs)
+static u64 hub_rt_read(struct clocksource *cs)
{
return REMOTE_HUB_L(cputonasid(0), PI_RT_COUNT);
}
--- a/arch/mn10300/kernel/csrc-mn10300.c
+++ b/arch/mn10300/kernel/csrc-mn10300.c
@@ -13,7 +13,7 @@
#include <asm/timex.h>
#include "internal.h"
-static cycle_t mn10300_read(struct clocksource *cs)
+static u64 mn10300_read(struct clocksource *cs)
{
return read_timestamp_counter();
}
--- a/arch/nios2/kernel/time.c
+++ b/arch/nios2/kernel/time.c
@@ -81,7 +81,7 @@ static inline unsigned long read_timersn
return count;
}
-static cycle_t nios2_timer_read(struct clocksource *cs)
+static u64 nios2_timer_read(struct clocksource *cs)
{
struct nios2_clocksource *nios2_cs = to_nios2_clksource(cs);
unsigned long flags;
--- a/arch/openrisc/kernel/time.c
+++ b/arch/openrisc/kernel/time.c
@@ -117,9 +117,9 @@ static __init void openrisc_clockevent_i
* is 32 bits wide and runs at the CPU clock frequency.
*/
-static cycle_t openrisc_timer_read(struct clocksource *cs)
+static u64 openrisc_timer_read(struct clocksource *cs)
{
- return (cycle_t) mfspr(SPR_TTCR);
+ return (u64) mfspr(SPR_TTCR);
}
static struct clocksource openrisc_timer = {
--- a/arch/parisc/kernel/time.c
+++ b/arch/parisc/kernel/time.c
@@ -191,7 +191,7 @@ EXPORT_SYMBOL(profile_pc);
/* clock source code */
-static cycle_t notrace read_cr16(struct clocksource *cs)
+static u64 notrace read_cr16(struct clocksource *cs)
{
return get_cycles();
}
--- a/arch/powerpc/kernel/time.c
+++ b/arch/powerpc/kernel/time.c
@@ -80,7 +80,7 @@
#include <linux/clockchips.h>
#include <linux/timekeeper_internal.h>
-static cycle_t rtc_read(struct clocksource *);
+static u64 rtc_read(struct clocksource *);
static struct clocksource clocksource_rtc = {
.name = "rtc",
.rating = 400,
@@ -89,7 +89,7 @@ static struct clocksource clocksource_rt
.read = rtc_read,
};
-static cycle_t timebase_read(struct clocksource *);
+static u64 timebase_read(struct clocksource *);
static struct clocksource clocksource_timebase = {
.name = "timebase",
.rating = 400,
@@ -802,18 +802,18 @@ void read_persistent_clock(struct timesp
}
/* clocksource code */
-static cycle_t rtc_read(struct clocksource *cs)
+static u64 rtc_read(struct clocksource *cs)
{
- return (cycle_t)get_rtc();
+ return (u64)get_rtc();
}
-static cycle_t timebase_read(struct clocksource *cs)
+static u64 timebase_read(struct clocksource *cs)
{
- return (cycle_t)get_tb();
+ return (u64)get_tb();
}
void update_vsyscall_old(struct timespec *wall_time, struct timespec *wtm,
- struct clocksource *clock, u32 mult, cycle_t cycle_last)
+ struct clocksource *clock, u32 mult, u64 cycle_last)
{
u64 new_tb_to_xs, new_stamp_xsec;
u32 frac_sec;
--- a/arch/s390/kernel/time.c
+++ b/arch/s390/kernel/time.c
@@ -213,7 +213,7 @@ void read_boot_clock64(struct timespec64
tod_to_timeval(clock - TOD_UNIX_EPOCH, ts);
}
-static cycle_t read_tod_clock(struct clocksource *cs)
+static u64 read_tod_clock(struct clocksource *cs)
{
return get_tod_clock();
}
--- a/arch/sparc/kernel/time_32.c
+++ b/arch/sparc/kernel/time_32.c
@@ -148,7 +148,7 @@ static unsigned int sbus_cycles_offset(v
return offset;
}
-static cycle_t timer_cs_read(struct clocksource *cs)
+static u64 timer_cs_read(struct clocksource *cs)
{
unsigned int seq, offset;
u64 cycles;
--- a/arch/sparc/kernel/time_64.c
+++ b/arch/sparc/kernel/time_64.c
@@ -770,7 +770,7 @@ void udelay(unsigned long usecs)
}
EXPORT_SYMBOL(udelay);
-static cycle_t clocksource_tick_read(struct clocksource *cs)
+static u64 clocksource_tick_read(struct clocksource *cs)
{
return tick_ops->get_tick();
}
--- a/arch/um/kernel/time.c
+++ b/arch/um/kernel/time.c
@@ -83,7 +83,7 @@ static irqreturn_t um_timer(int irq, voi
return IRQ_HANDLED;
}
-static cycle_t timer_read(struct clocksource *cs)
+static u64 timer_read(struct clocksource *cs)
{
return os_nsecs() / TIMER_MULTIPLIER;
}
--- a/arch/unicore32/kernel/time.c
+++ b/arch/unicore32/kernel/time.c
@@ -62,7 +62,7 @@ static struct clock_event_device ckevt_p
.set_state_oneshot = puv3_osmr0_shutdown,
};
-static cycle_t puv3_read_oscr(struct clocksource *cs)
+static u64 puv3_read_oscr(struct clocksource *cs)
{
return readl(OST_OSCR);
}
--- a/arch/x86/entry/vdso/vclock_gettime.c
+++ b/arch/x86/entry/vdso/vclock_gettime.c
@@ -92,10 +92,10 @@ static notrace const struct pvclock_vsys
return (const struct pvclock_vsyscall_time_info *)&pvclock_page;
}
-static notrace cycle_t vread_pvclock(int *mode)
+static notrace u64 vread_pvclock(int *mode)
{
const struct pvclock_vcpu_time_info *pvti = &get_pvti0()->pvti;
- cycle_t ret;
+ u64 ret;
u64 last;
u32 version;
@@ -142,9 +142,9 @@ static notrace cycle_t vread_pvclock(int
}
#endif
-notrace static cycle_t vread_tsc(void)
+notrace static u64 vread_tsc(void)
{
- cycle_t ret = (cycle_t)rdtsc_ordered();
+ u64 ret = (u64)rdtsc_ordered();
u64 last = gtod->cycle_last;
if (likely(ret >= last))
--- a/arch/x86/include/asm/kvm_host.h
+++ b/arch/x86/include/asm/kvm_host.h
@@ -758,7 +758,7 @@ struct kvm_arch {
spinlock_t pvclock_gtod_sync_lock;
bool use_master_clock;
u64 master_kernel_ns;
- cycle_t master_cycle_now;
+ u64 master_cycle_now;
struct delayed_work kvmclock_update_work;
struct delayed_work kvmclock_sync_work;
--- a/arch/x86/include/asm/pvclock.h
+++ b/arch/x86/include/asm/pvclock.h
@@ -14,7 +14,7 @@ static inline struct pvclock_vsyscall_ti
#endif
/* some helper functions for xen and kvm pv clock sources */
-cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src);
+u64 pvclock_clocksource_read(struct pvclock_vcpu_time_info *src);
u8 pvclock_read_flags(struct pvclock_vcpu_time_info *src);
void pvclock_set_flags(u8 flags);
unsigned long pvclock_tsc_khz(struct pvclock_vcpu_time_info *src);
@@ -87,11 +87,11 @@ static inline u64 pvclock_scale_delta(u6
}
static __always_inline
-cycle_t __pvclock_read_cycles(const struct pvclock_vcpu_time_info *src,
+u64 __pvclock_read_cycles(const struct pvclock_vcpu_time_info *src,
u64 tsc)
{
u64 delta = tsc - src->tsc_timestamp;
- cycle_t offset = pvclock_scale_delta(delta, src->tsc_to_system_mul,
+ u64 offset = pvclock_scale_delta(delta, src->tsc_to_system_mul,
src->tsc_shift);
return src->system_time + offset;
}
--- a/arch/x86/include/asm/tsc.h
+++ b/arch/x86/include/asm/tsc.h
@@ -29,7 +29,7 @@ static inline cycles_t get_cycles(void)
return rdtsc();
}
-extern struct system_counterval_t convert_art_to_tsc(cycle_t art);
+extern struct system_counterval_t convert_art_to_tsc(u64 art);
extern void tsc_init(void);
extern void mark_tsc_unstable(char *reason);
--- a/arch/x86/include/asm/vgtod.h
+++ b/arch/x86/include/asm/vgtod.h
@@ -17,8 +17,8 @@ struct vsyscall_gtod_data {
unsigned seq;
int vclock_mode;
- cycle_t cycle_last;
- cycle_t mask;
+ u64 cycle_last;
+ u64 mask;
u32 mult;
u32 shift;
--- a/arch/x86/kernel/apb_timer.c
+++ b/arch/x86/kernel/apb_timer.c
@@ -247,7 +247,7 @@ void apbt_setup_secondary_clock(void) {}
static int apbt_clocksource_register(void)
{
u64 start, now;
- cycle_t t1;
+ u64 t1;
/* Start the counter, use timer 2 as source, timer 0/1 for event */
dw_apb_clocksource_start(clocksource_apbt);
@@ -355,7 +355,7 @@ unsigned long apbt_quick_calibrate(void)
{
int i, scale;
u64 old, new;
- cycle_t t1, t2;
+ u64 t1, t2;
unsigned long khz = 0;
u32 loop, shift;
--- a/arch/x86/kernel/cpu/mshyperv.c
+++ b/arch/x86/kernel/cpu/mshyperv.c
@@ -133,9 +133,9 @@ static uint32_t __init ms_hyperv_platfo
return 0;
}
-static cycle_t read_hv_clock(struct clocksource *arg)
+static u64 read_hv_clock(struct clocksource *arg)
{
- cycle_t current_tick;
+ u64 current_tick;
/*
* Read the partition counter to get the current tick count. This count
* is set to 0 when the partition is created and is incremented in
--- a/arch/x86/kernel/hpet.c
+++ b/arch/x86/kernel/hpet.c
@@ -791,7 +791,7 @@ static union hpet_lock hpet __cacheline_
{ .lock = __ARCH_SPIN_LOCK_UNLOCKED, },
};
-static cycle_t read_hpet(struct clocksource *cs)
+static u64 read_hpet(struct clocksource *cs)
{
unsigned long flags;
union hpet_lock old, new;
@@ -802,7 +802,7 @@ static cycle_t read_hpet(struct clocksou
* Read HPET directly if in NMI.
*/
if (in_nmi())
- return (cycle_t)hpet_readl(HPET_COUNTER);
+ return (u64)hpet_readl(HPET_COUNTER);
/*
* Read the current state of the lock and HPET value atomically.
@@ -821,7 +821,7 @@ static cycle_t read_hpet(struct clocksou
WRITE_ONCE(hpet.value, new.value);
arch_spin_unlock(&hpet.lock);
local_irq_restore(flags);
- return (cycle_t)new.value;
+ return (u64)new.value;
}
local_irq_restore(flags);
@@ -843,15 +843,15 @@ static cycle_t read_hpet(struct clocksou
new.lockval = READ_ONCE(hpet.lockval);
} while ((new.value == old.value) && arch_spin_is_locked(&new.lock));
- return (cycle_t)new.value;
+ return (u64)new.value;
}
#else
/*
* For UP or 32-bit.
*/
-static cycle_t read_hpet(struct clocksource *cs)
+static u64 read_hpet(struct clocksource *cs)
{
- return (cycle_t)hpet_readl(HPET_COUNTER);
+ return (u64)hpet_readl(HPET_COUNTER);
}
#endif
@@ -867,7 +867,7 @@ static struct clocksource clocksource_hp
static int hpet_clocksource_register(void)
{
u64 start, now;
- cycle_t t1;
+ u64 t1;
/* Start the counter */
hpet_restart_counter();
--- a/arch/x86/kernel/kvmclock.c
+++ b/arch/x86/kernel/kvmclock.c
@@ -32,7 +32,7 @@
static int kvmclock __ro_after_init = 1;
static int msr_kvm_system_time = MSR_KVM_SYSTEM_TIME;
static int msr_kvm_wall_clock = MSR_KVM_WALL_CLOCK;
-static cycle_t kvm_sched_clock_offset;
+static u64 kvm_sched_clock_offset;
static int parse_no_kvmclock(char *arg)
{
@@ -79,10 +79,10 @@ static int kvm_set_wallclock(const struc
return -1;
}
-static cycle_t kvm_clock_read(void)
+static u64 kvm_clock_read(void)
{
struct pvclock_vcpu_time_info *src;
- cycle_t ret;
+ u64 ret;
int cpu;
preempt_disable_notrace();
@@ -93,12 +93,12 @@ static cycle_t kvm_clock_read(void)
return ret;
}
-static cycle_t kvm_clock_get_cycles(struct clocksource *cs)
+static u64 kvm_clock_get_cycles(struct clocksource *cs)
{
return kvm_clock_read();
}
-static cycle_t kvm_sched_clock_read(void)
+static u64 kvm_sched_clock_read(void)
{
return kvm_clock_read() - kvm_sched_clock_offset;
}
--- a/arch/x86/kernel/pvclock.c
+++ b/arch/x86/kernel/pvclock.c
@@ -71,10 +71,10 @@ u8 pvclock_read_flags(struct pvclock_vcp
return flags & valid_flags;
}
-cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src)
+u64 pvclock_clocksource_read(struct pvclock_vcpu_time_info *src)
{
unsigned version;
- cycle_t ret;
+ u64 ret;
u64 last;
u8 flags;
--- a/arch/x86/kernel/tsc.c
+++ b/arch/x86/kernel/tsc.c
@@ -1080,9 +1080,9 @@ static struct clocksource clocksource_ts
* checking the result of read_tsc() - cycle_last for being negative.
* That works because CLOCKSOURCE_MASK(64) does not mask out any bit.
*/
-static cycle_t read_tsc(struct clocksource *cs)
+static u64 read_tsc(struct clocksource *cs)
{
- return (cycle_t)rdtsc_ordered();
+ return (u64)rdtsc_ordered();
}
/*
@@ -1170,7 +1170,7 @@ int unsynchronized_tsc(void)
/*
* Convert ART to TSC given numerator/denominator found in detect_art()
*/
-struct system_counterval_t convert_art_to_tsc(cycle_t art)
+struct system_counterval_t convert_art_to_tsc(u64 art)
{
u64 tmp, res, rem;
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -1128,8 +1128,8 @@ struct pvclock_gtod_data {
struct { /* extract of a clocksource struct */
int vclock_mode;
- cycle_t cycle_last;
- cycle_t mask;
+ u64 cycle_last;
+ u64 mask;
u32 mult;
u32 shift;
} clock;
@@ -1569,9 +1569,9 @@ static inline void adjust_tsc_offset_hos
#ifdef CONFIG_X86_64
-static cycle_t read_tsc(void)
+static u64 read_tsc(void)
{
- cycle_t ret = (cycle_t)rdtsc_ordered();
+ u64 ret = (u64)rdtsc_ordered();
u64 last = pvclock_gtod_data.clock.cycle_last;
if (likely(ret >= last))
@@ -1589,7 +1589,7 @@ static cycle_t read_tsc(void)
return last;
}
-static inline u64 vgettsc(cycle_t *cycle_now)
+static inline u64 vgettsc(u64 *cycle_now)
{
long v;
struct pvclock_gtod_data *gtod = &pvclock_gtod_data;
@@ -1600,7 +1600,7 @@ static inline u64 vgettsc(cycle_t *cycle
return v * gtod->clock.mult;
}
-static int do_monotonic_boot(s64 *t, cycle_t *cycle_now)
+static int do_monotonic_boot(s64 *t, u64 *cycle_now)
{
struct pvclock_gtod_data *gtod = &pvclock_gtod_data;
unsigned long seq;
@@ -1621,7 +1621,7 @@ static int do_monotonic_boot(s64 *t, cyc
}
/* returns true if host is using tsc clocksource */
-static bool kvm_get_time_and_clockread(s64 *kernel_ns, cycle_t *cycle_now)
+static bool kvm_get_time_and_clockread(s64 *kernel_ns, u64 *cycle_now)
{
/* checked again under seqlock below */
if (pvclock_gtod_data.clock.vclock_mode != VCLOCK_TSC)
--- a/arch/x86/lguest/boot.c
+++ b/arch/x86/lguest/boot.c
@@ -930,7 +930,7 @@ static unsigned long lguest_tsc_khz(void
* If we can't use the TSC, the kernel falls back to our lower-priority
* "lguest_clock", where we read the time value given to us by the Host.
*/
-static cycle_t lguest_clock_read(struct clocksource *cs)
+static u64 lguest_clock_read(struct clocksource *cs)
{
unsigned long sec, nsec;
--- a/arch/x86/platform/uv/uv_time.c
+++ b/arch/x86/platform/uv/uv_time.c
@@ -30,7 +30,7 @@
#define RTC_NAME "sgi_rtc"
-static cycle_t uv_read_rtc(struct clocksource *cs);
+static u64 uv_read_rtc(struct clocksource *cs);
static int uv_rtc_next_event(unsigned long, struct clock_event_device *);
static int uv_rtc_shutdown(struct clock_event_device *evt);
@@ -38,7 +38,7 @@ static struct clocksource clocksource_uv
.name = RTC_NAME,
.rating = 299,
.read = uv_read_rtc,
- .mask = (cycle_t)UVH_RTC_REAL_TIME_CLOCK_MASK,
+ .mask = (u64)UVH_RTC_REAL_TIME_CLOCK_MASK,
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
};
@@ -296,7 +296,7 @@ static int uv_rtc_unset_timer(int cpu, i
* cachelines of it's own page. This allows faster simultaneous reads
* from a given socket.
*/
-static cycle_t uv_read_rtc(struct clocksource *cs)
+static u64 uv_read_rtc(struct clocksource *cs)
{
unsigned long offset;
@@ -305,7 +305,7 @@ static cycle_t uv_read_rtc(struct clocks
else
offset = (uv_blade_processor_id() * L1_CACHE_BYTES) % PAGE_SIZE;
- return (cycle_t)uv_read_local_mmr(UVH_RTC | offset);
+ return (u64)uv_read_local_mmr(UVH_RTC | offset);
}
/*
--- a/arch/x86/xen/time.c
+++ b/arch/x86/xen/time.c
@@ -39,10 +39,10 @@ static unsigned long xen_tsc_khz(void)
return pvclock_tsc_khz(info);
}
-cycle_t xen_clocksource_read(void)
+u64 xen_clocksource_read(void)
{
struct pvclock_vcpu_time_info *src;
- cycle_t ret;
+ u64 ret;
preempt_disable_notrace();
src = &__this_cpu_read(xen_vcpu)->time;
@@ -51,7 +51,7 @@ cycle_t xen_clocksource_read(void)
return ret;
}
-static cycle_t xen_clocksource_get_cycles(struct clocksource *cs)
+static u64 xen_clocksource_get_cycles(struct clocksource *cs)
{
return xen_clocksource_read();
}
--- a/arch/x86/xen/xen-ops.h
+++ b/arch/x86/xen/xen-ops.h
@@ -67,7 +67,7 @@ void xen_init_irq_ops(void);
void xen_setup_timer(int cpu);
void xen_setup_runstate_info(int cpu);
void xen_teardown_timer(int cpu);
-cycle_t xen_clocksource_read(void);
+u64 xen_clocksource_read(void);
void xen_setup_cpu_clockevents(void);
void __init xen_init_time_ops(void);
void __init xen_hvm_init_time_ops(void);
--- a/arch/xtensa/kernel/time.c
+++ b/arch/xtensa/kernel/time.c
@@ -34,9 +34,9 @@
unsigned long ccount_freq; /* ccount Hz */
EXPORT_SYMBOL(ccount_freq);
-static cycle_t ccount_read(struct clocksource *cs)
+static u64 ccount_read(struct clocksource *cs)
{
- return (cycle_t)get_ccount();
+ return (u64)get_ccount();
}
static u64 notrace ccount_sched_clock_read(void)
--- a/drivers/char/hpet.c
+++ b/drivers/char/hpet.c
@@ -69,9 +69,9 @@ static u32 hpet_nhpet, hpet_max_freq = H
#ifdef CONFIG_IA64
static void __iomem *hpet_mctr;
-static cycle_t read_hpet(struct clocksource *cs)
+static u64 read_hpet(struct clocksource *cs)
{
- return (cycle_t)read_counter((void __iomem *)hpet_mctr);
+ return (u64)read_counter((void __iomem *)hpet_mctr);
}
static struct clocksource clocksource_hpet = {
--- a/drivers/clocksource/acpi_pm.c
+++ b/drivers/clocksource/acpi_pm.c
@@ -58,16 +58,16 @@ u32 acpi_pm_read_verified(void)
return v2;
}
-static cycle_t acpi_pm_read(struct clocksource *cs)
+static u64 acpi_pm_read(struct clocksource *cs)
{
- return (cycle_t)read_pmtmr();
+ return (u64)read_pmtmr();
}
static struct clocksource clocksource_acpi_pm = {
.name = "acpi_pm",
.rating = 200,
.read = acpi_pm_read,
- .mask = (cycle_t)ACPI_PM_MASK,
+ .mask = (u64)ACPI_PM_MASK,
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
};
@@ -81,9 +81,9 @@ static int __init acpi_pm_good_setup(cha
}
__setup("acpi_pm_good", acpi_pm_good_setup);
-static cycle_t acpi_pm_read_slow(struct clocksource *cs)
+static u64 acpi_pm_read_slow(struct clocksource *cs)
{
- return (cycle_t)acpi_pm_read_verified();
+ return (u64)acpi_pm_read_verified();
}
static inline void acpi_pm_need_workaround(void)
@@ -145,7 +145,7 @@ DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_SE
*/
static int verify_pmtmr_rate(void)
{
- cycle_t value1, value2;
+ u64 value1, value2;
unsigned long count, delta;
mach_prepare_counter();
@@ -175,7 +175,7 @@ static int verify_pmtmr_rate(void)
static int __init init_acpi_pm_clocksource(void)
{
- cycle_t value1, value2;
+ u64 value1, value2;
unsigned int i, j = 0;
if (!pmtmr_ioport)
--- a/drivers/clocksource/arm_arch_timer.c
+++ b/drivers/clocksource/arm_arch_timer.c
@@ -561,12 +561,12 @@ static u64 arch_counter_get_cntvct_mem(v
*/
u64 (*arch_timer_read_counter)(void) = arch_counter_get_cntvct;
-static cycle_t arch_counter_read(struct clocksource *cs)
+static u64 arch_counter_read(struct clocksource *cs)
{
return arch_timer_read_counter();
}
-static cycle_t arch_counter_read_cc(const struct cyclecounter *cc)
+static u64 arch_counter_read_cc(const struct cyclecounter *cc)
{
return arch_timer_read_counter();
}
--- a/drivers/clocksource/arm_global_timer.c
+++ b/drivers/clocksource/arm_global_timer.c
@@ -195,7 +195,7 @@ static int gt_dying_cpu(unsigned int cpu
return 0;
}
-static cycle_t gt_clocksource_read(struct clocksource *cs)
+static u64 gt_clocksource_read(struct clocksource *cs)
{
return gt_counter_read();
}
--- a/drivers/clocksource/cadence_ttc_timer.c
+++ b/drivers/clocksource/cadence_ttc_timer.c
@@ -158,11 +158,11 @@ static irqreturn_t ttc_clock_event_inter
*
* returns: Current timer counter register value
**/
-static cycle_t __ttc_clocksource_read(struct clocksource *cs)
+static u64 __ttc_clocksource_read(struct clocksource *cs)
{
struct ttc_timer *timer = &to_ttc_timer_clksrc(cs)->ttc;
- return (cycle_t)readl_relaxed(timer->base_addr +
+ return (u64)readl_relaxed(timer->base_addr +
TTC_COUNT_VAL_OFFSET);
}
--- a/drivers/clocksource/clksrc-dbx500-prcmu.c
+++ b/drivers/clocksource/clksrc-dbx500-prcmu.c
@@ -30,7 +30,7 @@
static void __iomem *clksrc_dbx500_timer_base;
-static cycle_t notrace clksrc_dbx500_prcmu_read(struct clocksource *cs)
+static u64 notrace clksrc_dbx500_prcmu_read(struct clocksource *cs)
{
void __iomem *base = clksrc_dbx500_timer_base;
u32 count, count2;
--- a/drivers/clocksource/dw_apb_timer.c
+++ b/drivers/clocksource/dw_apb_timer.c
@@ -348,7 +348,7 @@ void dw_apb_clocksource_start(struct dw_
dw_apb_clocksource_read(dw_cs);
}
-static cycle_t __apbt_read_clocksource(struct clocksource *cs)
+static u64 __apbt_read_clocksource(struct clocksource *cs)
{
u32 current_count;
struct dw_apb_clocksource *dw_cs =
@@ -357,7 +357,7 @@ static cycle_t __apbt_read_clocksource(s
current_count = apbt_readl_relaxed(&dw_cs->timer,
APBTMR_N_CURRENT_VALUE);
- return (cycle_t)~current_count;
+ return (u64)~current_count;
}
static void apbt_restart_clocksource(struct clocksource *cs)
@@ -416,7 +416,7 @@ void dw_apb_clocksource_register(struct
*
* @dw_cs: The clocksource to read.
*/
-cycle_t dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs)
+u64 dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs)
{
- return (cycle_t)~apbt_readl(&dw_cs->timer, APBTMR_N_CURRENT_VALUE);
+ return (u64)~apbt_readl(&dw_cs->timer, APBTMR_N_CURRENT_VALUE);
}
--- a/drivers/clocksource/em_sti.c
+++ b/drivers/clocksource/em_sti.c
@@ -110,9 +110,9 @@ static void em_sti_disable(struct em_sti
clk_disable_unprepare(p->clk);
}
-static cycle_t em_sti_count(struct em_sti_priv *p)
+static u64 em_sti_count(struct em_sti_priv *p)
{
- cycle_t ticks;
+ u64 ticks;
unsigned long flags;
/* the STI hardware buffers the 48-bit count, but to
@@ -121,14 +121,14 @@ static cycle_t em_sti_count(struct em_st
* Always read STI_COUNT_H before STI_COUNT_L.
*/
raw_spin_lock_irqsave(&p->lock, flags);
- ticks = (cycle_t)(em_sti_read(p, STI_COUNT_H) & 0xffff) << 32;
+ ticks = (u64)(em_sti_read(p, STI_COUNT_H) & 0xffff) << 32;
ticks |= em_sti_read(p, STI_COUNT_L);
raw_spin_unlock_irqrestore(&p->lock, flags);
return ticks;
}
-static cycle_t em_sti_set_next(struct em_sti_priv *p, cycle_t next)
+static u64 em_sti_set_next(struct em_sti_priv *p, u64 next)
{
unsigned long flags;
@@ -198,7 +198,7 @@ static struct em_sti_priv *cs_to_em_sti(
return container_of(cs, struct em_sti_priv, cs);
}
-static cycle_t em_sti_clocksource_read(struct clocksource *cs)
+static u64 em_sti_clocksource_read(struct clocksource *cs)
{
return em_sti_count(cs_to_em_sti(cs));
}
@@ -271,7 +271,7 @@ static int em_sti_clock_event_next(unsig
struct clock_event_device *ced)
{
struct em_sti_priv *p = ced_to_em_sti(ced);
- cycle_t next;
+ u64 next;
int safe;
next = em_sti_set_next(p, em_sti_count(p) + delta);
--- a/drivers/clocksource/exynos_mct.c
+++ b/drivers/clocksource/exynos_mct.c
@@ -183,7 +183,7 @@ static u64 exynos4_read_count_64(void)
hi2 = readl_relaxed(reg_base + EXYNOS4_MCT_G_CNT_U);
} while (hi != hi2);
- return ((cycle_t)hi << 32) | lo;
+ return ((u64)hi << 32) | lo;
}
/**
@@ -199,7 +199,7 @@ static u32 notrace exynos4_read_count_32
return readl_relaxed(reg_base + EXYNOS4_MCT_G_CNT_L);
}
-static cycle_t exynos4_frc_read(struct clocksource *cs)
+static u64 exynos4_frc_read(struct clocksource *cs)
{
return exynos4_read_count_32();
}
@@ -266,7 +266,7 @@ static void exynos4_mct_comp0_stop(void)
static void exynos4_mct_comp0_start(bool periodic, unsigned long cycles)
{
unsigned int tcon;
- cycle_t comp_cycle;
+ u64 comp_cycle;
tcon = readl_relaxed(reg_base + EXYNOS4_MCT_G_TCON);
--- a/drivers/clocksource/h8300_timer16.c
+++ b/drivers/clocksource/h8300_timer16.c
@@ -72,7 +72,7 @@ static inline struct timer16_priv *cs_to
return container_of(cs, struct timer16_priv, cs);
}
-static cycle_t timer16_clocksource_read(struct clocksource *cs)
+static u64 timer16_clocksource_read(struct clocksource *cs)
{
struct timer16_priv *p = cs_to_priv(cs);
unsigned long raw, value;
--- a/drivers/clocksource/h8300_tpu.c
+++ b/drivers/clocksource/h8300_tpu.c
@@ -64,7 +64,7 @@ static inline struct tpu_priv *cs_to_pri
return container_of(cs, struct tpu_priv, cs);
}
-static cycle_t tpu_clocksource_read(struct clocksource *cs)
+static u64 tpu_clocksource_read(struct clocksource *cs)
{
struct tpu_priv *p = cs_to_priv(cs);
unsigned long flags;
--- a/drivers/clocksource/i8253.c
+++ b/drivers/clocksource/i8253.c
@@ -25,7 +25,7 @@ EXPORT_SYMBOL(i8253_lock);
* to just read by itself. So use jiffies to emulate a free
* running counter:
*/
-static cycle_t i8253_read(struct clocksource *cs)
+static u64 i8253_read(struct clocksource *cs)
{
static int old_count;
static u32 old_jifs;
@@ -83,7 +83,7 @@ static cycle_t i8253_read(struct clockso
count = (PIT_LATCH - 1) - count;
- return (cycle_t)(jifs * PIT_LATCH) + count;
+ return (u64)(jifs * PIT_LATCH) + count;
}
static struct clocksource i8253_cs = {
--- a/drivers/clocksource/jcore-pit.c
+++ b/drivers/clocksource/jcore-pit.c
@@ -57,7 +57,7 @@ static notrace u64 jcore_sched_clock_rea
return seclo * NSEC_PER_SEC + nsec;
}
-static cycle_t jcore_clocksource_read(struct clocksource *cs)
+static u64 jcore_clocksource_read(struct clocksource *cs)
{
return jcore_sched_clock_read();
}
--- a/drivers/clocksource/metag_generic.c
+++ b/drivers/clocksource/metag_generic.c
@@ -56,7 +56,7 @@ static int metag_timer_set_next_event(un
return 0;
}
-static cycle_t metag_clocksource_read(struct clocksource *cs)
+static u64 metag_clocksource_read(struct clocksource *cs)
{
return __core_reg_get(TXTIMER);
}
--- a/drivers/clocksource/mips-gic-timer.c
+++ b/drivers/clocksource/mips-gic-timer.c
@@ -125,7 +125,7 @@ static int gic_clockevent_init(void)
return 0;
}
-static cycle_t gic_hpt_read(struct clocksource *cs)
+static u64 gic_hpt_read(struct clocksource *cs)
{
return gic_read_count();
}
--- a/drivers/clocksource/mmio.c
+++ b/drivers/clocksource/mmio.c
@@ -20,24 +20,24 @@ static inline struct clocksource_mmio *t
return container_of(c, struct clocksource_mmio, clksrc);
}
-cycle_t clocksource_mmio_readl_up(struct clocksource *c)
+u64 clocksource_mmio_readl_up(struct clocksource *c)
{
- return (cycle_t)readl_relaxed(to_mmio_clksrc(c)->reg);
+ return (u64)readl_relaxed(to_mmio_clksrc(c)->reg);
}
-cycle_t clocksource_mmio_readl_down(struct clocksource *c)
+u64 clocksource_mmio_readl_down(struct clocksource *c)
{
- return ~(cycle_t)readl_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
+ return ~(u64)readl_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
}
-cycle_t clocksource_mmio_readw_up(struct clocksource *c)
+u64 clocksource_mmio_readw_up(struct clocksource *c)
{
- return (cycle_t)readw_relaxed(to_mmio_clksrc(c)->reg);
+ return (u64)readw_relaxed(to_mmio_clksrc(c)->reg);
}
-cycle_t clocksource_mmio_readw_down(struct clocksource *c)
+u64 clocksource_mmio_readw_down(struct clocksource *c)
{
- return ~(cycle_t)readw_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
+ return ~(u64)readw_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
}
/**
@@ -51,7 +51,7 @@ cycle_t clocksource_mmio_readw_down(stru
*/
int __init clocksource_mmio_init(void __iomem *base, const char *name,
unsigned long hz, int rating, unsigned bits,
- cycle_t (*read)(struct clocksource *))
+ u64 (*read)(struct clocksource *))
{
struct clocksource_mmio *cs;
--- a/drivers/clocksource/mxs_timer.c
+++ b/drivers/clocksource/mxs_timer.c
@@ -97,7 +97,7 @@ static void timrot_irq_acknowledge(void)
HW_TIMROT_TIMCTRLn(0) + STMP_OFFSET_REG_CLR);
}
-static cycle_t timrotv1_get_cycles(struct clocksource *cs)
+static u64 timrotv1_get_cycles(struct clocksource *cs)
{
return ~((__raw_readl(mxs_timrot_base + HW_TIMROT_TIMCOUNTn(1))
& 0xffff0000) >> 16);
--- a/drivers/clocksource/qcom-timer.c
+++ b/drivers/clocksource/qcom-timer.c
@@ -89,7 +89,7 @@ static struct clock_event_device __percp
static void __iomem *source_base;
-static notrace cycle_t msm_read_timer_count(struct clocksource *cs)
+static notrace u64 msm_read_timer_count(struct clocksource *cs)
{
return readl_relaxed(source_base + TIMER_COUNT_VAL);
}
--- a/drivers/clocksource/samsung_pwm_timer.c
+++ b/drivers/clocksource/samsung_pwm_timer.c
@@ -307,7 +307,7 @@ static void samsung_clocksource_resume(s
samsung_time_start(pwm.source_id, true);
}
-static cycle_t notrace samsung_clocksource_read(struct clocksource *c)
+static u64 notrace samsung_clocksource_read(struct clocksource *c)
{
return ~readl_relaxed(pwm.source_reg);
}
--- a/drivers/clocksource/scx200_hrt.c
+++ b/drivers/clocksource/scx200_hrt.c
@@ -43,10 +43,10 @@ MODULE_PARM_DESC(ppm, "+-adjust to actua
/* The base timer frequency, * 27 if selected */
#define HRT_FREQ 1000000
-static cycle_t read_hrt(struct clocksource *cs)
+static u64 read_hrt(struct clocksource *cs)
{
/* Read the timer value */
- return (cycle_t) inl(scx200_cb_base + SCx200_TIMER_OFFSET);
+ return (u64) inl(scx200_cb_base + SCx200_TIMER_OFFSET);
}
static struct clocksource cs_hrt = {
--- a/drivers/clocksource/sh_cmt.c
+++ b/drivers/clocksource/sh_cmt.c
@@ -612,7 +612,7 @@ static struct sh_cmt_channel *cs_to_sh_c
return container_of(cs, struct sh_cmt_channel, cs);
}
-static cycle_t sh_cmt_clocksource_read(struct clocksource *cs)
+static u64 sh_cmt_clocksource_read(struct clocksource *cs)
{
struct sh_cmt_channel *ch = cs_to_sh_cmt(cs);
unsigned long flags, raw;
--- a/drivers/clocksource/sh_tmu.c
+++ b/drivers/clocksource/sh_tmu.c
@@ -255,7 +255,7 @@ static struct sh_tmu_channel *cs_to_sh_t
return container_of(cs, struct sh_tmu_channel, cs);
}
-static cycle_t sh_tmu_clocksource_read(struct clocksource *cs)
+static u64 sh_tmu_clocksource_read(struct clocksource *cs)
{
struct sh_tmu_channel *ch = cs_to_sh_tmu(cs);
--- a/drivers/clocksource/tcb_clksrc.c
+++ b/drivers/clocksource/tcb_clksrc.c
@@ -41,7 +41,7 @@
static void __iomem *tcaddr;
-static cycle_t tc_get_cycles(struct clocksource *cs)
+static u64 tc_get_cycles(struct clocksource *cs)
{
unsigned long flags;
u32 lower, upper;
@@ -56,7 +56,7 @@ static cycle_t tc_get_cycles(struct cloc
return (upper << 16) | lower;
}
-static cycle_t tc_get_cycles32(struct clocksource *cs)
+static u64 tc_get_cycles32(struct clocksource *cs)
{
return __raw_readl(tcaddr + ATMEL_TC_REG(0, CV));
}
--- a/drivers/clocksource/time-pistachio.c
+++ b/drivers/clocksource/time-pistachio.c
@@ -67,7 +67,7 @@ static inline void gpt_writel(void __iom
writel(value, base + 0x20 * gpt_id + offset);
}
-static cycle_t notrace
+static u64 notrace
pistachio_clocksource_read_cycles(struct clocksource *cs)
{
struct pistachio_clocksource *pcs = to_pistachio_clocksource(cs);
@@ -84,7 +84,7 @@ pistachio_clocksource_read_cycles(struct
counter = gpt_readl(pcs->base, TIMER_CURRENT_VALUE, 0);
raw_spin_unlock_irqrestore(&pcs->lock, flags);
- return (cycle_t)~counter;
+ return (u64)~counter;
}
static u64 notrace pistachio_read_sched_clock(void)
--- a/drivers/clocksource/timer-atlas7.c
+++ b/drivers/clocksource/timer-atlas7.c
@@ -85,7 +85,7 @@ static irqreturn_t sirfsoc_timer_interru
}
/* read 64-bit timer counter */
-static cycle_t sirfsoc_timer_read(struct clocksource *cs)
+static u64 sirfsoc_timer_read(struct clocksource *cs)
{
u64 cycles;
--- a/drivers/clocksource/timer-atmel-pit.c
+++ b/drivers/clocksource/timer-atmel-pit.c
@@ -73,7 +73,7 @@ static inline void pit_write(void __iome
* Clocksource: just a monotonic counter of MCK/16 cycles.
* We don't care whether or not PIT irqs are enabled.
*/
-static cycle_t read_pit_clk(struct clocksource *cs)
+static u64 read_pit_clk(struct clocksource *cs)
{
struct pit_data *data = clksrc_to_pit_data(cs);
unsigned long flags;
--- a/drivers/clocksource/timer-atmel-st.c
+++ b/drivers/clocksource/timer-atmel-st.c
@@ -92,7 +92,7 @@ static irqreturn_t at91rm9200_timer_inte
return IRQ_NONE;
}
-static cycle_t read_clk32k(struct clocksource *cs)
+static u64 read_clk32k(struct clocksource *cs)
{
return read_CRTR();
}
--- a/drivers/clocksource/timer-nps.c
+++ b/drivers/clocksource/timer-nps.c
@@ -48,11 +48,11 @@ static void *nps_msu_reg_low_addr[NPS_CL
static unsigned long nps_timer_rate;
-static cycle_t nps_clksrc_read(struct clocksource *clksrc)
+static u64 nps_clksrc_read(struct clocksource *clksrc)
{
int cluster = raw_smp_processor_id() >> NPS_CLUSTER_OFFSET;
- return (cycle_t)ioread32be(nps_msu_reg_low_addr[cluster]);
+ return (u64)ioread32be(nps_msu_reg_low_addr[cluster]);
}
static int __init nps_setup_clocksource(struct device_node *node,
--- a/drivers/clocksource/timer-prima2.c
+++ b/drivers/clocksource/timer-prima2.c
@@ -72,7 +72,7 @@ static irqreturn_t sirfsoc_timer_interru
}
/* read 64-bit timer counter */
-static cycle_t notrace sirfsoc_timer_read(struct clocksource *cs)
+static u64 notrace sirfsoc_timer_read(struct clocksource *cs)
{
u64 cycles;
--- a/drivers/clocksource/timer-sun5i.c
+++ b/drivers/clocksource/timer-sun5i.c
@@ -152,7 +152,7 @@ static irqreturn_t sun5i_timer_interrupt
return IRQ_HANDLED;
}
-static cycle_t sun5i_clksrc_read(struct clocksource *clksrc)
+static u64 sun5i_clksrc_read(struct clocksource *clksrc)
{
struct sun5i_timer_clksrc *cs = to_sun5i_timer_clksrc(clksrc);
--- a/drivers/clocksource/timer-ti-32k.c
+++ b/drivers/clocksource/timer-ti-32k.c
@@ -65,11 +65,11 @@ static inline struct ti_32k *to_ti_32k(s
return container_of(cs, struct ti_32k, cs);
}
-static cycle_t notrace ti_32k_read_cycles(struct clocksource *cs)
+static u64 notrace ti_32k_read_cycles(struct clocksource *cs)
{
struct ti_32k *ti = to_ti_32k(cs);
- return (cycle_t)readl_relaxed(ti->counter);
+ return (u64)readl_relaxed(ti->counter);
}
static struct ti_32k ti_32k_timer = {
--- a/drivers/clocksource/vt8500_timer.c
+++ b/drivers/clocksource/vt8500_timer.c
@@ -53,7 +53,7 @@
static void __iomem *regbase;
-static cycle_t vt8500_timer_read(struct clocksource *cs)
+static u64 vt8500_timer_read(struct clocksource *cs)
{
int loops = msecs_to_loops(10);
writel(3, regbase + TIMER_CTRL_VAL);
@@ -75,7 +75,7 @@ static int vt8500_timer_set_next_event(u
struct clock_event_device *evt)
{
int loops = msecs_to_loops(10);
- cycle_t alarm = clocksource.read(&clocksource) + cycles;
+ u64 alarm = clocksource.read(&clocksource) + cycles;
while ((readl(regbase + TIMER_AS_VAL) & TIMER_MATCH_W_ACTIVE)
&& --loops)
cpu_relax();
--- a/drivers/hv/hv.c
+++ b/drivers/hv/hv.c
@@ -135,9 +135,9 @@ u64 hv_do_hypercall(u64 control, void *i
EXPORT_SYMBOL_GPL(hv_do_hypercall);
#ifdef CONFIG_X86_64
-static cycle_t read_hv_clock_tsc(struct clocksource *arg)
+static u64 read_hv_clock_tsc(struct clocksource *arg)
{
- cycle_t current_tick;
+ u64 current_tick;
struct ms_hyperv_tsc_page *tsc_pg = hv_context.tsc_page;
if (tsc_pg->tsc_sequence != 0) {
@@ -146,7 +146,7 @@ static cycle_t read_hv_clock_tsc(struct
*/
while (1) {
- cycle_t tmp;
+ u64 tmp;
u32 sequence = tsc_pg->tsc_sequence;
u64 cur_tsc;
u64 scale = tsc_pg->tsc_scale;
@@ -350,7 +350,7 @@ int hv_post_message(union hv_connection_
static int hv_ce_set_next_event(unsigned long delta,
struct clock_event_device *evt)
{
- cycle_t current_tick;
+ u64 current_tick;
WARN_ON(!clockevent_state_oneshot(evt));
--- a/drivers/irqchip/irq-mips-gic.c
+++ b/drivers/irqchip/irq-mips-gic.c
@@ -152,12 +152,12 @@ static inline void gic_map_to_vpe(unsign
}
#ifdef CONFIG_CLKSRC_MIPS_GIC
-cycle_t gic_read_count(void)
+u64 gic_read_count(void)
{
unsigned int hi, hi2, lo;
if (mips_cm_is64)
- return (cycle_t)gic_read(GIC_REG(SHARED, GIC_SH_COUNTER));
+ return (u64)gic_read(GIC_REG(SHARED, GIC_SH_COUNTER));
do {
hi = gic_read32(GIC_REG(SHARED, GIC_SH_COUNTER_63_32));
@@ -165,7 +165,7 @@ cycle_t gic_read_count(void)
hi2 = gic_read32(GIC_REG(SHARED, GIC_SH_COUNTER_63_32));
} while (hi2 != hi);
- return (((cycle_t) hi) << 32) + lo;
+ return (((u64) hi) << 32) + lo;
}
unsigned int gic_get_count_width(void)
@@ -179,7 +179,7 @@ unsigned int gic_get_count_width(void)
return bits;
}
-void gic_write_compare(cycle_t cnt)
+void gic_write_compare(u64 cnt)
{
if (mips_cm_is64) {
gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE), cnt);
@@ -191,7 +191,7 @@ void gic_write_compare(cycle_t cnt)
}
}
-void gic_write_cpu_compare(cycle_t cnt, int cpu)
+void gic_write_cpu_compare(u64 cnt, int cpu)
{
unsigned long flags;
@@ -211,17 +211,17 @@ void gic_write_cpu_compare(cycle_t cnt,
local_irq_restore(flags);
}
-cycle_t gic_read_compare(void)
+u64 gic_read_compare(void)
{
unsigned int hi, lo;
if (mips_cm_is64)
- return (cycle_t)gic_read(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE));
+ return (u64)gic_read(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE));
hi = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_HI));
lo = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_LO));
- return (((cycle_t) hi) << 32) + lo;
+ return (((u64) hi) << 32) + lo;
}
void gic_start_count(void)
--- a/drivers/net/ethernet/amd/xgbe/xgbe-ptp.c
+++ b/drivers/net/ethernet/amd/xgbe/xgbe-ptp.c
@@ -122,7 +122,7 @@
#include "xgbe.h"
#include "xgbe-common.h"
-static cycle_t xgbe_cc_read(const struct cyclecounter *cc)
+static u64 xgbe_cc_read(const struct cyclecounter *cc)
{
struct xgbe_prv_data *pdata = container_of(cc,
struct xgbe_prv_data,
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
@@ -15219,7 +15219,7 @@ void bnx2x_set_rx_ts(struct bnx2x *bp, s
}
/* Read the PHC */
-static cycle_t bnx2x_cyclecounter_read(const struct cyclecounter *cc)
+static u64 bnx2x_cyclecounter_read(const struct cyclecounter *cc)
{
struct bnx2x *bp = container_of(cc, struct bnx2x, cyclecounter);
int port = BP_PORT(bp);
--- a/drivers/net/ethernet/freescale/fec_ptp.c
+++ b/drivers/net/ethernet/freescale/fec_ptp.c
@@ -230,7 +230,7 @@ static int fec_ptp_enable_pps(struct fec
* cyclecounter structure used to construct a ns counter from the
* arbitrary fixed point registers
*/
-static cycle_t fec_ptp_read(const struct cyclecounter *cc)
+static u64 fec_ptp_read(const struct cyclecounter *cc)
{
struct fec_enet_private *fep =
container_of(cc, struct fec_enet_private, cc);
--- a/drivers/net/ethernet/intel/e1000e/netdev.c
+++ b/drivers/net/ethernet/intel/e1000e/netdev.c
@@ -4305,24 +4305,24 @@ void e1000e_reinit_locked(struct e1000_a
/**
* e1000e_sanitize_systim - sanitize raw cycle counter reads
* @hw: pointer to the HW structure
- * @systim: cycle_t value read, sanitized and returned
+ * @systim: u64 timestamp value read, sanitized and returned
*
* Errata for 82574/82583 possible bad bits read from SYSTIMH/L:
* check to see that the time is incrementing at a reasonable
* rate and is a multiple of incvalue.
**/
-static cycle_t e1000e_sanitize_systim(struct e1000_hw *hw, cycle_t systim)
+static u64 e1000e_sanitize_systim(struct e1000_hw *hw, u64 systim)
{
u64 time_delta, rem, temp;
- cycle_t systim_next;
+ u64 systim_next;
u32 incvalue;
int i;
incvalue = er32(TIMINCA) & E1000_TIMINCA_INCVALUE_MASK;
for (i = 0; i < E1000_MAX_82574_SYSTIM_REREADS; i++) {
/* latch SYSTIMH on read of SYSTIML */
- systim_next = (cycle_t)er32(SYSTIML);
- systim_next |= (cycle_t)er32(SYSTIMH) << 32;
+ systim_next = (u64)er32(SYSTIML);
+ systim_next |= (u64)er32(SYSTIMH) << 32;
time_delta = systim_next - systim;
temp = time_delta;
@@ -4342,13 +4342,13 @@ static cycle_t e1000e_sanitize_systim(st
* e1000e_cyclecounter_read - read raw cycle counter (used by time counter)
* @cc: cyclecounter structure
**/
-static cycle_t e1000e_cyclecounter_read(const struct cyclecounter *cc)
+static u64 e1000e_cyclecounter_read(const struct cyclecounter *cc)
{
struct e1000_adapter *adapter = container_of(cc, struct e1000_adapter,
cc);
struct e1000_hw *hw = &adapter->hw;
u32 systimel, systimeh;
- cycle_t systim;
+ u64 systim;
/* SYSTIMH latching upon SYSTIML read does not work well.
* This means that if SYSTIML overflows after we read it but before
* we read SYSTIMH, the value of SYSTIMH has been incremented and we
@@ -4368,8 +4368,8 @@ static cycle_t e1000e_cyclecounter_read(
systimel = systimel_2;
}
}
- systim = (cycle_t)systimel;
- systim |= (cycle_t)systimeh << 32;
+ systim = (u64)systimel;
+ systim |= (u64)systimeh << 32;
if (adapter->flags2 & FLAG2_CHECK_SYSTIM_OVERFLOW)
systim = e1000e_sanitize_systim(hw, systim);
--- a/drivers/net/ethernet/intel/e1000e/ptp.c
+++ b/drivers/net/ethernet/intel/e1000e/ptp.c
@@ -127,8 +127,8 @@ static int e1000e_phc_get_syncdevicetime
unsigned long flags;
int i;
u32 tsync_ctrl;
- cycle_t dev_cycles;
- cycle_t sys_cycles;
+ u64 dev_cycles;
+ u64 sys_cycles;
tsync_ctrl = er32(TSYNCTXCTL);
tsync_ctrl |= E1000_TSYNCTXCTL_START_SYNC |
--- a/drivers/net/ethernet/intel/igb/igb_ptp.c
+++ b/drivers/net/ethernet/intel/igb/igb_ptp.c
@@ -77,7 +77,7 @@
static void igb_ptp_tx_hwtstamp(struct igb_adapter *adapter);
/* SYSTIM read access for the 82576 */
-static cycle_t igb_ptp_read_82576(const struct cyclecounter *cc)
+static u64 igb_ptp_read_82576(const struct cyclecounter *cc)
{
struct igb_adapter *igb = container_of(cc, struct igb_adapter, cc);
struct e1000_hw *hw = &igb->hw;
@@ -94,7 +94,7 @@ static cycle_t igb_ptp_read_82576(const
}
/* SYSTIM read access for the 82580 */
-static cycle_t igb_ptp_read_82580(const struct cyclecounter *cc)
+static u64 igb_ptp_read_82580(const struct cyclecounter *cc)
{
struct igb_adapter *igb = container_of(cc, struct igb_adapter, cc);
struct e1000_hw *hw = &igb->hw;
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c
@@ -245,7 +245,7 @@ static void ixgbe_ptp_setup_sdp_x540(str
* result of SYSTIME is 32bits of "billions of cycles" and 32 bits of
* "cycles", rather than seconds and nanoseconds.
*/
-static cycle_t ixgbe_ptp_read_X550(const struct cyclecounter *hw_cc)
+static u64 ixgbe_ptp_read_X550(const struct cyclecounter *hw_cc)
{
struct ixgbe_adapter *adapter =
container_of(hw_cc, struct ixgbe_adapter, hw_cc);
@@ -282,7 +282,7 @@ static cycle_t ixgbe_ptp_read_X550(const
* cyclecounter structure used to construct a ns counter from the
* arbitrary fixed point registers
*/
-static cycle_t ixgbe_ptp_read_82599(const struct cyclecounter *cc)
+static u64 ixgbe_ptp_read_82599(const struct cyclecounter *cc)
{
struct ixgbe_adapter *adapter =
container_of(cc, struct ixgbe_adapter, hw_cc);
--- a/drivers/net/ethernet/mellanox/mlx4/en_clock.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_clock.c
@@ -38,7 +38,7 @@
/* mlx4_en_read_clock - read raw cycle counter (to be used by time counter)
*/
-static cycle_t mlx4_en_read_clock(const struct cyclecounter *tc)
+static u64 mlx4_en_read_clock(const struct cyclecounter *tc)
{
struct mlx4_en_dev *mdev =
container_of(tc, struct mlx4_en_dev, cycles);
--- a/drivers/net/ethernet/mellanox/mlx4/main.c
+++ b/drivers/net/ethernet/mellanox/mlx4/main.c
@@ -1823,10 +1823,10 @@ static void unmap_bf_area(struct mlx4_de
io_mapping_free(mlx4_priv(dev)->bf_mapping);
}
-cycle_t mlx4_read_clock(struct mlx4_dev *dev)
+u64 mlx4_read_clock(struct mlx4_dev *dev)
{
u32 clockhi, clocklo, clockhi1;
- cycle_t cycles;
+ u64 cycles;
int i;
struct mlx4_priv *priv = mlx4_priv(dev);
--- a/drivers/net/ethernet/mellanox/mlx5/core/en_clock.c
+++ b/drivers/net/ethernet/mellanox/mlx5/core/en_clock.c
@@ -49,7 +49,7 @@ void mlx5e_fill_hwstamp(struct mlx5e_tst
hwts->hwtstamp = ns_to_ktime(nsec);
}
-static cycle_t mlx5e_read_internal_timer(const struct cyclecounter *cc)
+static u64 mlx5e_read_internal_timer(const struct cyclecounter *cc)
{
struct mlx5e_tstamp *tstamp = container_of(cc, struct mlx5e_tstamp,
cycles);
--- a/drivers/net/ethernet/mellanox/mlx5/core/main.c
+++ b/drivers/net/ethernet/mellanox/mlx5/core/main.c
@@ -522,7 +522,7 @@ int mlx5_core_disable_hca(struct mlx5_co
return mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out));
}
-cycle_t mlx5_read_internal_timer(struct mlx5_core_dev *dev)
+u64 mlx5_read_internal_timer(struct mlx5_core_dev *dev)
{
u32 timer_h, timer_h1, timer_l;
@@ -532,7 +532,7 @@ cycle_t mlx5_read_internal_timer(struct
if (timer_h != timer_h1) /* wrap around */
timer_l = ioread32be(&dev->iseg->internal_timer_l);
- return (cycle_t)timer_l | (cycle_t)timer_h1 << 32;
+ return (u64)timer_l | (u64)timer_h1 << 32;
}
static int mlx5_irq_set_affinity_hint(struct mlx5_core_dev *mdev, int i)
--- a/drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
+++ b/drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
@@ -93,7 +93,7 @@ bool mlx5_sriov_is_enabled(struct mlx5_c
int mlx5_core_enable_hca(struct mlx5_core_dev *dev, u16 func_id);
int mlx5_core_disable_hca(struct mlx5_core_dev *dev, u16 func_id);
int mlx5_wait_for_vf_pages(struct mlx5_core_dev *dev);
-cycle_t mlx5_read_internal_timer(struct mlx5_core_dev *dev);
+u64 mlx5_read_internal_timer(struct mlx5_core_dev *dev);
u32 mlx5_get_msix_vec(struct mlx5_core_dev *dev, int vecidx);
struct mlx5_eq *mlx5_eqn2eq(struct mlx5_core_dev *dev, int eqn);
void mlx5_cq_tasklet_cb(unsigned long data);
--- a/drivers/net/ethernet/ti/cpts.c
+++ b/drivers/net/ethernet/ti/cpts.c
@@ -101,7 +101,7 @@ static int cpts_fifo_read(struct cpts *c
return type == match ? 0 : -1;
}
-static cycle_t cpts_systim_read(const struct cyclecounter *cc)
+static u64 cpts_systim_read(const struct cyclecounter *cc)
{
u64 val = 0;
struct cpts_event *event;
--- a/include/kvm/arm_arch_timer.h
+++ b/include/kvm/arm_arch_timer.h
@@ -25,13 +25,13 @@
struct arch_timer_kvm {
/* Virtual offset */
- cycle_t cntvoff;
+ u64 cntvoff;
};
struct arch_timer_cpu {
/* Registers: control register, timer value */
u32 cntv_ctl; /* Saved/restored */
- cycle_t cntv_cval; /* Saved/restored */
+ u64 cntv_cval; /* Saved/restored */
/*
* Anything that is not used directly from assembly code goes
--- a/include/linux/clocksource.h
+++ b/include/linux/clocksource.h
@@ -75,8 +75,8 @@ struct module;
* structure.
*/
struct clocksource {
- cycle_t (*read)(struct clocksource *cs);
- cycle_t mask;
+ u64 (*read)(struct clocksource *cs);
+ u64 mask;
u32 mult;
u32 shift;
u64 max_idle_ns;
@@ -98,8 +98,8 @@ struct clocksource {
#ifdef CONFIG_CLOCKSOURCE_WATCHDOG
/* Watchdog related data, used by the framework */
struct list_head wd_list;
- cycle_t cs_last;
- cycle_t wd_last;
+ u64 cs_last;
+ u64 wd_last;
#endif
struct module *owner;
};
@@ -117,7 +117,7 @@ struct clocksource {
#define CLOCK_SOURCE_RESELECT 0x100
/* simplify initialization of mask field */
-#define CLOCKSOURCE_MASK(bits) (cycle_t)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
+#define CLOCKSOURCE_MASK(bits) (u64)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
static inline u32 clocksource_freq2mult(u32 freq, u32 shift_constant, u64 from)
{
@@ -173,7 +173,7 @@ static inline u32 clocksource_hz2mult(u3
*
* XXX - This could use some mult_lxl_ll() asm optimization
*/
-static inline s64 clocksource_cyc2ns(cycle_t cycles, u32 mult, u32 shift)
+static inline s64 clocksource_cyc2ns(u64 cycles, u32 mult, u32 shift)
{
return ((u64) cycles * mult) >> shift;
}
@@ -233,13 +233,13 @@ static inline void __clocksource_update_
extern int timekeeping_notify(struct clocksource *clock);
-extern cycle_t clocksource_mmio_readl_up(struct clocksource *);
-extern cycle_t clocksource_mmio_readl_down(struct clocksource *);
-extern cycle_t clocksource_mmio_readw_up(struct clocksource *);
-extern cycle_t clocksource_mmio_readw_down(struct clocksource *);
+extern u64 clocksource_mmio_readl_up(struct clocksource *);
+extern u64 clocksource_mmio_readl_down(struct clocksource *);
+extern u64 clocksource_mmio_readw_up(struct clocksource *);
+extern u64 clocksource_mmio_readw_down(struct clocksource *);
extern int clocksource_mmio_init(void __iomem *, const char *,
- unsigned long, int, unsigned, cycle_t (*)(struct clocksource *));
+ unsigned long, int, unsigned, u64 (*)(struct clocksource *));
extern int clocksource_i8253_init(void);
--- a/include/linux/dw_apb_timer.h
+++ b/include/linux/dw_apb_timer.h
@@ -50,6 +50,6 @@ dw_apb_clocksource_init(unsigned rating,
unsigned long freq);
void dw_apb_clocksource_register(struct dw_apb_clocksource *dw_cs);
void dw_apb_clocksource_start(struct dw_apb_clocksource *dw_cs);
-cycle_t dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs);
+u64 dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs);
#endif /* __DW_APB_TIMER_H__ */
--- a/include/linux/irqchip/mips-gic.h
+++ b/include/linux/irqchip/mips-gic.h
@@ -259,11 +259,11 @@ extern void gic_init(unsigned long gic_b
unsigned long gic_addrspace_size, unsigned int cpu_vec,
unsigned int irqbase);
extern void gic_clocksource_init(unsigned int);
-extern cycle_t gic_read_count(void);
+extern u64 gic_read_count(void);
extern unsigned int gic_get_count_width(void);
-extern cycle_t gic_read_compare(void);
-extern void gic_write_compare(cycle_t cnt);
-extern void gic_write_cpu_compare(cycle_t cnt, int cpu);
+extern u64 gic_read_compare(void);
+extern void gic_write_compare(u64 cnt);
+extern void gic_write_cpu_compare(u64 cnt, int cpu);
extern void gic_start_count(void);
extern void gic_stop_count(void);
extern int gic_get_c0_compare_int(void);
--- a/include/linux/mlx4/device.h
+++ b/include/linux/mlx4/device.h
@@ -1461,7 +1461,7 @@ int mlx4_get_roce_gid_from_slave(struct
int mlx4_FLOW_STEERING_IB_UC_QP_RANGE(struct mlx4_dev *dev, u32 min_range_qpn,
u32 max_range_qpn);
-cycle_t mlx4_read_clock(struct mlx4_dev *dev);
+u64 mlx4_read_clock(struct mlx4_dev *dev);
struct mlx4_active_ports {
DECLARE_BITMAP(ports, MLX4_MAX_PORTS);
--- a/include/linux/timecounter.h
+++ b/include/linux/timecounter.h
@@ -20,7 +20,7 @@
#include <linux/types.h>
/* simplify initialization of mask field */
-#define CYCLECOUNTER_MASK(bits) (cycle_t)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
+#define CYCLECOUNTER_MASK(bits) (u64)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
/**
* struct cyclecounter - hardware abstraction for a free running counter
@@ -37,8 +37,8 @@
* @shift: cycle to nanosecond divisor (power of two)
*/
struct cyclecounter {
- cycle_t (*read)(const struct cyclecounter *cc);
- cycle_t mask;
+ u64 (*read)(const struct cyclecounter *cc);
+ u64 mask;
u32 mult;
u32 shift;
};
@@ -63,7 +63,7 @@ struct cyclecounter {
*/
struct timecounter {
const struct cyclecounter *cc;
- cycle_t cycle_last;
+ u64 cycle_last;
u64 nsec;
u64 mask;
u64 frac;
@@ -77,7 +77,7 @@ struct timecounter {
* @frac: pointer to storage for the fractional nanoseconds.
*/
static inline u64 cyclecounter_cyc2ns(const struct cyclecounter *cc,
- cycle_t cycles, u64 mask, u64 *frac)
+ u64 cycles, u64 mask, u64 *frac)
{
u64 ns = (u64) cycles;
@@ -134,6 +134,6 @@ extern u64 timecounter_read(struct timec
* in the past.
*/
extern u64 timecounter_cyc2time(struct timecounter *tc,
- cycle_t cycle_tstamp);
+ u64 cycle_tstamp);
#endif
--- a/include/linux/timekeeper_internal.h
+++ b/include/linux/timekeeper_internal.h
@@ -29,9 +29,9 @@
*/
struct tk_read_base {
struct clocksource *clock;
- cycle_t (*read)(struct clocksource *cs);
- cycle_t mask;
- cycle_t cycle_last;
+ u64 (*read)(struct clocksource *cs);
+ u64 mask;
+ u64 cycle_last;
u32 mult;
u32 shift;
u64 xtime_nsec;
@@ -97,7 +97,7 @@ struct timekeeper {
struct timespec64 raw_time;
/* The following members are for timekeeping internal use */
- cycle_t cycle_interval;
+ u64 cycle_interval;
u64 xtime_interval;
s64 xtime_remainder;
u32 raw_interval;
@@ -136,7 +136,7 @@ extern void update_vsyscall_tz(void);
extern void update_vsyscall_old(struct timespec *ts, struct timespec *wtm,
struct clocksource *c, u32 mult,
- cycle_t cycle_last);
+ u64 cycle_last);
extern void update_vsyscall_tz(void);
#else
--- a/include/linux/timekeeping.h
+++ b/include/linux/timekeeping.h
@@ -292,7 +292,7 @@ extern void ktime_get_raw_and_real_ts64(
* @cs_was_changed_seq: The sequence number of clocksource change events
*/
struct system_time_snapshot {
- cycle_t cycles;
+ u64 cycles;
ktime_t real;
ktime_t raw;
unsigned int clock_was_set_seq;
@@ -320,7 +320,7 @@ struct system_device_crosststamp {
* timekeeping code to verify comparibility of two cycle values
*/
struct system_counterval_t {
- cycle_t cycles;
+ u64 cycles;
struct clocksource *cs;
};
--- a/include/linux/types.h
+++ b/include/linux/types.h
@@ -228,8 +228,5 @@ struct callback_head {
typedef void (*rcu_callback_t)(struct rcu_head *head);
typedef void (*call_rcu_func_t)(struct rcu_head *head, rcu_callback_t func);
-/* clocksource cycle base type */
-typedef u64 cycle_t;
-
#endif /* __ASSEMBLY__ */
#endif /* _LINUX_TYPES_H */
--- a/kernel/time/clocksource.c
+++ b/kernel/time/clocksource.c
@@ -169,7 +169,7 @@ void clocksource_mark_unstable(struct cl
static void clocksource_watchdog(unsigned long data)
{
struct clocksource *cs;
- cycle_t csnow, wdnow, cslast, wdlast, delta;
+ u64 csnow, wdnow, cslast, wdlast, delta;
int64_t wd_nsec, cs_nsec;
int next_cpu, reset_pending;
--- a/kernel/time/jiffies.c
+++ b/kernel/time/jiffies.c
@@ -59,9 +59,9 @@
#define JIFFIES_SHIFT 8
#endif
-static cycle_t jiffies_read(struct clocksource *cs)
+static u64 jiffies_read(struct clocksource *cs)
{
- return (cycle_t) jiffies;
+ return (u64) jiffies;
}
static struct clocksource clocksource_jiffies = {
--- a/kernel/time/timecounter.c
+++ b/kernel/time/timecounter.c
@@ -43,7 +43,7 @@ EXPORT_SYMBOL_GPL(timecounter_init);
*/
static u64 timecounter_read_delta(struct timecounter *tc)
{
- cycle_t cycle_now, cycle_delta;
+ u64 cycle_now, cycle_delta;
u64 ns_offset;
/* read cycle counter: */
@@ -80,7 +80,7 @@ EXPORT_SYMBOL_GPL(timecounter_read);
* time previous to the time stored in the cycle counter.
*/
static u64 cc_cyc2ns_backwards(const struct cyclecounter *cc,
- cycle_t cycles, u64 mask, u64 frac)
+ u64 cycles, u64 mask, u64 frac)
{
u64 ns = (u64) cycles;
@@ -90,7 +90,7 @@ static u64 cc_cyc2ns_backwards(const str
}
u64 timecounter_cyc2time(struct timecounter *tc,
- cycle_t cycle_tstamp)
+ u64 cycle_tstamp)
{
u64 delta = (cycle_tstamp - tc->cycle_last) & tc->cc->mask;
u64 nsec = tc->nsec, frac = tc->frac;
--- a/kernel/time/timekeeping.c
+++ b/kernel/time/timekeeping.c
@@ -119,10 +119,10 @@ static inline void tk_update_sleep_time(
#ifdef CONFIG_DEBUG_TIMEKEEPING
#define WARNING_FREQ (HZ*300) /* 5 minute rate-limiting */
-static void timekeeping_check_update(struct timekeeper *tk, cycle_t offset)
+static void timekeeping_check_update(struct timekeeper *tk, u64 offset)
{
- cycle_t max_cycles = tk->tkr_mono.clock->max_cycles;
+ u64 max_cycles = tk->tkr_mono.clock->max_cycles;
const char *name = tk->tkr_mono.clock->name;
if (offset > max_cycles) {
@@ -158,10 +158,10 @@ static void timekeeping_check_update(str
}
}
-static inline cycle_t timekeeping_get_delta(struct tk_read_base *tkr)
+static inline u64 timekeeping_get_delta(struct tk_read_base *tkr)
{
struct timekeeper *tk = &tk_core.timekeeper;
- cycle_t now, last, mask, max, delta;
+ u64 now, last, mask, max, delta;
unsigned int seq;
/*
@@ -199,12 +199,12 @@ static inline cycle_t timekeeping_get_de
return delta;
}
#else
-static inline void timekeeping_check_update(struct timekeeper *tk, cycle_t offset)
+static inline void timekeeping_check_update(struct timekeeper *tk, u64 offset)
{
}
-static inline cycle_t timekeeping_get_delta(struct tk_read_base *tkr)
+static inline u64 timekeeping_get_delta(struct tk_read_base *tkr)
{
- cycle_t cycle_now, delta;
+ u64 cycle_now, delta;
/* read clocksource */
cycle_now = tkr->read(tkr->clock);
@@ -229,7 +229,7 @@ static inline cycle_t timekeeping_get_de
*/
static void tk_setup_internals(struct timekeeper *tk, struct clocksource *clock)
{
- cycle_t interval;
+ u64 interval;
u64 tmp, ntpinterval;
struct clocksource *old_clock;
@@ -254,7 +254,7 @@ static void tk_setup_internals(struct ti
if (tmp == 0)
tmp = 1;
- interval = (cycle_t) tmp;
+ interval = (u64) tmp;
tk->cycle_interval = interval;
/* Go back from cycles -> shifted ns */
@@ -346,16 +346,16 @@ static inline u64 timekeeping_delta_to_n
static inline u64 timekeeping_get_ns(struct tk_read_base *tkr)
{
- cycle_t delta;
+ u64 delta;
delta = timekeeping_get_delta(tkr);
return timekeeping_delta_to_ns(tkr, delta);
}
static inline u64 timekeeping_cycles_to_ns(struct tk_read_base *tkr,
- cycle_t cycles)
+ u64 cycles)
{
- cycle_t delta;
+ u64 delta;
/* calculate the delta since the last update_wall_time */
delta = clocksource_delta(cycles, tkr->cycle_last, tkr->mask);
@@ -459,9 +459,9 @@ u64 ktime_get_raw_fast_ns(void)
EXPORT_SYMBOL_GPL(ktime_get_raw_fast_ns);
/* Suspend-time cycles value for halted fast timekeeper. */
-static cycle_t cycles_at_suspend;
+static u64 cycles_at_suspend;
-static cycle_t dummy_clock_read(struct clocksource *cs)
+static u64 dummy_clock_read(struct clocksource *cs)
{
return cycles_at_suspend;
}
@@ -655,7 +655,7 @@ static void timekeeping_update(struct ti
static void timekeeping_forward_now(struct timekeeper *tk)
{
struct clocksource *clock = tk->tkr_mono.clock;
- cycle_t cycle_now, delta;
+ u64 cycle_now, delta;
u64 nsec;
cycle_now = tk->tkr_mono.read(clock);
@@ -928,7 +928,7 @@ void ktime_get_snapshot(struct system_ti
ktime_t base_real;
u64 nsec_raw;
u64 nsec_real;
- cycle_t now;
+ u64 now;
WARN_ON_ONCE(timekeeping_suspended);
@@ -987,8 +987,8 @@ static int scale64_check_overflow(u64 mu
* interval is partial_history_cycles.
*/
static int adjust_historical_crosststamp(struct system_time_snapshot *history,
- cycle_t partial_history_cycles,
- cycle_t total_history_cycles,
+ u64 partial_history_cycles,
+ u64 total_history_cycles,
bool discontinuity,
struct system_device_crosststamp *ts)
{
@@ -1052,7 +1052,7 @@ static int adjust_historical_crosststamp
/*
* cycle_between - true if test occurs chronologically between before and after
*/
-static bool cycle_between(cycle_t before, cycle_t test, cycle_t after)
+static bool cycle_between(u64 before, u64 test, u64 after)
{
if (test > before && test < after)
return true;
@@ -1082,7 +1082,7 @@ int get_device_system_crosststamp(int (*
{
struct system_counterval_t system_counterval;
struct timekeeper *tk = &tk_core.timekeeper;
- cycle_t cycles, now, interval_start;
+ u64 cycles, now, interval_start;
unsigned int clock_was_set_seq = 0;
ktime_t base_real, base_raw;
u64 nsec_real, nsec_raw;
@@ -1143,7 +1143,7 @@ int get_device_system_crosststamp(int (*
* current interval
*/
if (do_interp) {
- cycle_t partial_history_cycles, total_history_cycles;
+ u64 partial_history_cycles, total_history_cycles;
bool discontinuity;
/*
@@ -1649,7 +1649,7 @@ void timekeeping_resume(void)
struct clocksource *clock = tk->tkr_mono.clock;
unsigned long flags;
struct timespec64 ts_new, ts_delta;
- cycle_t cycle_now;
+ u64 cycle_now;
sleeptime_injected = false;
read_persistent_clock64(&ts_new);
@@ -2015,11 +2015,11 @@ static inline unsigned int accumulate_ns
*
* Returns the unconsumed cycles.
*/
-static cycle_t logarithmic_accumulation(struct timekeeper *tk, cycle_t offset,
+static u64 logarithmic_accumulation(struct timekeeper *tk, u64 offset,
u32 shift,
unsigned int *clock_set)
{
- cycle_t interval = tk->cycle_interval << shift;
+ u64 interval = tk->cycle_interval << shift;
u64 raw_nsecs;
/* If the offset is smaller than a shifted interval, do nothing */
@@ -2060,7 +2060,7 @@ void update_wall_time(void)
{
struct timekeeper *real_tk = &tk_core.timekeeper;
struct timekeeper *tk = &shadow_timekeeper;
- cycle_t offset;
+ u64 offset;
int shift = 0, maxshift;
unsigned int clock_set = 0;
unsigned long flags;
--- a/kernel/time/timekeeping_internal.h
+++ b/kernel/time/timekeeping_internal.h
@@ -13,9 +13,9 @@ extern void tk_debug_account_sleep_time(
#endif
#ifdef CONFIG_CLOCKSOURCE_VALIDATE_LAST_CYCLE
-static inline cycle_t clocksource_delta(cycle_t now, cycle_t last, cycle_t mask)
+static inline u64 clocksource_delta(u64 now, u64 last, u64 mask)
{
- cycle_t ret = (now - last) & mask;
+ u64 ret = (now - last) & mask;
/*
* Prevent time going backwards by checking the MSB of mask in
@@ -24,7 +24,7 @@ static inline cycle_t clocksource_delta(
return ret & ~(mask >> 1) ? 0 : ret;
}
#else
-static inline cycle_t clocksource_delta(cycle_t now, cycle_t last, cycle_t mask)
+static inline u64 clocksource_delta(u64 now, u64 last, u64 mask)
{
return (now - last) & mask;
}
--- a/kernel/trace/ftrace.c
+++ b/kernel/trace/ftrace.c
@@ -2847,7 +2847,7 @@ static void ftrace_shutdown_sysctl(void)
}
}
-static cycle_t ftrace_update_time;
+static u64 ftrace_update_time;
unsigned long ftrace_update_tot_cnt;
static inline int ops_traces_mod(struct ftrace_ops *ops)
@@ -2894,7 +2894,7 @@ static int ftrace_update_code(struct mod
{
struct ftrace_page *pg;
struct dyn_ftrace *p;
- cycle_t start, stop;
+ u64 start, stop;
unsigned long update_cnt = 0;
unsigned long rec_flags = 0;
int i;
--- a/kernel/trace/trace.c
+++ b/kernel/trace/trace.c
@@ -234,7 +234,7 @@ static int __init set_tracepoint_printk(
}
__setup("tp_printk", set_tracepoint_printk);
-unsigned long long ns2usecs(cycle_t nsec)
+unsigned long long ns2usecs(u64 nsec)
{
nsec += 500;
do_div(nsec, 1000);
@@ -571,7 +571,7 @@ int trace_pid_write(struct trace_pid_lis
return read;
}
-static cycle_t buffer_ftrace_now(struct trace_buffer *buf, int cpu)
+static u64 buffer_ftrace_now(struct trace_buffer *buf, int cpu)
{
u64 ts;
@@ -585,7 +585,7 @@ static cycle_t buffer_ftrace_now(struct
return ts;
}
-cycle_t ftrace_now(int cpu)
+u64 ftrace_now(int cpu)
{
return buffer_ftrace_now(&global_trace.trace_buffer, cpu);
}
--- a/kernel/trace/trace.h
+++ b/kernel/trace/trace.h
@@ -157,7 +157,7 @@ struct trace_array_cpu {
unsigned long policy;
unsigned long rt_priority;
unsigned long skipped_entries;
- cycle_t preempt_timestamp;
+ u64 preempt_timestamp;
pid_t pid;
kuid_t uid;
char comm[TASK_COMM_LEN];
@@ -175,7 +175,7 @@ struct trace_buffer {
struct trace_array *tr;
struct ring_buffer *buffer;
struct trace_array_cpu __percpu *data;
- cycle_t time_start;
+ u64 time_start;
int cpu;
};
@@ -686,7 +686,7 @@ static inline void __trace_stack(struct
}
#endif /* CONFIG_STACKTRACE */
-extern cycle_t ftrace_now(int cpu);
+extern u64 ftrace_now(int cpu);
extern void trace_find_cmdline(int pid, char comm[]);
extern void trace_event_follow_fork(struct trace_array *tr, bool enable);
@@ -733,7 +733,7 @@ extern int trace_selftest_startup_branch
#endif /* CONFIG_FTRACE_STARTUP_TEST */
extern void *head_page(struct trace_array_cpu *data);
-extern unsigned long long ns2usecs(cycle_t nsec);
+extern unsigned long long ns2usecs(u64 nsec);
extern int
trace_vbprintk(unsigned long ip, const char *fmt, va_list args);
extern int
--- a/kernel/trace/trace_irqsoff.c
+++ b/kernel/trace/trace_irqsoff.c
@@ -286,7 +286,7 @@ static void irqsoff_print_header(struct
/*
* Should this new latency be reported/recorded?
*/
-static bool report_latency(struct trace_array *tr, cycle_t delta)
+static bool report_latency(struct trace_array *tr, u64 delta)
{
if (tracing_thresh) {
if (delta < tracing_thresh)
@@ -304,7 +304,7 @@ check_critical_timing(struct trace_array
unsigned long parent_ip,
int cpu)
{
- cycle_t T0, T1, delta;
+ u64 T0, T1, delta;
unsigned long flags;
int pc;
--- a/kernel/trace/trace_sched_wakeup.c
+++ b/kernel/trace/trace_sched_wakeup.c
@@ -346,7 +346,7 @@ static void wakeup_print_header(struct s
/*
* Should this new latency be reported/recorded?
*/
-static bool report_latency(struct trace_array *tr, cycle_t delta)
+static bool report_latency(struct trace_array *tr, u64 delta)
{
if (tracing_thresh) {
if (delta < tracing_thresh)
@@ -428,7 +428,7 @@ probe_wakeup_sched_switch(void *ignore,
struct task_struct *prev, struct task_struct *next)
{
struct trace_array_cpu *data;
- cycle_t T0, T1, delta;
+ u64 T0, T1, delta;
unsigned long flags;
long disabled;
int cpu;
--- a/sound/hda/hdac_stream.c
+++ b/sound/hda/hdac_stream.c
@@ -465,7 +465,7 @@ int snd_hdac_stream_set_params(struct hd
}
EXPORT_SYMBOL_GPL(snd_hdac_stream_set_params);
-static cycle_t azx_cc_read(const struct cyclecounter *cc)
+static u64 azx_cc_read(const struct cyclecounter *cc)
{
struct hdac_stream *azx_dev = container_of(cc, struct hdac_stream, cc);
@@ -473,7 +473,7 @@ static cycle_t azx_cc_read(const struct
}
static void azx_timecounter_init(struct hdac_stream *azx_dev,
- bool force, cycle_t last)
+ bool force, u64 last)
{
struct timecounter *tc = &azx_dev->tc;
struct cyclecounter *cc = &azx_dev->cc;
@@ -523,7 +523,7 @@ void snd_hdac_stream_timecounter_init(st
struct snd_pcm_runtime *runtime = azx_dev->substream->runtime;
struct hdac_stream *s;
bool inited = false;
- cycle_t cycle_last = 0;
+ u64 cycle_last = 0;
int i = 0;
list_for_each_entry(s, &bus->stream_list, list) {
--- a/virt/kvm/arm/arch_timer.c
+++ b/virt/kvm/arm/arch_timer.c
@@ -39,7 +39,7 @@ void kvm_timer_vcpu_put(struct kvm_vcpu
vcpu->arch.timer_cpu.active_cleared_last = false;
}
-static cycle_t kvm_phys_timer_read(void)
+static u64 kvm_phys_timer_read(void)
{
return timecounter->cc->read(timecounter->cc);
}
@@ -102,7 +102,7 @@ static void kvm_timer_inject_irq_work(st
static u64 kvm_timer_compute_delta(struct kvm_vcpu *vcpu)
{
- cycle_t cval, now;
+ u64 cval, now;
cval = vcpu->arch.timer_cpu.cntv_cval;
now = kvm_phys_timer_read() - vcpu->kvm->arch.timer.cntvoff;
@@ -155,7 +155,7 @@ static bool kvm_timer_irq_can_fire(struc
bool kvm_timer_should_fire(struct kvm_vcpu *vcpu)
{
struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu;
- cycle_t cval, now;
+ u64 cval, now;
if (!kvm_timer_irq_can_fire(vcpu))
return false;
On Thu, Dec 08, 2016 at 08:49:45PM -0000, Thomas Gleixner wrote:
> Kill the ever confusing typedef and use u64.
>
> NOT FOR INCLUSION - Must be regenerated at some point via coccinelle
>
> Not-Signed-off-by: Thomas Gleixner <[email protected]>
Not-Reviewed-by: David Gibson <[email protected]>
The concept seems sensible to me, I haven't actually gone through and
verified that your script has done what you intended.
> ---
> arch/alpha/kernel/time.c | 4 -
> arch/arc/kernel/time.c | 12 ++--
> arch/arm/mach-davinci/time.c | 2
> arch/arm/mach-ep93xx/timer-ep93xx.c | 4 -
> arch/arm/mach-footbridge/dc21285-timer.c | 2
> arch/arm/mach-ixp4xx/common.c | 2
> arch/arm/mach-mmp/time.c | 2
> arch/arm/mach-omap2/timer.c | 4 -
> arch/arm/plat-iop/time.c | 2
> arch/avr32/kernel/time.c | 4 -
> arch/blackfin/kernel/time-ts.c | 4 -
> arch/c6x/kernel/time.c | 2
> arch/hexagon/kernel/time.c | 4 -
> arch/ia64/kernel/cyclone.c | 4 -
> arch/ia64/kernel/fsyscall_gtod_data.h | 6 +-
> arch/ia64/kernel/time.c | 6 +-
> arch/ia64/sn/kernel/sn2/timer.c | 4 -
> arch/m68k/68000/timers.c | 2
> arch/m68k/coldfire/dma_timer.c | 2
> arch/m68k/coldfire/pit.c | 2
> arch/m68k/coldfire/sltimers.c | 2
> arch/m68k/coldfire/timers.c | 2
> arch/microblaze/kernel/timer.c | 6 +-
> arch/mips/alchemy/common/time.c | 2
> arch/mips/cavium-octeon/csrc-octeon.c | 2
> arch/mips/jz4740/time.c | 2
> arch/mips/kernel/cevt-txx9.c | 2
> arch/mips/kernel/csrc-bcm1480.c | 4 -
> arch/mips/kernel/csrc-ioasic.c | 2
> arch/mips/kernel/csrc-r4k.c | 2
> arch/mips/kernel/csrc-sb1250.c | 4 -
> arch/mips/loongson32/common/time.c | 4 -
> arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c | 4 -
> arch/mips/loongson64/loongson-3/hpet.c | 4 -
> arch/mips/mti-malta/malta-time.c | 2
> arch/mips/netlogic/common/time.c | 4 -
> arch/mips/sgi-ip27/ip27-timer.c | 2
> arch/mn10300/kernel/csrc-mn10300.c | 2
> arch/nios2/kernel/time.c | 2
> arch/openrisc/kernel/time.c | 4 -
> arch/parisc/kernel/time.c | 2
> arch/powerpc/kernel/time.c | 14 ++---
> arch/s390/kernel/time.c | 2
> arch/sparc/kernel/time_32.c | 2
> arch/sparc/kernel/time_64.c | 2
> arch/um/kernel/time.c | 2
> arch/unicore32/kernel/time.c | 2
> arch/x86/entry/vdso/vclock_gettime.c | 8 +--
> arch/x86/include/asm/kvm_host.h | 2
> arch/x86/include/asm/pvclock.h | 6 +-
> arch/x86/include/asm/tsc.h | 2
> arch/x86/include/asm/vgtod.h | 4 -
> arch/x86/kernel/apb_timer.c | 4 -
> arch/x86/kernel/cpu/mshyperv.c | 4 -
> arch/x86/kernel/hpet.c | 14 ++---
> arch/x86/kernel/kvmclock.c | 10 ++--
> arch/x86/kernel/pvclock.c | 4 -
> arch/x86/kernel/tsc.c | 6 +-
> arch/x86/kvm/x86.c | 14 ++---
> arch/x86/lguest/boot.c | 2
> arch/x86/platform/uv/uv_time.c | 8 +--
> arch/x86/xen/time.c | 6 +-
> arch/x86/xen/xen-ops.h | 2
> arch/xtensa/kernel/time.c | 4 -
> drivers/char/hpet.c | 4 -
> drivers/clocksource/acpi_pm.c | 14 ++---
> drivers/clocksource/arm_arch_timer.c | 4 -
> drivers/clocksource/arm_global_timer.c | 2
> drivers/clocksource/cadence_ttc_timer.c | 4 -
> drivers/clocksource/clksrc-dbx500-prcmu.c | 2
> drivers/clocksource/dw_apb_timer.c | 8 +--
> drivers/clocksource/em_sti.c | 12 ++--
> drivers/clocksource/exynos_mct.c | 6 +-
> drivers/clocksource/h8300_timer16.c | 2
> drivers/clocksource/h8300_tpu.c | 2
> drivers/clocksource/i8253.c | 4 -
> drivers/clocksource/jcore-pit.c | 2
> drivers/clocksource/metag_generic.c | 2
> drivers/clocksource/mips-gic-timer.c | 2
> drivers/clocksource/mmio.c | 18 +++----
> drivers/clocksource/mxs_timer.c | 2
> drivers/clocksource/qcom-timer.c | 2
> drivers/clocksource/samsung_pwm_timer.c | 2
> drivers/clocksource/scx200_hrt.c | 4 -
> drivers/clocksource/sh_cmt.c | 2
> drivers/clocksource/sh_tmu.c | 2
> drivers/clocksource/tcb_clksrc.c | 4 -
> drivers/clocksource/time-pistachio.c | 4 -
> drivers/clocksource/timer-atlas7.c | 2
> drivers/clocksource/timer-atmel-pit.c | 2
> drivers/clocksource/timer-atmel-st.c | 2
> drivers/clocksource/timer-nps.c | 4 -
> drivers/clocksource/timer-prima2.c | 2
> drivers/clocksource/timer-sun5i.c | 2
> drivers/clocksource/timer-ti-32k.c | 4 -
> drivers/clocksource/vt8500_timer.c | 4 -
> drivers/hv/hv.c | 8 +--
> drivers/irqchip/irq-mips-gic.c | 16 +++---
> drivers/net/ethernet/amd/xgbe/xgbe-ptp.c | 2
> drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c | 2
> drivers/net/ethernet/freescale/fec_ptp.c | 2
> drivers/net/ethernet/intel/e1000e/netdev.c | 18 +++----
> drivers/net/ethernet/intel/e1000e/ptp.c | 4 -
> drivers/net/ethernet/intel/igb/igb_ptp.c | 4 -
> drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c | 4 -
> drivers/net/ethernet/mellanox/mlx4/en_clock.c | 2
> drivers/net/ethernet/mellanox/mlx4/main.c | 4 -
> drivers/net/ethernet/mellanox/mlx5/core/en_clock.c | 2
> drivers/net/ethernet/mellanox/mlx5/core/main.c | 4 -
> drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h | 2
> drivers/net/ethernet/ti/cpts.c | 2
> include/kvm/arm_arch_timer.h | 4 -
> include/linux/clocksource.h | 22 ++++----
> include/linux/dw_apb_timer.h | 2
> include/linux/irqchip/mips-gic.h | 8 +--
> include/linux/mlx4/device.h | 2
> include/linux/timecounter.h | 12 ++--
> include/linux/timekeeper_internal.h | 10 ++--
> include/linux/timekeeping.h | 4 -
> include/linux/types.h | 3 -
> kernel/time/clocksource.c | 2
> kernel/time/jiffies.c | 4 -
> kernel/time/timecounter.c | 6 +-
> kernel/time/timekeeping.c | 50 ++++++++++----------
> kernel/time/timekeeping_internal.h | 6 +-
> kernel/trace/ftrace.c | 4 -
> kernel/trace/trace.c | 6 +-
> kernel/trace/trace.h | 8 +--
> kernel/trace/trace_irqsoff.c | 4 -
> kernel/trace/trace_sched_wakeup.c | 4 -
> sound/hda/hdac_stream.c | 6 +-
> virt/kvm/arm/arch_timer.c | 6 +-
> 132 files changed, 318 insertions(+), 321 deletions(-)
>
> --- a/arch/alpha/kernel/time.c
> +++ b/arch/alpha/kernel/time.c
> @@ -133,7 +133,7 @@ init_rtc_clockevent(void)
> * The QEMU clock as a clocksource primitive.
> */
>
> -static cycle_t
> +static u64
> qemu_cs_read(struct clocksource *cs)
> {
> return qemu_get_vmtime();
> @@ -260,7 +260,7 @@ common_init_rtc(void)
> * use this method when WTINT is in use.
> */
>
> -static cycle_t read_rpcc(struct clocksource *cs)
> +static u64 read_rpcc(struct clocksource *cs)
> {
> return rpcc();
> }
> --- a/arch/arc/kernel/time.c
> +++ b/arch/arc/kernel/time.c
> @@ -83,7 +83,7 @@ static int noinline arc_get_timer_clk(st
>
> #ifdef CONFIG_ARC_HAS_GFRC
>
> -static cycle_t arc_read_gfrc(struct clocksource *cs)
> +static u64 arc_read_gfrc(struct clocksource *cs)
> {
> unsigned long flags;
> union {
> @@ -92,7 +92,7 @@ static cycle_t arc_read_gfrc(struct cloc
> #else
> struct { u32 l, h; };
> #endif
> - cycle_t full;
> + u64 full;
> } stamp;
>
> local_irq_save(flags);
> @@ -140,7 +140,7 @@ CLOCKSOURCE_OF_DECLARE(arc_gfrc, "snps,a
> #define AUX_RTC_LOW 0x104
> #define AUX_RTC_HIGH 0x105
>
> -static cycle_t arc_read_rtc(struct clocksource *cs)
> +static u64 arc_read_rtc(struct clocksource *cs)
> {
> unsigned long status;
> union {
> @@ -149,7 +149,7 @@ static cycle_t arc_read_rtc(struct clock
> #else
> struct { u32 low, high; };
> #endif
> - cycle_t full;
> + u64 full;
> } stamp;
>
> /*
> @@ -203,9 +203,9 @@ CLOCKSOURCE_OF_DECLARE(arc_rtc, "snps,ar
> * 32bit TIMER1 to keep counting monotonically and wraparound
> */
>
> -static cycle_t arc_read_timer1(struct clocksource *cs)
> +static u64 arc_read_timer1(struct clocksource *cs)
> {
> - return (cycle_t) read_aux_reg(ARC_REG_TIMER1_CNT);
> + return (u64) read_aux_reg(ARC_REG_TIMER1_CNT);
> }
>
> static struct clocksource arc_counter_timer1 = {
> --- a/arch/arm/mach-davinci/time.c
> +++ b/arch/arm/mach-davinci/time.c
> @@ -268,7 +268,7 @@ static void __init timer_init(void)
> /*
> * clocksource
> */
> -static cycle_t read_cycles(struct clocksource *cs)
> +static u64 read_cycles(struct clocksource *cs)
> {
> struct timer_s *t = &timers[TID_CLOCKSOURCE];
>
> --- a/arch/arm/mach-ep93xx/timer-ep93xx.c
> +++ b/arch/arm/mach-ep93xx/timer-ep93xx.c
> @@ -59,13 +59,13 @@ static u64 notrace ep93xx_read_sched_clo
> return ret;
> }
>
> -cycle_t ep93xx_clocksource_read(struct clocksource *c)
> +u64 ep93xx_clocksource_read(struct clocksource *c)
> {
> u64 ret;
>
> ret = readl(EP93XX_TIMER4_VALUE_LOW);
> ret |= ((u64) (readl(EP93XX_TIMER4_VALUE_HIGH) & 0xff) << 32);
> - return (cycle_t) ret;
> + return ret;
> }
>
> static int ep93xx_clkevt_set_next_event(unsigned long next,
> --- a/arch/arm/mach-footbridge/dc21285-timer.c
> +++ b/arch/arm/mach-footbridge/dc21285-timer.c
> @@ -19,7 +19,7 @@
>
> #include "common.h"
>
> -static cycle_t cksrc_dc21285_read(struct clocksource *cs)
> +static u64 cksrc_dc21285_read(struct clocksource *cs)
> {
> return cs->mask - *CSR_TIMER2_VALUE;
> }
> --- a/arch/arm/mach-ixp4xx/common.c
> +++ b/arch/arm/mach-ixp4xx/common.c
> @@ -493,7 +493,7 @@ static u64 notrace ixp4xx_read_sched_clo
> * clocksource
> */
>
> -static cycle_t ixp4xx_clocksource_read(struct clocksource *c)
> +static u64 ixp4xx_clocksource_read(struct clocksource *c)
> {
> return *IXP4XX_OSTS;
> }
> --- a/arch/arm/mach-mmp/time.c
> +++ b/arch/arm/mach-mmp/time.c
> @@ -144,7 +144,7 @@ static struct clock_event_device ckevt =
> .set_state_oneshot = timer_set_shutdown,
> };
>
> -static cycle_t clksrc_read(struct clocksource *cs)
> +static u64 clksrc_read(struct clocksource *cs)
> {
> return timer_read();
> }
> --- a/arch/arm/mach-omap2/timer.c
> +++ b/arch/arm/mach-omap2/timer.c
> @@ -369,9 +369,9 @@ static bool use_gptimer_clksrc __initdat
> /*
> * clocksource
> */
> -static cycle_t clocksource_read_cycles(struct clocksource *cs)
> +static u64 clocksource_read_cycles(struct clocksource *cs)
> {
> - return (cycle_t)__omap_dm_timer_read_counter(&clksrc,
> + return (u64)__omap_dm_timer_read_counter(&clksrc,
> OMAP_TIMER_NONPOSTED);
> }
>
> --- a/arch/arm/plat-iop/time.c
> +++ b/arch/arm/plat-iop/time.c
> @@ -38,7 +38,7 @@
> /*
> * IOP clocksource (free-running timer 1).
> */
> -static cycle_t notrace iop_clocksource_read(struct clocksource *unused)
> +static u64 notrace iop_clocksource_read(struct clocksource *unused)
> {
> return 0xffffffffu - read_tcr1();
> }
> --- a/arch/avr32/kernel/time.c
> +++ b/arch/avr32/kernel/time.c
> @@ -20,9 +20,9 @@
>
> static bool disable_cpu_idle_poll;
>
> -static cycle_t read_cycle_count(struct clocksource *cs)
> +static u64 read_cycle_count(struct clocksource *cs)
> {
> - return (cycle_t)sysreg_read(COUNT);
> + return (u64)sysreg_read(COUNT);
> }
>
> /*
> --- a/arch/blackfin/kernel/time-ts.c
> +++ b/arch/blackfin/kernel/time-ts.c
> @@ -26,7 +26,7 @@
>
> #if defined(CONFIG_CYCLES_CLOCKSOURCE)
>
> -static notrace cycle_t bfin_read_cycles(struct clocksource *cs)
> +static notrace u64 bfin_read_cycles(struct clocksource *cs)
> {
> #ifdef CONFIG_CPU_FREQ
> return __bfin_cycles_off + (get_cycles() << __bfin_cycles_mod);
> @@ -80,7 +80,7 @@ void __init setup_gptimer0(void)
> enable_gptimers(TIMER0bit);
> }
>
> -static cycle_t bfin_read_gptimer0(struct clocksource *cs)
> +static u64 bfin_read_gptimer0(struct clocksource *cs)
> {
> return bfin_read_TIMER0_COUNTER();
> }
> --- a/arch/c6x/kernel/time.c
> +++ b/arch/c6x/kernel/time.c
> @@ -26,7 +26,7 @@
> static u32 sched_clock_multiplier;
> #define SCHED_CLOCK_SHIFT 16
>
> -static cycle_t tsc_read(struct clocksource *cs)
> +static u64 tsc_read(struct clocksource *cs)
> {
> return get_cycles();
> }
> --- a/arch/hexagon/kernel/time.c
> +++ b/arch/hexagon/kernel/time.c
> @@ -72,9 +72,9 @@ struct adsp_hw_timer_struct {
> /* Look for "TCX0" for related constants. */
> static __iomem struct adsp_hw_timer_struct *rtos_timer;
>
> -static cycle_t timer_get_cycles(struct clocksource *cs)
> +static u64 timer_get_cycles(struct clocksource *cs)
> {
> - return (cycle_t) __vmgettime();
> + return (u64) __vmgettime();
> }
>
> static struct clocksource hexagon_clocksource = {
> --- a/arch/ia64/kernel/cyclone.c
> +++ b/arch/ia64/kernel/cyclone.c
> @@ -21,9 +21,9 @@ void __init cyclone_setup(void)
>
> static void __iomem *cyclone_mc;
>
> -static cycle_t read_cyclone(struct clocksource *cs)
> +static u64 read_cyclone(struct clocksource *cs)
> {
> - return (cycle_t)readq((void __iomem *)cyclone_mc);
> + return (u64)readq((void __iomem *)cyclone_mc);
> }
>
> static struct clocksource clocksource_cyclone = {
> --- a/arch/ia64/kernel/fsyscall_gtod_data.h
> +++ b/arch/ia64/kernel/fsyscall_gtod_data.h
> @@ -9,15 +9,15 @@ struct fsyscall_gtod_data_t {
> seqcount_t seq;
> struct timespec wall_time;
> struct timespec monotonic_time;
> - cycle_t clk_mask;
> + u64 clk_mask;
> u32 clk_mult;
> u32 clk_shift;
> void *clk_fsys_mmio;
> - cycle_t clk_cycle_last;
> + u64 clk_cycle_last;
> } ____cacheline_aligned;
>
> struct itc_jitter_data_t {
> int itc_jitter;
> - cycle_t itc_lastcycle;
> + u64 itc_lastcycle;
> } ____cacheline_aligned;
>
> --- a/arch/ia64/kernel/time.c
> +++ b/arch/ia64/kernel/time.c
> @@ -31,7 +31,7 @@
>
> #include "fsyscall_gtod_data.h"
>
> -static cycle_t itc_get_cycles(struct clocksource *cs);
> +static u64 itc_get_cycles(struct clocksource *cs);
>
> struct fsyscall_gtod_data_t fsyscall_gtod_data;
>
> @@ -323,7 +323,7 @@ void ia64_init_itm(void)
> }
> }
>
> -static cycle_t itc_get_cycles(struct clocksource *cs)
> +static u64 itc_get_cycles(struct clocksource *cs)
> {
> unsigned long lcycle, now, ret;
>
> @@ -397,7 +397,7 @@ void update_vsyscall_tz(void)
> }
>
> void update_vsyscall_old(struct timespec *wall, struct timespec *wtm,
> - struct clocksource *c, u32 mult, cycle_t cycle_last)
> + struct clocksource *c, u32 mult, u64 cycle_last)
> {
> write_seqcount_begin(&fsyscall_gtod_data.seq);
>
> --- a/arch/ia64/sn/kernel/sn2/timer.c
> +++ b/arch/ia64/sn/kernel/sn2/timer.c
> @@ -22,9 +22,9 @@
>
> extern unsigned long sn_rtc_cycles_per_second;
>
> -static cycle_t read_sn2(struct clocksource *cs)
> +static u64 read_sn2(struct clocksource *cs)
> {
> - return (cycle_t)readq(RTC_COUNTER_ADDR);
> + return (u64)readq(RTC_COUNTER_ADDR);
> }
>
> static struct clocksource clocksource_sn2 = {
> --- a/arch/m68k/68000/timers.c
> +++ b/arch/m68k/68000/timers.c
> @@ -76,7 +76,7 @@ static struct irqaction m68328_timer_irq
>
> /***************************************************************************/
>
> -static cycle_t m68328_read_clk(struct clocksource *cs)
> +static u64 m68328_read_clk(struct clocksource *cs)
> {
> unsigned long flags;
> u32 cycles;
> --- a/arch/m68k/coldfire/dma_timer.c
> +++ b/arch/m68k/coldfire/dma_timer.c
> @@ -34,7 +34,7 @@
> #define DMA_DTMR_CLK_DIV_16 (2 << 1)
> #define DMA_DTMR_ENABLE (1 << 0)
>
> -static cycle_t cf_dt_get_cycles(struct clocksource *cs)
> +static u64 cf_dt_get_cycles(struct clocksource *cs)
> {
> return __raw_readl(DTCN0);
> }
> --- a/arch/m68k/coldfire/pit.c
> +++ b/arch/m68k/coldfire/pit.c
> @@ -118,7 +118,7 @@ static struct irqaction pit_irq = {
>
> /***************************************************************************/
>
> -static cycle_t pit_read_clk(struct clocksource *cs)
> +static u64 pit_read_clk(struct clocksource *cs)
> {
> unsigned long flags;
> u32 cycles;
> --- a/arch/m68k/coldfire/sltimers.c
> +++ b/arch/m68k/coldfire/sltimers.c
> @@ -97,7 +97,7 @@ static struct irqaction mcfslt_timer_irq
> .handler = mcfslt_tick,
> };
>
> -static cycle_t mcfslt_read_clk(struct clocksource *cs)
> +static u64 mcfslt_read_clk(struct clocksource *cs)
> {
> unsigned long flags;
> u32 cycles, scnt;
> --- a/arch/m68k/coldfire/timers.c
> +++ b/arch/m68k/coldfire/timers.c
> @@ -89,7 +89,7 @@ static struct irqaction mcftmr_timer_irq
>
> /***************************************************************************/
>
> -static cycle_t mcftmr_read_clk(struct clocksource *cs)
> +static u64 mcftmr_read_clk(struct clocksource *cs)
> {
> unsigned long flags;
> u32 cycles;
> --- a/arch/microblaze/kernel/timer.c
> +++ b/arch/microblaze/kernel/timer.c
> @@ -190,17 +190,17 @@ static u64 xilinx_clock_read(void)
> return read_fn(timer_baseaddr + TCR1);
> }
>
> -static cycle_t xilinx_read(struct clocksource *cs)
> +static u64 xilinx_read(struct clocksource *cs)
> {
> /* reading actual value of timer 1 */
> - return (cycle_t)xilinx_clock_read();
> + return (u64)xilinx_clock_read();
> }
>
> static struct timecounter xilinx_tc = {
> .cc = NULL,
> };
>
> -static cycle_t xilinx_cc_read(const struct cyclecounter *cc)
> +static u64 xilinx_cc_read(const struct cyclecounter *cc)
> {
> return xilinx_read(NULL);
> }
> --- a/arch/mips/alchemy/common/time.c
> +++ b/arch/mips/alchemy/common/time.c
> @@ -44,7 +44,7 @@
> /* 32kHz clock enabled and detected */
> #define CNTR_OK (SYS_CNTRL_E0 | SYS_CNTRL_32S)
>
> -static cycle_t au1x_counter1_read(struct clocksource *cs)
> +static u64 au1x_counter1_read(struct clocksource *cs)
> {
> return alchemy_rdsys(AU1000_SYS_RTCREAD);
> }
> --- a/arch/mips/cavium-octeon/csrc-octeon.c
> +++ b/arch/mips/cavium-octeon/csrc-octeon.c
> @@ -98,7 +98,7 @@ void octeon_init_cvmcount(void)
> local_irq_restore(flags);
> }
>
> -static cycle_t octeon_cvmcount_read(struct clocksource *cs)
> +static u64 octeon_cvmcount_read(struct clocksource *cs)
> {
> return read_c0_cvmcount();
> }
> --- a/arch/mips/jz4740/time.c
> +++ b/arch/mips/jz4740/time.c
> @@ -34,7 +34,7 @@
>
> static uint16_t jz4740_jiffies_per_tick;
>
> -static cycle_t jz4740_clocksource_read(struct clocksource *cs)
> +static u64 jz4740_clocksource_read(struct clocksource *cs)
> {
> return jz4740_timer_get_count(TIMER_CLOCKSOURCE);
> }
> --- a/arch/mips/kernel/cevt-txx9.c
> +++ b/arch/mips/kernel/cevt-txx9.c
> @@ -27,7 +27,7 @@ struct txx9_clocksource {
> struct txx9_tmr_reg __iomem *tmrptr;
> };
>
> -static cycle_t txx9_cs_read(struct clocksource *cs)
> +static u64 txx9_cs_read(struct clocksource *cs)
> {
> struct txx9_clocksource *txx9_cs =
> container_of(cs, struct txx9_clocksource, cs);
> --- a/arch/mips/kernel/csrc-bcm1480.c
> +++ b/arch/mips/kernel/csrc-bcm1480.c
> @@ -25,9 +25,9 @@
>
> #include <asm/sibyte/sb1250.h>
>
> -static cycle_t bcm1480_hpt_read(struct clocksource *cs)
> +static u64 bcm1480_hpt_read(struct clocksource *cs)
> {
> - return (cycle_t) __raw_readq(IOADDR(A_SCD_ZBBUS_CYCLE_COUNT));
> + return (u64) __raw_readq(IOADDR(A_SCD_ZBBUS_CYCLE_COUNT));
> }
>
> struct clocksource bcm1480_clocksource = {
> --- a/arch/mips/kernel/csrc-ioasic.c
> +++ b/arch/mips/kernel/csrc-ioasic.c
> @@ -22,7 +22,7 @@
> #include <asm/dec/ioasic.h>
> #include <asm/dec/ioasic_addrs.h>
>
> -static cycle_t dec_ioasic_hpt_read(struct clocksource *cs)
> +static u64 dec_ioasic_hpt_read(struct clocksource *cs)
> {
> return ioasic_read(IO_REG_FCTR);
> }
> --- a/arch/mips/kernel/csrc-r4k.c
> +++ b/arch/mips/kernel/csrc-r4k.c
> @@ -11,7 +11,7 @@
>
> #include <asm/time.h>
>
> -static cycle_t c0_hpt_read(struct clocksource *cs)
> +static u64 c0_hpt_read(struct clocksource *cs)
> {
> return read_c0_count();
> }
> --- a/arch/mips/kernel/csrc-sb1250.c
> +++ b/arch/mips/kernel/csrc-sb1250.c
> @@ -30,7 +30,7 @@
> * The HPT is free running from SB1250_HPT_VALUE down to 0 then starts over
> * again.
> */
> -static inline cycle_t sb1250_hpt_get_cycles(void)
> +static inline u64 sb1250_hpt_get_cycles(void)
> {
> unsigned int count;
> void __iomem *addr;
> @@ -41,7 +41,7 @@ static inline cycle_t sb1250_hpt_get_cyc
> return SB1250_HPT_VALUE - count;
> }
>
> -static cycle_t sb1250_hpt_read(struct clocksource *cs)
> +static u64 sb1250_hpt_read(struct clocksource *cs)
> {
> return sb1250_hpt_get_cycles();
> }
> --- a/arch/mips/loongson32/common/time.c
> +++ b/arch/mips/loongson32/common/time.c
> @@ -63,7 +63,7 @@ void __init ls1x_pwmtimer_init(void)
> ls1x_pwmtimer_restart();
> }
>
> -static cycle_t ls1x_clocksource_read(struct clocksource *cs)
> +static u64 ls1x_clocksource_read(struct clocksource *cs)
> {
> unsigned long flags;
> int count;
> @@ -107,7 +107,7 @@ static cycle_t ls1x_clocksource_read(str
>
> raw_spin_unlock_irqrestore(&ls1x_timer_lock, flags);
>
> - return (cycle_t) (jifs * ls1x_jiffies_per_tick) + count;
> + return (u64) (jifs * ls1x_jiffies_per_tick) + count;
> }
>
> static struct clocksource ls1x_clocksource = {
> --- a/arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c
> +++ b/arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c
> @@ -144,7 +144,7 @@ void __init setup_mfgpt0_timer(void)
> * to just read by itself. So use jiffies to emulate a free
> * running counter:
> */
> -static cycle_t mfgpt_read(struct clocksource *cs)
> +static u64 mfgpt_read(struct clocksource *cs)
> {
> unsigned long flags;
> int count;
> @@ -188,7 +188,7 @@ static cycle_t mfgpt_read(struct clockso
>
> raw_spin_unlock_irqrestore(&mfgpt_lock, flags);
>
> - return (cycle_t) (jifs * COMPARE) + count;
> + return (u64) (jifs * COMPARE) + count;
> }
>
> static struct clocksource clocksource_mfgpt = {
> --- a/arch/mips/loongson64/loongson-3/hpet.c
> +++ b/arch/mips/loongson64/loongson-3/hpet.c
> @@ -248,9 +248,9 @@ void __init setup_hpet_timer(void)
> pr_info("hpet clock event device register\n");
> }
>
> -static cycle_t hpet_read_counter(struct clocksource *cs)
> +static u64 hpet_read_counter(struct clocksource *cs)
> {
> - return (cycle_t)hpet_read(HPET_COUNTER);
> + return (u64)hpet_read(HPET_COUNTER);
> }
>
> static void hpet_suspend(struct clocksource *cs)
> --- a/arch/mips/mti-malta/malta-time.c
> +++ b/arch/mips/mti-malta/malta-time.c
> @@ -75,7 +75,7 @@ static void __init estimate_frequencies(
> unsigned int count, start;
> unsigned char secs1, secs2, ctrl;
> int secs;
> - cycle_t giccount = 0, gicstart = 0;
> + u64 giccount = 0, gicstart = 0;
>
> #if defined(CONFIG_KVM_GUEST) && CONFIG_KVM_GUEST_TIMER_FREQ
> mips_hpt_frequency = CONFIG_KVM_GUEST_TIMER_FREQ * 1000000;
> --- a/arch/mips/netlogic/common/time.c
> +++ b/arch/mips/netlogic/common/time.c
> @@ -59,14 +59,14 @@ unsigned int get_c0_compare_int(void)
> return IRQ_TIMER;
> }
>
> -static cycle_t nlm_get_pic_timer(struct clocksource *cs)
> +static u64 nlm_get_pic_timer(struct clocksource *cs)
> {
> uint64_t picbase = nlm_get_node(0)->picbase;
>
> return ~nlm_pic_read_timer(picbase, PIC_CLOCK_TIMER);
> }
>
> -static cycle_t nlm_get_pic_timer32(struct clocksource *cs)
> +static u64 nlm_get_pic_timer32(struct clocksource *cs)
> {
> uint64_t picbase = nlm_get_node(0)->picbase;
>
> --- a/arch/mips/sgi-ip27/ip27-timer.c
> +++ b/arch/mips/sgi-ip27/ip27-timer.c
> @@ -140,7 +140,7 @@ static void __init hub_rt_clock_event_gl
> setup_irq(irq, &hub_rt_irqaction);
> }
>
> -static cycle_t hub_rt_read(struct clocksource *cs)
> +static u64 hub_rt_read(struct clocksource *cs)
> {
> return REMOTE_HUB_L(cputonasid(0), PI_RT_COUNT);
> }
> --- a/arch/mn10300/kernel/csrc-mn10300.c
> +++ b/arch/mn10300/kernel/csrc-mn10300.c
> @@ -13,7 +13,7 @@
> #include <asm/timex.h>
> #include "internal.h"
>
> -static cycle_t mn10300_read(struct clocksource *cs)
> +static u64 mn10300_read(struct clocksource *cs)
> {
> return read_timestamp_counter();
> }
> --- a/arch/nios2/kernel/time.c
> +++ b/arch/nios2/kernel/time.c
> @@ -81,7 +81,7 @@ static inline unsigned long read_timersn
> return count;
> }
>
> -static cycle_t nios2_timer_read(struct clocksource *cs)
> +static u64 nios2_timer_read(struct clocksource *cs)
> {
> struct nios2_clocksource *nios2_cs = to_nios2_clksource(cs);
> unsigned long flags;
> --- a/arch/openrisc/kernel/time.c
> +++ b/arch/openrisc/kernel/time.c
> @@ -117,9 +117,9 @@ static __init void openrisc_clockevent_i
> * is 32 bits wide and runs at the CPU clock frequency.
> */
>
> -static cycle_t openrisc_timer_read(struct clocksource *cs)
> +static u64 openrisc_timer_read(struct clocksource *cs)
> {
> - return (cycle_t) mfspr(SPR_TTCR);
> + return (u64) mfspr(SPR_TTCR);
> }
>
> static struct clocksource openrisc_timer = {
> --- a/arch/parisc/kernel/time.c
> +++ b/arch/parisc/kernel/time.c
> @@ -191,7 +191,7 @@ EXPORT_SYMBOL(profile_pc);
>
> /* clock source code */
>
> -static cycle_t notrace read_cr16(struct clocksource *cs)
> +static u64 notrace read_cr16(struct clocksource *cs)
> {
> return get_cycles();
> }
> --- a/arch/powerpc/kernel/time.c
> +++ b/arch/powerpc/kernel/time.c
> @@ -80,7 +80,7 @@
> #include <linux/clockchips.h>
> #include <linux/timekeeper_internal.h>
>
> -static cycle_t rtc_read(struct clocksource *);
> +static u64 rtc_read(struct clocksource *);
> static struct clocksource clocksource_rtc = {
> .name = "rtc",
> .rating = 400,
> @@ -89,7 +89,7 @@ static struct clocksource clocksource_rt
> .read = rtc_read,
> };
>
> -static cycle_t timebase_read(struct clocksource *);
> +static u64 timebase_read(struct clocksource *);
> static struct clocksource clocksource_timebase = {
> .name = "timebase",
> .rating = 400,
> @@ -802,18 +802,18 @@ void read_persistent_clock(struct timesp
> }
>
> /* clocksource code */
> -static cycle_t rtc_read(struct clocksource *cs)
> +static u64 rtc_read(struct clocksource *cs)
> {
> - return (cycle_t)get_rtc();
> + return (u64)get_rtc();
> }
>
> -static cycle_t timebase_read(struct clocksource *cs)
> +static u64 timebase_read(struct clocksource *cs)
> {
> - return (cycle_t)get_tb();
> + return (u64)get_tb();
> }
>
> void update_vsyscall_old(struct timespec *wall_time, struct timespec *wtm,
> - struct clocksource *clock, u32 mult, cycle_t cycle_last)
> + struct clocksource *clock, u32 mult, u64 cycle_last)
> {
> u64 new_tb_to_xs, new_stamp_xsec;
> u32 frac_sec;
> --- a/arch/s390/kernel/time.c
> +++ b/arch/s390/kernel/time.c
> @@ -213,7 +213,7 @@ void read_boot_clock64(struct timespec64
> tod_to_timeval(clock - TOD_UNIX_EPOCH, ts);
> }
>
> -static cycle_t read_tod_clock(struct clocksource *cs)
> +static u64 read_tod_clock(struct clocksource *cs)
> {
> return get_tod_clock();
> }
> --- a/arch/sparc/kernel/time_32.c
> +++ b/arch/sparc/kernel/time_32.c
> @@ -148,7 +148,7 @@ static unsigned int sbus_cycles_offset(v
> return offset;
> }
>
> -static cycle_t timer_cs_read(struct clocksource *cs)
> +static u64 timer_cs_read(struct clocksource *cs)
> {
> unsigned int seq, offset;
> u64 cycles;
> --- a/arch/sparc/kernel/time_64.c
> +++ b/arch/sparc/kernel/time_64.c
> @@ -770,7 +770,7 @@ void udelay(unsigned long usecs)
> }
> EXPORT_SYMBOL(udelay);
>
> -static cycle_t clocksource_tick_read(struct clocksource *cs)
> +static u64 clocksource_tick_read(struct clocksource *cs)
> {
> return tick_ops->get_tick();
> }
> --- a/arch/um/kernel/time.c
> +++ b/arch/um/kernel/time.c
> @@ -83,7 +83,7 @@ static irqreturn_t um_timer(int irq, voi
> return IRQ_HANDLED;
> }
>
> -static cycle_t timer_read(struct clocksource *cs)
> +static u64 timer_read(struct clocksource *cs)
> {
> return os_nsecs() / TIMER_MULTIPLIER;
> }
> --- a/arch/unicore32/kernel/time.c
> +++ b/arch/unicore32/kernel/time.c
> @@ -62,7 +62,7 @@ static struct clock_event_device ckevt_p
> .set_state_oneshot = puv3_osmr0_shutdown,
> };
>
> -static cycle_t puv3_read_oscr(struct clocksource *cs)
> +static u64 puv3_read_oscr(struct clocksource *cs)
> {
> return readl(OST_OSCR);
> }
> --- a/arch/x86/entry/vdso/vclock_gettime.c
> +++ b/arch/x86/entry/vdso/vclock_gettime.c
> @@ -92,10 +92,10 @@ static notrace const struct pvclock_vsys
> return (const struct pvclock_vsyscall_time_info *)&pvclock_page;
> }
>
> -static notrace cycle_t vread_pvclock(int *mode)
> +static notrace u64 vread_pvclock(int *mode)
> {
> const struct pvclock_vcpu_time_info *pvti = &get_pvti0()->pvti;
> - cycle_t ret;
> + u64 ret;
> u64 last;
> u32 version;
>
> @@ -142,9 +142,9 @@ static notrace cycle_t vread_pvclock(int
> }
> #endif
>
> -notrace static cycle_t vread_tsc(void)
> +notrace static u64 vread_tsc(void)
> {
> - cycle_t ret = (cycle_t)rdtsc_ordered();
> + u64 ret = (u64)rdtsc_ordered();
> u64 last = gtod->cycle_last;
>
> if (likely(ret >= last))
> --- a/arch/x86/include/asm/kvm_host.h
> +++ b/arch/x86/include/asm/kvm_host.h
> @@ -758,7 +758,7 @@ struct kvm_arch {
> spinlock_t pvclock_gtod_sync_lock;
> bool use_master_clock;
> u64 master_kernel_ns;
> - cycle_t master_cycle_now;
> + u64 master_cycle_now;
> struct delayed_work kvmclock_update_work;
> struct delayed_work kvmclock_sync_work;
>
> --- a/arch/x86/include/asm/pvclock.h
> +++ b/arch/x86/include/asm/pvclock.h
> @@ -14,7 +14,7 @@ static inline struct pvclock_vsyscall_ti
> #endif
>
> /* some helper functions for xen and kvm pv clock sources */
> -cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src);
> +u64 pvclock_clocksource_read(struct pvclock_vcpu_time_info *src);
> u8 pvclock_read_flags(struct pvclock_vcpu_time_info *src);
> void pvclock_set_flags(u8 flags);
> unsigned long pvclock_tsc_khz(struct pvclock_vcpu_time_info *src);
> @@ -87,11 +87,11 @@ static inline u64 pvclock_scale_delta(u6
> }
>
> static __always_inline
> -cycle_t __pvclock_read_cycles(const struct pvclock_vcpu_time_info *src,
> +u64 __pvclock_read_cycles(const struct pvclock_vcpu_time_info *src,
> u64 tsc)
> {
> u64 delta = tsc - src->tsc_timestamp;
> - cycle_t offset = pvclock_scale_delta(delta, src->tsc_to_system_mul,
> + u64 offset = pvclock_scale_delta(delta, src->tsc_to_system_mul,
> src->tsc_shift);
> return src->system_time + offset;
> }
> --- a/arch/x86/include/asm/tsc.h
> +++ b/arch/x86/include/asm/tsc.h
> @@ -29,7 +29,7 @@ static inline cycles_t get_cycles(void)
> return rdtsc();
> }
>
> -extern struct system_counterval_t convert_art_to_tsc(cycle_t art);
> +extern struct system_counterval_t convert_art_to_tsc(u64 art);
>
> extern void tsc_init(void);
> extern void mark_tsc_unstable(char *reason);
> --- a/arch/x86/include/asm/vgtod.h
> +++ b/arch/x86/include/asm/vgtod.h
> @@ -17,8 +17,8 @@ struct vsyscall_gtod_data {
> unsigned seq;
>
> int vclock_mode;
> - cycle_t cycle_last;
> - cycle_t mask;
> + u64 cycle_last;
> + u64 mask;
> u32 mult;
> u32 shift;
>
> --- a/arch/x86/kernel/apb_timer.c
> +++ b/arch/x86/kernel/apb_timer.c
> @@ -247,7 +247,7 @@ void apbt_setup_secondary_clock(void) {}
> static int apbt_clocksource_register(void)
> {
> u64 start, now;
> - cycle_t t1;
> + u64 t1;
>
> /* Start the counter, use timer 2 as source, timer 0/1 for event */
> dw_apb_clocksource_start(clocksource_apbt);
> @@ -355,7 +355,7 @@ unsigned long apbt_quick_calibrate(void)
> {
> int i, scale;
> u64 old, new;
> - cycle_t t1, t2;
> + u64 t1, t2;
> unsigned long khz = 0;
> u32 loop, shift;
>
> --- a/arch/x86/kernel/cpu/mshyperv.c
> +++ b/arch/x86/kernel/cpu/mshyperv.c
> @@ -133,9 +133,9 @@ static uint32_t __init ms_hyperv_platfo
> return 0;
> }
>
> -static cycle_t read_hv_clock(struct clocksource *arg)
> +static u64 read_hv_clock(struct clocksource *arg)
> {
> - cycle_t current_tick;
> + u64 current_tick;
> /*
> * Read the partition counter to get the current tick count. This count
> * is set to 0 when the partition is created and is incremented in
> --- a/arch/x86/kernel/hpet.c
> +++ b/arch/x86/kernel/hpet.c
> @@ -791,7 +791,7 @@ static union hpet_lock hpet __cacheline_
> { .lock = __ARCH_SPIN_LOCK_UNLOCKED, },
> };
>
> -static cycle_t read_hpet(struct clocksource *cs)
> +static u64 read_hpet(struct clocksource *cs)
> {
> unsigned long flags;
> union hpet_lock old, new;
> @@ -802,7 +802,7 @@ static cycle_t read_hpet(struct clocksou
> * Read HPET directly if in NMI.
> */
> if (in_nmi())
> - return (cycle_t)hpet_readl(HPET_COUNTER);
> + return (u64)hpet_readl(HPET_COUNTER);
>
> /*
> * Read the current state of the lock and HPET value atomically.
> @@ -821,7 +821,7 @@ static cycle_t read_hpet(struct clocksou
> WRITE_ONCE(hpet.value, new.value);
> arch_spin_unlock(&hpet.lock);
> local_irq_restore(flags);
> - return (cycle_t)new.value;
> + return (u64)new.value;
> }
> local_irq_restore(flags);
>
> @@ -843,15 +843,15 @@ static cycle_t read_hpet(struct clocksou
> new.lockval = READ_ONCE(hpet.lockval);
> } while ((new.value == old.value) && arch_spin_is_locked(&new.lock));
>
> - return (cycle_t)new.value;
> + return (u64)new.value;
> }
> #else
> /*
> * For UP or 32-bit.
> */
> -static cycle_t read_hpet(struct clocksource *cs)
> +static u64 read_hpet(struct clocksource *cs)
> {
> - return (cycle_t)hpet_readl(HPET_COUNTER);
> + return (u64)hpet_readl(HPET_COUNTER);
> }
> #endif
>
> @@ -867,7 +867,7 @@ static struct clocksource clocksource_hp
> static int hpet_clocksource_register(void)
> {
> u64 start, now;
> - cycle_t t1;
> + u64 t1;
>
> /* Start the counter */
> hpet_restart_counter();
> --- a/arch/x86/kernel/kvmclock.c
> +++ b/arch/x86/kernel/kvmclock.c
> @@ -32,7 +32,7 @@
> static int kvmclock __ro_after_init = 1;
> static int msr_kvm_system_time = MSR_KVM_SYSTEM_TIME;
> static int msr_kvm_wall_clock = MSR_KVM_WALL_CLOCK;
> -static cycle_t kvm_sched_clock_offset;
> +static u64 kvm_sched_clock_offset;
>
> static int parse_no_kvmclock(char *arg)
> {
> @@ -79,10 +79,10 @@ static int kvm_set_wallclock(const struc
> return -1;
> }
>
> -static cycle_t kvm_clock_read(void)
> +static u64 kvm_clock_read(void)
> {
> struct pvclock_vcpu_time_info *src;
> - cycle_t ret;
> + u64 ret;
> int cpu;
>
> preempt_disable_notrace();
> @@ -93,12 +93,12 @@ static cycle_t kvm_clock_read(void)
> return ret;
> }
>
> -static cycle_t kvm_clock_get_cycles(struct clocksource *cs)
> +static u64 kvm_clock_get_cycles(struct clocksource *cs)
> {
> return kvm_clock_read();
> }
>
> -static cycle_t kvm_sched_clock_read(void)
> +static u64 kvm_sched_clock_read(void)
> {
> return kvm_clock_read() - kvm_sched_clock_offset;
> }
> --- a/arch/x86/kernel/pvclock.c
> +++ b/arch/x86/kernel/pvclock.c
> @@ -71,10 +71,10 @@ u8 pvclock_read_flags(struct pvclock_vcp
> return flags & valid_flags;
> }
>
> -cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src)
> +u64 pvclock_clocksource_read(struct pvclock_vcpu_time_info *src)
> {
> unsigned version;
> - cycle_t ret;
> + u64 ret;
> u64 last;
> u8 flags;
>
> --- a/arch/x86/kernel/tsc.c
> +++ b/arch/x86/kernel/tsc.c
> @@ -1080,9 +1080,9 @@ static struct clocksource clocksource_ts
> * checking the result of read_tsc() - cycle_last for being negative.
> * That works because CLOCKSOURCE_MASK(64) does not mask out any bit.
> */
> -static cycle_t read_tsc(struct clocksource *cs)
> +static u64 read_tsc(struct clocksource *cs)
> {
> - return (cycle_t)rdtsc_ordered();
> + return (u64)rdtsc_ordered();
> }
>
> /*
> @@ -1170,7 +1170,7 @@ int unsynchronized_tsc(void)
> /*
> * Convert ART to TSC given numerator/denominator found in detect_art()
> */
> -struct system_counterval_t convert_art_to_tsc(cycle_t art)
> +struct system_counterval_t convert_art_to_tsc(u64 art)
> {
> u64 tmp, res, rem;
>
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -1128,8 +1128,8 @@ struct pvclock_gtod_data {
>
> struct { /* extract of a clocksource struct */
> int vclock_mode;
> - cycle_t cycle_last;
> - cycle_t mask;
> + u64 cycle_last;
> + u64 mask;
> u32 mult;
> u32 shift;
> } clock;
> @@ -1569,9 +1569,9 @@ static inline void adjust_tsc_offset_hos
>
> #ifdef CONFIG_X86_64
>
> -static cycle_t read_tsc(void)
> +static u64 read_tsc(void)
> {
> - cycle_t ret = (cycle_t)rdtsc_ordered();
> + u64 ret = (u64)rdtsc_ordered();
> u64 last = pvclock_gtod_data.clock.cycle_last;
>
> if (likely(ret >= last))
> @@ -1589,7 +1589,7 @@ static cycle_t read_tsc(void)
> return last;
> }
>
> -static inline u64 vgettsc(cycle_t *cycle_now)
> +static inline u64 vgettsc(u64 *cycle_now)
> {
> long v;
> struct pvclock_gtod_data *gtod = &pvclock_gtod_data;
> @@ -1600,7 +1600,7 @@ static inline u64 vgettsc(cycle_t *cycle
> return v * gtod->clock.mult;
> }
>
> -static int do_monotonic_boot(s64 *t, cycle_t *cycle_now)
> +static int do_monotonic_boot(s64 *t, u64 *cycle_now)
> {
> struct pvclock_gtod_data *gtod = &pvclock_gtod_data;
> unsigned long seq;
> @@ -1621,7 +1621,7 @@ static int do_monotonic_boot(s64 *t, cyc
> }
>
> /* returns true if host is using tsc clocksource */
> -static bool kvm_get_time_and_clockread(s64 *kernel_ns, cycle_t *cycle_now)
> +static bool kvm_get_time_and_clockread(s64 *kernel_ns, u64 *cycle_now)
> {
> /* checked again under seqlock below */
> if (pvclock_gtod_data.clock.vclock_mode != VCLOCK_TSC)
> --- a/arch/x86/lguest/boot.c
> +++ b/arch/x86/lguest/boot.c
> @@ -930,7 +930,7 @@ static unsigned long lguest_tsc_khz(void
> * If we can't use the TSC, the kernel falls back to our lower-priority
> * "lguest_clock", where we read the time value given to us by the Host.
> */
> -static cycle_t lguest_clock_read(struct clocksource *cs)
> +static u64 lguest_clock_read(struct clocksource *cs)
> {
> unsigned long sec, nsec;
>
> --- a/arch/x86/platform/uv/uv_time.c
> +++ b/arch/x86/platform/uv/uv_time.c
> @@ -30,7 +30,7 @@
>
> #define RTC_NAME "sgi_rtc"
>
> -static cycle_t uv_read_rtc(struct clocksource *cs);
> +static u64 uv_read_rtc(struct clocksource *cs);
> static int uv_rtc_next_event(unsigned long, struct clock_event_device *);
> static int uv_rtc_shutdown(struct clock_event_device *evt);
>
> @@ -38,7 +38,7 @@ static struct clocksource clocksource_uv
> .name = RTC_NAME,
> .rating = 299,
> .read = uv_read_rtc,
> - .mask = (cycle_t)UVH_RTC_REAL_TIME_CLOCK_MASK,
> + .mask = (u64)UVH_RTC_REAL_TIME_CLOCK_MASK,
> .flags = CLOCK_SOURCE_IS_CONTINUOUS,
> };
>
> @@ -296,7 +296,7 @@ static int uv_rtc_unset_timer(int cpu, i
> * cachelines of it's own page. This allows faster simultaneous reads
> * from a given socket.
> */
> -static cycle_t uv_read_rtc(struct clocksource *cs)
> +static u64 uv_read_rtc(struct clocksource *cs)
> {
> unsigned long offset;
>
> @@ -305,7 +305,7 @@ static cycle_t uv_read_rtc(struct clocks
> else
> offset = (uv_blade_processor_id() * L1_CACHE_BYTES) % PAGE_SIZE;
>
> - return (cycle_t)uv_read_local_mmr(UVH_RTC | offset);
> + return (u64)uv_read_local_mmr(UVH_RTC | offset);
> }
>
> /*
> --- a/arch/x86/xen/time.c
> +++ b/arch/x86/xen/time.c
> @@ -39,10 +39,10 @@ static unsigned long xen_tsc_khz(void)
> return pvclock_tsc_khz(info);
> }
>
> -cycle_t xen_clocksource_read(void)
> +u64 xen_clocksource_read(void)
> {
> struct pvclock_vcpu_time_info *src;
> - cycle_t ret;
> + u64 ret;
>
> preempt_disable_notrace();
> src = &__this_cpu_read(xen_vcpu)->time;
> @@ -51,7 +51,7 @@ cycle_t xen_clocksource_read(void)
> return ret;
> }
>
> -static cycle_t xen_clocksource_get_cycles(struct clocksource *cs)
> +static u64 xen_clocksource_get_cycles(struct clocksource *cs)
> {
> return xen_clocksource_read();
> }
> --- a/arch/x86/xen/xen-ops.h
> +++ b/arch/x86/xen/xen-ops.h
> @@ -67,7 +67,7 @@ void xen_init_irq_ops(void);
> void xen_setup_timer(int cpu);
> void xen_setup_runstate_info(int cpu);
> void xen_teardown_timer(int cpu);
> -cycle_t xen_clocksource_read(void);
> +u64 xen_clocksource_read(void);
> void xen_setup_cpu_clockevents(void);
> void __init xen_init_time_ops(void);
> void __init xen_hvm_init_time_ops(void);
> --- a/arch/xtensa/kernel/time.c
> +++ b/arch/xtensa/kernel/time.c
> @@ -34,9 +34,9 @@
> unsigned long ccount_freq; /* ccount Hz */
> EXPORT_SYMBOL(ccount_freq);
>
> -static cycle_t ccount_read(struct clocksource *cs)
> +static u64 ccount_read(struct clocksource *cs)
> {
> - return (cycle_t)get_ccount();
> + return (u64)get_ccount();
> }
>
> static u64 notrace ccount_sched_clock_read(void)
> --- a/drivers/char/hpet.c
> +++ b/drivers/char/hpet.c
> @@ -69,9 +69,9 @@ static u32 hpet_nhpet, hpet_max_freq = H
> #ifdef CONFIG_IA64
> static void __iomem *hpet_mctr;
>
> -static cycle_t read_hpet(struct clocksource *cs)
> +static u64 read_hpet(struct clocksource *cs)
> {
> - return (cycle_t)read_counter((void __iomem *)hpet_mctr);
> + return (u64)read_counter((void __iomem *)hpet_mctr);
> }
>
> static struct clocksource clocksource_hpet = {
> --- a/drivers/clocksource/acpi_pm.c
> +++ b/drivers/clocksource/acpi_pm.c
> @@ -58,16 +58,16 @@ u32 acpi_pm_read_verified(void)
> return v2;
> }
>
> -static cycle_t acpi_pm_read(struct clocksource *cs)
> +static u64 acpi_pm_read(struct clocksource *cs)
> {
> - return (cycle_t)read_pmtmr();
> + return (u64)read_pmtmr();
> }
>
> static struct clocksource clocksource_acpi_pm = {
> .name = "acpi_pm",
> .rating = 200,
> .read = acpi_pm_read,
> - .mask = (cycle_t)ACPI_PM_MASK,
> + .mask = (u64)ACPI_PM_MASK,
> .flags = CLOCK_SOURCE_IS_CONTINUOUS,
> };
>
> @@ -81,9 +81,9 @@ static int __init acpi_pm_good_setup(cha
> }
> __setup("acpi_pm_good", acpi_pm_good_setup);
>
> -static cycle_t acpi_pm_read_slow(struct clocksource *cs)
> +static u64 acpi_pm_read_slow(struct clocksource *cs)
> {
> - return (cycle_t)acpi_pm_read_verified();
> + return (u64)acpi_pm_read_verified();
> }
>
> static inline void acpi_pm_need_workaround(void)
> @@ -145,7 +145,7 @@ DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_SE
> */
> static int verify_pmtmr_rate(void)
> {
> - cycle_t value1, value2;
> + u64 value1, value2;
> unsigned long count, delta;
>
> mach_prepare_counter();
> @@ -175,7 +175,7 @@ static int verify_pmtmr_rate(void)
>
> static int __init init_acpi_pm_clocksource(void)
> {
> - cycle_t value1, value2;
> + u64 value1, value2;
> unsigned int i, j = 0;
>
> if (!pmtmr_ioport)
> --- a/drivers/clocksource/arm_arch_timer.c
> +++ b/drivers/clocksource/arm_arch_timer.c
> @@ -561,12 +561,12 @@ static u64 arch_counter_get_cntvct_mem(v
> */
> u64 (*arch_timer_read_counter)(void) = arch_counter_get_cntvct;
>
> -static cycle_t arch_counter_read(struct clocksource *cs)
> +static u64 arch_counter_read(struct clocksource *cs)
> {
> return arch_timer_read_counter();
> }
>
> -static cycle_t arch_counter_read_cc(const struct cyclecounter *cc)
> +static u64 arch_counter_read_cc(const struct cyclecounter *cc)
> {
> return arch_timer_read_counter();
> }
> --- a/drivers/clocksource/arm_global_timer.c
> +++ b/drivers/clocksource/arm_global_timer.c
> @@ -195,7 +195,7 @@ static int gt_dying_cpu(unsigned int cpu
> return 0;
> }
>
> -static cycle_t gt_clocksource_read(struct clocksource *cs)
> +static u64 gt_clocksource_read(struct clocksource *cs)
> {
> return gt_counter_read();
> }
> --- a/drivers/clocksource/cadence_ttc_timer.c
> +++ b/drivers/clocksource/cadence_ttc_timer.c
> @@ -158,11 +158,11 @@ static irqreturn_t ttc_clock_event_inter
> *
> * returns: Current timer counter register value
> **/
> -static cycle_t __ttc_clocksource_read(struct clocksource *cs)
> +static u64 __ttc_clocksource_read(struct clocksource *cs)
> {
> struct ttc_timer *timer = &to_ttc_timer_clksrc(cs)->ttc;
>
> - return (cycle_t)readl_relaxed(timer->base_addr +
> + return (u64)readl_relaxed(timer->base_addr +
> TTC_COUNT_VAL_OFFSET);
> }
>
> --- a/drivers/clocksource/clksrc-dbx500-prcmu.c
> +++ b/drivers/clocksource/clksrc-dbx500-prcmu.c
> @@ -30,7 +30,7 @@
>
> static void __iomem *clksrc_dbx500_timer_base;
>
> -static cycle_t notrace clksrc_dbx500_prcmu_read(struct clocksource *cs)
> +static u64 notrace clksrc_dbx500_prcmu_read(struct clocksource *cs)
> {
> void __iomem *base = clksrc_dbx500_timer_base;
> u32 count, count2;
> --- a/drivers/clocksource/dw_apb_timer.c
> +++ b/drivers/clocksource/dw_apb_timer.c
> @@ -348,7 +348,7 @@ void dw_apb_clocksource_start(struct dw_
> dw_apb_clocksource_read(dw_cs);
> }
>
> -static cycle_t __apbt_read_clocksource(struct clocksource *cs)
> +static u64 __apbt_read_clocksource(struct clocksource *cs)
> {
> u32 current_count;
> struct dw_apb_clocksource *dw_cs =
> @@ -357,7 +357,7 @@ static cycle_t __apbt_read_clocksource(s
> current_count = apbt_readl_relaxed(&dw_cs->timer,
> APBTMR_N_CURRENT_VALUE);
>
> - return (cycle_t)~current_count;
> + return (u64)~current_count;
> }
>
> static void apbt_restart_clocksource(struct clocksource *cs)
> @@ -416,7 +416,7 @@ void dw_apb_clocksource_register(struct
> *
> * @dw_cs: The clocksource to read.
> */
> -cycle_t dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs)
> +u64 dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs)
> {
> - return (cycle_t)~apbt_readl(&dw_cs->timer, APBTMR_N_CURRENT_VALUE);
> + return (u64)~apbt_readl(&dw_cs->timer, APBTMR_N_CURRENT_VALUE);
> }
> --- a/drivers/clocksource/em_sti.c
> +++ b/drivers/clocksource/em_sti.c
> @@ -110,9 +110,9 @@ static void em_sti_disable(struct em_sti
> clk_disable_unprepare(p->clk);
> }
>
> -static cycle_t em_sti_count(struct em_sti_priv *p)
> +static u64 em_sti_count(struct em_sti_priv *p)
> {
> - cycle_t ticks;
> + u64 ticks;
> unsigned long flags;
>
> /* the STI hardware buffers the 48-bit count, but to
> @@ -121,14 +121,14 @@ static cycle_t em_sti_count(struct em_st
> * Always read STI_COUNT_H before STI_COUNT_L.
> */
> raw_spin_lock_irqsave(&p->lock, flags);
> - ticks = (cycle_t)(em_sti_read(p, STI_COUNT_H) & 0xffff) << 32;
> + ticks = (u64)(em_sti_read(p, STI_COUNT_H) & 0xffff) << 32;
> ticks |= em_sti_read(p, STI_COUNT_L);
> raw_spin_unlock_irqrestore(&p->lock, flags);
>
> return ticks;
> }
>
> -static cycle_t em_sti_set_next(struct em_sti_priv *p, cycle_t next)
> +static u64 em_sti_set_next(struct em_sti_priv *p, u64 next)
> {
> unsigned long flags;
>
> @@ -198,7 +198,7 @@ static struct em_sti_priv *cs_to_em_sti(
> return container_of(cs, struct em_sti_priv, cs);
> }
>
> -static cycle_t em_sti_clocksource_read(struct clocksource *cs)
> +static u64 em_sti_clocksource_read(struct clocksource *cs)
> {
> return em_sti_count(cs_to_em_sti(cs));
> }
> @@ -271,7 +271,7 @@ static int em_sti_clock_event_next(unsig
> struct clock_event_device *ced)
> {
> struct em_sti_priv *p = ced_to_em_sti(ced);
> - cycle_t next;
> + u64 next;
> int safe;
>
> next = em_sti_set_next(p, em_sti_count(p) + delta);
> --- a/drivers/clocksource/exynos_mct.c
> +++ b/drivers/clocksource/exynos_mct.c
> @@ -183,7 +183,7 @@ static u64 exynos4_read_count_64(void)
> hi2 = readl_relaxed(reg_base + EXYNOS4_MCT_G_CNT_U);
> } while (hi != hi2);
>
> - return ((cycle_t)hi << 32) | lo;
> + return ((u64)hi << 32) | lo;
> }
>
> /**
> @@ -199,7 +199,7 @@ static u32 notrace exynos4_read_count_32
> return readl_relaxed(reg_base + EXYNOS4_MCT_G_CNT_L);
> }
>
> -static cycle_t exynos4_frc_read(struct clocksource *cs)
> +static u64 exynos4_frc_read(struct clocksource *cs)
> {
> return exynos4_read_count_32();
> }
> @@ -266,7 +266,7 @@ static void exynos4_mct_comp0_stop(void)
> static void exynos4_mct_comp0_start(bool periodic, unsigned long cycles)
> {
> unsigned int tcon;
> - cycle_t comp_cycle;
> + u64 comp_cycle;
>
> tcon = readl_relaxed(reg_base + EXYNOS4_MCT_G_TCON);
>
> --- a/drivers/clocksource/h8300_timer16.c
> +++ b/drivers/clocksource/h8300_timer16.c
> @@ -72,7 +72,7 @@ static inline struct timer16_priv *cs_to
> return container_of(cs, struct timer16_priv, cs);
> }
>
> -static cycle_t timer16_clocksource_read(struct clocksource *cs)
> +static u64 timer16_clocksource_read(struct clocksource *cs)
> {
> struct timer16_priv *p = cs_to_priv(cs);
> unsigned long raw, value;
> --- a/drivers/clocksource/h8300_tpu.c
> +++ b/drivers/clocksource/h8300_tpu.c
> @@ -64,7 +64,7 @@ static inline struct tpu_priv *cs_to_pri
> return container_of(cs, struct tpu_priv, cs);
> }
>
> -static cycle_t tpu_clocksource_read(struct clocksource *cs)
> +static u64 tpu_clocksource_read(struct clocksource *cs)
> {
> struct tpu_priv *p = cs_to_priv(cs);
> unsigned long flags;
> --- a/drivers/clocksource/i8253.c
> +++ b/drivers/clocksource/i8253.c
> @@ -25,7 +25,7 @@ EXPORT_SYMBOL(i8253_lock);
> * to just read by itself. So use jiffies to emulate a free
> * running counter:
> */
> -static cycle_t i8253_read(struct clocksource *cs)
> +static u64 i8253_read(struct clocksource *cs)
> {
> static int old_count;
> static u32 old_jifs;
> @@ -83,7 +83,7 @@ static cycle_t i8253_read(struct clockso
>
> count = (PIT_LATCH - 1) - count;
>
> - return (cycle_t)(jifs * PIT_LATCH) + count;
> + return (u64)(jifs * PIT_LATCH) + count;
> }
>
> static struct clocksource i8253_cs = {
> --- a/drivers/clocksource/jcore-pit.c
> +++ b/drivers/clocksource/jcore-pit.c
> @@ -57,7 +57,7 @@ static notrace u64 jcore_sched_clock_rea
> return seclo * NSEC_PER_SEC + nsec;
> }
>
> -static cycle_t jcore_clocksource_read(struct clocksource *cs)
> +static u64 jcore_clocksource_read(struct clocksource *cs)
> {
> return jcore_sched_clock_read();
> }
> --- a/drivers/clocksource/metag_generic.c
> +++ b/drivers/clocksource/metag_generic.c
> @@ -56,7 +56,7 @@ static int metag_timer_set_next_event(un
> return 0;
> }
>
> -static cycle_t metag_clocksource_read(struct clocksource *cs)
> +static u64 metag_clocksource_read(struct clocksource *cs)
> {
> return __core_reg_get(TXTIMER);
> }
> --- a/drivers/clocksource/mips-gic-timer.c
> +++ b/drivers/clocksource/mips-gic-timer.c
> @@ -125,7 +125,7 @@ static int gic_clockevent_init(void)
> return 0;
> }
>
> -static cycle_t gic_hpt_read(struct clocksource *cs)
> +static u64 gic_hpt_read(struct clocksource *cs)
> {
> return gic_read_count();
> }
> --- a/drivers/clocksource/mmio.c
> +++ b/drivers/clocksource/mmio.c
> @@ -20,24 +20,24 @@ static inline struct clocksource_mmio *t
> return container_of(c, struct clocksource_mmio, clksrc);
> }
>
> -cycle_t clocksource_mmio_readl_up(struct clocksource *c)
> +u64 clocksource_mmio_readl_up(struct clocksource *c)
> {
> - return (cycle_t)readl_relaxed(to_mmio_clksrc(c)->reg);
> + return (u64)readl_relaxed(to_mmio_clksrc(c)->reg);
> }
>
> -cycle_t clocksource_mmio_readl_down(struct clocksource *c)
> +u64 clocksource_mmio_readl_down(struct clocksource *c)
> {
> - return ~(cycle_t)readl_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
> + return ~(u64)readl_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
> }
>
> -cycle_t clocksource_mmio_readw_up(struct clocksource *c)
> +u64 clocksource_mmio_readw_up(struct clocksource *c)
> {
> - return (cycle_t)readw_relaxed(to_mmio_clksrc(c)->reg);
> + return (u64)readw_relaxed(to_mmio_clksrc(c)->reg);
> }
>
> -cycle_t clocksource_mmio_readw_down(struct clocksource *c)
> +u64 clocksource_mmio_readw_down(struct clocksource *c)
> {
> - return ~(cycle_t)readw_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
> + return ~(u64)readw_relaxed(to_mmio_clksrc(c)->reg) & c->mask;
> }
>
> /**
> @@ -51,7 +51,7 @@ cycle_t clocksource_mmio_readw_down(stru
> */
> int __init clocksource_mmio_init(void __iomem *base, const char *name,
> unsigned long hz, int rating, unsigned bits,
> - cycle_t (*read)(struct clocksource *))
> + u64 (*read)(struct clocksource *))
> {
> struct clocksource_mmio *cs;
>
> --- a/drivers/clocksource/mxs_timer.c
> +++ b/drivers/clocksource/mxs_timer.c
> @@ -97,7 +97,7 @@ static void timrot_irq_acknowledge(void)
> HW_TIMROT_TIMCTRLn(0) + STMP_OFFSET_REG_CLR);
> }
>
> -static cycle_t timrotv1_get_cycles(struct clocksource *cs)
> +static u64 timrotv1_get_cycles(struct clocksource *cs)
> {
> return ~((__raw_readl(mxs_timrot_base + HW_TIMROT_TIMCOUNTn(1))
> & 0xffff0000) >> 16);
> --- a/drivers/clocksource/qcom-timer.c
> +++ b/drivers/clocksource/qcom-timer.c
> @@ -89,7 +89,7 @@ static struct clock_event_device __percp
>
> static void __iomem *source_base;
>
> -static notrace cycle_t msm_read_timer_count(struct clocksource *cs)
> +static notrace u64 msm_read_timer_count(struct clocksource *cs)
> {
> return readl_relaxed(source_base + TIMER_COUNT_VAL);
> }
> --- a/drivers/clocksource/samsung_pwm_timer.c
> +++ b/drivers/clocksource/samsung_pwm_timer.c
> @@ -307,7 +307,7 @@ static void samsung_clocksource_resume(s
> samsung_time_start(pwm.source_id, true);
> }
>
> -static cycle_t notrace samsung_clocksource_read(struct clocksource *c)
> +static u64 notrace samsung_clocksource_read(struct clocksource *c)
> {
> return ~readl_relaxed(pwm.source_reg);
> }
> --- a/drivers/clocksource/scx200_hrt.c
> +++ b/drivers/clocksource/scx200_hrt.c
> @@ -43,10 +43,10 @@ MODULE_PARM_DESC(ppm, "+-adjust to actua
> /* The base timer frequency, * 27 if selected */
> #define HRT_FREQ 1000000
>
> -static cycle_t read_hrt(struct clocksource *cs)
> +static u64 read_hrt(struct clocksource *cs)
> {
> /* Read the timer value */
> - return (cycle_t) inl(scx200_cb_base + SCx200_TIMER_OFFSET);
> + return (u64) inl(scx200_cb_base + SCx200_TIMER_OFFSET);
> }
>
> static struct clocksource cs_hrt = {
> --- a/drivers/clocksource/sh_cmt.c
> +++ b/drivers/clocksource/sh_cmt.c
> @@ -612,7 +612,7 @@ static struct sh_cmt_channel *cs_to_sh_c
> return container_of(cs, struct sh_cmt_channel, cs);
> }
>
> -static cycle_t sh_cmt_clocksource_read(struct clocksource *cs)
> +static u64 sh_cmt_clocksource_read(struct clocksource *cs)
> {
> struct sh_cmt_channel *ch = cs_to_sh_cmt(cs);
> unsigned long flags, raw;
> --- a/drivers/clocksource/sh_tmu.c
> +++ b/drivers/clocksource/sh_tmu.c
> @@ -255,7 +255,7 @@ static struct sh_tmu_channel *cs_to_sh_t
> return container_of(cs, struct sh_tmu_channel, cs);
> }
>
> -static cycle_t sh_tmu_clocksource_read(struct clocksource *cs)
> +static u64 sh_tmu_clocksource_read(struct clocksource *cs)
> {
> struct sh_tmu_channel *ch = cs_to_sh_tmu(cs);
>
> --- a/drivers/clocksource/tcb_clksrc.c
> +++ b/drivers/clocksource/tcb_clksrc.c
> @@ -41,7 +41,7 @@
>
> static void __iomem *tcaddr;
>
> -static cycle_t tc_get_cycles(struct clocksource *cs)
> +static u64 tc_get_cycles(struct clocksource *cs)
> {
> unsigned long flags;
> u32 lower, upper;
> @@ -56,7 +56,7 @@ static cycle_t tc_get_cycles(struct cloc
> return (upper << 16) | lower;
> }
>
> -static cycle_t tc_get_cycles32(struct clocksource *cs)
> +static u64 tc_get_cycles32(struct clocksource *cs)
> {
> return __raw_readl(tcaddr + ATMEL_TC_REG(0, CV));
> }
> --- a/drivers/clocksource/time-pistachio.c
> +++ b/drivers/clocksource/time-pistachio.c
> @@ -67,7 +67,7 @@ static inline void gpt_writel(void __iom
> writel(value, base + 0x20 * gpt_id + offset);
> }
>
> -static cycle_t notrace
> +static u64 notrace
> pistachio_clocksource_read_cycles(struct clocksource *cs)
> {
> struct pistachio_clocksource *pcs = to_pistachio_clocksource(cs);
> @@ -84,7 +84,7 @@ pistachio_clocksource_read_cycles(struct
> counter = gpt_readl(pcs->base, TIMER_CURRENT_VALUE, 0);
> raw_spin_unlock_irqrestore(&pcs->lock, flags);
>
> - return (cycle_t)~counter;
> + return (u64)~counter;
> }
>
> static u64 notrace pistachio_read_sched_clock(void)
> --- a/drivers/clocksource/timer-atlas7.c
> +++ b/drivers/clocksource/timer-atlas7.c
> @@ -85,7 +85,7 @@ static irqreturn_t sirfsoc_timer_interru
> }
>
> /* read 64-bit timer counter */
> -static cycle_t sirfsoc_timer_read(struct clocksource *cs)
> +static u64 sirfsoc_timer_read(struct clocksource *cs)
> {
> u64 cycles;
>
> --- a/drivers/clocksource/timer-atmel-pit.c
> +++ b/drivers/clocksource/timer-atmel-pit.c
> @@ -73,7 +73,7 @@ static inline void pit_write(void __iome
> * Clocksource: just a monotonic counter of MCK/16 cycles.
> * We don't care whether or not PIT irqs are enabled.
> */
> -static cycle_t read_pit_clk(struct clocksource *cs)
> +static u64 read_pit_clk(struct clocksource *cs)
> {
> struct pit_data *data = clksrc_to_pit_data(cs);
> unsigned long flags;
> --- a/drivers/clocksource/timer-atmel-st.c
> +++ b/drivers/clocksource/timer-atmel-st.c
> @@ -92,7 +92,7 @@ static irqreturn_t at91rm9200_timer_inte
> return IRQ_NONE;
> }
>
> -static cycle_t read_clk32k(struct clocksource *cs)
> +static u64 read_clk32k(struct clocksource *cs)
> {
> return read_CRTR();
> }
> --- a/drivers/clocksource/timer-nps.c
> +++ b/drivers/clocksource/timer-nps.c
> @@ -48,11 +48,11 @@ static void *nps_msu_reg_low_addr[NPS_CL
>
> static unsigned long nps_timer_rate;
>
> -static cycle_t nps_clksrc_read(struct clocksource *clksrc)
> +static u64 nps_clksrc_read(struct clocksource *clksrc)
> {
> int cluster = raw_smp_processor_id() >> NPS_CLUSTER_OFFSET;
>
> - return (cycle_t)ioread32be(nps_msu_reg_low_addr[cluster]);
> + return (u64)ioread32be(nps_msu_reg_low_addr[cluster]);
> }
>
> static int __init nps_setup_clocksource(struct device_node *node,
> --- a/drivers/clocksource/timer-prima2.c
> +++ b/drivers/clocksource/timer-prima2.c
> @@ -72,7 +72,7 @@ static irqreturn_t sirfsoc_timer_interru
> }
>
> /* read 64-bit timer counter */
> -static cycle_t notrace sirfsoc_timer_read(struct clocksource *cs)
> +static u64 notrace sirfsoc_timer_read(struct clocksource *cs)
> {
> u64 cycles;
>
> --- a/drivers/clocksource/timer-sun5i.c
> +++ b/drivers/clocksource/timer-sun5i.c
> @@ -152,7 +152,7 @@ static irqreturn_t sun5i_timer_interrupt
> return IRQ_HANDLED;
> }
>
> -static cycle_t sun5i_clksrc_read(struct clocksource *clksrc)
> +static u64 sun5i_clksrc_read(struct clocksource *clksrc)
> {
> struct sun5i_timer_clksrc *cs = to_sun5i_timer_clksrc(clksrc);
>
> --- a/drivers/clocksource/timer-ti-32k.c
> +++ b/drivers/clocksource/timer-ti-32k.c
> @@ -65,11 +65,11 @@ static inline struct ti_32k *to_ti_32k(s
> return container_of(cs, struct ti_32k, cs);
> }
>
> -static cycle_t notrace ti_32k_read_cycles(struct clocksource *cs)
> +static u64 notrace ti_32k_read_cycles(struct clocksource *cs)
> {
> struct ti_32k *ti = to_ti_32k(cs);
>
> - return (cycle_t)readl_relaxed(ti->counter);
> + return (u64)readl_relaxed(ti->counter);
> }
>
> static struct ti_32k ti_32k_timer = {
> --- a/drivers/clocksource/vt8500_timer.c
> +++ b/drivers/clocksource/vt8500_timer.c
> @@ -53,7 +53,7 @@
>
> static void __iomem *regbase;
>
> -static cycle_t vt8500_timer_read(struct clocksource *cs)
> +static u64 vt8500_timer_read(struct clocksource *cs)
> {
> int loops = msecs_to_loops(10);
> writel(3, regbase + TIMER_CTRL_VAL);
> @@ -75,7 +75,7 @@ static int vt8500_timer_set_next_event(u
> struct clock_event_device *evt)
> {
> int loops = msecs_to_loops(10);
> - cycle_t alarm = clocksource.read(&clocksource) + cycles;
> + u64 alarm = clocksource.read(&clocksource) + cycles;
> while ((readl(regbase + TIMER_AS_VAL) & TIMER_MATCH_W_ACTIVE)
> && --loops)
> cpu_relax();
> --- a/drivers/hv/hv.c
> +++ b/drivers/hv/hv.c
> @@ -135,9 +135,9 @@ u64 hv_do_hypercall(u64 control, void *i
> EXPORT_SYMBOL_GPL(hv_do_hypercall);
>
> #ifdef CONFIG_X86_64
> -static cycle_t read_hv_clock_tsc(struct clocksource *arg)
> +static u64 read_hv_clock_tsc(struct clocksource *arg)
> {
> - cycle_t current_tick;
> + u64 current_tick;
> struct ms_hyperv_tsc_page *tsc_pg = hv_context.tsc_page;
>
> if (tsc_pg->tsc_sequence != 0) {
> @@ -146,7 +146,7 @@ static cycle_t read_hv_clock_tsc(struct
> */
>
> while (1) {
> - cycle_t tmp;
> + u64 tmp;
> u32 sequence = tsc_pg->tsc_sequence;
> u64 cur_tsc;
> u64 scale = tsc_pg->tsc_scale;
> @@ -350,7 +350,7 @@ int hv_post_message(union hv_connection_
> static int hv_ce_set_next_event(unsigned long delta,
> struct clock_event_device *evt)
> {
> - cycle_t current_tick;
> + u64 current_tick;
>
> WARN_ON(!clockevent_state_oneshot(evt));
>
> --- a/drivers/irqchip/irq-mips-gic.c
> +++ b/drivers/irqchip/irq-mips-gic.c
> @@ -152,12 +152,12 @@ static inline void gic_map_to_vpe(unsign
> }
>
> #ifdef CONFIG_CLKSRC_MIPS_GIC
> -cycle_t gic_read_count(void)
> +u64 gic_read_count(void)
> {
> unsigned int hi, hi2, lo;
>
> if (mips_cm_is64)
> - return (cycle_t)gic_read(GIC_REG(SHARED, GIC_SH_COUNTER));
> + return (u64)gic_read(GIC_REG(SHARED, GIC_SH_COUNTER));
>
> do {
> hi = gic_read32(GIC_REG(SHARED, GIC_SH_COUNTER_63_32));
> @@ -165,7 +165,7 @@ cycle_t gic_read_count(void)
> hi2 = gic_read32(GIC_REG(SHARED, GIC_SH_COUNTER_63_32));
> } while (hi2 != hi);
>
> - return (((cycle_t) hi) << 32) + lo;
> + return (((u64) hi) << 32) + lo;
> }
>
> unsigned int gic_get_count_width(void)
> @@ -179,7 +179,7 @@ unsigned int gic_get_count_width(void)
> return bits;
> }
>
> -void gic_write_compare(cycle_t cnt)
> +void gic_write_compare(u64 cnt)
> {
> if (mips_cm_is64) {
> gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE), cnt);
> @@ -191,7 +191,7 @@ void gic_write_compare(cycle_t cnt)
> }
> }
>
> -void gic_write_cpu_compare(cycle_t cnt, int cpu)
> +void gic_write_cpu_compare(u64 cnt, int cpu)
> {
> unsigned long flags;
>
> @@ -211,17 +211,17 @@ void gic_write_cpu_compare(cycle_t cnt,
> local_irq_restore(flags);
> }
>
> -cycle_t gic_read_compare(void)
> +u64 gic_read_compare(void)
> {
> unsigned int hi, lo;
>
> if (mips_cm_is64)
> - return (cycle_t)gic_read(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE));
> + return (u64)gic_read(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE));
>
> hi = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_HI));
> lo = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_LO));
>
> - return (((cycle_t) hi) << 32) + lo;
> + return (((u64) hi) << 32) + lo;
> }
>
> void gic_start_count(void)
> --- a/drivers/net/ethernet/amd/xgbe/xgbe-ptp.c
> +++ b/drivers/net/ethernet/amd/xgbe/xgbe-ptp.c
> @@ -122,7 +122,7 @@
> #include "xgbe.h"
> #include "xgbe-common.h"
>
> -static cycle_t xgbe_cc_read(const struct cyclecounter *cc)
> +static u64 xgbe_cc_read(const struct cyclecounter *cc)
> {
> struct xgbe_prv_data *pdata = container_of(cc,
> struct xgbe_prv_data,
> --- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
> +++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
> @@ -15219,7 +15219,7 @@ void bnx2x_set_rx_ts(struct bnx2x *bp, s
> }
>
> /* Read the PHC */
> -static cycle_t bnx2x_cyclecounter_read(const struct cyclecounter *cc)
> +static u64 bnx2x_cyclecounter_read(const struct cyclecounter *cc)
> {
> struct bnx2x *bp = container_of(cc, struct bnx2x, cyclecounter);
> int port = BP_PORT(bp);
> --- a/drivers/net/ethernet/freescale/fec_ptp.c
> +++ b/drivers/net/ethernet/freescale/fec_ptp.c
> @@ -230,7 +230,7 @@ static int fec_ptp_enable_pps(struct fec
> * cyclecounter structure used to construct a ns counter from the
> * arbitrary fixed point registers
> */
> -static cycle_t fec_ptp_read(const struct cyclecounter *cc)
> +static u64 fec_ptp_read(const struct cyclecounter *cc)
> {
> struct fec_enet_private *fep =
> container_of(cc, struct fec_enet_private, cc);
> --- a/drivers/net/ethernet/intel/e1000e/netdev.c
> +++ b/drivers/net/ethernet/intel/e1000e/netdev.c
> @@ -4305,24 +4305,24 @@ void e1000e_reinit_locked(struct e1000_a
> /**
> * e1000e_sanitize_systim - sanitize raw cycle counter reads
> * @hw: pointer to the HW structure
> - * @systim: cycle_t value read, sanitized and returned
> + * @systim: u64 timestamp value read, sanitized and returned
> *
> * Errata for 82574/82583 possible bad bits read from SYSTIMH/L:
> * check to see that the time is incrementing at a reasonable
> * rate and is a multiple of incvalue.
> **/
> -static cycle_t e1000e_sanitize_systim(struct e1000_hw *hw, cycle_t systim)
> +static u64 e1000e_sanitize_systim(struct e1000_hw *hw, u64 systim)
> {
> u64 time_delta, rem, temp;
> - cycle_t systim_next;
> + u64 systim_next;
> u32 incvalue;
> int i;
>
> incvalue = er32(TIMINCA) & E1000_TIMINCA_INCVALUE_MASK;
> for (i = 0; i < E1000_MAX_82574_SYSTIM_REREADS; i++) {
> /* latch SYSTIMH on read of SYSTIML */
> - systim_next = (cycle_t)er32(SYSTIML);
> - systim_next |= (cycle_t)er32(SYSTIMH) << 32;
> + systim_next = (u64)er32(SYSTIML);
> + systim_next |= (u64)er32(SYSTIMH) << 32;
>
> time_delta = systim_next - systim;
> temp = time_delta;
> @@ -4342,13 +4342,13 @@ static cycle_t e1000e_sanitize_systim(st
> * e1000e_cyclecounter_read - read raw cycle counter (used by time counter)
> * @cc: cyclecounter structure
> **/
> -static cycle_t e1000e_cyclecounter_read(const struct cyclecounter *cc)
> +static u64 e1000e_cyclecounter_read(const struct cyclecounter *cc)
> {
> struct e1000_adapter *adapter = container_of(cc, struct e1000_adapter,
> cc);
> struct e1000_hw *hw = &adapter->hw;
> u32 systimel, systimeh;
> - cycle_t systim;
> + u64 systim;
> /* SYSTIMH latching upon SYSTIML read does not work well.
> * This means that if SYSTIML overflows after we read it but before
> * we read SYSTIMH, the value of SYSTIMH has been incremented and we
> @@ -4368,8 +4368,8 @@ static cycle_t e1000e_cyclecounter_read(
> systimel = systimel_2;
> }
> }
> - systim = (cycle_t)systimel;
> - systim |= (cycle_t)systimeh << 32;
> + systim = (u64)systimel;
> + systim |= (u64)systimeh << 32;
>
> if (adapter->flags2 & FLAG2_CHECK_SYSTIM_OVERFLOW)
> systim = e1000e_sanitize_systim(hw, systim);
> --- a/drivers/net/ethernet/intel/e1000e/ptp.c
> +++ b/drivers/net/ethernet/intel/e1000e/ptp.c
> @@ -127,8 +127,8 @@ static int e1000e_phc_get_syncdevicetime
> unsigned long flags;
> int i;
> u32 tsync_ctrl;
> - cycle_t dev_cycles;
> - cycle_t sys_cycles;
> + u64 dev_cycles;
> + u64 sys_cycles;
>
> tsync_ctrl = er32(TSYNCTXCTL);
> tsync_ctrl |= E1000_TSYNCTXCTL_START_SYNC |
> --- a/drivers/net/ethernet/intel/igb/igb_ptp.c
> +++ b/drivers/net/ethernet/intel/igb/igb_ptp.c
> @@ -77,7 +77,7 @@
> static void igb_ptp_tx_hwtstamp(struct igb_adapter *adapter);
>
> /* SYSTIM read access for the 82576 */
> -static cycle_t igb_ptp_read_82576(const struct cyclecounter *cc)
> +static u64 igb_ptp_read_82576(const struct cyclecounter *cc)
> {
> struct igb_adapter *igb = container_of(cc, struct igb_adapter, cc);
> struct e1000_hw *hw = &igb->hw;
> @@ -94,7 +94,7 @@ static cycle_t igb_ptp_read_82576(const
> }
>
> /* SYSTIM read access for the 82580 */
> -static cycle_t igb_ptp_read_82580(const struct cyclecounter *cc)
> +static u64 igb_ptp_read_82580(const struct cyclecounter *cc)
> {
> struct igb_adapter *igb = container_of(cc, struct igb_adapter, cc);
> struct e1000_hw *hw = &igb->hw;
> --- a/drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c
> +++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c
> @@ -245,7 +245,7 @@ static void ixgbe_ptp_setup_sdp_x540(str
> * result of SYSTIME is 32bits of "billions of cycles" and 32 bits of
> * "cycles", rather than seconds and nanoseconds.
> */
> -static cycle_t ixgbe_ptp_read_X550(const struct cyclecounter *hw_cc)
> +static u64 ixgbe_ptp_read_X550(const struct cyclecounter *hw_cc)
> {
> struct ixgbe_adapter *adapter =
> container_of(hw_cc, struct ixgbe_adapter, hw_cc);
> @@ -282,7 +282,7 @@ static cycle_t ixgbe_ptp_read_X550(const
> * cyclecounter structure used to construct a ns counter from the
> * arbitrary fixed point registers
> */
> -static cycle_t ixgbe_ptp_read_82599(const struct cyclecounter *cc)
> +static u64 ixgbe_ptp_read_82599(const struct cyclecounter *cc)
> {
> struct ixgbe_adapter *adapter =
> container_of(cc, struct ixgbe_adapter, hw_cc);
> --- a/drivers/net/ethernet/mellanox/mlx4/en_clock.c
> +++ b/drivers/net/ethernet/mellanox/mlx4/en_clock.c
> @@ -38,7 +38,7 @@
>
> /* mlx4_en_read_clock - read raw cycle counter (to be used by time counter)
> */
> -static cycle_t mlx4_en_read_clock(const struct cyclecounter *tc)
> +static u64 mlx4_en_read_clock(const struct cyclecounter *tc)
> {
> struct mlx4_en_dev *mdev =
> container_of(tc, struct mlx4_en_dev, cycles);
> --- a/drivers/net/ethernet/mellanox/mlx4/main.c
> +++ b/drivers/net/ethernet/mellanox/mlx4/main.c
> @@ -1823,10 +1823,10 @@ static void unmap_bf_area(struct mlx4_de
> io_mapping_free(mlx4_priv(dev)->bf_mapping);
> }
>
> -cycle_t mlx4_read_clock(struct mlx4_dev *dev)
> +u64 mlx4_read_clock(struct mlx4_dev *dev)
> {
> u32 clockhi, clocklo, clockhi1;
> - cycle_t cycles;
> + u64 cycles;
> int i;
> struct mlx4_priv *priv = mlx4_priv(dev);
>
> --- a/drivers/net/ethernet/mellanox/mlx5/core/en_clock.c
> +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_clock.c
> @@ -49,7 +49,7 @@ void mlx5e_fill_hwstamp(struct mlx5e_tst
> hwts->hwtstamp = ns_to_ktime(nsec);
> }
>
> -static cycle_t mlx5e_read_internal_timer(const struct cyclecounter *cc)
> +static u64 mlx5e_read_internal_timer(const struct cyclecounter *cc)
> {
> struct mlx5e_tstamp *tstamp = container_of(cc, struct mlx5e_tstamp,
> cycles);
> --- a/drivers/net/ethernet/mellanox/mlx5/core/main.c
> +++ b/drivers/net/ethernet/mellanox/mlx5/core/main.c
> @@ -522,7 +522,7 @@ int mlx5_core_disable_hca(struct mlx5_co
> return mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out));
> }
>
> -cycle_t mlx5_read_internal_timer(struct mlx5_core_dev *dev)
> +u64 mlx5_read_internal_timer(struct mlx5_core_dev *dev)
> {
> u32 timer_h, timer_h1, timer_l;
>
> @@ -532,7 +532,7 @@ cycle_t mlx5_read_internal_timer(struct
> if (timer_h != timer_h1) /* wrap around */
> timer_l = ioread32be(&dev->iseg->internal_timer_l);
>
> - return (cycle_t)timer_l | (cycle_t)timer_h1 << 32;
> + return (u64)timer_l | (u64)timer_h1 << 32;
> }
>
> static int mlx5_irq_set_affinity_hint(struct mlx5_core_dev *mdev, int i)
> --- a/drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
> +++ b/drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
> @@ -93,7 +93,7 @@ bool mlx5_sriov_is_enabled(struct mlx5_c
> int mlx5_core_enable_hca(struct mlx5_core_dev *dev, u16 func_id);
> int mlx5_core_disable_hca(struct mlx5_core_dev *dev, u16 func_id);
> int mlx5_wait_for_vf_pages(struct mlx5_core_dev *dev);
> -cycle_t mlx5_read_internal_timer(struct mlx5_core_dev *dev);
> +u64 mlx5_read_internal_timer(struct mlx5_core_dev *dev);
> u32 mlx5_get_msix_vec(struct mlx5_core_dev *dev, int vecidx);
> struct mlx5_eq *mlx5_eqn2eq(struct mlx5_core_dev *dev, int eqn);
> void mlx5_cq_tasklet_cb(unsigned long data);
> --- a/drivers/net/ethernet/ti/cpts.c
> +++ b/drivers/net/ethernet/ti/cpts.c
> @@ -101,7 +101,7 @@ static int cpts_fifo_read(struct cpts *c
> return type == match ? 0 : -1;
> }
>
> -static cycle_t cpts_systim_read(const struct cyclecounter *cc)
> +static u64 cpts_systim_read(const struct cyclecounter *cc)
> {
> u64 val = 0;
> struct cpts_event *event;
> --- a/include/kvm/arm_arch_timer.h
> +++ b/include/kvm/arm_arch_timer.h
> @@ -25,13 +25,13 @@
>
> struct arch_timer_kvm {
> /* Virtual offset */
> - cycle_t cntvoff;
> + u64 cntvoff;
> };
>
> struct arch_timer_cpu {
> /* Registers: control register, timer value */
> u32 cntv_ctl; /* Saved/restored */
> - cycle_t cntv_cval; /* Saved/restored */
> + u64 cntv_cval; /* Saved/restored */
>
> /*
> * Anything that is not used directly from assembly code goes
> --- a/include/linux/clocksource.h
> +++ b/include/linux/clocksource.h
> @@ -75,8 +75,8 @@ struct module;
> * structure.
> */
> struct clocksource {
> - cycle_t (*read)(struct clocksource *cs);
> - cycle_t mask;
> + u64 (*read)(struct clocksource *cs);
> + u64 mask;
> u32 mult;
> u32 shift;
> u64 max_idle_ns;
> @@ -98,8 +98,8 @@ struct clocksource {
> #ifdef CONFIG_CLOCKSOURCE_WATCHDOG
> /* Watchdog related data, used by the framework */
> struct list_head wd_list;
> - cycle_t cs_last;
> - cycle_t wd_last;
> + u64 cs_last;
> + u64 wd_last;
> #endif
> struct module *owner;
> };
> @@ -117,7 +117,7 @@ struct clocksource {
> #define CLOCK_SOURCE_RESELECT 0x100
>
> /* simplify initialization of mask field */
> -#define CLOCKSOURCE_MASK(bits) (cycle_t)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
> +#define CLOCKSOURCE_MASK(bits) (u64)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
>
> static inline u32 clocksource_freq2mult(u32 freq, u32 shift_constant, u64 from)
> {
> @@ -173,7 +173,7 @@ static inline u32 clocksource_hz2mult(u3
> *
> * XXX - This could use some mult_lxl_ll() asm optimization
> */
> -static inline s64 clocksource_cyc2ns(cycle_t cycles, u32 mult, u32 shift)
> +static inline s64 clocksource_cyc2ns(u64 cycles, u32 mult, u32 shift)
> {
> return ((u64) cycles * mult) >> shift;
> }
> @@ -233,13 +233,13 @@ static inline void __clocksource_update_
>
> extern int timekeeping_notify(struct clocksource *clock);
>
> -extern cycle_t clocksource_mmio_readl_up(struct clocksource *);
> -extern cycle_t clocksource_mmio_readl_down(struct clocksource *);
> -extern cycle_t clocksource_mmio_readw_up(struct clocksource *);
> -extern cycle_t clocksource_mmio_readw_down(struct clocksource *);
> +extern u64 clocksource_mmio_readl_up(struct clocksource *);
> +extern u64 clocksource_mmio_readl_down(struct clocksource *);
> +extern u64 clocksource_mmio_readw_up(struct clocksource *);
> +extern u64 clocksource_mmio_readw_down(struct clocksource *);
>
> extern int clocksource_mmio_init(void __iomem *, const char *,
> - unsigned long, int, unsigned, cycle_t (*)(struct clocksource *));
> + unsigned long, int, unsigned, u64 (*)(struct clocksource *));
>
> extern int clocksource_i8253_init(void);
>
> --- a/include/linux/dw_apb_timer.h
> +++ b/include/linux/dw_apb_timer.h
> @@ -50,6 +50,6 @@ dw_apb_clocksource_init(unsigned rating,
> unsigned long freq);
> void dw_apb_clocksource_register(struct dw_apb_clocksource *dw_cs);
> void dw_apb_clocksource_start(struct dw_apb_clocksource *dw_cs);
> -cycle_t dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs);
> +u64 dw_apb_clocksource_read(struct dw_apb_clocksource *dw_cs);
>
> #endif /* __DW_APB_TIMER_H__ */
> --- a/include/linux/irqchip/mips-gic.h
> +++ b/include/linux/irqchip/mips-gic.h
> @@ -259,11 +259,11 @@ extern void gic_init(unsigned long gic_b
> unsigned long gic_addrspace_size, unsigned int cpu_vec,
> unsigned int irqbase);
> extern void gic_clocksource_init(unsigned int);
> -extern cycle_t gic_read_count(void);
> +extern u64 gic_read_count(void);
> extern unsigned int gic_get_count_width(void);
> -extern cycle_t gic_read_compare(void);
> -extern void gic_write_compare(cycle_t cnt);
> -extern void gic_write_cpu_compare(cycle_t cnt, int cpu);
> +extern u64 gic_read_compare(void);
> +extern void gic_write_compare(u64 cnt);
> +extern void gic_write_cpu_compare(u64 cnt, int cpu);
> extern void gic_start_count(void);
> extern void gic_stop_count(void);
> extern int gic_get_c0_compare_int(void);
> --- a/include/linux/mlx4/device.h
> +++ b/include/linux/mlx4/device.h
> @@ -1461,7 +1461,7 @@ int mlx4_get_roce_gid_from_slave(struct
> int mlx4_FLOW_STEERING_IB_UC_QP_RANGE(struct mlx4_dev *dev, u32 min_range_qpn,
> u32 max_range_qpn);
>
> -cycle_t mlx4_read_clock(struct mlx4_dev *dev);
> +u64 mlx4_read_clock(struct mlx4_dev *dev);
>
> struct mlx4_active_ports {
> DECLARE_BITMAP(ports, MLX4_MAX_PORTS);
> --- a/include/linux/timecounter.h
> +++ b/include/linux/timecounter.h
> @@ -20,7 +20,7 @@
> #include <linux/types.h>
>
> /* simplify initialization of mask field */
> -#define CYCLECOUNTER_MASK(bits) (cycle_t)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
> +#define CYCLECOUNTER_MASK(bits) (u64)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
>
> /**
> * struct cyclecounter - hardware abstraction for a free running counter
> @@ -37,8 +37,8 @@
> * @shift: cycle to nanosecond divisor (power of two)
> */
> struct cyclecounter {
> - cycle_t (*read)(const struct cyclecounter *cc);
> - cycle_t mask;
> + u64 (*read)(const struct cyclecounter *cc);
> + u64 mask;
> u32 mult;
> u32 shift;
> };
> @@ -63,7 +63,7 @@ struct cyclecounter {
> */
> struct timecounter {
> const struct cyclecounter *cc;
> - cycle_t cycle_last;
> + u64 cycle_last;
> u64 nsec;
> u64 mask;
> u64 frac;
> @@ -77,7 +77,7 @@ struct timecounter {
> * @frac: pointer to storage for the fractional nanoseconds.
> */
> static inline u64 cyclecounter_cyc2ns(const struct cyclecounter *cc,
> - cycle_t cycles, u64 mask, u64 *frac)
> + u64 cycles, u64 mask, u64 *frac)
> {
> u64 ns = (u64) cycles;
>
> @@ -134,6 +134,6 @@ extern u64 timecounter_read(struct timec
> * in the past.
> */
> extern u64 timecounter_cyc2time(struct timecounter *tc,
> - cycle_t cycle_tstamp);
> + u64 cycle_tstamp);
>
> #endif
> --- a/include/linux/timekeeper_internal.h
> +++ b/include/linux/timekeeper_internal.h
> @@ -29,9 +29,9 @@
> */
> struct tk_read_base {
> struct clocksource *clock;
> - cycle_t (*read)(struct clocksource *cs);
> - cycle_t mask;
> - cycle_t cycle_last;
> + u64 (*read)(struct clocksource *cs);
> + u64 mask;
> + u64 cycle_last;
> u32 mult;
> u32 shift;
> u64 xtime_nsec;
> @@ -97,7 +97,7 @@ struct timekeeper {
> struct timespec64 raw_time;
>
> /* The following members are for timekeeping internal use */
> - cycle_t cycle_interval;
> + u64 cycle_interval;
> u64 xtime_interval;
> s64 xtime_remainder;
> u32 raw_interval;
> @@ -136,7 +136,7 @@ extern void update_vsyscall_tz(void);
>
> extern void update_vsyscall_old(struct timespec *ts, struct timespec *wtm,
> struct clocksource *c, u32 mult,
> - cycle_t cycle_last);
> + u64 cycle_last);
> extern void update_vsyscall_tz(void);
>
> #else
> --- a/include/linux/timekeeping.h
> +++ b/include/linux/timekeeping.h
> @@ -292,7 +292,7 @@ extern void ktime_get_raw_and_real_ts64(
> * @cs_was_changed_seq: The sequence number of clocksource change events
> */
> struct system_time_snapshot {
> - cycle_t cycles;
> + u64 cycles;
> ktime_t real;
> ktime_t raw;
> unsigned int clock_was_set_seq;
> @@ -320,7 +320,7 @@ struct system_device_crosststamp {
> * timekeeping code to verify comparibility of two cycle values
> */
> struct system_counterval_t {
> - cycle_t cycles;
> + u64 cycles;
> struct clocksource *cs;
> };
>
> --- a/include/linux/types.h
> +++ b/include/linux/types.h
> @@ -228,8 +228,5 @@ struct callback_head {
> typedef void (*rcu_callback_t)(struct rcu_head *head);
> typedef void (*call_rcu_func_t)(struct rcu_head *head, rcu_callback_t func);
>
> -/* clocksource cycle base type */
> -typedef u64 cycle_t;
> -
> #endif /* __ASSEMBLY__ */
> #endif /* _LINUX_TYPES_H */
> --- a/kernel/time/clocksource.c
> +++ b/kernel/time/clocksource.c
> @@ -169,7 +169,7 @@ void clocksource_mark_unstable(struct cl
> static void clocksource_watchdog(unsigned long data)
> {
> struct clocksource *cs;
> - cycle_t csnow, wdnow, cslast, wdlast, delta;
> + u64 csnow, wdnow, cslast, wdlast, delta;
> int64_t wd_nsec, cs_nsec;
> int next_cpu, reset_pending;
>
> --- a/kernel/time/jiffies.c
> +++ b/kernel/time/jiffies.c
> @@ -59,9 +59,9 @@
> #define JIFFIES_SHIFT 8
> #endif
>
> -static cycle_t jiffies_read(struct clocksource *cs)
> +static u64 jiffies_read(struct clocksource *cs)
> {
> - return (cycle_t) jiffies;
> + return (u64) jiffies;
> }
>
> static struct clocksource clocksource_jiffies = {
> --- a/kernel/time/timecounter.c
> +++ b/kernel/time/timecounter.c
> @@ -43,7 +43,7 @@ EXPORT_SYMBOL_GPL(timecounter_init);
> */
> static u64 timecounter_read_delta(struct timecounter *tc)
> {
> - cycle_t cycle_now, cycle_delta;
> + u64 cycle_now, cycle_delta;
> u64 ns_offset;
>
> /* read cycle counter: */
> @@ -80,7 +80,7 @@ EXPORT_SYMBOL_GPL(timecounter_read);
> * time previous to the time stored in the cycle counter.
> */
> static u64 cc_cyc2ns_backwards(const struct cyclecounter *cc,
> - cycle_t cycles, u64 mask, u64 frac)
> + u64 cycles, u64 mask, u64 frac)
> {
> u64 ns = (u64) cycles;
>
> @@ -90,7 +90,7 @@ static u64 cc_cyc2ns_backwards(const str
> }
>
> u64 timecounter_cyc2time(struct timecounter *tc,
> - cycle_t cycle_tstamp)
> + u64 cycle_tstamp)
> {
> u64 delta = (cycle_tstamp - tc->cycle_last) & tc->cc->mask;
> u64 nsec = tc->nsec, frac = tc->frac;
> --- a/kernel/time/timekeeping.c
> +++ b/kernel/time/timekeeping.c
> @@ -119,10 +119,10 @@ static inline void tk_update_sleep_time(
> #ifdef CONFIG_DEBUG_TIMEKEEPING
> #define WARNING_FREQ (HZ*300) /* 5 minute rate-limiting */
>
> -static void timekeeping_check_update(struct timekeeper *tk, cycle_t offset)
> +static void timekeeping_check_update(struct timekeeper *tk, u64 offset)
> {
>
> - cycle_t max_cycles = tk->tkr_mono.clock->max_cycles;
> + u64 max_cycles = tk->tkr_mono.clock->max_cycles;
> const char *name = tk->tkr_mono.clock->name;
>
> if (offset > max_cycles) {
> @@ -158,10 +158,10 @@ static void timekeeping_check_update(str
> }
> }
>
> -static inline cycle_t timekeeping_get_delta(struct tk_read_base *tkr)
> +static inline u64 timekeeping_get_delta(struct tk_read_base *tkr)
> {
> struct timekeeper *tk = &tk_core.timekeeper;
> - cycle_t now, last, mask, max, delta;
> + u64 now, last, mask, max, delta;
> unsigned int seq;
>
> /*
> @@ -199,12 +199,12 @@ static inline cycle_t timekeeping_get_de
> return delta;
> }
> #else
> -static inline void timekeeping_check_update(struct timekeeper *tk, cycle_t offset)
> +static inline void timekeeping_check_update(struct timekeeper *tk, u64 offset)
> {
> }
> -static inline cycle_t timekeeping_get_delta(struct tk_read_base *tkr)
> +static inline u64 timekeeping_get_delta(struct tk_read_base *tkr)
> {
> - cycle_t cycle_now, delta;
> + u64 cycle_now, delta;
>
> /* read clocksource */
> cycle_now = tkr->read(tkr->clock);
> @@ -229,7 +229,7 @@ static inline cycle_t timekeeping_get_de
> */
> static void tk_setup_internals(struct timekeeper *tk, struct clocksource *clock)
> {
> - cycle_t interval;
> + u64 interval;
> u64 tmp, ntpinterval;
> struct clocksource *old_clock;
>
> @@ -254,7 +254,7 @@ static void tk_setup_internals(struct ti
> if (tmp == 0)
> tmp = 1;
>
> - interval = (cycle_t) tmp;
> + interval = (u64) tmp;
> tk->cycle_interval = interval;
>
> /* Go back from cycles -> shifted ns */
> @@ -346,16 +346,16 @@ static inline u64 timekeeping_delta_to_n
>
> static inline u64 timekeeping_get_ns(struct tk_read_base *tkr)
> {
> - cycle_t delta;
> + u64 delta;
>
> delta = timekeeping_get_delta(tkr);
> return timekeeping_delta_to_ns(tkr, delta);
> }
>
> static inline u64 timekeeping_cycles_to_ns(struct tk_read_base *tkr,
> - cycle_t cycles)
> + u64 cycles)
> {
> - cycle_t delta;
> + u64 delta;
>
> /* calculate the delta since the last update_wall_time */
> delta = clocksource_delta(cycles, tkr->cycle_last, tkr->mask);
> @@ -459,9 +459,9 @@ u64 ktime_get_raw_fast_ns(void)
> EXPORT_SYMBOL_GPL(ktime_get_raw_fast_ns);
>
> /* Suspend-time cycles value for halted fast timekeeper. */
> -static cycle_t cycles_at_suspend;
> +static u64 cycles_at_suspend;
>
> -static cycle_t dummy_clock_read(struct clocksource *cs)
> +static u64 dummy_clock_read(struct clocksource *cs)
> {
> return cycles_at_suspend;
> }
> @@ -655,7 +655,7 @@ static void timekeeping_update(struct ti
> static void timekeeping_forward_now(struct timekeeper *tk)
> {
> struct clocksource *clock = tk->tkr_mono.clock;
> - cycle_t cycle_now, delta;
> + u64 cycle_now, delta;
> u64 nsec;
>
> cycle_now = tk->tkr_mono.read(clock);
> @@ -928,7 +928,7 @@ void ktime_get_snapshot(struct system_ti
> ktime_t base_real;
> u64 nsec_raw;
> u64 nsec_real;
> - cycle_t now;
> + u64 now;
>
> WARN_ON_ONCE(timekeeping_suspended);
>
> @@ -987,8 +987,8 @@ static int scale64_check_overflow(u64 mu
> * interval is partial_history_cycles.
> */
> static int adjust_historical_crosststamp(struct system_time_snapshot *history,
> - cycle_t partial_history_cycles,
> - cycle_t total_history_cycles,
> + u64 partial_history_cycles,
> + u64 total_history_cycles,
> bool discontinuity,
> struct system_device_crosststamp *ts)
> {
> @@ -1052,7 +1052,7 @@ static int adjust_historical_crosststamp
> /*
> * cycle_between - true if test occurs chronologically between before and after
> */
> -static bool cycle_between(cycle_t before, cycle_t test, cycle_t after)
> +static bool cycle_between(u64 before, u64 test, u64 after)
> {
> if (test > before && test < after)
> return true;
> @@ -1082,7 +1082,7 @@ int get_device_system_crosststamp(int (*
> {
> struct system_counterval_t system_counterval;
> struct timekeeper *tk = &tk_core.timekeeper;
> - cycle_t cycles, now, interval_start;
> + u64 cycles, now, interval_start;
> unsigned int clock_was_set_seq = 0;
> ktime_t base_real, base_raw;
> u64 nsec_real, nsec_raw;
> @@ -1143,7 +1143,7 @@ int get_device_system_crosststamp(int (*
> * current interval
> */
> if (do_interp) {
> - cycle_t partial_history_cycles, total_history_cycles;
> + u64 partial_history_cycles, total_history_cycles;
> bool discontinuity;
>
> /*
> @@ -1649,7 +1649,7 @@ void timekeeping_resume(void)
> struct clocksource *clock = tk->tkr_mono.clock;
> unsigned long flags;
> struct timespec64 ts_new, ts_delta;
> - cycle_t cycle_now;
> + u64 cycle_now;
>
> sleeptime_injected = false;
> read_persistent_clock64(&ts_new);
> @@ -2015,11 +2015,11 @@ static inline unsigned int accumulate_ns
> *
> * Returns the unconsumed cycles.
> */
> -static cycle_t logarithmic_accumulation(struct timekeeper *tk, cycle_t offset,
> +static u64 logarithmic_accumulation(struct timekeeper *tk, u64 offset,
> u32 shift,
> unsigned int *clock_set)
> {
> - cycle_t interval = tk->cycle_interval << shift;
> + u64 interval = tk->cycle_interval << shift;
> u64 raw_nsecs;
>
> /* If the offset is smaller than a shifted interval, do nothing */
> @@ -2060,7 +2060,7 @@ void update_wall_time(void)
> {
> struct timekeeper *real_tk = &tk_core.timekeeper;
> struct timekeeper *tk = &shadow_timekeeper;
> - cycle_t offset;
> + u64 offset;
> int shift = 0, maxshift;
> unsigned int clock_set = 0;
> unsigned long flags;
> --- a/kernel/time/timekeeping_internal.h
> +++ b/kernel/time/timekeeping_internal.h
> @@ -13,9 +13,9 @@ extern void tk_debug_account_sleep_time(
> #endif
>
> #ifdef CONFIG_CLOCKSOURCE_VALIDATE_LAST_CYCLE
> -static inline cycle_t clocksource_delta(cycle_t now, cycle_t last, cycle_t mask)
> +static inline u64 clocksource_delta(u64 now, u64 last, u64 mask)
> {
> - cycle_t ret = (now - last) & mask;
> + u64 ret = (now - last) & mask;
>
> /*
> * Prevent time going backwards by checking the MSB of mask in
> @@ -24,7 +24,7 @@ static inline cycle_t clocksource_delta(
> return ret & ~(mask >> 1) ? 0 : ret;
> }
> #else
> -static inline cycle_t clocksource_delta(cycle_t now, cycle_t last, cycle_t mask)
> +static inline u64 clocksource_delta(u64 now, u64 last, u64 mask)
> {
> return (now - last) & mask;
> }
> --- a/kernel/trace/ftrace.c
> +++ b/kernel/trace/ftrace.c
> @@ -2847,7 +2847,7 @@ static void ftrace_shutdown_sysctl(void)
> }
> }
>
> -static cycle_t ftrace_update_time;
> +static u64 ftrace_update_time;
> unsigned long ftrace_update_tot_cnt;
>
> static inline int ops_traces_mod(struct ftrace_ops *ops)
> @@ -2894,7 +2894,7 @@ static int ftrace_update_code(struct mod
> {
> struct ftrace_page *pg;
> struct dyn_ftrace *p;
> - cycle_t start, stop;
> + u64 start, stop;
> unsigned long update_cnt = 0;
> unsigned long rec_flags = 0;
> int i;
> --- a/kernel/trace/trace.c
> +++ b/kernel/trace/trace.c
> @@ -234,7 +234,7 @@ static int __init set_tracepoint_printk(
> }
> __setup("tp_printk", set_tracepoint_printk);
>
> -unsigned long long ns2usecs(cycle_t nsec)
> +unsigned long long ns2usecs(u64 nsec)
> {
> nsec += 500;
> do_div(nsec, 1000);
> @@ -571,7 +571,7 @@ int trace_pid_write(struct trace_pid_lis
> return read;
> }
>
> -static cycle_t buffer_ftrace_now(struct trace_buffer *buf, int cpu)
> +static u64 buffer_ftrace_now(struct trace_buffer *buf, int cpu)
> {
> u64 ts;
>
> @@ -585,7 +585,7 @@ static cycle_t buffer_ftrace_now(struct
> return ts;
> }
>
> -cycle_t ftrace_now(int cpu)
> +u64 ftrace_now(int cpu)
> {
> return buffer_ftrace_now(&global_trace.trace_buffer, cpu);
> }
> --- a/kernel/trace/trace.h
> +++ b/kernel/trace/trace.h
> @@ -157,7 +157,7 @@ struct trace_array_cpu {
> unsigned long policy;
> unsigned long rt_priority;
> unsigned long skipped_entries;
> - cycle_t preempt_timestamp;
> + u64 preempt_timestamp;
> pid_t pid;
> kuid_t uid;
> char comm[TASK_COMM_LEN];
> @@ -175,7 +175,7 @@ struct trace_buffer {
> struct trace_array *tr;
> struct ring_buffer *buffer;
> struct trace_array_cpu __percpu *data;
> - cycle_t time_start;
> + u64 time_start;
> int cpu;
> };
>
> @@ -686,7 +686,7 @@ static inline void __trace_stack(struct
> }
> #endif /* CONFIG_STACKTRACE */
>
> -extern cycle_t ftrace_now(int cpu);
> +extern u64 ftrace_now(int cpu);
>
> extern void trace_find_cmdline(int pid, char comm[]);
> extern void trace_event_follow_fork(struct trace_array *tr, bool enable);
> @@ -733,7 +733,7 @@ extern int trace_selftest_startup_branch
> #endif /* CONFIG_FTRACE_STARTUP_TEST */
>
> extern void *head_page(struct trace_array_cpu *data);
> -extern unsigned long long ns2usecs(cycle_t nsec);
> +extern unsigned long long ns2usecs(u64 nsec);
> extern int
> trace_vbprintk(unsigned long ip, const char *fmt, va_list args);
> extern int
> --- a/kernel/trace/trace_irqsoff.c
> +++ b/kernel/trace/trace_irqsoff.c
> @@ -286,7 +286,7 @@ static void irqsoff_print_header(struct
> /*
> * Should this new latency be reported/recorded?
> */
> -static bool report_latency(struct trace_array *tr, cycle_t delta)
> +static bool report_latency(struct trace_array *tr, u64 delta)
> {
> if (tracing_thresh) {
> if (delta < tracing_thresh)
> @@ -304,7 +304,7 @@ check_critical_timing(struct trace_array
> unsigned long parent_ip,
> int cpu)
> {
> - cycle_t T0, T1, delta;
> + u64 T0, T1, delta;
> unsigned long flags;
> int pc;
>
> --- a/kernel/trace/trace_sched_wakeup.c
> +++ b/kernel/trace/trace_sched_wakeup.c
> @@ -346,7 +346,7 @@ static void wakeup_print_header(struct s
> /*
> * Should this new latency be reported/recorded?
> */
> -static bool report_latency(struct trace_array *tr, cycle_t delta)
> +static bool report_latency(struct trace_array *tr, u64 delta)
> {
> if (tracing_thresh) {
> if (delta < tracing_thresh)
> @@ -428,7 +428,7 @@ probe_wakeup_sched_switch(void *ignore,
> struct task_struct *prev, struct task_struct *next)
> {
> struct trace_array_cpu *data;
> - cycle_t T0, T1, delta;
> + u64 T0, T1, delta;
> unsigned long flags;
> long disabled;
> int cpu;
> --- a/sound/hda/hdac_stream.c
> +++ b/sound/hda/hdac_stream.c
> @@ -465,7 +465,7 @@ int snd_hdac_stream_set_params(struct hd
> }
> EXPORT_SYMBOL_GPL(snd_hdac_stream_set_params);
>
> -static cycle_t azx_cc_read(const struct cyclecounter *cc)
> +static u64 azx_cc_read(const struct cyclecounter *cc)
> {
> struct hdac_stream *azx_dev = container_of(cc, struct hdac_stream, cc);
>
> @@ -473,7 +473,7 @@ static cycle_t azx_cc_read(const struct
> }
>
> static void azx_timecounter_init(struct hdac_stream *azx_dev,
> - bool force, cycle_t last)
> + bool force, u64 last)
> {
> struct timecounter *tc = &azx_dev->tc;
> struct cyclecounter *cc = &azx_dev->cc;
> @@ -523,7 +523,7 @@ void snd_hdac_stream_timecounter_init(st
> struct snd_pcm_runtime *runtime = azx_dev->substream->runtime;
> struct hdac_stream *s;
> bool inited = false;
> - cycle_t cycle_last = 0;
> + u64 cycle_last = 0;
> int i = 0;
>
> list_for_each_entry(s, &bus->stream_list, list) {
> --- a/virt/kvm/arm/arch_timer.c
> +++ b/virt/kvm/arm/arch_timer.c
> @@ -39,7 +39,7 @@ void kvm_timer_vcpu_put(struct kvm_vcpu
> vcpu->arch.timer_cpu.active_cleared_last = false;
> }
>
> -static cycle_t kvm_phys_timer_read(void)
> +static u64 kvm_phys_timer_read(void)
> {
> return timecounter->cc->read(timecounter->cc);
> }
> @@ -102,7 +102,7 @@ static void kvm_timer_inject_irq_work(st
>
> static u64 kvm_timer_compute_delta(struct kvm_vcpu *vcpu)
> {
> - cycle_t cval, now;
> + u64 cval, now;
>
> cval = vcpu->arch.timer_cpu.cntv_cval;
> now = kvm_phys_timer_read() - vcpu->kvm->arch.timer.cntvoff;
> @@ -155,7 +155,7 @@ static bool kvm_timer_irq_can_fire(struc
> bool kvm_timer_should_fire(struct kvm_vcpu *vcpu)
> {
> struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu;
> - cycle_t cval, now;
> + u64 cval, now;
>
> if (!kvm_timer_irq_can_fire(vcpu))
> return false;
>
>
--
David Gibson | I'll have my music baroque, and my code
david AT gibson.dropbear.id.au | minimalist, thank you. NOT _the_ _other_
| _way_ _around_!
http://www.ozlabs.org/~dgibson