|  | /* | 
|  | * This file is subject to the terms and conditions of the GNU General Public | 
|  | * License.  See the file "COPYING" in the main directory of this archive | 
|  | * for more details. | 
|  | * | 
|  | * Copyright (C) 1998, 1999, 2003 by Ralf Baechle | 
|  | * Copyright (C) 2014 by Maciej W. Rozycki | 
|  | */ | 
|  | #ifndef _ASM_TIMEX_H | 
|  | #define _ASM_TIMEX_H | 
|  |  | 
|  | #ifdef __KERNEL__ | 
|  |  | 
|  | #include <linux/compiler.h> | 
|  |  | 
|  | #include <asm/cpu.h> | 
|  | #include <asm/cpu-features.h> | 
|  | #include <asm/mipsregs.h> | 
|  | #include <asm/cpu-type.h> | 
|  |  | 
|  | /* | 
|  | * This is the clock rate of the i8253 PIT.  A MIPS system may not have | 
|  | * a PIT by the symbol is used all over the kernel including some APIs. | 
|  | * So keeping it defined to the number for the PIT is the only sane thing | 
|  | * for now. | 
|  | */ | 
|  | #define CLOCK_TICK_RATE 1193182 | 
|  |  | 
|  | /* | 
|  | * Standard way to access the cycle counter. | 
|  | * Currently only used on SMP for scheduling. | 
|  | * | 
|  | * Only the low 32 bits are available as a continuously counting entity. | 
|  | * But this only means we'll force a reschedule every 8 seconds or so, | 
|  | * which isn't an evil thing. | 
|  | * | 
|  | * We know that all SMP capable CPUs have cycle counters. | 
|  | */ | 
|  |  | 
|  | typedef unsigned int cycles_t; | 
|  |  | 
|  | /* | 
|  | * On R4000/R4400 before version 5.0 an erratum exists such that if the | 
|  | * cycle counter is read in the exact moment that it is matching the | 
|  | * compare register, no interrupt will be generated. | 
|  | * | 
|  | * There is a suggested workaround and also the erratum can't strike if | 
|  | * the compare interrupt isn't being used as the clock source device. | 
|  | * However for now the implementaton of this function doesn't get these | 
|  | * fine details right. | 
|  | */ | 
|  | static inline int can_use_mips_counter(unsigned int prid) | 
|  | { | 
|  | int comp = (prid & PRID_COMP_MASK) != PRID_COMP_LEGACY; | 
|  |  | 
|  | if (__builtin_constant_p(cpu_has_counter) && !cpu_has_counter) | 
|  | return 0; | 
|  | else if (__builtin_constant_p(cpu_has_mips_r) && cpu_has_mips_r) | 
|  | return 1; | 
|  | else if (likely(!__builtin_constant_p(cpu_has_mips_r) && comp)) | 
|  | return 1; | 
|  | /* Make sure we don't peek at cpu_data[0].options in the fast path! */ | 
|  | if (!__builtin_constant_p(cpu_has_counter)) | 
|  | asm volatile("" : "=m" (cpu_data[0].options)); | 
|  | if (likely(cpu_has_counter && | 
|  | prid >= (PRID_IMP_R4000 | PRID_REV_ENCODE_44(5, 0)))) | 
|  | return 1; | 
|  | else | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static inline cycles_t get_cycles(void) | 
|  | { | 
|  | if (can_use_mips_counter(read_c0_prid())) | 
|  | return read_c0_count(); | 
|  | else | 
|  | return 0;	/* no usable counter */ | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Like get_cycles - but where c0_count is not available we desperately | 
|  | * use c0_random in an attempt to get at least a little bit of entropy. | 
|  | * | 
|  | * R6000 and R6000A neither have a count register nor a random register. | 
|  | * That leaves no entropy source in the CPU itself. | 
|  | */ | 
|  | static inline unsigned long random_get_entropy(void) | 
|  | { | 
|  | unsigned int prid = read_c0_prid(); | 
|  | unsigned int imp = prid & PRID_IMP_MASK; | 
|  |  | 
|  | if (can_use_mips_counter(prid)) | 
|  | return read_c0_count(); | 
|  | else if (likely(imp != PRID_IMP_R6000 && imp != PRID_IMP_R6000A)) | 
|  | return read_c0_random(); | 
|  | else | 
|  | return 0;	/* no usable register */ | 
|  | } | 
|  | #define random_get_entropy random_get_entropy | 
|  |  | 
|  | #endif /* __KERNEL__ */ | 
|  |  | 
|  | #endif /*  _ASM_TIMEX_H */ |