@@ -186,12 +186,12 @@ static inline unsigned int generic_hweight64(uint64_t w)
if ( BITS_PER_LONG < 64 )
return generic_hweight32(w >> 32) + generic_hweight32(w);
- w -= (w >> 1) & 0x5555555555555555ul;
- w = (w & 0x3333333333333333ul) + ((w >> 2) & 0x3333333333333333ul);
- w = (w + (w >> 4)) & 0x0f0f0f0f0f0f0f0ful;
+ w -= (w >> 1) & 0x5555555555555555UL;
+ w = (w & 0x3333333333333333UL) + ((w >> 2) & 0x3333333333333333UL);
+ w = (w + (w >> 4)) & 0x0f0f0f0f0f0f0f0fUL;
if ( IS_ENABLED(CONFIG_HAS_FAST_MULTIPLY) )
- return (w * 0x0101010101010101ul) >> 56;
+ return (w * 0x0101010101010101UL) >> 56;
w += w >> 8;
w += w >> 16;
@@ -37,8 +37,8 @@ struct tasklet
DECLARE_PER_CPU(unsigned long, tasklet_work_to_do);
#define _TASKLET_enqueued 0 /* Tasklet work is enqueued for this CPU. */
#define _TASKLET_scheduled 1 /* Scheduler has scheduled do_tasklet(). */
-#define TASKLET_enqueued (1ul << _TASKLET_enqueued)
-#define TASKLET_scheduled (1ul << _TASKLET_scheduled)
+#define TASKLET_enqueued (1UL << _TASKLET_enqueued)
+#define TASKLET_scheduled (1UL << _TASKLET_scheduled)
static inline bool tasklet_work_to_do(unsigned int cpu)
{
@@ -57,9 +57,9 @@ struct tm wallclock_time(uint64_t *ns);
#define SECONDS(_s) ((s_time_t)((_s) * 1000000000ULL))
#define MILLISECS(_ms) ((s_time_t)((_ms) * 1000000ULL))
#define MICROSECS(_us) ((s_time_t)((_us) * 1000ULL))
-#define STIME_MAX ((s_time_t)((uint64_t)~0ull>>1))
+#define STIME_MAX ((s_time_t)((uint64_t)~0ULL>>1))
/* Chosen so (NOW() + delta) wont overflow without an uptime of 200 years */
-#define STIME_DELTA_MAX ((s_time_t)((uint64_t)~0ull>>2))
+#define STIME_DELTA_MAX ((s_time_t)((uint64_t)~0ULL>>2))
/* Explicitly OR with 1 just in case version number gets out of sync. */
#define version_update_begin(v) (((v) + 1) | 1)