blob: dd558589cb523c31099a10d8dcf763cc87776278 [file] [log] [blame]
ths05f778c2007-10-27 13:05:54 +00001/*
2 * Utility compute operations used by translated code.
3 *
4 * Copyright (c) 2007 Thiemo Seufer
5 * Copyright (c) 2007 Jocelyn Mayer
6 *
7 * Permission is hereby granted, free of charge, to any person obtaining a copy
8 * of this software and associated documentation files (the "Software"), to deal
9 * in the Software without restriction, including without limitation the rights
10 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
11 * copies of the Software, and to permit persons to whom the Software is
12 * furnished to do so, subject to the following conditions:
13 *
14 * The above copyright notice and this permission notice shall be included in
15 * all copies or substantial portions of the Software.
16 *
17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
18 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
20 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
21 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
22 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
23 * THE SOFTWARE.
24 */
Markus Armbruster175de522016-06-29 15:29:06 +020025
Luis Pires8ac2d6c2021-10-25 16:11:37 -030026/* Portions of this work are licensed under the terms of the GNU GPL,
27 * version 2 or later. See the COPYING file in the top-level directory.
28 */
29
Paolo Bonzinicb9c3772012-12-06 12:15:58 +010030#ifndef HOST_UTILS_H
Markus Armbruster175de522016-06-29 15:29:06 +020031#define HOST_UTILS_H
ths05f778c2007-10-27 13:05:54 +000032
Richard Henderson652a4b72015-09-14 13:00:34 -070033#include "qemu/bswap.h"
Lucas Mateus Castro (alqotel)4724bbd2022-05-25 10:49:50 -030034#include "qemu/int128.h"
thscebdff72008-06-05 22:55:54 +000035
Richard Hendersonf5401662013-02-16 12:46:59 -080036#ifdef CONFIG_INT128
Blue Swirlfacd2852009-08-16 08:03:26 +000037static inline void mulu64(uint64_t *plow, uint64_t *phigh,
38 uint64_t a, uint64_t b)
j_mayer7a51ad82007-11-04 02:24:58 +000039{
Richard Hendersonf5401662013-02-16 12:46:59 -080040 __uint128_t r = (__uint128_t)a * b;
41 *plow = r;
42 *phigh = r >> 64;
j_mayer7a51ad82007-11-04 02:24:58 +000043}
Richard Hendersonf5401662013-02-16 12:46:59 -080044
Blue Swirlfacd2852009-08-16 08:03:26 +000045static inline void muls64(uint64_t *plow, uint64_t *phigh,
46 int64_t a, int64_t b)
j_mayer7a51ad82007-11-04 02:24:58 +000047{
Richard Hendersonf5401662013-02-16 12:46:59 -080048 __int128_t r = (__int128_t)a * b;
49 *plow = r;
50 *phigh = r >> 64;
j_mayer7a51ad82007-11-04 02:24:58 +000051}
Tom Musta98d1eb22014-01-07 10:05:51 -060052
Peter Maydell49caffe2015-08-19 16:20:20 +010053/* compute with 96 bit intermediate result: (a*b)/c */
54static inline uint64_t muldiv64(uint64_t a, uint32_t b, uint32_t c)
55{
56 return (__int128_t)a * b / c;
57}
58
Nicholas Piggin47de6c42023-08-08 14:19:47 +100059static inline uint64_t muldiv64_round_up(uint64_t a, uint32_t b, uint32_t c)
60{
61 return ((__int128_t)a * b + c - 1) / c;
62}
63
Luis Pires40f3e792021-10-25 16:11:38 -030064static inline uint64_t divu128(uint64_t *plow, uint64_t *phigh,
65 uint64_t divisor)
Tom Musta98d1eb22014-01-07 10:05:51 -060066{
Luis Pires9276a312021-10-25 16:11:36 -030067 __uint128_t dividend = ((__uint128_t)*phigh << 64) | *plow;
68 __uint128_t result = dividend / divisor;
Luis Pires40f3e792021-10-25 16:11:38 -030069
Luis Pires9276a312021-10-25 16:11:36 -030070 *plow = result;
Luis Pires40f3e792021-10-25 16:11:38 -030071 *phigh = result >> 64;
72 return dividend % divisor;
Tom Musta98d1eb22014-01-07 10:05:51 -060073}
Tom Mustae44259b2014-01-07 10:05:52 -060074
Luis Pires40f3e792021-10-25 16:11:38 -030075static inline int64_t divs128(uint64_t *plow, int64_t *phigh,
76 int64_t divisor)
Tom Mustae44259b2014-01-07 10:05:52 -060077{
Luis Pires40f3e792021-10-25 16:11:38 -030078 __int128_t dividend = ((__int128_t)*phigh << 64) | *plow;
Luis Pires9276a312021-10-25 16:11:36 -030079 __int128_t result = dividend / divisor;
Luis Pires40f3e792021-10-25 16:11:38 -030080
Luis Pires9276a312021-10-25 16:11:36 -030081 *plow = result;
Luis Pires40f3e792021-10-25 16:11:38 -030082 *phigh = result >> 64;
83 return dividend % divisor;
Tom Mustae44259b2014-01-07 10:05:52 -060084}
j_mayer7a51ad82007-11-04 02:24:58 +000085#else
Lijun Pandb7b62e2020-07-01 18:43:44 -050086void muls64(uint64_t *plow, uint64_t *phigh, int64_t a, int64_t b);
87void mulu64(uint64_t *plow, uint64_t *phigh, uint64_t a, uint64_t b);
Luis Pires40f3e792021-10-25 16:11:38 -030088uint64_t divu128(uint64_t *plow, uint64_t *phigh, uint64_t divisor);
89int64_t divs128(uint64_t *plow, int64_t *phigh, int64_t divisor);
Peter Maydell49caffe2015-08-19 16:20:20 +010090
Nicholas Piggin47de6c42023-08-08 14:19:47 +100091static inline uint64_t muldiv64_rounding(uint64_t a, uint32_t b, uint32_t c,
92 bool round_up)
Peter Maydell49caffe2015-08-19 16:20:20 +010093{
94 union {
95 uint64_t ll;
96 struct {
Marc-André Lureaue03b5682022-03-23 19:57:17 +040097#if HOST_BIG_ENDIAN
Peter Maydell49caffe2015-08-19 16:20:20 +010098 uint32_t high, low;
99#else
100 uint32_t low, high;
101#endif
102 } l;
103 } u, res;
104 uint64_t rl, rh;
105
106 u.ll = a;
107 rl = (uint64_t)u.l.low * (uint64_t)b;
Nicholas Piggin47de6c42023-08-08 14:19:47 +1000108 if (round_up) {
109 rl += c - 1;
110 }
Peter Maydell49caffe2015-08-19 16:20:20 +0100111 rh = (uint64_t)u.l.high * (uint64_t)b;
112 rh += (rl >> 32);
113 res.l.high = rh / c;
114 res.l.low = (((rh % c) << 32) + (rl & 0xffffffff)) / c;
115 return res.ll;
116}
Nicholas Piggin47de6c42023-08-08 14:19:47 +1000117
118static inline uint64_t muldiv64(uint64_t a, uint32_t b, uint32_t c)
119{
120 return muldiv64_rounding(a, b, c, false);
121}
122
123static inline uint64_t muldiv64_round_up(uint64_t a, uint32_t b, uint32_t c)
124{
125 return muldiv64_rounding(a, b, c, true);
126}
j_mayer7a51ad82007-11-04 02:24:58 +0000127#endif
128
Richard Henderson72d81152013-02-13 17:47:35 -0800129/**
Kiran Ostrolenk31fe2562023-04-28 15:47:48 +0100130 * clz8 - count leading zeros in a 8-bit value.
131 * @val: The value to search
132 *
133 * Returns 8 if the value is zero. Note that the GCC builtin is
134 * undefined if the value is zero.
135 *
136 * Note that the GCC builtin will upcast its argument to an `unsigned int`
137 * so this function subtracts off the number of prepended zeroes.
138 */
139static inline int clz8(uint8_t val)
140{
141 return val ? __builtin_clz(val) - 24 : 8;
142}
143
144/**
145 * clz16 - count leading zeros in a 16-bit value.
146 * @val: The value to search
147 *
148 * Returns 16 if the value is zero. Note that the GCC builtin is
149 * undefined if the value is zero.
150 *
151 * Note that the GCC builtin will upcast its argument to an `unsigned int`
152 * so this function subtracts off the number of prepended zeroes.
153 */
154static inline int clz16(uint16_t val)
155{
156 return val ? __builtin_clz(val) - 16 : 16;
157}
158
159/**
Richard Henderson72d81152013-02-13 17:47:35 -0800160 * clz32 - count leading zeros in a 32-bit value.
161 * @val: The value to search
162 *
163 * Returns 32 if the value is zero. Note that the GCC builtin is
164 * undefined if the value is zero.
165 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000166static inline int clz32(uint32_t val)
ths05f778c2007-10-27 13:05:54 +0000167{
Richard Henderson72d81152013-02-13 17:47:35 -0800168 return val ? __builtin_clz(val) : 32;
ths05f778c2007-10-27 13:05:54 +0000169}
170
Richard Henderson72d81152013-02-13 17:47:35 -0800171/**
172 * clo32 - count leading ones in a 32-bit value.
173 * @val: The value to search
174 *
175 * Returns 32 if the value is -1.
176 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000177static inline int clo32(uint32_t val)
ths05f778c2007-10-27 13:05:54 +0000178{
179 return clz32(~val);
180}
181
Richard Henderson72d81152013-02-13 17:47:35 -0800182/**
183 * clz64 - count leading zeros in a 64-bit value.
184 * @val: The value to search
185 *
186 * Returns 64 if the value is zero. Note that the GCC builtin is
187 * undefined if the value is zero.
188 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000189static inline int clz64(uint64_t val)
ths05f778c2007-10-27 13:05:54 +0000190{
Richard Henderson72d81152013-02-13 17:47:35 -0800191 return val ? __builtin_clzll(val) : 64;
ths05f778c2007-10-27 13:05:54 +0000192}
193
Richard Henderson72d81152013-02-13 17:47:35 -0800194/**
195 * clo64 - count leading ones in a 64-bit value.
196 * @val: The value to search
197 *
198 * Returns 64 if the value is -1.
199 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000200static inline int clo64(uint64_t val)
ths05f778c2007-10-27 13:05:54 +0000201{
202 return clz64(~val);
203}
j_mayerb9ef45f2007-10-28 12:52:38 +0000204
Richard Henderson72d81152013-02-13 17:47:35 -0800205/**
Kiran Ostrolenk31fe2562023-04-28 15:47:48 +0100206 * ctz8 - count trailing zeros in a 8-bit value.
207 * @val: The value to search
208 *
209 * Returns 8 if the value is zero. Note that the GCC builtin is
210 * undefined if the value is zero.
211 */
212static inline int ctz8(uint8_t val)
213{
214 return val ? __builtin_ctz(val) : 8;
215}
216
217/**
218 * ctz16 - count trailing zeros in a 16-bit value.
219 * @val: The value to search
220 *
221 * Returns 16 if the value is zero. Note that the GCC builtin is
222 * undefined if the value is zero.
223 */
224static inline int ctz16(uint16_t val)
225{
226 return val ? __builtin_ctz(val) : 16;
227}
228
229/**
Richard Henderson72d81152013-02-13 17:47:35 -0800230 * ctz32 - count trailing zeros in a 32-bit value.
231 * @val: The value to search
232 *
233 * Returns 32 if the value is zero. Note that the GCC builtin is
234 * undefined if the value is zero.
235 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000236static inline int ctz32(uint32_t val)
j_mayerb9ef45f2007-10-28 12:52:38 +0000237{
Richard Henderson72d81152013-02-13 17:47:35 -0800238 return val ? __builtin_ctz(val) : 32;
balrogc8906842008-11-12 17:18:41 +0000239}
240
Richard Henderson72d81152013-02-13 17:47:35 -0800241/**
242 * cto32 - count trailing ones in a 32-bit value.
243 * @val: The value to search
244 *
245 * Returns 32 if the value is -1.
246 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000247static inline int cto32(uint32_t val)
balrogc8906842008-11-12 17:18:41 +0000248{
j_mayerb9ef45f2007-10-28 12:52:38 +0000249 return ctz32(~val);
250}
251
Richard Henderson72d81152013-02-13 17:47:35 -0800252/**
253 * ctz64 - count trailing zeros in a 64-bit value.
254 * @val: The value to search
255 *
256 * Returns 64 if the value is zero. Note that the GCC builtin is
257 * undefined if the value is zero.
258 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000259static inline int ctz64(uint64_t val)
j_mayerb9ef45f2007-10-28 12:52:38 +0000260{
Richard Henderson72d81152013-02-13 17:47:35 -0800261 return val ? __builtin_ctzll(val) : 64;
j_mayerb9ef45f2007-10-28 12:52:38 +0000262}
263
Richard Henderson72d81152013-02-13 17:47:35 -0800264/**
Dr. David Alan Gilbert1c884ab2014-02-12 17:14:33 +0000265 * cto64 - count trailing ones in a 64-bit value.
Richard Henderson72d81152013-02-13 17:47:35 -0800266 * @val: The value to search
267 *
268 * Returns 64 if the value is -1.
269 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000270static inline int cto64(uint64_t val)
j_mayerb9ef45f2007-10-28 12:52:38 +0000271{
272 return ctz64(~val);
273}
274
Richard Henderson72d81152013-02-13 17:47:35 -0800275/**
Claudio Fontanaafd3fe42013-12-17 19:42:35 +0000276 * clrsb32 - count leading redundant sign bits in a 32-bit value.
277 * @val: The value to search
278 *
279 * Returns the number of bits following the sign bit that are equal to it.
280 * No special cases; output range is [0-31].
281 */
282static inline int clrsb32(uint32_t val)
283{
Thomas Huthf773b422018-12-03 14:33:12 +0100284#if __has_builtin(__builtin_clrsb) || !defined(__clang__)
Claudio Fontanaafd3fe42013-12-17 19:42:35 +0000285 return __builtin_clrsb(val);
286#else
287 return clz32(val ^ ((int32_t)val >> 1)) - 1;
288#endif
289}
290
291/**
292 * clrsb64 - count leading redundant sign bits in a 64-bit value.
293 * @val: The value to search
294 *
295 * Returns the number of bits following the sign bit that are equal to it.
296 * No special cases; output range is [0-63].
297 */
298static inline int clrsb64(uint64_t val)
299{
Thomas Huthf773b422018-12-03 14:33:12 +0100300#if __has_builtin(__builtin_clrsbll) || !defined(__clang__)
Claudio Fontanaafd3fe42013-12-17 19:42:35 +0000301 return __builtin_clrsbll(val);
302#else
303 return clz64(val ^ ((int64_t)val >> 1)) - 1;
304#endif
305}
306
307/**
Richard Henderson72d81152013-02-13 17:47:35 -0800308 * ctpop8 - count the population of one bits in an 8-bit value.
309 * @val: The value to search
310 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000311static inline int ctpop8(uint8_t val)
j_mayerb9ef45f2007-10-28 12:52:38 +0000312{
Richard Henderson72d81152013-02-13 17:47:35 -0800313 return __builtin_popcount(val);
j_mayerb9ef45f2007-10-28 12:52:38 +0000314}
315
Paolo Bonzini24899cd2024-05-31 10:52:42 +0200316/*
317 * parity8 - return the parity (1 = odd) of an 8-bit value.
318 * @val: The value to search
319 */
320static inline int parity8(uint8_t val)
321{
322 return __builtin_parity(val);
323}
324
Richard Henderson72d81152013-02-13 17:47:35 -0800325/**
326 * ctpop16 - count the population of one bits in a 16-bit value.
327 * @val: The value to search
328 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000329static inline int ctpop16(uint16_t val)
j_mayerb9ef45f2007-10-28 12:52:38 +0000330{
Richard Henderson72d81152013-02-13 17:47:35 -0800331 return __builtin_popcount(val);
j_mayerb9ef45f2007-10-28 12:52:38 +0000332}
333
Richard Henderson72d81152013-02-13 17:47:35 -0800334/**
335 * ctpop32 - count the population of one bits in a 32-bit value.
336 * @val: The value to search
337 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000338static inline int ctpop32(uint32_t val)
j_mayerb9ef45f2007-10-28 12:52:38 +0000339{
aurel327d019982008-10-12 00:53:08 +0000340 return __builtin_popcount(val);
j_mayerb9ef45f2007-10-28 12:52:38 +0000341}
342
Richard Henderson72d81152013-02-13 17:47:35 -0800343/**
344 * ctpop64 - count the population of one bits in a 64-bit value.
345 * @val: The value to search
346 */
Blue Swirlfacd2852009-08-16 08:03:26 +0000347static inline int ctpop64(uint64_t val)
j_mayerb9ef45f2007-10-28 12:52:38 +0000348{
aurel327d019982008-10-12 00:53:08 +0000349 return __builtin_popcountll(val);
ths3800af92007-12-18 01:58:05 +0000350}
Paolo Bonzinicb9c3772012-12-06 12:15:58 +0100351
Richard Henderson652a4b72015-09-14 13:00:34 -0700352/**
353 * revbit8 - reverse the bits in an 8-bit value.
354 * @x: The value to modify.
355 */
356static inline uint8_t revbit8(uint8_t x)
357{
Richard Henderson5140d6b2020-11-06 10:59:36 -0800358#if __has_builtin(__builtin_bitreverse8)
359 return __builtin_bitreverse8(x);
360#else
Richard Henderson652a4b72015-09-14 13:00:34 -0700361 /* Assign the correct nibble position. */
362 x = ((x & 0xf0) >> 4)
363 | ((x & 0x0f) << 4);
364 /* Assign the correct bit position. */
365 x = ((x & 0x88) >> 3)
366 | ((x & 0x44) >> 1)
367 | ((x & 0x22) << 1)
368 | ((x & 0x11) << 3);
369 return x;
Richard Henderson5140d6b2020-11-06 10:59:36 -0800370#endif
Richard Henderson652a4b72015-09-14 13:00:34 -0700371}
372
373/**
374 * revbit16 - reverse the bits in a 16-bit value.
375 * @x: The value to modify.
376 */
377static inline uint16_t revbit16(uint16_t x)
378{
Richard Henderson5140d6b2020-11-06 10:59:36 -0800379#if __has_builtin(__builtin_bitreverse16)
380 return __builtin_bitreverse16(x);
381#else
Richard Henderson652a4b72015-09-14 13:00:34 -0700382 /* Assign the correct byte position. */
383 x = bswap16(x);
384 /* Assign the correct nibble position. */
385 x = ((x & 0xf0f0) >> 4)
386 | ((x & 0x0f0f) << 4);
387 /* Assign the correct bit position. */
388 x = ((x & 0x8888) >> 3)
389 | ((x & 0x4444) >> 1)
390 | ((x & 0x2222) << 1)
391 | ((x & 0x1111) << 3);
392 return x;
Richard Henderson5140d6b2020-11-06 10:59:36 -0800393#endif
Richard Henderson652a4b72015-09-14 13:00:34 -0700394}
395
396/**
397 * revbit32 - reverse the bits in a 32-bit value.
398 * @x: The value to modify.
399 */
400static inline uint32_t revbit32(uint32_t x)
401{
Richard Henderson5140d6b2020-11-06 10:59:36 -0800402#if __has_builtin(__builtin_bitreverse32)
403 return __builtin_bitreverse32(x);
404#else
Richard Henderson652a4b72015-09-14 13:00:34 -0700405 /* Assign the correct byte position. */
406 x = bswap32(x);
407 /* Assign the correct nibble position. */
408 x = ((x & 0xf0f0f0f0u) >> 4)
409 | ((x & 0x0f0f0f0fu) << 4);
410 /* Assign the correct bit position. */
411 x = ((x & 0x88888888u) >> 3)
412 | ((x & 0x44444444u) >> 1)
413 | ((x & 0x22222222u) << 1)
414 | ((x & 0x11111111u) << 3);
415 return x;
Richard Henderson5140d6b2020-11-06 10:59:36 -0800416#endif
Richard Henderson652a4b72015-09-14 13:00:34 -0700417}
418
419/**
420 * revbit64 - reverse the bits in a 64-bit value.
421 * @x: The value to modify.
422 */
423static inline uint64_t revbit64(uint64_t x)
424{
Richard Henderson5140d6b2020-11-06 10:59:36 -0800425#if __has_builtin(__builtin_bitreverse64)
426 return __builtin_bitreverse64(x);
427#else
Richard Henderson652a4b72015-09-14 13:00:34 -0700428 /* Assign the correct byte position. */
429 x = bswap64(x);
430 /* Assign the correct nibble position. */
431 x = ((x & 0xf0f0f0f0f0f0f0f0ull) >> 4)
432 | ((x & 0x0f0f0f0f0f0f0f0full) << 4);
433 /* Assign the correct bit position. */
434 x = ((x & 0x8888888888888888ull) >> 3)
435 | ((x & 0x4444444444444444ull) >> 1)
436 | ((x & 0x2222222222222222ull) << 1)
437 | ((x & 0x1111111111111111ull) << 3);
438 return x;
Richard Henderson5140d6b2020-11-06 10:59:36 -0800439#endif
Richard Henderson652a4b72015-09-14 13:00:34 -0700440}
441
Richard Hendersoncec07c02020-11-06 17:42:36 -0800442/**
Luis Piresd03bba02021-09-10 08:26:05 -0300443 * Return the absolute value of a 64-bit integer as an unsigned 64-bit value
444 */
445static inline uint64_t uabs64(int64_t v)
446{
447 return v < 0 ? -v : v;
448}
449
450/**
Richard Hendersoncec07c02020-11-06 17:42:36 -0800451 * sadd32_overflow - addition with overflow indication
452 * @x, @y: addends
453 * @ret: Output for sum
454 *
455 * Computes *@ret = @x + @y, and returns true if and only if that
456 * value has been truncated.
457 */
458static inline bool sadd32_overflow(int32_t x, int32_t y, int32_t *ret)
459{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800460 return __builtin_add_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800461}
462
463/**
464 * sadd64_overflow - addition with overflow indication
465 * @x, @y: addends
466 * @ret: Output for sum
467 *
468 * Computes *@ret = @x + @y, and returns true if and only if that
469 * value has been truncated.
470 */
471static inline bool sadd64_overflow(int64_t x, int64_t y, int64_t *ret)
472{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800473 return __builtin_add_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800474}
475
476/**
477 * uadd32_overflow - addition with overflow indication
478 * @x, @y: addends
479 * @ret: Output for sum
480 *
481 * Computes *@ret = @x + @y, and returns true if and only if that
482 * value has been truncated.
483 */
484static inline bool uadd32_overflow(uint32_t x, uint32_t y, uint32_t *ret)
485{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800486 return __builtin_add_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800487}
488
489/**
490 * uadd64_overflow - addition with overflow indication
491 * @x, @y: addends
492 * @ret: Output for sum
493 *
494 * Computes *@ret = @x + @y, and returns true if and only if that
495 * value has been truncated.
496 */
497static inline bool uadd64_overflow(uint64_t x, uint64_t y, uint64_t *ret)
498{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800499 return __builtin_add_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800500}
501
502/**
503 * ssub32_overflow - subtraction with overflow indication
504 * @x: Minuend
505 * @y: Subtrahend
506 * @ret: Output for difference
507 *
508 * Computes *@ret = @x - @y, and returns true if and only if that
509 * value has been truncated.
510 */
511static inline bool ssub32_overflow(int32_t x, int32_t y, int32_t *ret)
512{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800513 return __builtin_sub_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800514}
515
516/**
517 * ssub64_overflow - subtraction with overflow indication
518 * @x: Minuend
519 * @y: Subtrahend
520 * @ret: Output for sum
521 *
522 * Computes *@ret = @x - @y, and returns true if and only if that
523 * value has been truncated.
524 */
525static inline bool ssub64_overflow(int64_t x, int64_t y, int64_t *ret)
526{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800527 return __builtin_sub_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800528}
529
530/**
531 * usub32_overflow - subtraction with overflow indication
532 * @x: Minuend
533 * @y: Subtrahend
534 * @ret: Output for sum
535 *
536 * Computes *@ret = @x - @y, and returns true if and only if that
537 * value has been truncated.
538 */
539static inline bool usub32_overflow(uint32_t x, uint32_t y, uint32_t *ret)
540{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800541 return __builtin_sub_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800542}
543
544/**
545 * usub64_overflow - subtraction with overflow indication
546 * @x: Minuend
547 * @y: Subtrahend
548 * @ret: Output for sum
549 *
550 * Computes *@ret = @x - @y, and returns true if and only if that
551 * value has been truncated.
552 */
553static inline bool usub64_overflow(uint64_t x, uint64_t y, uint64_t *ret)
554{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800555 return __builtin_sub_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800556}
557
558/**
559 * smul32_overflow - multiplication with overflow indication
560 * @x, @y: Input multipliers
561 * @ret: Output for product
562 *
563 * Computes *@ret = @x * @y, and returns true if and only if that
564 * value has been truncated.
565 */
566static inline bool smul32_overflow(int32_t x, int32_t y, int32_t *ret)
567{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800568 return __builtin_mul_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800569}
570
571/**
572 * smul64_overflow - multiplication with overflow indication
573 * @x, @y: Input multipliers
574 * @ret: Output for product
575 *
576 * Computes *@ret = @x * @y, and returns true if and only if that
577 * value has been truncated.
578 */
579static inline bool smul64_overflow(int64_t x, int64_t y, int64_t *ret)
580{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800581 return __builtin_mul_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800582}
583
584/**
585 * umul32_overflow - multiplication with overflow indication
586 * @x, @y: Input multipliers
587 * @ret: Output for product
588 *
589 * Computes *@ret = @x * @y, and returns true if and only if that
590 * value has been truncated.
591 */
592static inline bool umul32_overflow(uint32_t x, uint32_t y, uint32_t *ret)
593{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800594 return __builtin_mul_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800595}
596
597/**
598 * umul64_overflow - multiplication with overflow indication
599 * @x, @y: Input multipliers
600 * @ret: Output for product
601 *
602 * Computes *@ret = @x * @y, and returns true if and only if that
603 * value has been truncated.
604 */
605static inline bool umul64_overflow(uint64_t x, uint64_t y, uint64_t *ret)
606{
Richard Hendersoncec07c02020-11-06 17:42:36 -0800607 return __builtin_mul_overflow(x, y, ret);
Richard Hendersoncec07c02020-11-06 17:42:36 -0800608}
609
Luis Pirese06049f2021-10-29 16:24:07 -0300610/*
611 * Unsigned 128x64 multiplication.
612 * Returns true if the result got truncated to 128 bits.
613 * Otherwise, returns false and the multiplication result via plow and phigh.
614 */
615static inline bool mulu128(uint64_t *plow, uint64_t *phigh, uint64_t factor)
616{
Thomas Huth21d4e552022-07-21 09:48:09 +0200617#if defined(CONFIG_INT128)
Luis Pirese06049f2021-10-29 16:24:07 -0300618 bool res;
619 __uint128_t r;
620 __uint128_t f = ((__uint128_t)*phigh << 64) | *plow;
621 res = __builtin_mul_overflow(f, factor, &r);
622
623 *plow = r;
624 *phigh = r >> 64;
625
626 return res;
627#else
628 uint64_t dhi = *phigh;
629 uint64_t dlo = *plow;
630 uint64_t ahi;
631 uint64_t blo, bhi;
632
633 if (dhi == 0) {
634 mulu64(plow, phigh, dlo, factor);
635 return false;
636 }
637
638 mulu64(plow, &ahi, dlo, factor);
639 mulu64(&blo, &bhi, dhi, factor);
640
641 return uadd64_overflow(ahi, blo, phigh) || bhi != 0;
642#endif
643}
644
Richard Henderson1ec80702020-11-13 03:22:23 +0000645/**
646 * uadd64_carry - addition with carry-in and carry-out
647 * @x, @y: addends
648 * @pcarry: in-out carry value
649 *
650 * Computes @x + @y + *@pcarry, placing the carry-out back
651 * into *@pcarry and returning the 64-bit sum.
652 */
653static inline uint64_t uadd64_carry(uint64_t x, uint64_t y, bool *pcarry)
654{
655#if __has_builtin(__builtin_addcll)
656 unsigned long long c = *pcarry;
657 x = __builtin_addcll(x, y, c, &c);
658 *pcarry = c & 1;
659 return x;
660#else
661 bool c = *pcarry;
662 /* This is clang's internal expansion of __builtin_addc. */
663 c = uadd64_overflow(x, c, &x);
664 c |= uadd64_overflow(x, y, &x);
665 *pcarry = c;
666 return x;
667#endif
668}
669
670/**
671 * usub64_borrow - subtraction with borrow-in and borrow-out
672 * @x, @y: addends
673 * @pborrow: in-out borrow value
674 *
675 * Computes @x - @y - *@pborrow, placing the borrow-out back
676 * into *@pborrow and returning the 64-bit sum.
677 */
678static inline uint64_t usub64_borrow(uint64_t x, uint64_t y, bool *pborrow)
679{
Peter Maydell5726d5d2025-07-14 15:40:02 +0100680#if __has_builtin(__builtin_subcll)
Richard Henderson1ec80702020-11-13 03:22:23 +0000681 unsigned long long b = *pborrow;
682 x = __builtin_subcll(x, y, b, &b);
683 *pborrow = b & 1;
684 return x;
685#else
686 bool b = *pborrow;
687 b = usub64_overflow(x, b, &x);
688 b |= usub64_overflow(x, y, &x);
689 *pborrow = b;
690 return x;
691#endif
692}
693
Richard Henderson01654372013-02-13 17:47:34 -0800694/* Host type specific sizes of these routines. */
695
696#if ULONG_MAX == UINT32_MAX
697# define clzl clz32
698# define ctzl ctz32
699# define clol clo32
700# define ctol cto32
701# define ctpopl ctpop32
Richard Henderson652a4b72015-09-14 13:00:34 -0700702# define revbitl revbit32
Richard Henderson01654372013-02-13 17:47:34 -0800703#elif ULONG_MAX == UINT64_MAX
704# define clzl clz64
705# define ctzl ctz64
706# define clol clo64
707# define ctol cto64
708# define ctpopl ctpop64
Richard Henderson652a4b72015-09-14 13:00:34 -0700709# define revbitl revbit64
Richard Henderson01654372013-02-13 17:47:34 -0800710#else
711# error Unknown sizeof long
712#endif
713
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100714static inline bool is_power_of_2(uint64_t value)
715{
716 if (!value) {
Eric Blakee52eeb42016-05-31 12:33:31 -0600717 return false;
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100718 }
719
720 return !(value & (value - 1));
721}
722
Markus Armbruster43c64a02017-07-27 11:46:15 +0200723/**
724 * Return @value rounded down to the nearest power of two or zero.
725 */
726static inline uint64_t pow2floor(uint64_t value)
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100727{
Markus Armbruster43c64a02017-07-27 11:46:15 +0200728 if (!value) {
729 /* Avoid undefined shift by 64 */
730 return 0;
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100731 }
Markus Armbruster43c64a02017-07-27 11:46:15 +0200732 return 0x8000000000000000ull >> clz64(value);
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100733}
734
Markus Armbruster362aaf12017-07-27 11:46:16 +0200735/*
736 * Return @value rounded up to the nearest power of two modulo 2^64.
737 * This is *zero* for @value > 2^63, so be careful.
738 */
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100739static inline uint64_t pow2ceil(uint64_t value)
740{
Markus Armbruster362aaf12017-07-27 11:46:16 +0200741 int n = clz64(value - 1);
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100742
Markus Armbruster362aaf12017-07-27 11:46:16 +0200743 if (!n) {
744 /*
745 * @value - 1 has no leading zeroes, thus @value - 1 >= 2^63
746 * Therefore, either @value == 0 or @value > 2^63.
747 * If it's 0, return 1, else return 0.
748 */
749 return !value;
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100750 }
Markus Armbruster362aaf12017-07-27 11:46:16 +0200751 return 0x8000000000000000ull >> (n - 1);
Peter Maydell8f1ed5f2015-07-24 13:33:12 +0100752}
753
Yuval Shaia37e626c2018-01-14 11:01:43 +0200754static inline uint32_t pow2roundup32(uint32_t x)
755{
756 x |= (x >> 1);
757 x |= (x >> 2);
758 x |= (x >> 4);
759 x |= (x >> 8);
760 x |= (x >> 16);
761 return x + 1;
762}
763
Jose Ricardo Zivianif539fbe2017-01-10 00:10:09 -0200764/**
765 * urshift - 128-bit Unsigned Right Shift.
766 * @plow: in/out - lower 64-bit integer.
767 * @phigh: in/out - higher 64-bit integer.
768 * @shift: in - bytes to shift, between 0 and 127.
769 *
770 * Result is zero-extended and stored in plow/phigh, which are
771 * input/output variables. Shift values outside the range will
772 * be mod to 128. In other words, the caller is responsible to
773 * verify/assert both the shift range and plow/phigh pointers.
774 */
775void urshift(uint64_t *plow, uint64_t *phigh, int32_t shift);
776
777/**
778 * ulshift - 128-bit Unsigned Left Shift.
779 * @plow: in/out - lower 64-bit integer.
780 * @phigh: in/out - higher 64-bit integer.
781 * @shift: in - bytes to shift, between 0 and 127.
782 * @overflow: out - true if any 1-bit is shifted out.
783 *
784 * Result is zero-extended and stored in plow/phigh, which are
785 * input/output variables. Shift values outside the range will
786 * be mod to 128. In other words, the caller is responsible to
787 * verify/assert both the shift range and plow/phigh pointers.
788 */
789void ulshift(uint64_t *plow, uint64_t *phigh, int32_t shift, bool *overflow);
790
Luis Pires8ac2d6c2021-10-25 16:11:37 -0300791/* From the GNU Multi Precision Library - longlong.h __udiv_qrnnd
792 * (https://gmplib.org/repo/gmp/file/tip/longlong.h)
793 *
794 * Licensed under the GPLv2/LGPLv3
795 */
796static inline uint64_t udiv_qrnnd(uint64_t *r, uint64_t n1,
797 uint64_t n0, uint64_t d)
798{
799#if defined(__x86_64__)
800 uint64_t q;
801 asm("divq %4" : "=a"(q), "=d"(*r) : "0"(n0), "1"(n1), "rm"(d));
802 return q;
803#elif defined(__s390x__) && !defined(__clang__)
804 /* Need to use a TImode type to get an even register pair for DLGR. */
805 unsigned __int128 n = (unsigned __int128)n1 << 64 | n0;
806 asm("dlgr %0, %1" : "+r"(n) : "r"(d));
807 *r = n >> 64;
808 return n;
809#elif defined(_ARCH_PPC64) && defined(_ARCH_PWR7)
810 /* From Power ISA 2.06, programming note for divdeu. */
811 uint64_t q1, q2, Q, r1, r2, R;
812 asm("divdeu %0,%2,%4; divdu %1,%3,%4"
813 : "=&r"(q1), "=r"(q2)
814 : "r"(n1), "r"(n0), "r"(d));
815 r1 = -(q1 * d); /* low part of (n1<<64) - (q1 * d) */
816 r2 = n0 - (q2 * d);
817 Q = q1 + q2;
818 R = r1 + r2;
819 if (R >= d || R < r2) { /* overflow implies R > d */
820 Q += 1;
821 R -= d;
822 }
823 *r = R;
824 return Q;
825#else
826 uint64_t d0, d1, q0, q1, r1, r0, m;
827
828 d0 = (uint32_t)d;
829 d1 = d >> 32;
830
831 r1 = n1 % d1;
832 q1 = n1 / d1;
833 m = q1 * d0;
834 r1 = (r1 << 32) | (n0 >> 32);
835 if (r1 < m) {
836 q1 -= 1;
837 r1 += d;
838 if (r1 >= d) {
839 if (r1 < m) {
840 q1 -= 1;
841 r1 += d;
842 }
843 }
844 }
845 r1 -= m;
846
847 r0 = r1 % d1;
848 q0 = r1 / d1;
849 m = q0 * d0;
850 r0 = (r0 << 32) | (uint32_t)n0;
851 if (r0 < m) {
852 q0 -= 1;
853 r0 += d;
854 if (r0 >= d) {
855 if (r0 < m) {
856 q0 -= 1;
857 r0 += d;
858 }
859 }
860 }
861 r0 -= m;
862
863 *r = r0;
864 return (q1 << 32) | q0;
865#endif
866}
867
Lucas Mateus Castro (alqotel)4724bbd2022-05-25 10:49:50 -0300868Int128 divu256(Int128 *plow, Int128 *phigh, Int128 divisor);
Lucas Mateus Castro (alqotel)62c99472022-05-25 10:49:51 -0300869Int128 divs256(Int128 *plow, Int128 *phigh, Int128 divisor);
Paolo Bonzinicb9c3772012-12-06 12:15:58 +0100870#endif