Server IP : 172.67.216.182 / Your IP : 172.71.81.182 Web Server : Apache System : Linux krdc-ubuntu-s-2vcpu-4gb-amd-blr1-01.localdomain 5.15.0-142-generic #152-Ubuntu SMP Mon May 19 10:54:31 UTC 2025 x86_64 User : www ( 1000) PHP Version : 7.4.33 Disable Function : passthru,exec,system,putenv,chroot,chgrp,chown,shell_exec,popen,proc_open,pcntl_exec,ini_alter,ini_restore,dl,openlog,syslog,readlink,symlink,popepassthru,pcntl_alarm,pcntl_fork,pcntl_waitpid,pcntl_wait,pcntl_wifexited,pcntl_wifstopped,pcntl_wifsignaled,pcntl_wifcontinued,pcntl_wexitstatus,pcntl_wtermsig,pcntl_wstopsig,pcntl_signal,pcntl_signal_dispatch,pcntl_get_last_error,pcntl_strerror,pcntl_sigprocmask,pcntl_sigwaitinfo,pcntl_sigtimedwait,pcntl_exec,pcntl_getpriority,pcntl_setpriority,imap_open,apache_setenv MySQL : OFF | cURL : ON | WGET : ON | Perl : ON | Python : OFF | Sudo : ON | Pkexec : ON Directory : /usr/src/linux-headers-5.15.0-142/arch/h8300/include/asm/ |
Upload File : |
/* SPDX-License-Identifier: GPL-2.0 */ #ifndef _ASM_HASH_H #define _ASM_HASH_H /* * The later H8SX models have a 32x32-bit multiply, but the H8/300H * and H8S have only 16x16->32. Since it's tolerably compact, this is * basically an inlined version of the __mulsi3 code. Since the inputs * are not expected to be small, it's also simplfied by skipping the * early-out checks. * * (Since neither CPU has any multi-bit shift instructions, a * shift-and-add version is a non-starter.) * * TODO: come up with an arch-specific version of the hashing in fs/namei.c, * since that is heavily dependent on rotates. Which, as mentioned, suck * horribly on H8. */ #if defined(CONFIG_CPU_H300H) || defined(CONFIG_CPU_H8S) #define HAVE_ARCH__HASH_32 1 /* * Multiply by k = 0x61C88647. Fitting this into three registers requires * one extra instruction, but reducing register pressure will probably * make that back and then some. * * GCC asm note: %e1 is the high half of operand %1, while %f1 is the * low half. So if %1 is er4, then %e1 is e4 and %f1 is r4. * * This has been designed to modify x in place, since that's the most * common usage, but preserve k, since hash_64() makes two calls in * quick succession. */ static inline u32 __attribute_const__ __hash_32(u32 x) { u32 temp; asm( "mov.w %e1,%f0" "\n mulxu.w %f2,%0" /* klow * xhigh */ "\n mov.w %f0,%e1" /* The extra instruction */ "\n mov.w %f1,%f0" "\n mulxu.w %e2,%0" /* khigh * xlow */ "\n add.w %e1,%f0" "\n mulxu.w %f2,%1" /* klow * xlow */ "\n add.w %f0,%e1" : "=&r" (temp), "=r" (x) : "%r" (GOLDEN_RATIO_32), "1" (x)); return x; } #endif #endif /* _ASM_HASH_H */